code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" def _a ( _snake_case , _snake_case ): """simple docstring""" if len(_snake_case ) != len(_snake_case ): raise ValueError("""String lengths must match!""" ) UpperCAmelCase = 0 for chara, chara in zip(_snake_case , _snake_case ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( _snake_case = 0.1 ): """simple docstring""" UpperCAmelCase = 3 UpperCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup _UpperCamelCase = logging.get_logger(__name__) class lowerCamelCase__ ( snake_case ): def __init__( self ,**A ): requires_backends(self ,["""bs4"""] ) super().__init__(**A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = [] UpperCAmelCase = [] UpperCAmelCase = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag UpperCAmelCase = parent.find_all(child.name ,recursive=A ) xpath_tags.append(child.name ) xpath_subscripts.append( 0 if 1 == len(A ) else next(i for i, s in enumerate(A ,1 ) if s is child ) ) UpperCAmelCase = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def _UpperCamelCase ( self ,A ): UpperCAmelCase = BeautifulSoup(A ,"""html.parser""" ) UpperCAmelCase = [] UpperCAmelCase = [] UpperCAmelCase = [] for element in html_code.descendants: if type(A ) == bsa.element.NavigableString: if type(element.parent ) != bsa.element.Tag: continue UpperCAmelCase = html.unescape(A ).strip() if not text_in_this_tag: continue all_doc_strings.append(A ) UpperCAmelCase , UpperCAmelCase = self.xpath_soup(A ) stringaxtag_seq.append(A ) stringaxsubs_seq.append(A ) if len(A ) != len(A ): raise ValueError("""Number of doc strings and xtags does not correspond""" ) if len(A ) != len(A ): raise ValueError("""Number of doc strings and xsubs does not correspond""" ) return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = """""" for tagname, subs in zip(A ,A ): xpath += F'''/{tagname}''' if subs != 0: xpath += F'''[{subs}]''' return xpath def __call__( self ,A ): UpperCAmelCase = False # Check that strings has a valid type if isinstance(A ,A ): UpperCAmelCase = True elif isinstance(A ,(list, tuple) ): if len(A ) == 0 or isinstance(html_strings[0] ,A ): UpperCAmelCase = True if not valid_strings: raise ValueError( """HTML strings must of type `str`, `List[str]` (batch of examples), """ F'''but is of type {type(A )}.''' ) UpperCAmelCase = bool(isinstance(A ,(list, tuple) ) and (isinstance(html_strings[0] ,A )) ) if not is_batched: UpperCAmelCase = [html_strings] # Get nodes + xpaths UpperCAmelCase = [] UpperCAmelCase = [] for html_string in html_strings: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.get_three_from_single(A ) nodes.append(A ) UpperCAmelCase = [] for node, tag_list, sub_list in zip(A ,A ,A ): UpperCAmelCase = self.construct_xpath(A ,A ) xpath_strings.append(A ) xpaths.append(A ) # return as Dict UpperCAmelCase = {"""nodes""": nodes, """xpaths""": xpaths} UpperCAmelCase = BatchFeature(data=A ,tensor_type=A ) return encoded_inputs
74
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE = '''CLIPImageProcessor''' SCREAMING_SNAKE_CASE = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self ,A=None ,A=None ,**A ): UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) def __call__( self ,A=None ,A=None ,A=None ,**A ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.batch_decode(*A ,**A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.decode(*A ,**A ) @property def _UpperCamelCase ( self ): UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,A ,) return self.image_processor
74
1
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase__ : def __init__( self ,A ,A=13 ,A=7 ,A=True ,A=True ,A=False ,A=True ,A=99 ,A=32 ,A=5 ,A=4 ,A=37 ,A="gelu" ,A=0.1 ,A=0.1 ,A=512 ,A=16 ,A=2 ,A=0.02 ,A=3 ,A=4 ,A=None ,): UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self ): return BioGptConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A ,initializer_range=self.initializer_range ,) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ): UpperCAmelCase = BioGptModel(config=A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = BioGptForCausalLM(config=A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ,token_type_ids=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,*A ): UpperCAmelCase = BioGptModel(config=A ) model.to(A ) model.eval() # create attention mask UpperCAmelCase = torch.ones(input_ids.shape ,dtype=torch.long ,device=A ) UpperCAmelCase = self.seq_length // 2 UpperCAmelCase = 0 # first forward pass UpperCAmelCase , UpperCAmelCase = model(A ,attention_mask=A ).to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase = ids_tensor((self.batch_size, 1) ,config.vocab_size ) # change a random masked slice from input_ids UpperCAmelCase = ids_tensor((1,) ,A ).item() + 1 UpperCAmelCase = ids_tensor((self.batch_size, 1) ,config.vocab_size ).squeeze(-1 ) UpperCAmelCase = random_other_next_tokens # append to next input_ids and attn_mask UpperCAmelCase = torch.cat([input_ids, next_tokens] ,dim=-1 ) UpperCAmelCase = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) ,dtype=torch.long ,device=A )] ,dim=1 ,) # get two different outputs UpperCAmelCase = model(A ,attention_mask=A )["""last_hidden_state"""] UpperCAmelCase = model(A ,past_key_values=A ,attention_mask=A )["""last_hidden_state"""] # select random slice UpperCAmelCase = ids_tensor((1,) ,output_from_past.shape[-1] ).item() UpperCAmelCase = output_from_no_past[:, -1, random_slice_idx].detach() UpperCAmelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A ,A ,atol=1e-3 ) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,*A ): UpperCAmelCase = BioGptModel(config=A ).to(A ).eval() UpperCAmelCase = torch.ones(input_ids.shape ,dtype=torch.long ,device=A ) # first forward pass UpperCAmelCase = model(A ,attention_mask=A ,use_cache=A ) UpperCAmelCase , UpperCAmelCase = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase = ids_tensor((self.batch_size, 3) ,config.vocab_size ) UpperCAmelCase = ids_tensor((self.batch_size, 3) ,2 ) # append to next input_ids and UpperCAmelCase = torch.cat([input_ids, next_tokens] ,dim=-1 ) UpperCAmelCase = torch.cat([attention_mask, next_attn_mask] ,dim=-1 ) UpperCAmelCase = model(A ,attention_mask=A )["""last_hidden_state"""] UpperCAmelCase = model(A ,attention_mask=A ,past_key_values=A )[ """last_hidden_state""" ] # select random slice UpperCAmelCase = ids_tensor((1,) ,output_from_past.shape[-1] ).item() UpperCAmelCase = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A ,A ,atol=1e-3 ) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,*A ,A=False ): UpperCAmelCase = BioGptForCausalLM(A ) model.to(A ) if gradient_checkpointing: model.gradient_checkpointing_enable() UpperCAmelCase = model(A ,labels=A ) self.parent.assertEqual(result.loss.shape ,() ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _UpperCamelCase ( self ,A ,*A ): UpperCAmelCase = BioGptModel(A ) UpperCAmelCase = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) ,0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) ,0.01 ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,*A ): UpperCAmelCase = self.num_labels UpperCAmelCase = BioGptForTokenClassification(A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ,token_type_ids=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCamelCase__ ( snake_case , snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) SCREAMING_SNAKE_CASE = (BioGptForCausalLM,) if is_torch_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': BioGptModel, '''text-classification''': BioGptForSequenceClassification, '''text-generation''': BioGptForCausalLM, '''token-classification''': BioGptForTokenClassification, '''zero-shot''': BioGptForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ): UpperCAmelCase = BioGptModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,hidden_size=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase = type self.model_tester.create_and_check_model(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*A ,gradient_checkpointing=A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*A ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(A ) UpperCAmelCase = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) UpperCAmelCase = """left""" # Define PAD Token = EOS Token = 50256 UpperCAmelCase = tokenizer.eos_token UpperCAmelCase = model.config.eos_token_id # use different length sentences to test batching UpperCAmelCase = [ """Hello, my dog is a little""", """Today, I""", ] UpperCAmelCase = tokenizer(A ,return_tensors="""pt""" ,padding=A ) UpperCAmelCase = inputs["""input_ids"""].to(A ) UpperCAmelCase = model.generate( input_ids=A ,attention_mask=inputs["""attention_mask"""].to(A ) ,) UpperCAmelCase = tokenizer(sentences[0] ,return_tensors="""pt""" ).input_ids.to(A ) UpperCAmelCase = model.generate(input_ids=A ) UpperCAmelCase = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() UpperCAmelCase = tokenizer(sentences[1] ,return_tensors="""pt""" ).input_ids.to(A ) UpperCAmelCase = model.generate(input_ids=A ,max_length=model.config.max_length - num_paddings ) UpperCAmelCase = tokenizer.batch_decode(A ,skip_special_tokens=A ) UpperCAmelCase = tokenizer.decode(output_non_padded[0] ,skip_special_tokens=A ) UpperCAmelCase = tokenizer.decode(output_padded[0] ,skip_special_tokens=A ) UpperCAmelCase = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(A ,A ) self.assertListEqual(A ,[non_padded_sentence, padded_sentence] ) @slow def _UpperCamelCase ( self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = BioGptModel.from_pretrained(A ) self.assertIsNotNone(A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = 3 UpperCAmelCase = input_dict["""input_ids"""] UpperCAmelCase = input_ids.ne(1 ).to(A ) UpperCAmelCase = ids_tensor([self.model_tester.batch_size] ,self.model_tester.type_sequence_label_size ) UpperCAmelCase = BioGptForSequenceClassification(A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ,labels=A ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = 3 UpperCAmelCase = """multi_label_classification""" UpperCAmelCase = input_dict["""input_ids"""] UpperCAmelCase = input_ids.ne(1 ).to(A ) UpperCAmelCase = ids_tensor( [self.model_tester.batch_size, config.num_labels] ,self.model_tester.type_sequence_label_size ).to(torch.float ) UpperCAmelCase = BioGptForSequenceClassification(A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ,labels=A ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class lowerCamelCase__ ( unittest.TestCase ): @slow def _UpperCamelCase ( self ): UpperCAmelCase = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) UpperCAmelCase = torch.tensor([[2, 4_805, 9, 656, 21]] ) UpperCAmelCase = model(A )[0] UpperCAmelCase = 42_384 UpperCAmelCase = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape ,A ) UpperCAmelCase = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,A ,atol=1e-4 ) ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) UpperCAmelCase = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(A ) torch.manual_seed(0 ) UpperCAmelCase = tokenizer("""COVID-19 is""" ,return_tensors="""pt""" ).to(A ) UpperCAmelCase = model.generate( **A ,min_length=100 ,max_length=1_024 ,num_beams=5 ,early_stopping=A ,) UpperCAmelCase = tokenizer.decode(output_ids[0] ,skip_special_tokens=A ) UpperCAmelCase = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(A ,A )
74
"""simple docstring""" from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def _a ( _snake_case = "mumbai" ): """simple docstring""" UpperCAmelCase = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): UpperCAmelCase = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() UpperCAmelCase = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("""Bangalore"""), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
74
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _UpperCamelCase = {"""configuration_encoder_decoder""": ["""EncoderDecoderConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""EncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""TFEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""FlaxEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): UpperCAmelCase = ["""a""", """b""", """c"""] # Defaults to last layer if both are None UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,A ,A ) self.assertEqual(A ,["""c"""] ) self.assertEqual(A ,[2] ) # Out indices set to match out features UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(["""a""", """c"""] ,A ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features set to match out indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[0, 2] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features selected from negative indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[-3, -1] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[-3, -1] ) def _UpperCamelCase ( self ): # Stage names must be set with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,A ) # Out features must be a list with self.assertRaises(A ): verify_out_features_out_indices(("""a""", """b""") ,(0, 1) ,["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(A ): verify_out_features_out_indices(A ,0 ,["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(A ,(0, 1) ,["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0,) ,["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 2) ,["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(A ): verify_out_features_out_indices(["""b""", """a"""] ,(0, 1) ,["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] ,(0, 1, -1) ,["""a""", """b""", """c""", """d"""] ) def _UpperCamelCase ( self ): UpperCAmelCase = BackboneMixin() UpperCAmelCase = ["""a""", """b""", """c"""] UpperCAmelCase = ["""a""", """c"""] UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[0, 2] ) # Check out features and indices are updated correctly UpperCAmelCase = ["""a""", """b"""] self.assertEqual(backbone.out_features ,["""a""", """b"""] ) self.assertEqual(backbone.out_indices ,[0, 1] ) UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[-3, -1] )
74
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCamelCase = { """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCamelCase__ : def __init__( self ,A = 6 ): UpperCAmelCase = None UpperCAmelCase = None self.create_linked_list(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = current_node UpperCAmelCase = current_node for _ in range(1 ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = previous_node UpperCAmelCase = current_node UpperCAmelCase = self.front UpperCAmelCase = previous_node def _UpperCamelCase ( self ): return ( self.front == self.rear and self.front is not None and self.front.data is None ) def _UpperCamelCase ( self ): self.check_can_perform_operation() return self.front.data if self.front else None def _UpperCamelCase ( self ,A ): if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase = self.rear.next if self.rear: UpperCAmelCase = data def _UpperCamelCase ( self ): self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase = self.front.data UpperCAmelCase = None return data UpperCAmelCase = self.front UpperCAmelCase = old_front.next UpperCAmelCase = old_front.data UpperCAmelCase = None return data def _UpperCamelCase ( self ): if self.is_empty(): raise Exception("""Empty Queue""" ) def _UpperCamelCase ( self ): if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=snake_case ) class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = field(default='''image-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) SCREAMING_SNAKE_CASE = Features({'''image''': Image()} ) SCREAMING_SNAKE_CASE = Features({'''labels''': ClassLabel} ) SCREAMING_SNAKE_CASE = "image" SCREAMING_SNAKE_CASE = "labels" def _UpperCamelCase ( self ,A ): if self.label_column not in features: raise ValueError(F'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] ,A ): raise ValueError(F'''Column {self.label_column} is not a ClassLabel.''' ) UpperCAmelCase = copy.deepcopy(self ) UpperCAmelCase = self.label_schema.copy() UpperCAmelCase = features[self.label_column] UpperCAmelCase = label_schema return task_template @property def _UpperCamelCase ( self ): return { self.image_column: "image", self.label_column: "labels", }
74
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("""transformers.models.speecht5""") def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" hf_model.apply_weight_norm() UpperCAmelCase = checkpoint["""input_conv.weight_g"""] UpperCAmelCase = checkpoint["""input_conv.weight_v"""] UpperCAmelCase = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , ): """simple docstring""" if config_path is not None: UpperCAmelCase = SpeechTaHifiGanConfig.from_pretrained(_snake_case ) else: UpperCAmelCase = SpeechTaHifiGanConfig() UpperCAmelCase = SpeechTaHifiGan(_snake_case ) UpperCAmelCase = torch.load(_snake_case ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _snake_case , _snake_case ) UpperCAmelCase = np.load(_snake_case ) UpperCAmelCase = stats[0].reshape(-1 ) UpperCAmelCase = stats[1].reshape(-1 ) UpperCAmelCase = torch.from_numpy(_snake_case ).float() UpperCAmelCase = torch.from_numpy(_snake_case ).float() model.save_pretrained(_snake_case ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
74
1
"""simple docstring""" import argparse import struct import unittest class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = data # Initialize hash values UpperCAmelCase = [ 0x6A_09_E6_67, 0xBB_67_AE_85, 0x3C_6E_F3_72, 0xA5_4F_F5_3A, 0x51_0E_52_7F, 0x9B_05_68_8C, 0x1F_83_D9_AB, 0x5B_E0_CD_19, ] # Initialize round constants UpperCAmelCase = [ 0x42_8A_2F_98, 0x71_37_44_91, 0xB5_C0_FB_CF, 0xE9_B5_DB_A5, 0x39_56_C2_5B, 0x59_F1_11_F1, 0x92_3F_82_A4, 0xAB_1C_5E_D5, 0xD8_07_AA_98, 0x12_83_5B_01, 0x24_31_85_BE, 0x55_0C_7D_C3, 0x72_BE_5D_74, 0x80_DE_B1_FE, 0x9B_DC_06_A7, 0xC1_9B_F1_74, 0xE4_9B_69_C1, 0xEF_BE_47_86, 0x0F_C1_9D_C6, 0x24_0C_A1_CC, 0x2D_E9_2C_6F, 0x4A_74_84_AA, 0x5C_B0_A9_DC, 0x76_F9_88_DA, 0x98_3E_51_52, 0xA8_31_C6_6D, 0xB0_03_27_C8, 0xBF_59_7F_C7, 0xC6_E0_0B_F3, 0xD5_A7_91_47, 0x06_CA_63_51, 0x14_29_29_67, 0x27_B7_0A_85, 0x2E_1B_21_38, 0x4D_2C_6D_FC, 0x53_38_0D_13, 0x65_0A_73_54, 0x76_6A_0A_BB, 0x81_C2_C9_2E, 0x92_72_2C_85, 0xA2_BF_E8_A1, 0xA8_1A_66_4B, 0xC2_4B_8B_70, 0xC7_6C_51_A3, 0xD1_92_E8_19, 0xD6_99_06_24, 0xF4_0E_35_85, 0x10_6A_A0_70, 0x19_A4_C1_16, 0x1E_37_6C_08, 0x27_48_77_4C, 0x34_B0_BC_B5, 0x39_1C_0C_B3, 0x4E_D8_AA_4A, 0x5B_9C_CA_4F, 0x68_2E_6F_F3, 0x74_8F_82_EE, 0x78_A5_63_6F, 0x84_C8_78_14, 0x8C_C7_02_08, 0x90_BE_FF_FA, 0xA4_50_6C_EB, 0xBE_F9_A3_F7, 0xC6_71_78_F2, ] UpperCAmelCase = self.preprocessing(self.data ) self.final_hash() @staticmethod def _UpperCamelCase ( A ): UpperCAmelCase = b"""\x80""" + (b"""\x00""" * (63 - (len(A ) + 8) % 64)) UpperCAmelCase = struct.pack(""">Q""" ,(len(A ) * 8) ) return data + padding + big_endian_integer def _UpperCamelCase ( self ): # Convert into blocks of 64 bytes UpperCAmelCase = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers UpperCAmelCase = list(struct.unpack(""">16L""" ,A ) ) # add 48 0-ed integers words += [0] * 48 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array UpperCAmelCase = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) UpperCAmelCase = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) UpperCAmelCase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_00_00_00_00 # Compression UpperCAmelCase = self.ror(A ,6 ) ^ self.ror(A ,11 ) ^ self.ror(A ,25 ) UpperCAmelCase = (e & f) ^ ((~e & 0xFF_FF_FF_FF) & g) UpperCAmelCase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_00_00_00_00 UpperCAmelCase = self.ror(A ,2 ) ^ self.ror(A ,13 ) ^ self.ror(A ,22 ) UpperCAmelCase = (a & b) ^ (a & c) ^ (b & c) UpperCAmelCase = (sa + maj) % 0x1_00_00_00_00 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = ( g, f, e, ((d + tempa) % 0x1_00_00_00_00), c, b, a, ((tempa + tempa) % 0x1_00_00_00_00), ) UpperCAmelCase = [a, b, c, d, e, f, g, h] # Modify final values UpperCAmelCase = [ ((element + mutated_hash_values[index]) % 0x1_00_00_00_00) for index, element in enumerate(self.hashes ) ] UpperCAmelCase = """""".join([hex(A )[2:].zfill(8 ) for value in self.hashes] ) def _UpperCamelCase ( self ,A ,A ): return 0xFF_FF_FF_FF & (value << (32 - rotations)) | (value >> rotations) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): import hashlib UpperCAmelCase = bytes("""Test String""" ,"""utf-8""" ) self.assertEqual(SHAaaa(A ).hash ,hashlib.shaaaa(A ).hexdigest() ) def _a ( ): """simple docstring""" import doctest doctest.testmod() UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase = parser.parse_args() UpperCAmelCase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase = f.read() else: UpperCAmelCase = bytes(_snake_case , """utf-8""" ) print(SHAaaa(_snake_case ).hash ) if __name__ == "__main__": main()
74
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase = concatenate_datasets _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadManager _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
74
1
"""simple docstring""" def _a ( _snake_case ): # noqa: E741 """simple docstring""" UpperCAmelCase = len(_snake_case ) UpperCAmelCase = 0 UpperCAmelCase = [0] * n UpperCAmelCase = [False] * n UpperCAmelCase = [False] * n def dfs(_snake_case , _snake_case , _snake_case , _snake_case ): if parent == root: out_edge_count += 1 UpperCAmelCase = True UpperCAmelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase = dfs(_snake_case , _snake_case , _snake_case , _snake_case ) UpperCAmelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase = True # AP found via cycle if at == low[to]: UpperCAmelCase = True else: UpperCAmelCase = min(low[at] , _snake_case ) return out_edge_count for i in range(_snake_case ): if not visited[i]: UpperCAmelCase = 0 UpperCAmelCase = dfs(_snake_case , _snake_case , -1 , _snake_case ) UpperCAmelCase = out_edge_count > 1 for x in range(len(_snake_case ) ): if is_art[x] is True: print(_snake_case ) # Adjacency list of graph _UpperCamelCase = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
74
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowerCamelCase__ : def _UpperCamelCase ( self ): torch.manual_seed(0 ) UpperCAmelCase = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) UpperCAmelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( sample_size=32 ,layers_per_block=1 ,block_out_channels=[32, 64] ,down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] ,mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" ,up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] ,in_channels=3 ,out_channels=6 ,cross_attention_dim=32 ,encoder_hid_dim=32 ,attention_head_dim=8 ,addition_embed_type="""text""" ,addition_embed_type_num_heads=2 ,cross_attention_norm="""group_norm""" ,resnet_time_scale_shift="""scale_shift""" ,act_fn="""gelu""" ,) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) UpperCAmelCase = DDPMScheduler( num_train_timesteps=1_000 ,beta_schedule="""squaredcos_cap_v2""" ,beta_start=0.0001 ,beta_end=0.02 ,thresholding=A ,dynamic_thresholding_ratio=0.95 ,sample_max_value=1.0 ,prediction_type="""epsilon""" ,variance_type="""learned_range""" ,) torch.manual_seed(0 ) UpperCAmelCase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def _UpperCamelCase ( self ): torch.manual_seed(0 ) UpperCAmelCase = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) UpperCAmelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( sample_size=32 ,layers_per_block=[1, 2] ,block_out_channels=[32, 64] ,down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] ,mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" ,up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] ,in_channels=6 ,out_channels=6 ,cross_attention_dim=32 ,encoder_hid_dim=32 ,attention_head_dim=8 ,addition_embed_type="""text""" ,addition_embed_type_num_heads=2 ,cross_attention_norm="""group_norm""" ,resnet_time_scale_shift="""scale_shift""" ,act_fn="""gelu""" ,class_embed_type="""timestep""" ,mid_block_scale_factor=1.414 ,time_embedding_act_fn="""gelu""" ,time_embedding_dim=32 ,) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) UpperCAmelCase = DDPMScheduler( num_train_timesteps=1_000 ,beta_schedule="""squaredcos_cap_v2""" ,beta_start=0.0001 ,beta_end=0.02 ,thresholding=A ,dynamic_thresholding_ratio=0.95 ,sample_max_value=1.0 ,prediction_type="""epsilon""" ,variance_type="""learned_range""" ,) torch.manual_seed(0 ) UpperCAmelCase = DDPMScheduler( num_train_timesteps=1_000 ,beta_schedule="""squaredcos_cap_v2""" ,beta_start=0.0001 ,beta_end=0.02 ,) torch.manual_seed(0 ) UpperCAmelCase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def _UpperCamelCase ( self ): UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**A ) pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase = self.get_dummy_inputs(A ) UpperCAmelCase = inputs["""prompt"""] UpperCAmelCase = inputs["""generator"""] UpperCAmelCase = inputs["""num_inference_steps"""] UpperCAmelCase = inputs["""output_type"""] if "image" in inputs: UpperCAmelCase = inputs["""image"""] else: UpperCAmelCase = None if "mask_image" in inputs: UpperCAmelCase = inputs["""mask_image"""] else: UpperCAmelCase = None if "original_image" in inputs: UpperCAmelCase = inputs["""original_image"""] else: UpperCAmelCase = None UpperCAmelCase , UpperCAmelCase = pipe.encode_prompt(A ) # inputs with prompt converted to embeddings UpperCAmelCase = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: UpperCAmelCase = image if mask_image is not None: UpperCAmelCase = mask_image if original_image is not None: UpperCAmelCase = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(A ,A ,A ) UpperCAmelCase = pipe(**A )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(A ) UpperCAmelCase = self.pipeline_class.from_pretrained(A ) pipe_loaded.to(A ) pipe_loaded.set_progress_bar_config(disable=A ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(A ,A ) is None ,F'''`{optional_component}` did not stay set to None after loading.''' ,) UpperCAmelCase = self.get_dummy_inputs(A ) UpperCAmelCase = inputs["""generator"""] UpperCAmelCase = inputs["""num_inference_steps"""] UpperCAmelCase = inputs["""output_type"""] # inputs with prompt converted to embeddings UpperCAmelCase = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: UpperCAmelCase = image if mask_image is not None: UpperCAmelCase = mask_image if original_image is not None: UpperCAmelCase = original_image UpperCAmelCase = pipe_loaded(**A )[0] UpperCAmelCase = np.abs(to_np(A ) - to_np(A ) ).max() self.assertLess(A ,1e-4 ) def _UpperCamelCase ( self ): UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**A ) pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase = self.get_dummy_inputs(A ) UpperCAmelCase = pipe(**A )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(A ) UpperCAmelCase = self.pipeline_class.from_pretrained(A ) pipe_loaded.to(A ) pipe_loaded.set_progress_bar_config(disable=A ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests UpperCAmelCase = self.get_dummy_inputs(A ) UpperCAmelCase = pipe_loaded(**A )[0] UpperCAmelCase = np.abs(to_np(A ) - to_np(A ) ).max() self.assertLess(A ,1e-4 )
74
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCamelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") _UpperCamelCase = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) _UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the training data.'''} ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the validation data.'''} ) SCREAMING_SNAKE_CASE = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) SCREAMING_SNAKE_CASE = field(default=32 , metadata={'''help''': '''The size of the square patches to use for masking.'''} ) SCREAMING_SNAKE_CASE = field( default=0.6 , metadata={'''help''': '''Percentage of patches to mask.'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def _UpperCamelCase ( self ): UpperCAmelCase = {} if self.train_dir is not None: UpperCAmelCase = self.train_dir if self.validation_dir is not None: UpperCAmelCase = self.validation_dir UpperCAmelCase = data_files if data_files else None @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ''' '''checkpoint identifier on the hub. ''' '''Don\'t set if you want to train a model from scratch.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(snake_case )} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'''} , ) SCREAMING_SNAKE_CASE = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''Name or path of preprocessor config.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each image. If not specified, will use `image_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Stride to use for the encoder.'''} , ) class lowerCamelCase__ : def __init__( self ,A=192 ,A=32 ,A=4 ,A=0.6 ): UpperCAmelCase = input_size UpperCAmelCase = mask_patch_size UpperCAmelCase = model_patch_size UpperCAmelCase = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError("""Input size must be divisible by mask patch size""" ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError("""Mask patch size must be divisible by model patch size""" ) UpperCAmelCase = self.input_size // self.mask_patch_size UpperCAmelCase = self.mask_patch_size // self.model_patch_size UpperCAmelCase = self.rand_size**2 UpperCAmelCase = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self ): UpperCAmelCase = np.random.permutation(self.token_count )[: self.mask_count] UpperCAmelCase = np.zeros(self.token_count ,dtype=A ) UpperCAmelCase = 1 UpperCAmelCase = mask.reshape((self.rand_size, self.rand_size) ) UpperCAmelCase = mask.repeat(self.scale ,axis=0 ).repeat(self.scale ,axis=1 ) return torch.tensor(mask.flatten() ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = torch.stack([example["""pixel_values"""] for example in examples] ) UpperCAmelCase = torch.stack([example["""mask"""] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mim""" , _snake_case , _snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_snake_case ) transformers.utils.logging.set_verbosity(_snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. UpperCAmelCase = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _snake_case ) and data_args.train_val_split > 0.0: UpperCAmelCase = ds["""train"""].train_test_split(data_args.train_val_split ) UpperCAmelCase = split["""train"""] UpperCAmelCase = split["""test"""] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.config_name_or_path , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_snake_case , """decoder_type""" ): UpperCAmelCase = """simmim""" # adapt config UpperCAmelCase = model_args.image_size if model_args.image_size is not None else config.image_size UpperCAmelCase = model_args.patch_size if model_args.patch_size is not None else config.patch_size UpperCAmelCase = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { """image_size""": model_args.image_size, """patch_size""": model_args.patch_size, """encoder_stride""": model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } UpperCAmelCase = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: UpperCAmelCase = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) UpperCAmelCase = AutoModelForMaskedImageModeling.from_config(_snake_case ) if training_args.do_train: UpperCAmelCase = ds["""train"""].column_names else: UpperCAmelCase = ds["""validation"""].column_names if data_args.image_column_name is not None: UpperCAmelCase = data_args.image_column_name elif "image" in column_names: UpperCAmelCase = """image""" elif "img" in column_names: UpperCAmelCase = """img""" else: UpperCAmelCase = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py UpperCAmelCase = Compose( [ Lambda(lambda _snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator UpperCAmelCase = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_snake_case ): UpperCAmelCase = [transforms(_snake_case ) for image in examples[image_column_name]] UpperCAmelCase = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: UpperCAmelCase = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: UpperCAmelCase = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_snake_case ) # Initialize our trainer UpperCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase = last_checkpoint UpperCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase = trainer.evaluate() trainer.log_metrics("""eval""" , _snake_case ) trainer.save_metrics("""eval""" , _snake_case ) # Write model card and (optionally) push to hub UpperCAmelCase = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """masked-image-modeling""", """dataset""": data_args.dataset_name, """tags""": ["""masked-image-modeling"""], } if training_args.push_to_hub: trainer.push_to_hub(**_snake_case ) else: trainer.create_model_card(**_snake_case ) if __name__ == "__main__": main()
74
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _UpperCamelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""GPTSw3Tokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): _UpperCamelCase = True from torch.cuda.amp import autocast _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to log verbose messages or not.'''} , ) SCREAMING_SNAKE_CASE = field( default=2.0 , metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.5 , metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.99_99_95 , metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def _a ( _snake_case , _snake_case ): """simple docstring""" logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase = logging.WARNING if model_args.verbose_logging: UpperCAmelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase = logging.INFO logger.setLevel(_snake_case ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default='''train''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) SCREAMING_SNAKE_CASE = field( default='''validation''' , metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) } , ) SCREAMING_SNAKE_CASE = field( default='''file''' , metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) SCREAMING_SNAKE_CASE = field( default=1 , metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) SCREAMING_SNAKE_CASE = field( default=20.0 , metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = "longest" SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __call__( self ,A ): # reformat list to dict and set to pytorch format UpperCAmelCase = self.feature_extractor.pad( A ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="""pt""" ,) UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""input_values"""].shape[-1] ) UpperCAmelCase = batch["""input_values"""].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""attention_mask"""].sum(-1 ) ).to( torch.long ) UpperCAmelCase = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["""input_values"""].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase = 1 UpperCAmelCase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=A ,min_masks=2 ,) return batch class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=1 ,A=0 ,A=1.0 ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = 0 UpperCAmelCase = max_gumbel_temp UpperCAmelCase = min_gumbel_temp UpperCAmelCase = gumbel_temp_decay def _UpperCamelCase ( self ,A ,A ): model.train() UpperCAmelCase = self._prepare_inputs(A ) if self.use_amp: with autocast(): UpperCAmelCase = self.compute_loss(A ,A ) else: UpperCAmelCase = self.compute_loss(A ,A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase = loss.sum() / (inputs["""mask_time_indices"""]).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() configure_logger(_snake_case , _snake_case ) # Downloading and loading a dataset from the hub. UpperCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="""validation""" , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=_snake_case ) def prepare_dataset(_snake_case ): # check that all files have the correct sampling rate UpperCAmelCase , UpperCAmelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase = datasets.map( _snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["""train"""].column_names ) # filter audio files that are too long UpperCAmelCase = vectorized_datasets.filter( lambda _snake_case : len(data["""speech"""] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(_snake_case ): return feature_extractor(batch["""speech"""] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase = vectorized_datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["""train"""].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( """PreTraining is only supported for ``config.do_stable_layer_norm=True`` and""" """ ``config.feat_extract_norm='layer'""" ) UpperCAmelCase = WavaVecaForPreTraining(_snake_case ) UpperCAmelCase = DataCollatorForWavaVecaPretraining(model=_snake_case , feature_extractor=_snake_case ) UpperCAmelCase = WavaVecaPreTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , train_dataset=vectorized_datasets["""train"""] , eval_dataset=vectorized_datasets["""validation"""] , tokenizer=_snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
74
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer _UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name _UpperCamelCase = """ Examples: ```py >>> from PIL import Image >>> import torch >>> from diffusers import DiffusionPipeline >>> from diffusers.utils import export_to_gif, load_image >>> device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\") >>> repo = \"openai/shap-e-img2img\" >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16) >>> pipe = pipe.to(device) >>> guidance_scale = 3.0 >>> image_url = \"https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png\" >>> image = load_image(image_url).convert(\"RGB\") >>> images = pipe( ... image, ... guidance_scale=guidance_scale, ... num_inference_steps=64, ... frame_size=256, ... ).images >>> gif_path = export_to_gif(images[0], \"corgi_3d.gif\") ``` """ @dataclass class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = 42 class lowerCamelCase__ ( snake_case ): def __init__( self ,A ,A ,A ,A ,A ,): super().__init__() self.register_modules( prior=A ,image_encoder=A ,image_processor=A ,scheduler=A ,renderer=A ,) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ): if latents is None: UpperCAmelCase = randn_tensor(A ,generator=A ,device=A ,dtype=A ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) UpperCAmelCase = latents.to(A ) UpperCAmelCase = latents * scheduler.init_noise_sigma return latents def _UpperCamelCase ( self ,A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) UpperCAmelCase = torch.device(F'''cuda:{gpu_id}''' ) UpperCAmelCase = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(A ,A ) @property def _UpperCamelCase ( self ): if self.device != torch.device("""meta""" ) or not hasattr(self.image_encoder ,"""_hf_hook""" ): return self.device for module in self.image_encoder.modules(): if ( hasattr(A ,"""_hf_hook""" ) and hasattr(module._hf_hook ,"""execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def _UpperCamelCase ( self ,A ,A ,A ,A ,): if isinstance(A ,A ) and isinstance(image[0] ,torch.Tensor ): UpperCAmelCase = torch.cat(A ,axis=0 ) if image[0].ndim == 4 else torch.stack(A ,axis=0 ) if not isinstance(A ,torch.Tensor ): UpperCAmelCase = self.image_processor(A ,return_tensors="""pt""" ).pixel_values[0].unsqueeze(0 ) UpperCAmelCase = image.to(dtype=self.image_encoder.dtype ,device=A ) UpperCAmelCase = self.image_encoder(A )["""last_hidden_state"""] UpperCAmelCase = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 UpperCAmelCase = image_embeds.repeat_interleave(A ,dim=0 ) if do_classifier_free_guidance: UpperCAmelCase = torch.zeros_like(A ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(A ) def __call__( self ,A ,A = 1 ,A = 25 ,A = None ,A = None ,A = 4.0 ,A = 64 ,A = "pil" ,A = True ,): if isinstance(A ,PIL.Image.Image ): UpperCAmelCase = 1 elif isinstance(A ,torch.Tensor ): UpperCAmelCase = image.shape[0] elif isinstance(A ,A ) and isinstance(image[0] ,(torch.Tensor, PIL.Image.Image) ): UpperCAmelCase = len(A ) else: raise ValueError( F'''`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(A )}''' ) UpperCAmelCase = self._execution_device UpperCAmelCase = batch_size * num_images_per_prompt UpperCAmelCase = guidance_scale > 1.0 UpperCAmelCase = self._encode_image(A ,A ,A ,A ) # prior self.scheduler.set_timesteps(A ,device=A ) UpperCAmelCase = self.scheduler.timesteps UpperCAmelCase = self.prior.config.num_embeddings UpperCAmelCase = self.prior.config.embedding_dim UpperCAmelCase = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) ,image_embeds.dtype ,A ,A ,A ,self.scheduler ,) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim UpperCAmelCase = latents.reshape(latents.shape[0] ,A ,A ) for i, t in enumerate(self.progress_bar(A ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase = self.scheduler.scale_model_input(A ,A ) UpperCAmelCase = self.prior( A ,timestep=A ,proj_embedding=A ,).predicted_image_embedding # remove the variance UpperCAmelCase , UpperCAmelCase = noise_pred.split( scaled_model_input.shape[2] ,dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: UpperCAmelCase , UpperCAmelCase = noise_pred.chunk(2 ) UpperCAmelCase = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) UpperCAmelCase = self.scheduler.step( A ,timestep=A ,sample=A ,).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=A ) UpperCAmelCase = [] for i, latent in enumerate(A ): print() UpperCAmelCase = self.renderer.decode( latent[None, :] ,A ,size=A ,ray_batch_size=4_096 ,n_coarse_samples=64 ,n_fine_samples=128 ,) images.append(A ) UpperCAmelCase = torch.stack(A ) if output_type not in ["np", "pil"]: raise ValueError(F'''Only the output types `pil` and `np` are supported not output_type={output_type}''' ) UpperCAmelCase = images.cpu().numpy() if output_type == "pil": UpperCAmelCase = [self.numpy_to_pil(A ) for image in images] # Offload last model to CPU if hasattr(self ,"""final_offload_hook""" ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=A )
74
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = 30 UpperCAmelCase = self.seq_length + self.mem_len UpperCAmelCase = 15 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 99 UpperCAmelCase = [10, 50, 80] UpperCAmelCase = 32 UpperCAmelCase = 32 UpperCAmelCase = 4 UpperCAmelCase = 8 UpperCAmelCase = 128 UpperCAmelCase = 2 UpperCAmelCase = 2 UpperCAmelCase = None UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 3 UpperCAmelCase = self.vocab_size - 1 UpperCAmelCase = 0.01 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCamelCase ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLLMHeadModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase , UpperCAmelCase = model([input_ids_a, mems_a] ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLForSequenceClassification(A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids_a} return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = () if is_tf_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,d_embed=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A ) def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: UpperCAmelCase = model_class(A ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: UpperCAmelCase = model.get_output_embeddings() assert isinstance(A ,tf.keras.layers.Layer ) UpperCAmelCase = model.get_bias() assert name is None else: UpperCAmelCase = model.get_output_embeddings() assert x is None UpperCAmelCase = model.get_bias() assert name is None def _UpperCamelCase ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _UpperCamelCase ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFTransfoXLModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""This model doesn't play well with fit() due to not returning a single loss.""" ) def _UpperCamelCase ( self ): pass @require_tf class lowerCamelCase__ ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off UpperCAmelCase = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off UpperCAmelCase = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> UpperCAmelCase = model.generate(A ,max_length=200 ,do_sample=A ) self.assertListEqual(output_ids[0].numpy().tolist() ,A )
74
1
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 0 while num > 0: UpperCAmelCase = num % 8 UpperCAmelCase = octal + (remainder * math.floor(math.pow(10 , _snake_case ) )) counter += 1 UpperCAmelCase = math.floor(num / 8 ) # basically /= 8 without remainder if any # This formatting removes trailing '.0' from `octal`. return F'''0o{int(_snake_case )}''' def _a ( ): """simple docstring""" print("""\n2 in octal is:""" ) print(decimal_to_octal(2 ) ) # = 2 print("""\n8 in octal is:""" ) print(decimal_to_octal(8 ) ) # = 10 print("""\n65 in octal is:""" ) print(decimal_to_octal(65 ) ) # = 101 print("""\n216 in octal is:""" ) print(decimal_to_octal(216 ) ) # = 330 print("""\n512 in octal is:""" ) print(decimal_to_octal(512 ) ) # = 1000 print("""\n""" ) if __name__ == "__main__": main()
74
"""simple docstring""" from math import sqrt def _a ( _snake_case = 100_0000 ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_snake_case , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
74
1
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = nn.functional.normalize(_snake_case ) UpperCAmelCase = nn.functional.normalize(_snake_case ) return torch.mm(_snake_case , normalized_text_embeds.t() ) class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = CLIPConfig SCREAMING_SNAKE_CASE = ['''CLIPEncoderLayer'''] def __init__( self ,A ): super().__init__(A ) UpperCAmelCase = CLIPVisionModel(config.vision_config ) UpperCAmelCase = nn.Linear(config.vision_config.hidden_size ,config.projection_dim ,bias=A ) UpperCAmelCase = nn.Parameter(torch.ones(17 ,config.projection_dim ) ,requires_grad=A ) UpperCAmelCase = nn.Parameter(torch.ones(3 ,config.projection_dim ) ,requires_grad=A ) UpperCAmelCase = nn.Parameter(torch.ones(17 ) ,requires_grad=A ) UpperCAmelCase = nn.Parameter(torch.ones(3 ) ,requires_grad=A ) @torch.no_grad() def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = self.vision_model(A )[1] # pooled_output UpperCAmelCase = self.visual_projection(A ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 UpperCAmelCase = cosine_distance(A ,self.special_care_embeds ).cpu().float().numpy() UpperCAmelCase = cosine_distance(A ,self.concept_embeds ).cpu().float().numpy() UpperCAmelCase = [] UpperCAmelCase = image_embeds.shape[0] for i in range(A ): UpperCAmelCase = {"""special_scores""": {}, """special_care""": [], """concept_scores""": {}, """bad_concepts""": []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images UpperCAmelCase = 0.0 for concept_idx in range(len(special_cos_dist[0] ) ): UpperCAmelCase = special_cos_dist[i][concept_idx] UpperCAmelCase = self.special_care_embeds_weights[concept_idx].item() UpperCAmelCase = round(concept_cos - concept_threshold + adjustment ,3 ) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img["""special_scores"""][concept_idx]} ) UpperCAmelCase = 0.01 for concept_idx in range(len(cos_dist[0] ) ): UpperCAmelCase = cos_dist[i][concept_idx] UpperCAmelCase = self.concept_embeds_weights[concept_idx].item() UpperCAmelCase = round(concept_cos - concept_threshold + adjustment ,3 ) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(A ) result.append(A ) UpperCAmelCase = [len(res["""bad_concepts"""] ) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = self.vision_model(A )[1] # pooled_output UpperCAmelCase = self.visual_projection(A ) UpperCAmelCase = cosine_distance(A ,self.special_care_embeds ) UpperCAmelCase = cosine_distance(A ,self.concept_embeds ) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images UpperCAmelCase = 0.0 UpperCAmelCase = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) UpperCAmelCase = torch.any(special_scores > 0 ,dim=1 ) UpperCAmelCase = special_care * 0.01 UpperCAmelCase = special_adjustment.unsqueeze(1 ).expand(-1 ,cos_dist.shape[1] ) UpperCAmelCase = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) UpperCAmelCase = torch.any(concept_scores > 0 ,dim=1 ) return images, has_nsfw_concepts
74
"""simple docstring""" import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _UpperCamelCase = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } _UpperCamelCase = AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def _a ( _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = create_model( """HTSAT-tiny""" , """roberta""" , _snake_case , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=_snake_case , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = {} UpperCAmelCase = R""".*sequential.(\d+).*""" UpperCAmelCase = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase = key.replace(_snake_case , _snake_case ) if re.match(_snake_case , _snake_case ): # replace sequential layers with list UpperCAmelCase = re.match(_snake_case , _snake_case ).group(1 ) UpperCAmelCase = key.replace(F'''sequential.{sequential_layer}.''' , F'''layers.{int(_snake_case )//3}.linear.''' ) elif re.match(_snake_case , _snake_case ): UpperCAmelCase = int(re.match(_snake_case , _snake_case ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... UpperCAmelCase = 1 if projecton_layer == 0 else 2 UpperCAmelCase = key.replace(F'''_projection.{projecton_layer}.''' , F'''_projection.linear{transformers_projection_layer}.''' ) if "audio" and "qkv" in key: # split qkv into query key and value UpperCAmelCase = value UpperCAmelCase = mixed_qkv.size(0 ) // 3 UpperCAmelCase = mixed_qkv[:qkv_dim] UpperCAmelCase = mixed_qkv[qkv_dim : qkv_dim * 2] UpperCAmelCase = mixed_qkv[qkv_dim * 2 :] UpperCAmelCase = query_layer UpperCAmelCase = key_layer UpperCAmelCase = value_layer else: UpperCAmelCase = value return model_state_dict def _a ( _snake_case , _snake_case , _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = init_clap(_snake_case , enable_fusion=_snake_case ) clap_model.eval() UpperCAmelCase = clap_model.state_dict() UpperCAmelCase = rename_state_dict(_snake_case ) UpperCAmelCase = ClapConfig() UpperCAmelCase = enable_fusion UpperCAmelCase = ClapModel(_snake_case ) # ignore the spectrogram embedding layer model.load_state_dict(_snake_case , strict=_snake_case ) model.save_pretrained(_snake_case ) transformers_config.save_pretrained(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") _UpperCamelCase = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
74
1
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if isinstance(_snake_case , _snake_case ): raise TypeError("""'float' object cannot be interpreted as an integer""" ) if isinstance(_snake_case , _snake_case ): raise TypeError("""'str' object cannot be interpreted as an integer""" ) if num == 0: return "0b0" UpperCAmelCase = False if num < 0: UpperCAmelCase = True UpperCAmelCase = -num UpperCAmelCase = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(_snake_case ) for e in binary ) return "0b" + "".join(str(_snake_case ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _UpperCamelCase = abspath(join(dirname(dirname(__file__)), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_snake_case ) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_terminal_summary_main UpperCAmelCase = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(_snake_case , id=_snake_case )
74
1
"""simple docstring""" from scipy.stats import spearmanr import datasets _UpperCamelCase = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ _UpperCamelCase = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ _UpperCamelCase = R"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase__ ( datasets.Metric ): def _UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) ,reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] ,) def _UpperCamelCase ( self ,A ,A ,A=False ): UpperCAmelCase = spearmanr(A ,A ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
74
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class lowerCamelCase__ : def __init__( self ,A ,A ): if len(A ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) UpperCAmelCase = list(A ) UpperCAmelCase = degree def __add__( self ,A ): if self.degree > polynomial_a.degree: UpperCAmelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,A ) else: UpperCAmelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,A ) def __sub__( self ,A ): return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self ): return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self ,A ): UpperCAmelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): UpperCAmelCase = """""" for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(A ) return polynomial def __repr__( self ): return self.__str__() def _UpperCamelCase ( self ): UpperCAmelCase = [0] * self.degree for i in range(self.degree ): UpperCAmelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,A ) def _UpperCamelCase ( self ,A = 0 ): UpperCAmelCase = [0] * (self.degree + 2) UpperCAmelCase = constant for i in range(self.degree + 1 ): UpperCAmelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,A ) def __eq__( self ,A ): if not isinstance(A ,A ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self ,A ): return not self.__eq__(A )
74
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class lowerCamelCase__ ( unittest.TestCase ): def __init__( self ,A ,A=7 ,A=3 ,A=30 ,A=400 ,A=True ,A=None ,A=0.9 ,A=None ,A=True ,A=[0.5, 0.5, 0.5] ,A=[0.5, 0.5, 0.5] ,): UpperCAmelCase = size if size is not None else {"""shortest_edge""": 30} UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 30, """width""": 30} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize_and_center_crop UpperCAmelCase = size UpperCAmelCase = crop_pct UpperCAmelCase = crop_size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std def _UpperCamelCase ( self ): return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class lowerCamelCase__ ( snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = PoolFormerImageProcessor if is_vision_available() else None def _UpperCamelCase ( self ): UpperCAmelCase = PoolFormerImageProcessingTester(self ) @property def _UpperCamelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self ): UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize_and_center_crop""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""crop_pct""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""shortest_edge""": 30} ) self.assertEqual(image_processor.crop_size ,{"""height""": 30, """width""": 30} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ,crop_size=84 ) self.assertEqual(image_processor.size ,{"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size ,{"""height""": 84, """width""": 84} ) def _UpperCamelCase ( self ): pass def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,)
74
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy as np import tensorflow as tf from transformers import ( TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST, FlaubertConfig, TFFlaubertForMultipleChoice, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForSequenceClassification, TFFlaubertForTokenClassification, TFFlaubertModel, TFFlaubertWithLMHeadModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = 2 UpperCAmelCase = 99 UpperCAmelCase = 0 UpperCAmelCase = 32 UpperCAmelCase = 2 UpperCAmelCase = 4 UpperCAmelCase = 0.1 UpperCAmelCase = 0.1 UpperCAmelCase = 512 UpperCAmelCase = 16 UpperCAmelCase = 2 UpperCAmelCase = 0.02 UpperCAmelCase = 3 UpperCAmelCase = 4 UpperCAmelCase = """last""" UpperCAmelCase = True UpperCAmelCase = None UpperCAmelCase = 0 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ,dtype=tf.floataa ) UpperCAmelCase = None if self.use_input_lengths: UpperCAmelCase = ( ids_tensor([self.batch_size] ,vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.n_langs ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] ,2 ,dtype=tf.floataa ) UpperCAmelCase = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase = FlaubertConfig( vocab_size=self.vocab_size ,n_special=self.n_special ,emb_dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,gelu_activation=self.gelu_activation ,sinusoidal_embeddings=self.sinusoidal_embeddings ,asm=self.asm ,causal=self.causal ,n_langs=self.n_langs ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,summary_type=self.summary_type ,use_proj=self.use_proj ,bos_token_id=self.bos_token_id ,) return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertModel(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) UpperCAmelCase = [input_ids, input_mask] UpperCAmelCase = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertWithLMHeadModel(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForQuestionAnsweringSimple(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForSequenceClassification(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_labels UpperCAmelCase = TFFlaubertForTokenClassification(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_choices UpperCAmelCase = TFFlaubertForMultipleChoice(config=A ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """langs""": token_type_ids, """lengths""": input_lengths, } return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( ( TFFlaubertModel, TFFlaubertWithLMHeadModel, TFFlaubertForSequenceClassification, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForTokenClassification, TFFlaubertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = ( (TFFlaubertWithLMHeadModel,) if is_tf_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFFlaubertModel, '''fill-mask''': TFFlaubertWithLMHeadModel, '''question-answering''': TFFlaubertForQuestionAnsweringSimple, '''text-classification''': TFFlaubertForSequenceClassification, '''token-classification''': TFFlaubertForTokenClassification, '''zero-shot''': TFFlaubertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,emb_dim=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_token_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_multiple_choice(*A ) @slow def _UpperCamelCase ( self ): for model_name in TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFFlaubertModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_tf @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( unittest.TestCase ): @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModel.from_pretrained("""jplu/tf-flaubert-small-cased""" ) UpperCAmelCase = tf.convert_to_tensor( [[0, 158, 735, 2_592, 1_424, 6_727, 82, 1]] ,dtype=tf.intaa ,) # "J'aime flaubert !" UpperCAmelCase = model(A )[0] UpperCAmelCase = tf.TensorShape((1, 8, 512) ) self.assertEqual(output.shape ,A ) # compare the actual values for a slice. UpperCAmelCase = tf.convert_to_tensor( [ [ [-1.8768773, -1.566555, 0.27072418], [-1.6920038, -0.5873505, 1.9329599], [-2.9563985, -1.6993835, 1.7972052], ] ] ,dtype=tf.floataa ,) self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1e-4 ) )
74
1
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _UpperCamelCase = { """vocab_file""": { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json""" ), }, } _UpperCamelCase = { """vocab_file""": { """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json""" ), }, } _UpperCamelCase = { """vocab_file""": { """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json""" ), }, } _UpperCamelCase = { """facebook/dpr-ctx_encoder-single-nq-base""": 512, """facebook/dpr-ctx_encoder-multiset-base""": 512, } _UpperCamelCase = { """facebook/dpr-question_encoder-single-nq-base""": 512, """facebook/dpr-question_encoder-multiset-base""": 512, } _UpperCamelCase = { """facebook/dpr-reader-single-nq-base""": 512, """facebook/dpr-reader-multiset-base""": 512, } _UpperCamelCase = { """facebook/dpr-ctx_encoder-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-ctx_encoder-multiset-base""": {"""do_lower_case""": True}, } _UpperCamelCase = { """facebook/dpr-question_encoder-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-question_encoder-multiset-base""": {"""do_lower_case""": True}, } _UpperCamelCase = { """facebook/dpr-reader-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-reader-multiset-base""": {"""do_lower_case""": True}, } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE = DPRContextEncoderTokenizer class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE = DPRQuestionEncoderTokenizer _UpperCamelCase = collections.namedtuple( """DPRSpanPrediction""", ["""span_score""", """relevance_score""", """doc_id""", """start_index""", """end_index""", """text"""] ) _UpperCamelCase = collections.namedtuple("""DPRReaderOutput""", ["""start_logits""", """end_logits""", """relevance_logits"""]) _UpperCamelCase = R""" Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`. It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers), using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)` with the format: [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids> Args: questions (`str` or `List[str]`): The questions to be encoded. You can specify one question for many passages. In this case, the question will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in `titles` or `texts`. titles (`str` or `List[str]`): The passages titles to be encoded. This can be a string or a list of strings if there are several passages. texts (`str` or `List[str]`): The passages texts to be encoded. This can be a string or a list of strings if there are several passages. padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`): Activates and controls padding. Accepts the following values: - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`): Activates and controls truncation. Accepts the following values: - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will truncate token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch of pairs) is provided. - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the first sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the second sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths greater than the model maximum admissible input size). max_length (`int`, *optional*): Controls the maximum length to use by one of the truncation/padding parameters. If left unset or set to `None`, this will use the predefined model maximum length if a maximum length is required by one of the truncation/padding parameters. If the model has no specific maximum input length (like XLNet) truncation/padding to a maximum length will be deactivated. return_tensors (`str` or [`~utils.TensorType`], *optional*): If set, will return tensors instead of list of python integers. Acceptable values are: - `'tf'`: Return TensorFlow `tf.constant` objects. - `'pt'`: Return PyTorch `torch.Tensor` objects. - `'np'`: Return Numpy `np.ndarray` objects. return_attention_mask (`bool`, *optional*): Whether or not to return the attention mask. If not set, will return the attention mask according to the specific tokenizer's default, defined by the `return_outputs` attribute. [What are attention masks?](../glossary#attention-mask) Return: `Dict[str, List[List[int]]]`: A dictionary with the following keys: - `input_ids`: List of token ids to be fed to a model. - `attention_mask`: List of indices specifying which tokens should be attended to by the model. """ @add_start_docstrings(snake_case ) class lowerCamelCase__ : def __call__( self ,A ,A = None ,A = None ,A = False ,A = False ,A = None ,A = None ,A = None ,**A ,): if titles is None and texts is None: return super().__call__( A ,padding=A ,truncation=A ,max_length=A ,return_tensors=A ,return_attention_mask=A ,**A ,) elif titles is None or texts is None: UpperCAmelCase = titles if texts is None else texts return super().__call__( A ,A ,padding=A ,truncation=A ,max_length=A ,return_tensors=A ,return_attention_mask=A ,**A ,) UpperCAmelCase = titles if not isinstance(A ,A ) else [titles] UpperCAmelCase = texts if not isinstance(A ,A ) else [texts] UpperCAmelCase = len(A ) UpperCAmelCase = questions if not isinstance(A ,A ) else [questions] * n_passages assert len(A ) == len( A ), F'''There should be as many titles than texts but got {len(A )} titles and {len(A )} texts.''' UpperCAmelCase = super().__call__(A ,A ,padding=A ,truncation=A )["""input_ids"""] UpperCAmelCase = super().__call__(A ,add_special_tokens=A ,padding=A ,truncation=A )["""input_ids"""] UpperCAmelCase = { """input_ids""": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(A ,A ) ] } if return_attention_mask is not False: UpperCAmelCase = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) UpperCAmelCase = attention_mask return self.pad(A ,padding=A ,max_length=A ,return_tensors=A ) def _UpperCamelCase ( self ,A ,A ,A = 16 ,A = 64 ,A = 4 ,): UpperCAmelCase = reader_input["""input_ids"""] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = reader_output[:3] UpperCAmelCase = len(A ) UpperCAmelCase = sorted(range(A ) ,reverse=A ,key=relevance_logits.__getitem__ ) UpperCAmelCase = [] for doc_id in sorted_docs: UpperCAmelCase = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence UpperCAmelCase = sequence_ids.index(self.sep_token_id ,2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: UpperCAmelCase = sequence_ids.index(self.pad_token_id ) else: UpperCAmelCase = len(A ) UpperCAmelCase = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] ,end_logits=end_logits[doc_id][passage_offset:sequence_len] ,max_answer_length=A ,top_spans=A ,) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] ,relevance_score=relevance_logits[doc_id] ,doc_id=A ,start_index=A ,end_index=A ,text=self.decode(sequence_ids[start_index : end_index + 1] ) ,) ) if len(A ) >= num_spans: break return nbest_spans_predictions[:num_spans] def _UpperCamelCase ( self ,A ,A ,A ,A ,): UpperCAmelCase = [] for start_index, start_score in enumerate(A ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) UpperCAmelCase = sorted(A ,key=lambda A : x[1] ,reverse=A ) UpperCAmelCase = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' UpperCAmelCase = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(A ) == top_spans: break return chosen_span_intervals @add_end_docstrings(snake_case ) class lowerCamelCase__ ( snake_case , snake_case ): SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = READER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = READER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE = ['''input_ids''', '''attention_mask'''] SCREAMING_SNAKE_CASE = DPRReaderTokenizer
74
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): UpperCAmelCase = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): UpperCAmelCase = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] UpperCAmelCase = key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(_snake_case )-1}''' ) if "norm" in key: UpperCAmelCase = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] UpperCAmelCase = key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(_snake_case )-1}''' ) if "layer_norm1" in key: UpperCAmelCase = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: UpperCAmelCase = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase = key[key.find("""block""" ) + len("""block""" )] UpperCAmelCase = key.replace(F'''block{idx}''' , F'''block.{int(_snake_case )-1}''' ) if "attn.q" in key: UpperCAmelCase = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: UpperCAmelCase = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: UpperCAmelCase = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: UpperCAmelCase = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: UpperCAmelCase = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: UpperCAmelCase = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: UpperCAmelCase = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) UpperCAmelCase = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase = key[key.find("""linear_c""" ) + len("""linear_c""" )] UpperCAmelCase = key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(_snake_case )-1}''' ) if "bot_conv" in key: UpperCAmelCase = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: UpperCAmelCase = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: UpperCAmelCase = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: UpperCAmelCase = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: UpperCAmelCase = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: UpperCAmelCase = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: UpperCAmelCase = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): UpperCAmelCase = key.replace("""module.last_layer_depth""" , """head.head""" ) UpperCAmelCase = value return new_state_dict def _a ( _snake_case , _snake_case ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict UpperCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase = kv_bias[config.hidden_sizes[i] :] def _a ( ): """simple docstring""" UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case=False , _snake_case=None ): """simple docstring""" UpperCAmelCase = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) UpperCAmelCase = GLPNImageProcessor() # prepare image UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=_snake_case , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict UpperCAmelCase = torch.load(_snake_case , map_location=torch.device("""cpu""" ) ) # rename keys UpperCAmelCase = rename_keys(_snake_case ) # key and value matrices need special treatment read_in_k_v(_snake_case , _snake_case ) # create HuggingFace model and load state dict UpperCAmelCase = GLPNForDepthEstimation(_snake_case ) model.load_state_dict(_snake_case ) model.eval() # forward pass UpperCAmelCase = model(_snake_case ) UpperCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: UpperCAmelCase = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: UpperCAmelCase = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) UpperCAmelCase = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _snake_case , atol=1E-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=_snake_case , ) image_processor.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=_snake_case , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCamelCase = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
74
1
"""simple docstring""" import os from collections import deque import torch from torch.utils.data import Dataset class lowerCamelCase__ ( snake_case ): def __init__( self ,A="" ,A="train" ): assert os.path.isdir(A ) UpperCAmelCase = [] UpperCAmelCase = os.listdir(A ) for story_filename in story_filenames_list: if "summary" in story_filename: continue UpperCAmelCase = os.path.join(A ,A ) if not os.path.isfile(A ): continue self.documents.append(A ) def __len__( self ): return len(self.documents ) def __getitem__( self ,A ): UpperCAmelCase = self.documents[idx] UpperCAmelCase = document_path.split("""/""" )[-1] with open(A ,encoding="""utf-8""" ) as source: UpperCAmelCase = source.read() UpperCAmelCase , UpperCAmelCase = process_story(A ) return document_name, story_lines, summary_lines def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = list(filter(lambda _snake_case : len(_snake_case ) != 0 , [line.strip() for line in raw_story.split("""\n""" )] ) ) # for some unknown reason some lines miss a period, add it UpperCAmelCase = [_add_missing_period(_snake_case ) for line in nonempty_lines] # gather article lines UpperCAmelCase = [] UpperCAmelCase = deque(_snake_case ) while True: try: UpperCAmelCase = lines.popleft() if element.startswith("""@highlight""" ): break story_lines.append(_snake_case ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines UpperCAmelCase = list(filter(lambda _snake_case : not t.startswith("""@highlight""" ) , _snake_case ) ) return story_lines, summary_lines def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = [""".""", """!""", """?""", """...""", """'""", """`""", """\"""", """\u2019""", """\u2019""", """)"""] if line.startswith("""@highlight""" ): return line if line[-1] in END_TOKENS: return line return line + "." def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" if len(_snake_case ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(_snake_case )) ) return sequence def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = torch.ones_like(_snake_case ) UpperCAmelCase = sequence == pad_token_id UpperCAmelCase = 0 return mask def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = [tokenizer.encode(_snake_case ) for line in story_lines] UpperCAmelCase = [token for sentence in story_lines_token_ids for token in sentence] UpperCAmelCase = [tokenizer.encode(_snake_case ) for line in summary_lines] UpperCAmelCase = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = [] for sequence in batch: UpperCAmelCase = -1 UpperCAmelCase = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(_snake_case ) return torch.tensor(_snake_case )
74
"""simple docstring""" def _a ( _snake_case ): # noqa: E741 """simple docstring""" UpperCAmelCase = len(_snake_case ) UpperCAmelCase = 0 UpperCAmelCase = [0] * n UpperCAmelCase = [False] * n UpperCAmelCase = [False] * n def dfs(_snake_case , _snake_case , _snake_case , _snake_case ): if parent == root: out_edge_count += 1 UpperCAmelCase = True UpperCAmelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase = dfs(_snake_case , _snake_case , _snake_case , _snake_case ) UpperCAmelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase = True # AP found via cycle if at == low[to]: UpperCAmelCase = True else: UpperCAmelCase = min(low[at] , _snake_case ) return out_edge_count for i in range(_snake_case ): if not visited[i]: UpperCAmelCase = 0 UpperCAmelCase = dfs(_snake_case , _snake_case , -1 , _snake_case ) UpperCAmelCase = out_edge_count > 1 for x in range(len(_snake_case ) ): if is_art[x] is True: print(_snake_case ) # Adjacency list of graph _UpperCamelCase = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
74
1
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """Helsinki-NLP/opus-mt-en-de""": """https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json""", # See all Marian models at https://huggingface.co/models?filter=marian } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''marian''' SCREAMING_SNAKE_CASE = ['''past_key_values'''] SCREAMING_SNAKE_CASE = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self ,A=58_101 ,A=None ,A=1_024 ,A=12 ,A=4_096 ,A=16 ,A=12 ,A=4_096 ,A=16 ,A=0.0 ,A=0.0 ,A=True ,A=True ,A="gelu" ,A=1_024 ,A=0.1 ,A=0.0 ,A=0.0 ,A=0.02 ,A=58_100 ,A=False ,A=58_100 ,A=0 ,A=0 ,A=True ,**A ,): UpperCAmelCase = vocab_size UpperCAmelCase = decoder_vocab_size or vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = share_encoder_decoder_embeddings super().__init__( pad_token_id=A ,eos_token_id=A ,is_encoder_decoder=A ,decoder_start_token_id=A ,forced_eos_token_id=A ,**A ,) class lowerCamelCase__ ( snake_case ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _UpperCamelCase ( self ): if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase = {0: """batch"""} UpperCAmelCase = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: UpperCAmelCase = {0: """batch""", 1: """decoder_sequence"""} UpperCAmelCase = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(A ,direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase , UpperCAmelCase = self.num_layers for i in range(A ): UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} else: UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _UpperCamelCase ( self ): if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = super().outputs else: UpperCAmelCase = super(A ,self ).outputs if self.use_past: UpperCAmelCase , UpperCAmelCase = self.num_layers for i in range(A ): UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def _UpperCamelCase ( self ,A ,A = -1 ,A = -1 ,A = False ,A = None ,): UpperCAmelCase = self._generate_dummy_inputs_for_encoder_and_decoder( A ,A ,A ,A ,A ) # Generate decoder inputs UpperCAmelCase = seq_length if not self.use_past else 1 UpperCAmelCase = self._generate_dummy_inputs_for_encoder_and_decoder( A ,A ,A ,A ,A ) UpperCAmelCase = {F'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} UpperCAmelCase = dict(**A ,**A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch UpperCAmelCase , UpperCAmelCase = common_inputs["""input_ids"""].shape UpperCAmelCase = common_inputs["""decoder_input_ids"""].shape[1] UpperCAmelCase , UpperCAmelCase = self.num_attention_heads UpperCAmelCase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase = decoder_seq_length + 3 UpperCAmelCase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCAmelCase = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(A ,A )] ,dim=1 ) UpperCAmelCase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCAmelCase , UpperCAmelCase = self.num_layers UpperCAmelCase = min(A ,A ) UpperCAmelCase = max(A ,A ) - min_num_layers UpperCAmelCase = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(A ): common_inputs["past_key_values"].append( ( torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), ) ) # TODO: test this. UpperCAmelCase = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(A ,A ): common_inputs["past_key_values"].append((torch.zeros(A ), torch.zeros(A )) ) return common_inputs def _UpperCamelCase ( self ,A ,A = -1 ,A = -1 ,A = False ,A = None ,): UpperCAmelCase = self._generate_dummy_inputs_for_encoder_and_decoder( A ,A ,A ,A ,A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch UpperCAmelCase , UpperCAmelCase = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values UpperCAmelCase = seqlen + 2 UpperCAmelCase , UpperCAmelCase = self.num_layers UpperCAmelCase , UpperCAmelCase = self.num_attention_heads UpperCAmelCase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase = common_inputs["""attention_mask"""].dtype UpperCAmelCase = torch.cat( [common_inputs["""attention_mask"""], torch.ones(A ,A ,dtype=A )] ,dim=1 ) UpperCAmelCase = [ (torch.zeros(A ), torch.zeros(A )) for _ in range(A ) ] return common_inputs def _UpperCamelCase ( self ,A ,A = -1 ,A = -1 ,A = False ,A = None ,): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase = compute_effective_axis_dimension( A ,fixed_dimension=OnnxConfig.default_fixed_batch ,num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase = tokenizer.num_special_tokens_to_add(A ) UpperCAmelCase = compute_effective_axis_dimension( A ,fixed_dimension=OnnxConfig.default_fixed_sequence ,num_token_to_add=A ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCAmelCase = dict(tokenizer(A ,return_tensors=A ) ) return common_inputs def _UpperCamelCase ( self ,A ,A = -1 ,A = -1 ,A = False ,A = None ,): if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) else: UpperCAmelCase = self._generate_dummy_inputs_for_causal_lm( A ,batch_size=A ,seq_length=A ,is_pair=A ,framework=A ) return common_inputs def _UpperCamelCase ( self ,A ,A ,A ,A ): if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = super()._flatten_past_key_values_(A ,A ,A ,A ) else: UpperCAmelCase = super(A ,self )._flatten_past_key_values_( A ,A ,A ,A ) @property def _UpperCamelCase ( self ): return 1e-4
74
"""simple docstring""" _UpperCamelCase = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ _UpperCamelCase = [{"""type""": """code""", """content""": INSTALL_CONTENT}] _UpperCamelCase = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
74
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = """▁""" _UpperCamelCase = {"""vocab_file""": """sentencepiece.bpe.model"""} _UpperCamelCase = { """vocab_file""": { """facebook/xglm-564M""": """https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model""", } } _UpperCamelCase = { """facebook/xglm-564M""": 2048, } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = ['''input_ids''', '''attention_mask'''] def __init__( self ,A ,A="<s>" ,A="</s>" ,A="</s>" ,A="<s>" ,A="<unk>" ,A="<pad>" ,A = None ,**A ,): UpperCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer UpperCAmelCase = 7 UpperCAmelCase = [F'''<madeupword{i}>''' for i in range(self.num_madeup_words )] UpperCAmelCase = kwargs.get("""additional_special_tokens""" ,[] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=A ,eos_token=A ,unk_token=A ,sep_token=A ,cls_token=A ,pad_token=A ,sp_model_kwargs=self.sp_model_kwargs ,**A ,) UpperCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(A ) ) UpperCAmelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase = 1 # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} UpperCAmelCase = len(self.sp_model ) UpperCAmelCase = {F'''<madeupword{i}>''': sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(A ) UpperCAmelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): UpperCAmelCase = self.__dict__.copy() UpperCAmelCase = None UpperCAmelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self ,A ): UpperCAmelCase = d # for backward compatibility if not hasattr(self ,"""sp_model_kwargs""" ): UpperCAmelCase = {} UpperCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _UpperCamelCase ( self ,A ,A = None ): if token_ids_a is None: return [self.sep_token_id] + token_ids_a UpperCAmelCase = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _UpperCamelCase ( self ,A ,A = None ,A = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A ,token_ids_a=A ,already_has_special_tokens=A ) if token_ids_a is None: return [1] + ([0] * len(A )) return [1] + ([0] * len(A )) + [1, 1] + ([0] * len(A )) def _UpperCamelCase ( self ,A ,A = None ): UpperCAmelCase = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def _UpperCamelCase ( self ): return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def _UpperCamelCase ( self ): UpperCAmelCase = {self.convert_ids_to_tokens(A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _UpperCamelCase ( self ,A ): return self.sp_model.encode(A ,out_type=A ) def _UpperCamelCase ( self ,A ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase = self.sp_model.PieceToId(A ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _UpperCamelCase ( self ,A ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = """""".join(A ).replace(A ,""" """ ).strip() return out_string def _UpperCamelCase ( self ,A ,A = None ): if not os.path.isdir(A ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase = os.path.join( A ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,A ) elif not os.path.isfile(self.vocab_file ): with open(A ,"""wb""" ) as fi: UpperCAmelCase = self.sp_model.serialized_model_proto() fi.write(A ) return (out_vocab_file,)
74
"""simple docstring""" import argparse import struct import unittest class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = data # Initialize hash values UpperCAmelCase = [ 0x6A_09_E6_67, 0xBB_67_AE_85, 0x3C_6E_F3_72, 0xA5_4F_F5_3A, 0x51_0E_52_7F, 0x9B_05_68_8C, 0x1F_83_D9_AB, 0x5B_E0_CD_19, ] # Initialize round constants UpperCAmelCase = [ 0x42_8A_2F_98, 0x71_37_44_91, 0xB5_C0_FB_CF, 0xE9_B5_DB_A5, 0x39_56_C2_5B, 0x59_F1_11_F1, 0x92_3F_82_A4, 0xAB_1C_5E_D5, 0xD8_07_AA_98, 0x12_83_5B_01, 0x24_31_85_BE, 0x55_0C_7D_C3, 0x72_BE_5D_74, 0x80_DE_B1_FE, 0x9B_DC_06_A7, 0xC1_9B_F1_74, 0xE4_9B_69_C1, 0xEF_BE_47_86, 0x0F_C1_9D_C6, 0x24_0C_A1_CC, 0x2D_E9_2C_6F, 0x4A_74_84_AA, 0x5C_B0_A9_DC, 0x76_F9_88_DA, 0x98_3E_51_52, 0xA8_31_C6_6D, 0xB0_03_27_C8, 0xBF_59_7F_C7, 0xC6_E0_0B_F3, 0xD5_A7_91_47, 0x06_CA_63_51, 0x14_29_29_67, 0x27_B7_0A_85, 0x2E_1B_21_38, 0x4D_2C_6D_FC, 0x53_38_0D_13, 0x65_0A_73_54, 0x76_6A_0A_BB, 0x81_C2_C9_2E, 0x92_72_2C_85, 0xA2_BF_E8_A1, 0xA8_1A_66_4B, 0xC2_4B_8B_70, 0xC7_6C_51_A3, 0xD1_92_E8_19, 0xD6_99_06_24, 0xF4_0E_35_85, 0x10_6A_A0_70, 0x19_A4_C1_16, 0x1E_37_6C_08, 0x27_48_77_4C, 0x34_B0_BC_B5, 0x39_1C_0C_B3, 0x4E_D8_AA_4A, 0x5B_9C_CA_4F, 0x68_2E_6F_F3, 0x74_8F_82_EE, 0x78_A5_63_6F, 0x84_C8_78_14, 0x8C_C7_02_08, 0x90_BE_FF_FA, 0xA4_50_6C_EB, 0xBE_F9_A3_F7, 0xC6_71_78_F2, ] UpperCAmelCase = self.preprocessing(self.data ) self.final_hash() @staticmethod def _UpperCamelCase ( A ): UpperCAmelCase = b"""\x80""" + (b"""\x00""" * (63 - (len(A ) + 8) % 64)) UpperCAmelCase = struct.pack(""">Q""" ,(len(A ) * 8) ) return data + padding + big_endian_integer def _UpperCamelCase ( self ): # Convert into blocks of 64 bytes UpperCAmelCase = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers UpperCAmelCase = list(struct.unpack(""">16L""" ,A ) ) # add 48 0-ed integers words += [0] * 48 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array UpperCAmelCase = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) UpperCAmelCase = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) UpperCAmelCase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_00_00_00_00 # Compression UpperCAmelCase = self.ror(A ,6 ) ^ self.ror(A ,11 ) ^ self.ror(A ,25 ) UpperCAmelCase = (e & f) ^ ((~e & 0xFF_FF_FF_FF) & g) UpperCAmelCase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_00_00_00_00 UpperCAmelCase = self.ror(A ,2 ) ^ self.ror(A ,13 ) ^ self.ror(A ,22 ) UpperCAmelCase = (a & b) ^ (a & c) ^ (b & c) UpperCAmelCase = (sa + maj) % 0x1_00_00_00_00 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = ( g, f, e, ((d + tempa) % 0x1_00_00_00_00), c, b, a, ((tempa + tempa) % 0x1_00_00_00_00), ) UpperCAmelCase = [a, b, c, d, e, f, g, h] # Modify final values UpperCAmelCase = [ ((element + mutated_hash_values[index]) % 0x1_00_00_00_00) for index, element in enumerate(self.hashes ) ] UpperCAmelCase = """""".join([hex(A )[2:].zfill(8 ) for value in self.hashes] ) def _UpperCamelCase ( self ,A ,A ): return 0xFF_FF_FF_FF & (value << (32 - rotations)) | (value >> rotations) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): import hashlib UpperCAmelCase = bytes("""Test String""" ,"""utf-8""" ) self.assertEqual(SHAaaa(A ).hash ,hashlib.shaaaa(A ).hexdigest() ) def _a ( ): """simple docstring""" import doctest doctest.testmod() UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase = parser.parse_args() UpperCAmelCase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase = f.read() else: UpperCAmelCase = bytes(_snake_case , """utf-8""" ) print(SHAaaa(_snake_case ).hash ) if __name__ == "__main__": main()
74
1
"""simple docstring""" import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''train''' SCREAMING_SNAKE_CASE = '''dev''' SCREAMING_SNAKE_CASE = '''test''' class lowerCamelCase__ : @staticmethod def _UpperCamelCase ( A ,A ): raise NotImplementedError @staticmethod def _UpperCamelCase ( A ): raise NotImplementedError @staticmethod def _UpperCamelCase ( A ,A ,A ,A ,A=False ,A="[CLS]" ,A=1 ,A="[SEP]" ,A=False ,A=False ,A=0 ,A=0 ,A=-100 ,A=0 ,A=True ,): UpperCAmelCase = {label: i for i, label in enumerate(A )} UpperCAmelCase = [] for ex_index, example in enumerate(A ): if ex_index % 10_000 == 0: logger.info("""Writing example %d of %d""" ,A ,len(A ) ) UpperCAmelCase = [] UpperCAmelCase = [] for word, label in zip(example.words ,example.labels ): UpperCAmelCase = tokenizer.tokenize(A ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(A ) > 0: tokens.extend(A ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(A ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. UpperCAmelCase = tokenizer.num_special_tokens_to_add() if len(A ) > max_seq_length - special_tokens_count: UpperCAmelCase = tokens[: (max_seq_length - special_tokens_count)] UpperCAmelCase = label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] UpperCAmelCase = [sequence_a_segment_id] * len(A ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: UpperCAmelCase = [cls_token] + tokens UpperCAmelCase = [pad_token_label_id] + label_ids UpperCAmelCase = [cls_token_segment_id] + segment_ids UpperCAmelCase = tokenizer.convert_tokens_to_ids(A ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. UpperCAmelCase = [1 if mask_padding_with_zero else 0] * len(A ) # Zero-pad up to the sequence length. UpperCAmelCase = max_seq_length - len(A ) if pad_on_left: UpperCAmelCase = ([pad_token] * padding_length) + input_ids UpperCAmelCase = ([0 if mask_padding_with_zero else 1] * padding_length) + input_mask UpperCAmelCase = ([pad_token_segment_id] * padding_length) + segment_ids UpperCAmelCase = ([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(A ) == max_seq_length assert len(A ) == max_seq_length assert len(A ) == max_seq_length assert len(A ) == max_seq_length if ex_index < 5: logger.info("""*** Example ***""" ) logger.info("""guid: %s""" ,example.guid ) logger.info("""tokens: %s""" ,""" """.join([str(A ) for x in tokens] ) ) logger.info("""input_ids: %s""" ,""" """.join([str(A ) for x in input_ids] ) ) logger.info("""input_mask: %s""" ,""" """.join([str(A ) for x in input_mask] ) ) logger.info("""segment_ids: %s""" ,""" """.join([str(A ) for x in segment_ids] ) ) logger.info("""label_ids: %s""" ,""" """.join([str(A ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: UpperCAmelCase = None features.append( InputFeatures( input_ids=A ,attention_mask=A ,token_type_ids=A ,label_ids=A ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = nn.CrossEntropyLoss().ignore_index def __init__( self ,A ,A ,A ,A ,A ,A = None ,A=False ,A = Split.train ,): # Load data features from cache or dataset file UpperCAmelCase = os.path.join( A ,"""cached_{}_{}_{}""".format(mode.value ,tokenizer.__class__.__name__ ,str(A ) ) ,) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. UpperCAmelCase = cached_features_file + """.lock""" with FileLock(A ): if os.path.exists(A ) and not overwrite_cache: logger.info(F'''Loading features from cached file {cached_features_file}''' ) UpperCAmelCase = torch.load(A ) else: logger.info(F'''Creating features from dataset file at {data_dir}''' ) UpperCAmelCase = token_classification_task.read_examples_from_file(A ,A ) # TODO clean up all this to leverage built-in features of tokenizers UpperCAmelCase = token_classification_task.convert_examples_to_features( A ,A ,A ,A ,cls_token_at_end=bool(model_type in ["""xlnet"""] ) ,cls_token=tokenizer.cls_token ,cls_token_segment_id=2 if model_type in ["""xlnet"""] else 0 ,sep_token=tokenizer.sep_token ,sep_token_extra=A ,pad_on_left=bool(tokenizer.padding_side == """left""" ) ,pad_token=tokenizer.pad_token_id ,pad_token_segment_id=tokenizer.pad_token_type_id ,pad_token_label_id=self.pad_token_label_id ,) logger.info(F'''Saving features into cached file {cached_features_file}''' ) torch.save(self.features ,A ) def __len__( self ): return len(self.features ) def __getitem__( self ,A ): return self.features[i] if is_tf_available(): import tensorflow as tf class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = -100 def __init__( self ,A ,A ,A ,A ,A ,A = None ,A=False ,A = Split.train ,): UpperCAmelCase = token_classification_task.read_examples_from_file(A ,A ) # TODO clean up all this to leverage built-in features of tokenizers UpperCAmelCase = token_classification_task.convert_examples_to_features( A ,A ,A ,A ,cls_token_at_end=bool(model_type in ["""xlnet"""] ) ,cls_token=tokenizer.cls_token ,cls_token_segment_id=2 if model_type in ["""xlnet"""] else 0 ,sep_token=tokenizer.sep_token ,sep_token_extra=A ,pad_on_left=bool(tokenizer.padding_side == """left""" ) ,pad_token=tokenizer.pad_token_id ,pad_token_segment_id=tokenizer.pad_token_type_id ,pad_token_label_id=self.pad_token_label_id ,) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: UpperCAmelCase = tf.data.Dataset.from_generator( A ,({"""input_ids""": tf.intaa, """attention_mask""": tf.intaa}, tf.intaa) ,( {"""input_ids""": tf.TensorShape([None] ), """attention_mask""": tf.TensorShape([None] )}, tf.TensorShape([None] ), ) ,) else: UpperCAmelCase = tf.data.Dataset.from_generator( A ,({"""input_ids""": tf.intaa, """attention_mask""": tf.intaa, """token_type_ids""": tf.intaa}, tf.intaa) ,( { """input_ids""": tf.TensorShape([None] ), """attention_mask""": tf.TensorShape([None] ), """token_type_ids""": tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) ,) def _UpperCamelCase ( self ): UpperCAmelCase = self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self ): return len(self.features ) def __getitem__( self ,A ): return self.features[i]
74
"""simple docstring""" def _a ( _snake_case = 10 , _snake_case = 22 ): """simple docstring""" UpperCAmelCase = range(1 , _snake_case ) UpperCAmelCase = range(1 , _snake_case ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F"""{solution(10, 22) = }""")
74
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCamelCase = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" from __future__ import annotations def _a ( _snake_case ): """simple docstring""" return len(set(_snake_case ) ) == len(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( _snake_case = 0.1 ): """simple docstring""" UpperCAmelCase = 3 UpperCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("""transformers.models.speecht5""") def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" hf_model.apply_weight_norm() UpperCAmelCase = checkpoint["""input_conv.weight_g"""] UpperCAmelCase = checkpoint["""input_conv.weight_v"""] UpperCAmelCase = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , ): """simple docstring""" if config_path is not None: UpperCAmelCase = SpeechTaHifiGanConfig.from_pretrained(_snake_case ) else: UpperCAmelCase = SpeechTaHifiGanConfig() UpperCAmelCase = SpeechTaHifiGan(_snake_case ) UpperCAmelCase = torch.load(_snake_case ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _snake_case , _snake_case ) UpperCAmelCase = np.load(_snake_case ) UpperCAmelCase = stats[0].reshape(-1 ) UpperCAmelCase = stats[1].reshape(-1 ) UpperCAmelCase = torch.from_numpy(_snake_case ).float() UpperCAmelCase = torch.from_numpy(_snake_case ).float() model.save_pretrained(_snake_case ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
74
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE = '''CLIPImageProcessor''' SCREAMING_SNAKE_CASE = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self ,A=None ,A=None ,**A ): UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) def __call__( self ,A=None ,A=None ,A=None ,**A ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.batch_decode(*A ,**A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.decode(*A ,**A ) @property def _UpperCamelCase ( self ): UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,A ,) return self.image_processor
74
1
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''naver-clova-ix/donut-base-finetuned-docvqa''' SCREAMING_SNAKE_CASE = ( '''This is a tool that answers a question about an document (pdf). It takes an input named `document` which ''' '''should be the document containing the information, as well as a `question` that is the question about the ''' '''document. It returns a text that contains the answer to the question.''' ) SCREAMING_SNAKE_CASE = '''document_qa''' SCREAMING_SNAKE_CASE = AutoProcessor SCREAMING_SNAKE_CASE = VisionEncoderDecoderModel SCREAMING_SNAKE_CASE = ['''image''', '''text'''] SCREAMING_SNAKE_CASE = ['''text'''] def __init__( self ,*A ,**A ): if not is_vision_available(): raise ValueError("""Pillow must be installed to use the DocumentQuestionAnsweringTool.""" ) super().__init__(*A ,**A ) def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = """<s_docvqa><s_question>{user_input}</s_question><s_answer>""" UpperCAmelCase = task_prompt.replace("""{user_input}""" ,A ) UpperCAmelCase = self.pre_processor.tokenizer( A ,add_special_tokens=A ,return_tensors="""pt""" ).input_ids UpperCAmelCase = self.pre_processor(A ,return_tensors="""pt""" ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def _UpperCamelCase ( self ,A ): return self.model.generate( inputs["""pixel_values"""].to(self.device ) ,decoder_input_ids=inputs["""decoder_input_ids"""].to(self.device ) ,max_length=self.model.decoder.config.max_position_embeddings ,early_stopping=A ,pad_token_id=self.pre_processor.tokenizer.pad_token_id ,eos_token_id=self.pre_processor.tokenizer.eos_token_id ,use_cache=A ,num_beams=1 ,bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] ,return_dict_in_generate=A ,).sequences def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.pre_processor.batch_decode(A )[0] UpperCAmelCase = sequence.replace(self.pre_processor.tokenizer.eos_token ,"""""" ) UpperCAmelCase = sequence.replace(self.pre_processor.tokenizer.pad_token ,"""""" ) UpperCAmelCase = re.sub(r"""<.*?>""" ,"""""" ,A ,count=1 ).strip() # remove first task start token UpperCAmelCase = self.pre_processor.tokenajson(A ) return sequence["answer"]
74
"""simple docstring""" from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def _a ( _snake_case = "mumbai" ): """simple docstring""" UpperCAmelCase = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): UpperCAmelCase = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() UpperCAmelCase = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("""Bangalore"""), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
74
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class lowerCamelCase__ ( unittest.TestCase ): def __init__( self ,A ,A=7 ,A=3 ,A=18 ,A=30 ,A=400 ,A=True ,A=None ,A=True ,A=None ,A=True ,A=[0.48145466, 0.4578275, 0.40821073] ,A=[0.26862954, 0.26130258, 0.27577711] ,A=True ,): UpperCAmelCase = size if size is not None else {"""height""": 224, """width""": 224} UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = image_size UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_convert_rgb def _UpperCamelCase ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def _UpperCamelCase ( self ,A=False ,A=False ,A=False ): assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: UpperCAmelCase = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 255 ,size=(self.num_channels, self.max_resolution, self.max_resolution) ,dtype=np.uinta ) ) else: UpperCAmelCase = [] for i in range(self.batch_size ): UpperCAmelCase , UpperCAmelCase = np.random.choice(np.arange(self.min_resolution ,self.max_resolution ) ,2 ) image_inputs.append(np.random.randint(255 ,size=(self.num_channels, width, height) ,dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension UpperCAmelCase = [Image.fromarray(np.moveaxis(A ,0 ,-1 ) ) for x in image_inputs] if torchify: UpperCAmelCase = [torch.from_numpy(A ) for x in image_inputs] return image_inputs @require_torch @require_vision class lowerCamelCase__ ( snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ChineseCLIPImageProcessor if is_vision_available() else None def _UpperCamelCase ( self ): UpperCAmelCase = ChineseCLIPImageProcessingTester(self ,do_center_crop=A ) @property def _UpperCamelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self ): UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_center_crop""" ) ) self.assertTrue(hasattr(A ,"""center_crop""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) self.assertTrue(hasattr(A ,"""do_convert_rgb""" ) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 224, """width""": 224} ) self.assertEqual(image_processor.crop_size ,{"""height""": 18, """width""": 18} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ,crop_size=84 ) self.assertEqual(image_processor.size ,{"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size ,{"""height""": 84, """width""": 84} ) def _UpperCamelCase ( self ): pass def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=A ,numpify=A ) for image in image_inputs: self.assertIsInstance(A ,np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=A ,torchify=A ) for image in image_inputs: self.assertIsInstance(A ,torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) @require_torch @require_vision class lowerCamelCase__ ( snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ChineseCLIPImageProcessor if is_vision_available() else None def _UpperCamelCase ( self ): UpperCAmelCase = ChineseCLIPImageProcessingTester(self ,num_channels=4 ,do_center_crop=A ) UpperCAmelCase = 3 @property def _UpperCamelCase ( self ): return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self ): UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A ,"""do_resize""" ) ) self.assertTrue(hasattr(A ,"""size""" ) ) self.assertTrue(hasattr(A ,"""do_center_crop""" ) ) self.assertTrue(hasattr(A ,"""center_crop""" ) ) self.assertTrue(hasattr(A ,"""do_normalize""" ) ) self.assertTrue(hasattr(A ,"""image_mean""" ) ) self.assertTrue(hasattr(A ,"""image_std""" ) ) self.assertTrue(hasattr(A ,"""do_convert_rgb""" ) ) def _UpperCamelCase ( self ): pass def _UpperCamelCase ( self ): # Initialize image_processing UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=A ) for image in image_inputs: self.assertIsInstance(A ,Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,) # Test batched UpperCAmelCase = image_processing(A ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) ,)
74
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): UpperCAmelCase = ["""a""", """b""", """c"""] # Defaults to last layer if both are None UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,A ,A ) self.assertEqual(A ,["""c"""] ) self.assertEqual(A ,[2] ) # Out indices set to match out features UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(["""a""", """c"""] ,A ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features set to match out indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[0, 2] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features selected from negative indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[-3, -1] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[-3, -1] ) def _UpperCamelCase ( self ): # Stage names must be set with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,A ) # Out features must be a list with self.assertRaises(A ): verify_out_features_out_indices(("""a""", """b""") ,(0, 1) ,["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(A ): verify_out_features_out_indices(A ,0 ,["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(A ,(0, 1) ,["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0,) ,["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 2) ,["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(A ): verify_out_features_out_indices(["""b""", """a"""] ,(0, 1) ,["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] ,(0, 1, -1) ,["""a""", """b""", """c""", """d"""] ) def _UpperCamelCase ( self ): UpperCAmelCase = BackboneMixin() UpperCAmelCase = ["""a""", """b""", """c"""] UpperCAmelCase = ["""a""", """c"""] UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[0, 2] ) # Check out features and indices are updated correctly UpperCAmelCase = ["""a""", """b"""] self.assertEqual(backbone.out_features ,["""a""", """b"""] ) self.assertEqual(backbone.out_indices ,[0, 1] ) UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[-3, -1] )
74
1
"""simple docstring""" from __future__ import annotations def _a ( _snake_case ): """simple docstring""" return len(set(_snake_case ) ) == len(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCamelCase__ : def __init__( self ,A = 6 ): UpperCAmelCase = None UpperCAmelCase = None self.create_linked_list(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = current_node UpperCAmelCase = current_node for _ in range(1 ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = previous_node UpperCAmelCase = current_node UpperCAmelCase = self.front UpperCAmelCase = previous_node def _UpperCamelCase ( self ): return ( self.front == self.rear and self.front is not None and self.front.data is None ) def _UpperCamelCase ( self ): self.check_can_perform_operation() return self.front.data if self.front else None def _UpperCamelCase ( self ,A ): if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase = self.rear.next if self.rear: UpperCAmelCase = data def _UpperCamelCase ( self ): self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase = self.front.data UpperCAmelCase = None return data UpperCAmelCase = self.front UpperCAmelCase = old_front.next UpperCAmelCase = old_front.data UpperCAmelCase = None return data def _UpperCamelCase ( self ): if self.is_empty(): raise Exception("""Empty Queue""" ) def _UpperCamelCase ( self ): if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _UpperCamelCase = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } _UpperCamelCase = AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def _a ( _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = create_model( """HTSAT-tiny""" , """roberta""" , _snake_case , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=_snake_case , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = {} UpperCAmelCase = R""".*sequential.(\d+).*""" UpperCAmelCase = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase = key.replace(_snake_case , _snake_case ) if re.match(_snake_case , _snake_case ): # replace sequential layers with list UpperCAmelCase = re.match(_snake_case , _snake_case ).group(1 ) UpperCAmelCase = key.replace(F'''sequential.{sequential_layer}.''' , F'''layers.{int(_snake_case )//3}.linear.''' ) elif re.match(_snake_case , _snake_case ): UpperCAmelCase = int(re.match(_snake_case , _snake_case ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... UpperCAmelCase = 1 if projecton_layer == 0 else 2 UpperCAmelCase = key.replace(F'''_projection.{projecton_layer}.''' , F'''_projection.linear{transformers_projection_layer}.''' ) if "audio" and "qkv" in key: # split qkv into query key and value UpperCAmelCase = value UpperCAmelCase = mixed_qkv.size(0 ) // 3 UpperCAmelCase = mixed_qkv[:qkv_dim] UpperCAmelCase = mixed_qkv[qkv_dim : qkv_dim * 2] UpperCAmelCase = mixed_qkv[qkv_dim * 2 :] UpperCAmelCase = query_layer UpperCAmelCase = key_layer UpperCAmelCase = value_layer else: UpperCAmelCase = value return model_state_dict def _a ( _snake_case , _snake_case , _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = init_clap(_snake_case , enable_fusion=_snake_case ) clap_model.eval() UpperCAmelCase = clap_model.state_dict() UpperCAmelCase = rename_state_dict(_snake_case ) UpperCAmelCase = ClapConfig() UpperCAmelCase = enable_fusion UpperCAmelCase = ClapModel(_snake_case ) # ignore the spectrogram embedding layer model.load_state_dict(_snake_case , strict=_snake_case ) model.save_pretrained(_snake_case ) transformers_config.save_pretrained(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") _UpperCamelCase = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
74
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("""transformers.models.speecht5""") def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" hf_model.apply_weight_norm() UpperCAmelCase = checkpoint["""input_conv.weight_g"""] UpperCAmelCase = checkpoint["""input_conv.weight_v"""] UpperCAmelCase = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , ): """simple docstring""" if config_path is not None: UpperCAmelCase = SpeechTaHifiGanConfig.from_pretrained(_snake_case ) else: UpperCAmelCase = SpeechTaHifiGanConfig() UpperCAmelCase = SpeechTaHifiGan(_snake_case ) UpperCAmelCase = torch.load(_snake_case ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _snake_case , _snake_case ) UpperCAmelCase = np.load(_snake_case ) UpperCAmelCase = stats[0].reshape(-1 ) UpperCAmelCase = stats[1].reshape(-1 ) UpperCAmelCase = torch.from_numpy(_snake_case ).float() UpperCAmelCase = torch.from_numpy(_snake_case ).float() model.save_pretrained(_snake_case ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
74
1
"""simple docstring""" def _a ( _snake_case , _snake_case , _snake_case = 0 , _snake_case = 0 ): """simple docstring""" UpperCAmelCase = right or len(_snake_case ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(_snake_case , _snake_case , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase = concatenate_datasets _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadManager _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
74
1
"""simple docstring""" import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """stabilityai/stable-diffusion-2""" ,revision="""bf16""" ,dtype=jnp.bfloataa ,) UpperCAmelCase = """A painting of a squirrel eating a burger""" UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = sd_pipe.prepare_inputs(A ) UpperCAmelCase = replicate(A ) UpperCAmelCase = shard(A ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = jax.random.split(A ,jax.device_count() ) UpperCAmelCase = sd_pipe(A ,A ,A ,num_inference_steps=25 ,jit=A )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) UpperCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase = images[0, 253:256, 253:256, -1] UpperCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase = jnp.array([0.4238, 0.4414, 0.4395, 0.4453, 0.4629, 0.4590, 0.4531, 0.45508, 0.4512] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def _UpperCamelCase ( self ): UpperCAmelCase = """stabilityai/stable-diffusion-2""" UpperCAmelCase , UpperCAmelCase = FlaxDPMSolverMultistepScheduler.from_pretrained(A ,subfolder="""scheduler""" ) UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( A ,scheduler=A ,revision="""bf16""" ,dtype=jnp.bfloataa ,) UpperCAmelCase = scheduler_params UpperCAmelCase = """A painting of a squirrel eating a burger""" UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = sd_pipe.prepare_inputs(A ) UpperCAmelCase = replicate(A ) UpperCAmelCase = shard(A ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = jax.random.split(A ,jax.device_count() ) UpperCAmelCase = sd_pipe(A ,A ,A ,num_inference_steps=25 ,jit=A )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) UpperCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase = images[0, 253:256, 253:256, -1] UpperCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase = jnp.array([0.4336, 0.42969, 0.4453, 0.4199, 0.4297, 0.4531, 0.4434, 0.4434, 0.4297] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
74
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """ut/deta""": """https://huggingface.co/ut/deta/resolve/main/config.json""", } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''deta''' SCREAMING_SNAKE_CASE = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self ,A=None ,A=900 ,A=2_048 ,A=6 ,A=2_048 ,A=8 ,A=6 ,A=1_024 ,A=8 ,A=0.0 ,A=True ,A="relu" ,A=256 ,A=0.1 ,A=0.0 ,A=0.0 ,A=0.02 ,A=1.0 ,A=True ,A=False ,A="sine" ,A=5 ,A=4 ,A=4 ,A=True ,A=300 ,A=True ,A=True ,A=1 ,A=5 ,A=2 ,A=1 ,A=1 ,A=5 ,A=2 ,A=0.1 ,A=0.25 ,**A ,): if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) UpperCAmelCase = CONFIG_MAPPING["""resnet"""](out_features=["""stage2""", """stage3""", """stage4"""] ) else: if isinstance(A ,A ): UpperCAmelCase = backbone_config.pop("""model_type""" ) UpperCAmelCase = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase = config_class.from_dict(A ) UpperCAmelCase = backbone_config UpperCAmelCase = num_queries UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = init_xavier_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = auxiliary_loss UpperCAmelCase = position_embedding_type # deformable attributes UpperCAmelCase = num_feature_levels UpperCAmelCase = encoder_n_points UpperCAmelCase = decoder_n_points UpperCAmelCase = two_stage UpperCAmelCase = two_stage_num_proposals UpperCAmelCase = with_box_refine UpperCAmelCase = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("""If two_stage is True, with_box_refine must be True.""" ) # Hungarian matcher UpperCAmelCase = class_cost UpperCAmelCase = bbox_cost UpperCAmelCase = giou_cost # Loss coefficients UpperCAmelCase = mask_loss_coefficient UpperCAmelCase = dice_loss_coefficient UpperCAmelCase = bbox_loss_coefficient UpperCAmelCase = giou_loss_coefficient UpperCAmelCase = eos_coefficient UpperCAmelCase = focal_alpha super().__init__(is_encoder_decoder=A ,**A ) @property def _UpperCamelCase ( self ): return self.encoder_attention_heads @property def _UpperCamelCase ( self ): return self.d_model def _UpperCamelCase ( self ): UpperCAmelCase = copy.deepcopy(self.__dict__ ) UpperCAmelCase = self.backbone_config.to_dict() UpperCAmelCase = self.__class__.model_type return output
74
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCamelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") _UpperCamelCase = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) _UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the training data.'''} ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the validation data.'''} ) SCREAMING_SNAKE_CASE = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) SCREAMING_SNAKE_CASE = field(default=32 , metadata={'''help''': '''The size of the square patches to use for masking.'''} ) SCREAMING_SNAKE_CASE = field( default=0.6 , metadata={'''help''': '''Percentage of patches to mask.'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def _UpperCamelCase ( self ): UpperCAmelCase = {} if self.train_dir is not None: UpperCAmelCase = self.train_dir if self.validation_dir is not None: UpperCAmelCase = self.validation_dir UpperCAmelCase = data_files if data_files else None @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ''' '''checkpoint identifier on the hub. ''' '''Don\'t set if you want to train a model from scratch.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(snake_case )} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'''} , ) SCREAMING_SNAKE_CASE = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''Name or path of preprocessor config.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each image. If not specified, will use `image_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Stride to use for the encoder.'''} , ) class lowerCamelCase__ : def __init__( self ,A=192 ,A=32 ,A=4 ,A=0.6 ): UpperCAmelCase = input_size UpperCAmelCase = mask_patch_size UpperCAmelCase = model_patch_size UpperCAmelCase = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError("""Input size must be divisible by mask patch size""" ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError("""Mask patch size must be divisible by model patch size""" ) UpperCAmelCase = self.input_size // self.mask_patch_size UpperCAmelCase = self.mask_patch_size // self.model_patch_size UpperCAmelCase = self.rand_size**2 UpperCAmelCase = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self ): UpperCAmelCase = np.random.permutation(self.token_count )[: self.mask_count] UpperCAmelCase = np.zeros(self.token_count ,dtype=A ) UpperCAmelCase = 1 UpperCAmelCase = mask.reshape((self.rand_size, self.rand_size) ) UpperCAmelCase = mask.repeat(self.scale ,axis=0 ).repeat(self.scale ,axis=1 ) return torch.tensor(mask.flatten() ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = torch.stack([example["""pixel_values"""] for example in examples] ) UpperCAmelCase = torch.stack([example["""mask"""] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mim""" , _snake_case , _snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_snake_case ) transformers.utils.logging.set_verbosity(_snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. UpperCAmelCase = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _snake_case ) and data_args.train_val_split > 0.0: UpperCAmelCase = ds["""train"""].train_test_split(data_args.train_val_split ) UpperCAmelCase = split["""train"""] UpperCAmelCase = split["""test"""] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.config_name_or_path , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_snake_case , """decoder_type""" ): UpperCAmelCase = """simmim""" # adapt config UpperCAmelCase = model_args.image_size if model_args.image_size is not None else config.image_size UpperCAmelCase = model_args.patch_size if model_args.patch_size is not None else config.patch_size UpperCAmelCase = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { """image_size""": model_args.image_size, """patch_size""": model_args.patch_size, """encoder_stride""": model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } UpperCAmelCase = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: UpperCAmelCase = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) UpperCAmelCase = AutoModelForMaskedImageModeling.from_config(_snake_case ) if training_args.do_train: UpperCAmelCase = ds["""train"""].column_names else: UpperCAmelCase = ds["""validation"""].column_names if data_args.image_column_name is not None: UpperCAmelCase = data_args.image_column_name elif "image" in column_names: UpperCAmelCase = """image""" elif "img" in column_names: UpperCAmelCase = """img""" else: UpperCAmelCase = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py UpperCAmelCase = Compose( [ Lambda(lambda _snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator UpperCAmelCase = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_snake_case ): UpperCAmelCase = [transforms(_snake_case ) for image in examples[image_column_name]] UpperCAmelCase = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: UpperCAmelCase = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: UpperCAmelCase = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_snake_case ) # Initialize our trainer UpperCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase = last_checkpoint UpperCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase = trainer.evaluate() trainer.log_metrics("""eval""" , _snake_case ) trainer.save_metrics("""eval""" , _snake_case ) # Write model card and (optionally) push to hub UpperCAmelCase = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """masked-image-modeling""", """dataset""": data_args.dataset_name, """tags""": ["""masked-image-modeling"""], } if training_args.push_to_hub: trainer.push_to_hub(**_snake_case ) else: trainer.create_model_card(**_snake_case ) if __name__ == "__main__": main()
74
1
"""simple docstring""" import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class lowerCamelCase__ ( unittest.TestCase ): @parameterized.expand([(None,), ("""foo.json""",)] ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = GenerationConfig( do_sample=A ,temperature=0.7 ,length_penalty=1.0 ,bad_words_ids=[[1, 2, 3], [4, 5]] ,) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(A ,config_name=A ) UpperCAmelCase = GenerationConfig.from_pretrained(A ,config_name=A ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample ,A ) self.assertEqual(loaded_config.temperature ,0.7 ) self.assertEqual(loaded_config.length_penalty ,1.0 ) self.assertEqual(loaded_config.bad_words_ids ,[[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k ,50 ) self.assertEqual(loaded_config.max_length ,20 ) self.assertEqual(loaded_config.max_time ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = AutoConfig.from_pretrained("""gpt2""" ) UpperCAmelCase = GenerationConfig.from_model_config(A ) UpperCAmelCase = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(A ,A ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id ,default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id ,model_config.eos_token_id ) def _UpperCamelCase ( self ): UpperCAmelCase = GenerationConfig() UpperCAmelCase = { """max_new_tokens""": 1_024, """foo""": """bar""", } UpperCAmelCase = copy.deepcopy(A ) UpperCAmelCase = generation_config.update(**A ) # update_kwargs was not modified (no side effects) self.assertEqual(A ,A ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens ,1_024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(A ,{"""foo""": """bar"""} ) def _UpperCamelCase ( self ): UpperCAmelCase = GenerationConfig() UpperCAmelCase = """bar""" with tempfile.TemporaryDirectory("""test-generation-config""" ) as tmp_dir: generation_config.save_pretrained(A ) UpperCAmelCase = GenerationConfig.from_pretrained(A ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo ,"""bar""" ) UpperCAmelCase = GenerationConfig.from_model_config(A ) assert not hasattr(A ,"""foo""" ) # no new kwargs should be initialized if from config def _UpperCamelCase ( self ): UpperCAmelCase = GenerationConfig() self.assertEqual(default_config.temperature ,1.0 ) self.assertEqual(default_config.do_sample ,A ) self.assertEqual(default_config.num_beams ,1 ) UpperCAmelCase = GenerationConfig( do_sample=A ,temperature=0.7 ,length_penalty=1.0 ,bad_words_ids=[[1, 2, 3], [4, 5]] ,) self.assertEqual(config.temperature ,0.7 ) self.assertEqual(config.do_sample ,A ) self.assertEqual(config.num_beams ,1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(A ) UpperCAmelCase = GenerationConfig.from_pretrained(A ,temperature=1.0 ) self.assertEqual(loaded_config.temperature ,1.0 ) self.assertEqual(loaded_config.do_sample ,A ) self.assertEqual(loaded_config.num_beams ,1 ) # default value @is_staging_test class lowerCamelCase__ ( unittest.TestCase ): @classmethod def _UpperCamelCase ( cls ): UpperCAmelCase = TOKEN HfFolder.save_token(A ) @classmethod def _UpperCamelCase ( cls ): try: delete_repo(token=cls._token ,repo_id="""test-generation-config""" ) except HTTPError: pass try: delete_repo(token=cls._token ,repo_id="""valid_org/test-generation-config-org""" ) except HTTPError: pass def _UpperCamelCase ( self ): UpperCAmelCase = GenerationConfig( do_sample=A ,temperature=0.7 ,length_penalty=1.0 ,) config.push_to_hub("""test-generation-config""" ,use_auth_token=self._token ) UpperCAmelCase = GenerationConfig.from_pretrained(F'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A ,getattr(A ,A ) ) # Reset repo delete_repo(token=self._token ,repo_id="""test-generation-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( A ,repo_id="""test-generation-config""" ,push_to_hub=A ,use_auth_token=self._token ) UpperCAmelCase = GenerationConfig.from_pretrained(F'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A ,getattr(A ,A ) ) def _UpperCamelCase ( self ): UpperCAmelCase = GenerationConfig( do_sample=A ,temperature=0.7 ,length_penalty=1.0 ,) config.push_to_hub("""valid_org/test-generation-config-org""" ,use_auth_token=self._token ) UpperCAmelCase = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A ,getattr(A ,A ) ) # Reset repo delete_repo(token=self._token ,repo_id="""valid_org/test-generation-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( A ,repo_id="""valid_org/test-generation-config-org""" ,push_to_hub=A ,use_auth_token=self._token ) UpperCAmelCase = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A ,getattr(A ,A ) )
74
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): _UpperCamelCase = True from torch.cuda.amp import autocast _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to log verbose messages or not.'''} , ) SCREAMING_SNAKE_CASE = field( default=2.0 , metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.5 , metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.99_99_95 , metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def _a ( _snake_case , _snake_case ): """simple docstring""" logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase = logging.WARNING if model_args.verbose_logging: UpperCAmelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase = logging.INFO logger.setLevel(_snake_case ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default='''train''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) SCREAMING_SNAKE_CASE = field( default='''validation''' , metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) } , ) SCREAMING_SNAKE_CASE = field( default='''file''' , metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) SCREAMING_SNAKE_CASE = field( default=1 , metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) SCREAMING_SNAKE_CASE = field( default=20.0 , metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = "longest" SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __call__( self ,A ): # reformat list to dict and set to pytorch format UpperCAmelCase = self.feature_extractor.pad( A ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="""pt""" ,) UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""input_values"""].shape[-1] ) UpperCAmelCase = batch["""input_values"""].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""attention_mask"""].sum(-1 ) ).to( torch.long ) UpperCAmelCase = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["""input_values"""].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase = 1 UpperCAmelCase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=A ,min_masks=2 ,) return batch class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=1 ,A=0 ,A=1.0 ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = 0 UpperCAmelCase = max_gumbel_temp UpperCAmelCase = min_gumbel_temp UpperCAmelCase = gumbel_temp_decay def _UpperCamelCase ( self ,A ,A ): model.train() UpperCAmelCase = self._prepare_inputs(A ) if self.use_amp: with autocast(): UpperCAmelCase = self.compute_loss(A ,A ) else: UpperCAmelCase = self.compute_loss(A ,A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase = loss.sum() / (inputs["""mask_time_indices"""]).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() configure_logger(_snake_case , _snake_case ) # Downloading and loading a dataset from the hub. UpperCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="""validation""" , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=_snake_case ) def prepare_dataset(_snake_case ): # check that all files have the correct sampling rate UpperCAmelCase , UpperCAmelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase = datasets.map( _snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["""train"""].column_names ) # filter audio files that are too long UpperCAmelCase = vectorized_datasets.filter( lambda _snake_case : len(data["""speech"""] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(_snake_case ): return feature_extractor(batch["""speech"""] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase = vectorized_datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["""train"""].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( """PreTraining is only supported for ``config.do_stable_layer_norm=True`` and""" """ ``config.feat_extract_norm='layer'""" ) UpperCAmelCase = WavaVecaForPreTraining(_snake_case ) UpperCAmelCase = DataCollatorForWavaVecaPretraining(model=_snake_case , feature_extractor=_snake_case ) UpperCAmelCase = WavaVecaPreTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , train_dataset=vectorized_datasets["""train"""] , eval_dataset=vectorized_datasets["""validation"""] , tokenizer=_snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
74
1
"""simple docstring""" from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def _a ( _snake_case ): """simple docstring""" return {key.lstrip("""-""" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def _a ( ): """simple docstring""" UpperCAmelCase = ArgumentParser( """HuggingFace Datasets CLI tool""" , usage="""datasets-cli <command> [<args>]""" , allow_abbrev=_snake_case ) UpperCAmelCase = parser.add_subparsers(help="""datasets-cli command helpers""" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(_snake_case ) EnvironmentCommand.register_subcommand(_snake_case ) TestCommand.register_subcommand(_snake_case ) RunBeamCommand.register_subcommand(_snake_case ) DummyDataCommand.register_subcommand(_snake_case ) # Parse args UpperCAmelCase , UpperCAmelCase = parser.parse_known_args() if not hasattr(_snake_case , """func""" ): parser.print_help() exit(1 ) UpperCAmelCase = parse_unknown_args(_snake_case ) # Run UpperCAmelCase = args.func(_snake_case , **_snake_case ) service.run() if __name__ == "__main__": main()
74
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = 30 UpperCAmelCase = self.seq_length + self.mem_len UpperCAmelCase = 15 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 99 UpperCAmelCase = [10, 50, 80] UpperCAmelCase = 32 UpperCAmelCase = 32 UpperCAmelCase = 4 UpperCAmelCase = 8 UpperCAmelCase = 128 UpperCAmelCase = 2 UpperCAmelCase = 2 UpperCAmelCase = None UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 3 UpperCAmelCase = self.vocab_size - 1 UpperCAmelCase = 0.01 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCamelCase ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLLMHeadModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase , UpperCAmelCase = model([input_ids_a, mems_a] ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLForSequenceClassification(A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids_a} return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = () if is_tf_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,d_embed=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A ) def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: UpperCAmelCase = model_class(A ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: UpperCAmelCase = model.get_output_embeddings() assert isinstance(A ,tf.keras.layers.Layer ) UpperCAmelCase = model.get_bias() assert name is None else: UpperCAmelCase = model.get_output_embeddings() assert x is None UpperCAmelCase = model.get_bias() assert name is None def _UpperCamelCase ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _UpperCamelCase ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFTransfoXLModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""This model doesn't play well with fit() due to not returning a single loss.""" ) def _UpperCamelCase ( self ): pass @require_tf class lowerCamelCase__ ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off UpperCAmelCase = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off UpperCAmelCase = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> UpperCAmelCase = model.generate(A ,max_length=200 ,do_sample=A ) self.assertListEqual(output_ids[0].numpy().tolist() ,A )
74
1
"""simple docstring""" import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB_FILES_NAMES, save_json from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_sentencepiece_available(): import sentencepiece as sp _UpperCamelCase = 5 _UpperCamelCase = 10 @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = SpeechaTextTokenizer SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True def _UpperCamelCase ( self ): super().setUp() UpperCAmelCase = sp.SentencePieceProcessor() spm_model.Load(A ) UpperCAmelCase = ["""<s>""", """<pad>""", """</s>""", """<unk>"""] vocab += [spm_model.IdToPiece(id_ ) for id_ in range(len(A ) )] UpperCAmelCase = dict(zip(A ,range(len(A ) ) ) ) UpperCAmelCase = Path(self.tmpdirname ) save_json(A ,save_dir / VOCAB_FILES_NAMES["""vocab_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(A ,save_dir / VOCAB_FILES_NAMES["""spm_file"""] ) UpperCAmelCase = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _UpperCamelCase ( self ): UpperCAmelCase = """<pad>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""<s>""" ) self.assertEqual(vocab_keys[1] ,"""<pad>""" ) self.assertEqual(vocab_keys[-1] ,"""j""" ) self.assertEqual(len(A ) ,1_001 ) def _UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size ,1_001 ) def _UpperCamelCase ( self ): UpperCAmelCase = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(A ,["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A ) ,[289, 50, 14, 174, 386] ,) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( A ,[SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """."""] ,) UpperCAmelCase = tokenizer.convert_tokens_to_ids(A ) self.assertListEqual(A ,[12, 25, 88, 59, 28, 23, 11, 4, 606, 351, 351, 351, 7, 16, 70, 50, 76, 84, 10, 4, 8] ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(A ) self.assertListEqual( A ,[SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """."""] ,) @slow def _UpperCamelCase ( self ): # fmt: off UpperCAmelCase = {"""input_ids""": [[3_791, 797, 31, 11, 64, 797, 31, 2_429, 433, 12, 1_176, 12, 20, 786, 915, 142, 2_413, 240, 37, 3_238, 797, 31, 11, 35, 93, 915, 142, 2_413, 240, 37, 5_540, 567, 1_276, 93, 37, 610, 40, 62, 455, 657, 1_042, 123, 780, 177, 37, 309, 241, 1_298, 514, 20, 292, 2_737, 114, 2_469, 241, 85, 64, 302, 548, 528, 423, 4, 509, 406, 423, 37, 601, 4, 777, 302, 548, 528, 423, 284, 4, 3_388, 511, 459, 4, 3_555, 40, 321, 302, 705, 4, 3_388, 511, 583, 326, 5, 5, 5, 62, 3_310, 560, 177, 2_680, 217, 1_508, 32, 31, 853, 418, 64, 583, 511, 1_605, 62, 35, 93, 560, 177, 2_680, 217, 1_508, 1_521, 64, 583, 511, 519, 62, 20, 1_515, 764, 20, 149, 261, 5_625, 7_972, 20, 5_540, 567, 1_276, 93, 3_925, 1_675, 11, 15, 802, 7_972, 576, 217, 1_508, 11, 35, 93, 1_253, 2_441, 15, 289, 652, 31, 416, 321, 3_842, 115, 40, 911, 8, 476, 619, 4, 380, 142, 423, 335, 240, 35, 93, 264, 8, 11, 335, 569, 420, 163, 5, 2], [260, 548, 528, 423, 20, 451, 20, 2_681, 1_153, 3_434, 20, 5_540, 37, 567, 126, 1_253, 2_441, 3_376, 449, 210, 431, 1_563, 177, 767, 5_540, 11, 1_203, 472, 11, 2_953, 685, 285, 364, 706, 1_153, 20, 6_799, 20, 2_869, 20, 4_464, 126, 40, 2_429, 20, 1_040, 866, 2_664, 418, 20, 318, 20, 1_726, 186, 20, 265, 522, 35, 93, 2_191, 4_634, 20, 1_040, 12, 6_799, 15, 228, 2_356, 142, 31, 11, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [2_575, 2_666, 684, 1_582, 1_176, 12, 627, 149, 619, 20, 4_902, 563, 11, 20, 149, 261, 3_420, 2_356, 174, 142, 4_714, 131, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""facebook/s2t-small-mustc-en-de-st""" ,revision="""a14f04cf0776c02f62a8cb800cf7909e15ea23ad""" ,) @require_sentencepiece class lowerCamelCase__ ( unittest.TestCase ): SCREAMING_SNAKE_CASE = '''valhalla/s2t_mustc_multilinguial_medium''' SCREAMING_SNAKE_CASE = '''C\'est trop cool''' SCREAMING_SNAKE_CASE = '''Esto es genial''' @classmethod def _UpperCamelCase ( cls ): UpperCAmelCase = SpeechaTextTokenizer.from_pretrained(cls.checkpoint_name ) return cls def _UpperCamelCase ( self ): self.assertEqual(self.tokenizer.lang_code_to_id["""pt"""] ,4 ) self.assertEqual(self.tokenizer.lang_code_to_id["""ru"""] ,6 ) self.assertEqual(self.tokenizer.lang_code_to_id["""it"""] ,9 ) self.assertEqual(self.tokenizer.lang_code_to_id["""de"""] ,11 ) def _UpperCamelCase ( self ): self.assertEqual(self.tokenizer.vocab_size ,10_000 ) def _UpperCamelCase ( self ): self.assertIn(A ,self.tokenizer.all_special_ids ) UpperCAmelCase = [ES_CODE, 4, 1_601, 47, 7_647, 2] UpperCAmelCase = self.tokenizer.decode(A ,skip_special_tokens=A ) UpperCAmelCase = self.tokenizer.decode(generated_ids[1:] ,skip_special_tokens=A ) self.assertEqual(A ,A ) self.assertNotIn(self.tokenizer.eos_token ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = """fr""" UpperCAmelCase = self.tokenizer(self.french_text ).input_ids self.assertEqual(encoded[0] ,A ) self.assertEqual(encoded[-1] ,self.tokenizer.eos_token_id ) def _UpperCamelCase ( self ): UpperCAmelCase = """fr""" self.assertListEqual(self.tokenizer.prefix_tokens ,[FR_CODE] ) UpperCAmelCase = """es""" self.assertListEqual(self.tokenizer.prefix_tokens ,[ES_CODE] )
74
"""simple docstring""" from math import sqrt def _a ( _snake_case = 100_0000 ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_snake_case , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
74
1
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if n == 1 or not isinstance(_snake_case , _snake_case ): return 0 elif n == 2: return 1 else: UpperCAmelCase = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 2 while digits < n: index += 1 UpperCAmelCase = len(str(fibonacci(_snake_case ) ) ) return index def _a ( _snake_case = 1000 ): """simple docstring""" return fibonacci_digits_index(_snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
74
"""simple docstring""" import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _UpperCamelCase = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } _UpperCamelCase = AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def _a ( _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = create_model( """HTSAT-tiny""" , """roberta""" , _snake_case , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=_snake_case , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = {} UpperCAmelCase = R""".*sequential.(\d+).*""" UpperCAmelCase = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase = key.replace(_snake_case , _snake_case ) if re.match(_snake_case , _snake_case ): # replace sequential layers with list UpperCAmelCase = re.match(_snake_case , _snake_case ).group(1 ) UpperCAmelCase = key.replace(F'''sequential.{sequential_layer}.''' , F'''layers.{int(_snake_case )//3}.linear.''' ) elif re.match(_snake_case , _snake_case ): UpperCAmelCase = int(re.match(_snake_case , _snake_case ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... UpperCAmelCase = 1 if projecton_layer == 0 else 2 UpperCAmelCase = key.replace(F'''_projection.{projecton_layer}.''' , F'''_projection.linear{transformers_projection_layer}.''' ) if "audio" and "qkv" in key: # split qkv into query key and value UpperCAmelCase = value UpperCAmelCase = mixed_qkv.size(0 ) // 3 UpperCAmelCase = mixed_qkv[:qkv_dim] UpperCAmelCase = mixed_qkv[qkv_dim : qkv_dim * 2] UpperCAmelCase = mixed_qkv[qkv_dim * 2 :] UpperCAmelCase = query_layer UpperCAmelCase = key_layer UpperCAmelCase = value_layer else: UpperCAmelCase = value return model_state_dict def _a ( _snake_case , _snake_case , _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = init_clap(_snake_case , enable_fusion=_snake_case ) clap_model.eval() UpperCAmelCase = clap_model.state_dict() UpperCAmelCase = rename_state_dict(_snake_case ) UpperCAmelCase = ClapConfig() UpperCAmelCase = enable_fusion UpperCAmelCase = ClapModel(_snake_case ) # ignore the spectrogram embedding layer model.load_state_dict(_snake_case , strict=_snake_case ) model.save_pretrained(_snake_case ) transformers_config.save_pretrained(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") _UpperCamelCase = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
74
1
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCamelCase = {"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _UpperCamelCase = abspath(join(dirname(dirname(__file__)), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_snake_case ) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_terminal_summary_main UpperCAmelCase = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(_snake_case , id=_snake_case )
74
1
"""simple docstring""" import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed _UpperCamelCase = logging.getLogger(__name__) def _a ( _snake_case=2 , _snake_case=3 , _snake_case=16 , _snake_case = 10 , _snake_case = 2 ): """simple docstring""" def get_dataset(_snake_case ): UpperCAmelCase = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(_snake_case , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) UpperCAmelCase = get_dataset(_snake_case ) UpperCAmelCase = get_dataset(_snake_case ) UpperCAmelCase = DataLoader(_snake_case , shuffle=_snake_case , batch_size=_snake_case , num_workers=4 ) UpperCAmelCase = DataLoader(_snake_case , shuffle=_snake_case , batch_size=_snake_case , num_workers=4 ) return (train_dataloader, valid_dataloader) def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=None ): """simple docstring""" UpperCAmelCase = [] for epoch in range(_snake_case ): # Train quickly model.train() for batch in dataloader: UpperCAmelCase , UpperCAmelCase = batch UpperCAmelCase = model(_snake_case ) UpperCAmelCase = torch.nn.functional.mse_loss(_snake_case , _snake_case ) accelerator.backward(_snake_case ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class lowerCamelCase__ ( nn.Module ): def __init__( self ): super().__init__() UpperCAmelCase = nn.Parameter(torch.randn(1 ) ) UpperCAmelCase = nn.Parameter(torch.randn(1 ) ) def _UpperCamelCase ( self ,A ): return x * self.a + self.b class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) UpperCAmelCase , UpperCAmelCase = dummy_dataloaders() UpperCAmelCase = ProjectConfiguration(total_limit=1 ,project_dir=A ,automatic_checkpoint_naming=A ) # Train baseline UpperCAmelCase = Accelerator(project_config=A ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( A ,A ,A ,A ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) ,1 ) def _UpperCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) UpperCAmelCase , UpperCAmelCase = dummy_dataloaders() # Train baseline UpperCAmelCase = Accelerator() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( A ,A ,A ,A ) # Save initial UpperCAmelCase = os.path.join(A ,"""initial""" ) accelerator.save_state(A ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() UpperCAmelCase = train(3 ,A ,A ,A ,A ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() # Train partially set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) UpperCAmelCase , UpperCAmelCase = dummy_dataloaders() UpperCAmelCase = Accelerator() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( A ,A ,A ,A ) accelerator.load_state(A ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() self.assertEqual(A ,A ) self.assertEqual(A ,A ) self.assertEqual(A ,A ) UpperCAmelCase = train(2 ,A ,A ,A ,A ) # Save everything UpperCAmelCase = os.path.join(A ,"""checkpoint""" ) accelerator.save_state(A ) # Load everything back in and make sure all states work accelerator.load_state(A ) test_rands += train(1 ,A ,A ,A ,A ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() self.assertEqual(A ,A ) self.assertEqual(A ,A ) self.assertEqual(A ,A ) self.assertEqual(A ,A ) def _UpperCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) UpperCAmelCase , UpperCAmelCase = dummy_dataloaders() UpperCAmelCase = ProjectConfiguration(automatic_checkpoint_naming=A ) # Train baseline UpperCAmelCase = Accelerator(project_dir=A ,project_config=A ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( A ,A ,A ,A ) # Save initial accelerator.save_state() ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() UpperCAmelCase = train(3 ,A ,A ,A ,A ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() # Train partially set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) UpperCAmelCase , UpperCAmelCase = dummy_dataloaders() UpperCAmelCase = ProjectConfiguration(iteration=1 ,automatic_checkpoint_naming=A ) UpperCAmelCase = Accelerator(project_dir=A ,project_config=A ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( A ,A ,A ,A ) accelerator.load_state(os.path.join(A ,"""checkpoints""" ,"""checkpoint_0""" ) ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() self.assertEqual(A ,A ) self.assertEqual(A ,A ) self.assertEqual(A ,A ) UpperCAmelCase = train(2 ,A ,A ,A ,A ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A ,"""checkpoints""" ,"""checkpoint_1""" ) ) test_rands += train(1 ,A ,A ,A ,A ) ((UpperCAmelCase) , (UpperCAmelCase)) = model.a.item(), model.b.item() UpperCAmelCase = optimizer.state_dict() self.assertEqual(A ,A ) self.assertEqual(A ,A ) self.assertEqual(A ,A ) self.assertEqual(A ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = torch.tensor([1, 2, 3] ) UpperCAmelCase = torch.tensor([2, 3, 4] ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(net.parameters() ) UpperCAmelCase = Accelerator() with self.assertRaises(A ) as ve: accelerator.register_for_checkpointing(A ,A ,A ,A ) UpperCAmelCase = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def _UpperCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = torch.optim.Adam(params=model.parameters() ,lr=1e-3 ) UpperCAmelCase = torch.optim.lr_scheduler.StepLR(A ,step_size=1 ,gamma=0.99 ) UpperCAmelCase , UpperCAmelCase = dummy_dataloaders() UpperCAmelCase = ProjectConfiguration(automatic_checkpoint_naming=A ) # Train baseline UpperCAmelCase = Accelerator(project_dir=A ,project_config=A ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( A ,A ,A ,A ,A ) # Save initial accelerator.save_state() UpperCAmelCase = scheduler.state_dict() train(3 ,A ,A ,A ,A ,A ) self.assertNotEqual(A ,scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(A ,"""checkpoints""" ,"""checkpoint_0""" ) ) self.assertEqual(A ,scheduler.state_dict() ) def _UpperCamelCase ( self ): with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase = DummyModel() UpperCAmelCase = ProjectConfiguration(automatic_checkpoint_naming=A ,total_limit=2 ) # Train baseline UpperCAmelCase = Accelerator(project_dir=A ,project_config=A ) UpperCAmelCase = accelerator.prepare(A ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(A ,"""checkpoints""" ,"""checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A ,"""checkpoints""" ,"""checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(A ,"""checkpoints""" ,"""checkpoint_10""" ) ) ) @require_cuda def _UpperCamelCase ( self ): UpperCAmelCase = ["""torchrun""", F'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] execute_subprocess_async(A ,env=os.environ.copy() ) if __name__ == "__main__": _UpperCamelCase = """/tmp/accelerate/state_checkpointing""" _UpperCamelCase = DummyModel() _UpperCamelCase = torch.optim.Adam(params=model.parameters(), lr=1E-3) _UpperCamelCase = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) _UpperCamelCase , _UpperCamelCase = dummy_dataloaders() _UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline _UpperCamelCase = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision="""no""") if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) _UpperCamelCase , _UpperCamelCase = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: _UpperCamelCase = group["""params"""][0].device break assert param_device.type == accelerator.device.type _UpperCamelCase = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, """checkpoints""", """checkpoint_0"""), map_location="""cpu""") for group in optimizer.param_groups: _UpperCamelCase = group["""params"""][0].device break assert ( param_device.type == torch.device("""cpu""").type ), F"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, """checkpoints""", """checkpoint_0"""), map_location="""on_device""") for group in optimizer.param_groups: _UpperCamelCase = group["""params"""][0].device break assert ( param_device.type == accelerator.device.type ), F"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match="""Unsupported optimizer map location passed"""): accelerator.load_state(os.path.join(savedir, """checkpoints""", """checkpoint_0"""), map_location="""invalid""") accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
74
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class lowerCamelCase__ : def __init__( self ,A ,A ): if len(A ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) UpperCAmelCase = list(A ) UpperCAmelCase = degree def __add__( self ,A ): if self.degree > polynomial_a.degree: UpperCAmelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,A ) else: UpperCAmelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,A ) def __sub__( self ,A ): return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self ): return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self ,A ): UpperCAmelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): UpperCAmelCase = """""" for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(A ) return polynomial def __repr__( self ): return self.__str__() def _UpperCamelCase ( self ): UpperCAmelCase = [0] * self.degree for i in range(self.degree ): UpperCAmelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,A ) def _UpperCamelCase ( self ,A = 0 ): UpperCAmelCase = [0] * (self.degree + 2) UpperCAmelCase = constant for i in range(self.degree + 1 ): UpperCAmelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,A ) def __eq__( self ,A ): if not isinstance(A ,A ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self ,A ): return not self.__eq__(A )
74
1
"""simple docstring""" import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = set() UpperCAmelCase = [] def parse_line(_snake_case ): for line in fp: if isinstance(_snake_case , _snake_case ): UpperCAmelCase = line.decode("""UTF-8""" ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(""" """ ): # process a single warning and move it to `selected_warnings`. if len(_snake_case ) > 0: UpperCAmelCase = """\n""".join(_snake_case ) # Only keep the warnings specified in `targets` if any(F''': {x}: ''' in warning for x in targets ): selected_warnings.add(_snake_case ) buffer.clear() continue else: UpperCAmelCase = line.strip() buffer.append(_snake_case ) if from_gh: for filename in os.listdir(_snake_case ): UpperCAmelCase = os.path.join(_snake_case , _snake_case ) if not os.path.isdir(_snake_case ): # read the file if filename != "warnings.txt": continue with open(_snake_case ) as fp: parse_line(_snake_case ) else: try: with zipfile.ZipFile(_snake_case ) as z: for filename in z.namelist(): if not os.path.isdir(_snake_case ): # read the file if filename != "warnings.txt": continue with z.open(_snake_case ) as fp: parse_line(_snake_case ) except Exception: logger.warning( F'''{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.''' ) return selected_warnings def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = set() UpperCAmelCase = [os.path.join(_snake_case , _snake_case ) for p in os.listdir(_snake_case ) if (p.endswith(""".zip""" ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_snake_case , _snake_case ) ) return selected_warnings if __name__ == "__main__": def _a ( _snake_case ): """simple docstring""" return values.split(""",""" ) _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links _UpperCamelCase = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts _UpperCamelCase = extract_warnings(args.output_dir, args.targets) _UpperCamelCase = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
74
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy as np import tensorflow as tf from transformers import ( TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST, FlaubertConfig, TFFlaubertForMultipleChoice, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForSequenceClassification, TFFlaubertForTokenClassification, TFFlaubertModel, TFFlaubertWithLMHeadModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = 2 UpperCAmelCase = 99 UpperCAmelCase = 0 UpperCAmelCase = 32 UpperCAmelCase = 2 UpperCAmelCase = 4 UpperCAmelCase = 0.1 UpperCAmelCase = 0.1 UpperCAmelCase = 512 UpperCAmelCase = 16 UpperCAmelCase = 2 UpperCAmelCase = 0.02 UpperCAmelCase = 3 UpperCAmelCase = 4 UpperCAmelCase = """last""" UpperCAmelCase = True UpperCAmelCase = None UpperCAmelCase = 0 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ,dtype=tf.floataa ) UpperCAmelCase = None if self.use_input_lengths: UpperCAmelCase = ( ids_tensor([self.batch_size] ,vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.n_langs ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] ,2 ,dtype=tf.floataa ) UpperCAmelCase = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase = FlaubertConfig( vocab_size=self.vocab_size ,n_special=self.n_special ,emb_dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,gelu_activation=self.gelu_activation ,sinusoidal_embeddings=self.sinusoidal_embeddings ,asm=self.asm ,causal=self.causal ,n_langs=self.n_langs ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,summary_type=self.summary_type ,use_proj=self.use_proj ,bos_token_id=self.bos_token_id ,) return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertModel(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) UpperCAmelCase = [input_ids, input_mask] UpperCAmelCase = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertWithLMHeadModel(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForQuestionAnsweringSimple(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForSequenceClassification(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_labels UpperCAmelCase = TFFlaubertForTokenClassification(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_choices UpperCAmelCase = TFFlaubertForMultipleChoice(config=A ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """langs""": token_type_ids, """lengths""": input_lengths, } return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( ( TFFlaubertModel, TFFlaubertWithLMHeadModel, TFFlaubertForSequenceClassification, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForTokenClassification, TFFlaubertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = ( (TFFlaubertWithLMHeadModel,) if is_tf_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFFlaubertModel, '''fill-mask''': TFFlaubertWithLMHeadModel, '''question-answering''': TFFlaubertForQuestionAnsweringSimple, '''text-classification''': TFFlaubertForSequenceClassification, '''token-classification''': TFFlaubertForTokenClassification, '''zero-shot''': TFFlaubertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,emb_dim=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_token_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_multiple_choice(*A ) @slow def _UpperCamelCase ( self ): for model_name in TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFFlaubertModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_tf @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( unittest.TestCase ): @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModel.from_pretrained("""jplu/tf-flaubert-small-cased""" ) UpperCAmelCase = tf.convert_to_tensor( [[0, 158, 735, 2_592, 1_424, 6_727, 82, 1]] ,dtype=tf.intaa ,) # "J'aime flaubert !" UpperCAmelCase = model(A )[0] UpperCAmelCase = tf.TensorShape((1, 8, 512) ) self.assertEqual(output.shape ,A ) # compare the actual values for a slice. UpperCAmelCase = tf.convert_to_tensor( [ [ [-1.8768773, -1.566555, 0.27072418], [-1.6920038, -0.5873505, 1.9329599], [-2.9563985, -1.6993835, 1.7972052], ] ] ,dtype=tf.floataa ,) self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1e-4 ) )
74
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """s-JoL/Open-Llama-V1""": """https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json""", } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''open-llama''' def __init__( self ,A=100_000 ,A=4_096 ,A=11_008 ,A=32 ,A=32 ,A="silu" ,A=2_048 ,A=0.02 ,A=1e-6 ,A=True ,A=0 ,A=1 ,A=2 ,A=False ,A=True ,A=0.1 ,A=0.1 ,A=True ,A=True ,A=None ,**A ,): UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = hidden_size UpperCAmelCase = intermediate_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = initializer_range UpperCAmelCase = rms_norm_eps UpperCAmelCase = use_cache UpperCAmelCase = kwargs.pop( """use_memorry_efficient_attention""" ,A ) UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_dropout_prob UpperCAmelCase = use_stable_embedding UpperCAmelCase = shared_input_output_embedding UpperCAmelCase = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,tie_word_embeddings=A ,**A ,) def _UpperCamelCase ( self ): if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ F'''got {self.rope_scaling}''' ) UpperCAmelCase = self.rope_scaling.get("""type""" ,A ) UpperCAmelCase = self.rope_scaling.get("""factor""" ,A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(A ,A ) or rope_scaling_factor <= 1.0: raise ValueError(F'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
74
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): UpperCAmelCase = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): UpperCAmelCase = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] UpperCAmelCase = key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(_snake_case )-1}''' ) if "norm" in key: UpperCAmelCase = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] UpperCAmelCase = key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(_snake_case )-1}''' ) if "layer_norm1" in key: UpperCAmelCase = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: UpperCAmelCase = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase = key[key.find("""block""" ) + len("""block""" )] UpperCAmelCase = key.replace(F'''block{idx}''' , F'''block.{int(_snake_case )-1}''' ) if "attn.q" in key: UpperCAmelCase = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: UpperCAmelCase = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: UpperCAmelCase = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: UpperCAmelCase = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: UpperCAmelCase = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: UpperCAmelCase = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: UpperCAmelCase = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) UpperCAmelCase = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase = key[key.find("""linear_c""" ) + len("""linear_c""" )] UpperCAmelCase = key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(_snake_case )-1}''' ) if "bot_conv" in key: UpperCAmelCase = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: UpperCAmelCase = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: UpperCAmelCase = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: UpperCAmelCase = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: UpperCAmelCase = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: UpperCAmelCase = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: UpperCAmelCase = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): UpperCAmelCase = key.replace("""module.last_layer_depth""" , """head.head""" ) UpperCAmelCase = value return new_state_dict def _a ( _snake_case , _snake_case ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict UpperCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase = kv_bias[config.hidden_sizes[i] :] def _a ( ): """simple docstring""" UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case=False , _snake_case=None ): """simple docstring""" UpperCAmelCase = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) UpperCAmelCase = GLPNImageProcessor() # prepare image UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=_snake_case , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict UpperCAmelCase = torch.load(_snake_case , map_location=torch.device("""cpu""" ) ) # rename keys UpperCAmelCase = rename_keys(_snake_case ) # key and value matrices need special treatment read_in_k_v(_snake_case , _snake_case ) # create HuggingFace model and load state dict UpperCAmelCase = GLPNForDepthEstimation(_snake_case ) model.load_state_dict(_snake_case ) model.eval() # forward pass UpperCAmelCase = model(_snake_case ) UpperCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: UpperCAmelCase = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: UpperCAmelCase = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) UpperCAmelCase = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _snake_case , atol=1E-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=_snake_case , ) image_processor.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=_snake_case , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCamelCase = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
74
1
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class lowerCamelCase__ : def __init__( self ,A ,A ,A ,A ,A ,A=0.2 ,A=0.2 ): UpperCAmelCase = bp_numa UpperCAmelCase = bp_numa UpperCAmelCase = bp_numa UpperCAmelCase = conva_get[:2] UpperCAmelCase = conva_get[2] UpperCAmelCase = size_pa UpperCAmelCase = rate_w UpperCAmelCase = rate_t UpperCAmelCase = [ np.mat(-1 * np.random.rand(self.conva[0] ,self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] UpperCAmelCase = np.mat(-1 * np.random.rand(self.num_bpa ,self.num_bpa ) + 0.5 ) UpperCAmelCase = np.mat(-1 * np.random.rand(self.num_bpa ,self.num_bpa ) + 0.5 ) UpperCAmelCase = -2 * np.random.rand(self.conva[1] ) + 1 UpperCAmelCase = -2 * np.random.rand(self.num_bpa ) + 1 UpperCAmelCase = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCamelCase ( self ,A ): # save model dict with pickle UpperCAmelCase = { """num_bp1""": self.num_bpa, """num_bp2""": self.num_bpa, """num_bp3""": self.num_bpa, """conv1""": self.conva, """step_conv1""": self.step_conva, """size_pooling1""": self.size_poolinga, """rate_weight""": self.rate_weight, """rate_thre""": self.rate_thre, """w_conv1""": self.w_conva, """wkj""": self.wkj, """vji""": self.vji, """thre_conv1""": self.thre_conva, """thre_bp2""": self.thre_bpa, """thre_bp3""": self.thre_bpa, } with open(A ,"""wb""" ) as f: pickle.dump(A ,A ) print(F'''Model saved: {save_path}''' ) @classmethod def _UpperCamelCase ( cls ,A ): # read saved model with open(A ,"""rb""" ) as f: UpperCAmelCase = pickle.load(A ) # noqa: S301 UpperCAmelCase = model_dic.get("""conv1""" ) conv_get.append(model_dic.get("""step_conv1""" ) ) UpperCAmelCase = model_dic.get("""size_pooling1""" ) UpperCAmelCase = model_dic.get("""num_bp1""" ) UpperCAmelCase = model_dic.get("""num_bp2""" ) UpperCAmelCase = model_dic.get("""num_bp3""" ) UpperCAmelCase = model_dic.get("""rate_weight""" ) UpperCAmelCase = model_dic.get("""rate_thre""" ) # create model instance UpperCAmelCase = CNN(A ,A ,A ,A ,A ,A ,A ) # modify model parameter UpperCAmelCase = model_dic.get("""w_conv1""" ) UpperCAmelCase = model_dic.get("""wkj""" ) UpperCAmelCase = model_dic.get("""vji""" ) UpperCAmelCase = model_dic.get("""thre_conv1""" ) UpperCAmelCase = model_dic.get("""thre_bp2""" ) UpperCAmelCase = model_dic.get("""thre_bp3""" ) return conv_ins def _UpperCamelCase ( self ,A ): return 1 / (1 + np.exp(-1 * x )) def _UpperCamelCase ( self ,A ): return round(A ,3 ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): # convolution process UpperCAmelCase = convs[0] UpperCAmelCase = convs[1] UpperCAmelCase = np.shape(A )[0] # get the data slice of original image data, data_focus UpperCAmelCase = [] for i_focus in range(0 ,size_data - size_conv + 1 ,A ): for j_focus in range(0 ,size_data - size_conv + 1 ,A ): UpperCAmelCase = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(A ) # calculate the feature map of every single kernel, and saved as list of matrix UpperCAmelCase = [] UpperCAmelCase = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(A ): UpperCAmelCase = [] for i_focus in range(len(A ) ): UpperCAmelCase = ( np.sum(np.multiply(data_focus[i_focus] ,w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(A ) ) UpperCAmelCase = np.asmatrix(A ).reshape( A ,A ) data_featuremap.append(A ) # expanding the data slice to One dimenssion UpperCAmelCase = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(A ) ) UpperCAmelCase = np.asarray(A ) return focus_list, data_featuremap def _UpperCamelCase ( self ,A ,A ,A="average_pool" ): # pooling process UpperCAmelCase = len(featuremaps[0] ) UpperCAmelCase = int(size_map / size_pooling ) UpperCAmelCase = [] for i_map in range(len(A ) ): UpperCAmelCase = featuremaps[i_map] UpperCAmelCase = [] for i_focus in range(0 ,A ,A ): for j_focus in range(0 ,A ,A ): UpperCAmelCase = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(A ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(A ) ) UpperCAmelCase = np.asmatrix(A ).reshape(A ,A ) featuremap_pooled.append(A ) return featuremap_pooled def _UpperCamelCase ( self ,A ): # expanding three dimension data to one dimension list UpperCAmelCase = [] for i in range(len(A ) ): UpperCAmelCase = np.shape(data[i] ) UpperCAmelCase = data[i].reshape(1 ,shapes[0] * shapes[1] ) UpperCAmelCase = data_listed.getA().tolist()[0] data_expanded.extend(A ) UpperCAmelCase = np.asarray(A ) return data_expanded def _UpperCamelCase ( self ,A ): # expanding matrix to one dimension list UpperCAmelCase = np.asarray(A ) UpperCAmelCase = np.shape(A ) UpperCAmelCase = data_mat.reshape(1 ,shapes[0] * shapes[1] ) return data_expanded def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): UpperCAmelCase = [] UpperCAmelCase = 0 for i_map in range(A ): UpperCAmelCase = np.ones((size_map, size_map) ) for i in range(0 ,A ,A ): for j in range(0 ,A ,A ): UpperCAmelCase = pd_pool[ i_pool ] UpperCAmelCase = i_pool + 1 UpperCAmelCase = np.multiply( A ,np.multiply(out_map[i_map] ,(1 - out_map[i_map]) ) ) pd_all.append(A ) return pd_all def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A=bool ): # model traning print("""----------------------Start Training-------------------------""" ) print((""" - - Shape: Train_Data """, np.shape(A )) ) print((""" - - Shape: Teach_Data """, np.shape(A )) ) UpperCAmelCase = 0 UpperCAmelCase = [] UpperCAmelCase = 10_000 while rp < n_repeat and mse >= error_accuracy: UpperCAmelCase = 0 print(F'''-------------Learning Time {rp}--------------''' ) for p in range(len(A ) ): # print('------------Learning Image: %d--------------'%p) UpperCAmelCase = np.asmatrix(datas_train[p] ) UpperCAmelCase = np.asarray(datas_teach[p] ) UpperCAmelCase , UpperCAmelCase = self.convolute( A ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) UpperCAmelCase = self.pooling(A ,self.size_poolinga ) UpperCAmelCase = np.shape(A ) UpperCAmelCase = self._expand(A ) UpperCAmelCase = data_bp_input UpperCAmelCase = np.dot(A ,self.vji.T ) - self.thre_bpa UpperCAmelCase = self.sig(A ) UpperCAmelCase = np.dot(A ,self.wkj.T ) - self.thre_bpa UpperCAmelCase = self.sig(A ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- UpperCAmelCase = np.multiply( (data_teach - bp_outa) ,np.multiply(A ,(1 - bp_outa) ) ) UpperCAmelCase = np.multiply( np.dot(A ,self.wkj ) ,np.multiply(A ,(1 - bp_outa) ) ) UpperCAmelCase = np.dot(A ,self.vji ) UpperCAmelCase = pd_i_all / (self.size_poolinga * self.size_poolinga) UpperCAmelCase = pd_conva_pooled.T.getA().tolist() UpperCAmelCase = self._calculate_gradient_from_pool( A ,A ,shape_featuremapa[0] ,shape_featuremapa[1] ,self.size_poolinga ,) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): UpperCAmelCase = self._expand_mat(pd_conva_all[k_conv] ) UpperCAmelCase = self.rate_weight * np.dot(A ,A ) UpperCAmelCase = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) UpperCAmelCase = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer UpperCAmelCase = self.wkj + pd_k_all.T * bp_outa * self.rate_weight UpperCAmelCase = self.vji + pd_j_all.T * bp_outa * self.rate_weight UpperCAmelCase = self.thre_bpa - pd_k_all * self.rate_thre UpperCAmelCase = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image UpperCAmelCase = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) UpperCAmelCase = rp + 1 UpperCAmelCase = error_count / patterns all_mse.append(A ) def draw_error(): UpperCAmelCase = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(A ,"""+-""" ) plt.plot(A ,"""r--""" ) plt.xlabel("""Learning Times""" ) plt.ylabel("""All_mse""" ) plt.grid(A ,alpha=0.5 ) plt.show() print("""------------------Training Complished---------------------""" ) print((""" - - Training epoch: """, rp, F''' - - Mse: {mse:.6f}''') ) if draw_e: draw_error() return mse def _UpperCamelCase ( self ,A ): # model predict UpperCAmelCase = [] print("""-------------------Start Testing-------------------------""" ) print((""" - - Shape: Test_Data """, np.shape(A )) ) for p in range(len(A ) ): UpperCAmelCase = np.asmatrix(datas_test[p] ) UpperCAmelCase , UpperCAmelCase = self.convolute( A ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) UpperCAmelCase = self.pooling(A ,self.size_poolinga ) UpperCAmelCase = self._expand(A ) UpperCAmelCase = data_bp_input UpperCAmelCase = bp_outa * self.vji.T - self.thre_bpa UpperCAmelCase = self.sig(A ) UpperCAmelCase = bp_outa * self.wkj.T - self.thre_bpa UpperCAmelCase = self.sig(A ) produce_out.extend(bp_outa.getA().tolist() ) UpperCAmelCase = [list(map(self.do_round ,A ) ) for each in produce_out] return np.asarray(A ) def _UpperCamelCase ( self ,A ): # return the data of image after convoluting process so we can check it out UpperCAmelCase = np.asmatrix(A ) UpperCAmelCase , UpperCAmelCase = self.convolute( A ,self.conva ,self.w_conva ,self.thre_conva ,conv_step=self.step_conva ,) UpperCAmelCase = self.pooling(A ,self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
74
"""simple docstring""" def _a ( _snake_case ): # noqa: E741 """simple docstring""" UpperCAmelCase = len(_snake_case ) UpperCAmelCase = 0 UpperCAmelCase = [0] * n UpperCAmelCase = [False] * n UpperCAmelCase = [False] * n def dfs(_snake_case , _snake_case , _snake_case , _snake_case ): if parent == root: out_edge_count += 1 UpperCAmelCase = True UpperCAmelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase = dfs(_snake_case , _snake_case , _snake_case , _snake_case ) UpperCAmelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase = True # AP found via cycle if at == low[to]: UpperCAmelCase = True else: UpperCAmelCase = min(low[at] , _snake_case ) return out_edge_count for i in range(_snake_case ): if not visited[i]: UpperCAmelCase = 0 UpperCAmelCase = dfs(_snake_case , _snake_case , -1 , _snake_case ) UpperCAmelCase = out_edge_count > 1 for x in range(len(_snake_case ) ): if is_art[x] is True: print(_snake_case ) # Adjacency list of graph _UpperCamelCase = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
74
1
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): _UpperCamelCase = True from torch.cuda.amp import autocast _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to log verbose messages or not.'''} , ) SCREAMING_SNAKE_CASE = field( default=2.0 , metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.5 , metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.99_99_95 , metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def _a ( _snake_case , _snake_case ): """simple docstring""" logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase = logging.WARNING if model_args.verbose_logging: UpperCAmelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase = logging.INFO logger.setLevel(_snake_case ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default='''train''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) SCREAMING_SNAKE_CASE = field( default='''validation''' , metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) } , ) SCREAMING_SNAKE_CASE = field( default='''file''' , metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) SCREAMING_SNAKE_CASE = field( default=1 , metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) SCREAMING_SNAKE_CASE = field( default=20.0 , metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = "longest" SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __call__( self ,A ): # reformat list to dict and set to pytorch format UpperCAmelCase = self.feature_extractor.pad( A ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="""pt""" ,) UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""input_values"""].shape[-1] ) UpperCAmelCase = batch["""input_values"""].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""attention_mask"""].sum(-1 ) ).to( torch.long ) UpperCAmelCase = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["""input_values"""].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase = 1 UpperCAmelCase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=A ,min_masks=2 ,) return batch class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=1 ,A=0 ,A=1.0 ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = 0 UpperCAmelCase = max_gumbel_temp UpperCAmelCase = min_gumbel_temp UpperCAmelCase = gumbel_temp_decay def _UpperCamelCase ( self ,A ,A ): model.train() UpperCAmelCase = self._prepare_inputs(A ) if self.use_amp: with autocast(): UpperCAmelCase = self.compute_loss(A ,A ) else: UpperCAmelCase = self.compute_loss(A ,A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase = loss.sum() / (inputs["""mask_time_indices"""]).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() configure_logger(_snake_case , _snake_case ) # Downloading and loading a dataset from the hub. UpperCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="""validation""" , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=_snake_case ) def prepare_dataset(_snake_case ): # check that all files have the correct sampling rate UpperCAmelCase , UpperCAmelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase = datasets.map( _snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["""train"""].column_names ) # filter audio files that are too long UpperCAmelCase = vectorized_datasets.filter( lambda _snake_case : len(data["""speech"""] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(_snake_case ): return feature_extractor(batch["""speech"""] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase = vectorized_datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["""train"""].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( """PreTraining is only supported for ``config.do_stable_layer_norm=True`` and""" """ ``config.feat_extract_norm='layer'""" ) UpperCAmelCase = WavaVecaForPreTraining(_snake_case ) UpperCAmelCase = DataCollatorForWavaVecaPretraining(model=_snake_case , feature_extractor=_snake_case ) UpperCAmelCase = WavaVecaPreTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , train_dataset=vectorized_datasets["""train"""] , eval_dataset=vectorized_datasets["""validation"""] , tokenizer=_snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
74
"""simple docstring""" _UpperCamelCase = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ _UpperCamelCase = [{"""type""": """code""", """content""": INSTALL_CONTENT}] _UpperCamelCase = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
74
1
"""simple docstring""" import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _UpperCamelCase = get_tests_dir("""fixtures/test_sentencepiece_no_bos.model""") @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = PegasusTokenizer SCREAMING_SNAKE_CASE = PegasusTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _UpperCamelCase ( self ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = PegasusTokenizer(A ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _UpperCamelCase ( self ): return PegasusTokenizer.from_pretrained("""google/pegasus-large""" ) def _UpperCamelCase ( self ,**A ): return PegasusTokenizer.from_pretrained(self.tmpdirname ,**A ) def _UpperCamelCase ( self ,A ): return ("This is a test", "This is a test") def _UpperCamelCase ( self ): UpperCAmelCase = """</s>""" UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"""<pad>""" ) self.assertEqual(vocab_keys[1] ,"""</s>""" ) self.assertEqual(vocab_keys[-1] ,"""v""" ) self.assertEqual(len(A ) ,1_103 ) def _UpperCamelCase ( self ): self.assertEqual(self.get_tokenizer().vocab_size ,1_103 ) def _UpperCamelCase ( self ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = self.tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = ( """Let's see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important""" """ </s> <pad> <pad> <pad>""" ) UpperCAmelCase = rust_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] UpperCAmelCase = py_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] self.assertListEqual(A ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word UpperCAmelCase = """<mask_1> To ensure a <mask_2> flow of bank resolutions.""" UpperCAmelCase = [2, 413, 615, 114, 3, 1_971, 113, 1_679, 10_710, 107, 1] UpperCAmelCase = tokenizer([raw_input_str] ,return_tensors=A ).input_ids[0] self.assertListEqual(A ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 96_103 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 103 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 105 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1_024 UpperCAmelCase = """To ensure a smooth flow of bank resolutions.""" UpperCAmelCase = [413, 615, 114, 2_291, 1_971, 113, 1_679, 10_710, 107, 1] UpperCAmelCase = tokenizer([raw_input_str] ,return_tensors=A ).input_ids[0] self.assertListEqual(A ,A ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def _UpperCamelCase ( self ): UpperCAmelCase = ["""This is going to be way too long.""" * 150, """short example"""] UpperCAmelCase = ["""not super long but more than 5 tokens""", """tiny"""] UpperCAmelCase = self._large_tokenizer(A ,padding=A ,truncation=A ,return_tensors="""pt""" ) UpperCAmelCase = self._large_tokenizer( text_target=A ,max_length=5 ,padding=A ,truncation=A ,return_tensors="""pt""" ) assert batch.input_ids.shape == (2, 1_024) assert batch.attention_mask.shape == (2, 1_024) assert targets["input_ids"].shape == (2, 5) assert len(A ) == 2 # input_ids, attention_mask. @slow def _UpperCamelCase ( self ): # fmt: off UpperCAmelCase = {"""input_ids""": [[38_979, 143, 18_485, 606, 130, 26_669, 87_686, 121, 54_189, 1_129, 111, 26_669, 87_686, 121, 9_114, 14_787, 121, 13_249, 158, 592, 956, 121, 14_621, 31_576, 143, 62_613, 108, 9_688, 930, 43_430, 11_562, 62_613, 304, 108, 11_443, 897, 108, 9_314, 17_415, 63_399, 108, 11_443, 7_614, 18_316, 118, 4_284, 7_148, 12_430, 143, 1_400, 25_703, 158, 111, 4_284, 7_148, 11_772, 143, 21_297, 1_064, 158, 122, 204, 3_506, 1_754, 1_133, 14_787, 1_581, 115, 33_224, 4_482, 111, 1_355, 110, 29_173, 317, 50_833, 108, 20_147, 94_665, 111, 77_198, 107, 1], [110, 62_613, 117, 638, 112, 1_133, 121, 20_098, 1_355, 79_050, 13_872, 135, 1_596, 53_541, 1_352, 141, 13_039, 5_542, 124, 302, 518, 111, 268, 2_956, 115, 149, 4_427, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [139, 1_235, 2_799, 18_289, 17_780, 204, 109, 9_474, 1_296, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="""google/bigbird-pegasus-large-arxiv""" ,revision="""ba85d0851d708441f91440d509690f1ab6353415""" ,) @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = PegasusTokenizer SCREAMING_SNAKE_CASE = PegasusTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _UpperCamelCase ( self ): super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = PegasusTokenizer(A ,offset=0 ,mask_token_sent=A ,mask_token="""[MASK]""" ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _UpperCamelCase ( self ): return PegasusTokenizer.from_pretrained("""google/bigbird-pegasus-large-arxiv""" ) def _UpperCamelCase ( self ,**A ): return PegasusTokenizer.from_pretrained(self.tmpdirname ,**A ) def _UpperCamelCase ( self ,A ): return ("This is a test", "This is a test") def _UpperCamelCase ( self ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = self.tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = ( """Let's see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>""" """ <pad> <pad> <pad>""" ) UpperCAmelCase = rust_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] UpperCAmelCase = py_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] self.assertListEqual(A ,A ) @require_torch def _UpperCamelCase ( self ): UpperCAmelCase = ["""This is going to be way too long.""" * 1_000, """short example"""] UpperCAmelCase = ["""not super long but more than 5 tokens""", """tiny"""] UpperCAmelCase = self._large_tokenizer(A ,padding=A ,truncation=A ,return_tensors="""pt""" ) UpperCAmelCase = self._large_tokenizer( text_target=A ,max_length=5 ,padding=A ,truncation=A ,return_tensors="""pt""" ) assert batch.input_ids.shape == (2, 4_096) assert batch.attention_mask.shape == (2, 4_096) assert targets["input_ids"].shape == (2, 5) assert len(A ) == 2 # input_ids, attention_mask. def _UpperCamelCase ( self ): UpperCAmelCase = ( """This is an example string that is used to test the original TF implementation against the HF""" """ implementation""" ) UpperCAmelCase = self._large_tokenizer(A ).input_ids self.assertListEqual( A ,[182, 117, 142, 587, 4_211, 120, 117, 263, 112, 804, 109, 856, 25_016, 3_137, 464, 109, 26_955, 3_137, 1] ,)
74
"""simple docstring""" import argparse import struct import unittest class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = data # Initialize hash values UpperCAmelCase = [ 0x6A_09_E6_67, 0xBB_67_AE_85, 0x3C_6E_F3_72, 0xA5_4F_F5_3A, 0x51_0E_52_7F, 0x9B_05_68_8C, 0x1F_83_D9_AB, 0x5B_E0_CD_19, ] # Initialize round constants UpperCAmelCase = [ 0x42_8A_2F_98, 0x71_37_44_91, 0xB5_C0_FB_CF, 0xE9_B5_DB_A5, 0x39_56_C2_5B, 0x59_F1_11_F1, 0x92_3F_82_A4, 0xAB_1C_5E_D5, 0xD8_07_AA_98, 0x12_83_5B_01, 0x24_31_85_BE, 0x55_0C_7D_C3, 0x72_BE_5D_74, 0x80_DE_B1_FE, 0x9B_DC_06_A7, 0xC1_9B_F1_74, 0xE4_9B_69_C1, 0xEF_BE_47_86, 0x0F_C1_9D_C6, 0x24_0C_A1_CC, 0x2D_E9_2C_6F, 0x4A_74_84_AA, 0x5C_B0_A9_DC, 0x76_F9_88_DA, 0x98_3E_51_52, 0xA8_31_C6_6D, 0xB0_03_27_C8, 0xBF_59_7F_C7, 0xC6_E0_0B_F3, 0xD5_A7_91_47, 0x06_CA_63_51, 0x14_29_29_67, 0x27_B7_0A_85, 0x2E_1B_21_38, 0x4D_2C_6D_FC, 0x53_38_0D_13, 0x65_0A_73_54, 0x76_6A_0A_BB, 0x81_C2_C9_2E, 0x92_72_2C_85, 0xA2_BF_E8_A1, 0xA8_1A_66_4B, 0xC2_4B_8B_70, 0xC7_6C_51_A3, 0xD1_92_E8_19, 0xD6_99_06_24, 0xF4_0E_35_85, 0x10_6A_A0_70, 0x19_A4_C1_16, 0x1E_37_6C_08, 0x27_48_77_4C, 0x34_B0_BC_B5, 0x39_1C_0C_B3, 0x4E_D8_AA_4A, 0x5B_9C_CA_4F, 0x68_2E_6F_F3, 0x74_8F_82_EE, 0x78_A5_63_6F, 0x84_C8_78_14, 0x8C_C7_02_08, 0x90_BE_FF_FA, 0xA4_50_6C_EB, 0xBE_F9_A3_F7, 0xC6_71_78_F2, ] UpperCAmelCase = self.preprocessing(self.data ) self.final_hash() @staticmethod def _UpperCamelCase ( A ): UpperCAmelCase = b"""\x80""" + (b"""\x00""" * (63 - (len(A ) + 8) % 64)) UpperCAmelCase = struct.pack(""">Q""" ,(len(A ) * 8) ) return data + padding + big_endian_integer def _UpperCamelCase ( self ): # Convert into blocks of 64 bytes UpperCAmelCase = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers UpperCAmelCase = list(struct.unpack(""">16L""" ,A ) ) # add 48 0-ed integers words += [0] * 48 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array UpperCAmelCase = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) UpperCAmelCase = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) UpperCAmelCase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_00_00_00_00 # Compression UpperCAmelCase = self.ror(A ,6 ) ^ self.ror(A ,11 ) ^ self.ror(A ,25 ) UpperCAmelCase = (e & f) ^ ((~e & 0xFF_FF_FF_FF) & g) UpperCAmelCase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_00_00_00_00 UpperCAmelCase = self.ror(A ,2 ) ^ self.ror(A ,13 ) ^ self.ror(A ,22 ) UpperCAmelCase = (a & b) ^ (a & c) ^ (b & c) UpperCAmelCase = (sa + maj) % 0x1_00_00_00_00 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = ( g, f, e, ((d + tempa) % 0x1_00_00_00_00), c, b, a, ((tempa + tempa) % 0x1_00_00_00_00), ) UpperCAmelCase = [a, b, c, d, e, f, g, h] # Modify final values UpperCAmelCase = [ ((element + mutated_hash_values[index]) % 0x1_00_00_00_00) for index, element in enumerate(self.hashes ) ] UpperCAmelCase = """""".join([hex(A )[2:].zfill(8 ) for value in self.hashes] ) def _UpperCamelCase ( self ,A ,A ): return 0xFF_FF_FF_FF & (value << (32 - rotations)) | (value >> rotations) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): import hashlib UpperCAmelCase = bytes("""Test String""" ,"""utf-8""" ) self.assertEqual(SHAaaa(A ).hash ,hashlib.shaaaa(A ).hexdigest() ) def _a ( ): """simple docstring""" import doctest doctest.testmod() UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase = parser.parse_args() UpperCAmelCase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase = f.read() else: UpperCAmelCase = bytes(_snake_case , """utf-8""" ) print(SHAaaa(_snake_case ).hash ) if __name__ == "__main__": main()
74
1
"""simple docstring""" import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger() def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case = True ): """simple docstring""" print(F'''Converting {name}...''' ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": UpperCAmelCase = timm.create_model("""levit_128s""" , pretrained=_snake_case ) else: UpperCAmelCase = timm.create_model("""levit_128""" , pretrained=_snake_case ) if hidden_sizes == 192: UpperCAmelCase = timm.create_model("""levit_192""" , pretrained=_snake_case ) if hidden_sizes == 256: UpperCAmelCase = timm.create_model("""levit_256""" , pretrained=_snake_case ) if hidden_sizes == 384: UpperCAmelCase = timm.create_model("""levit_384""" , pretrained=_snake_case ) from_model.eval() UpperCAmelCase = LevitForImageClassificationWithTeacher(_snake_case ).eval() UpperCAmelCase = OrderedDict() UpperCAmelCase = from_model.state_dict() UpperCAmelCase = list(from_model.state_dict().keys() ) UpperCAmelCase = list(our_model.state_dict().keys() ) print(len(_snake_case ) , len(_snake_case ) ) for i in range(len(_snake_case ) ): UpperCAmelCase = weights[og_keys[i]] our_model.load_state_dict(_snake_case ) UpperCAmelCase = torch.randn((2, 3, 224, 224) ) UpperCAmelCase = from_model(_snake_case ) UpperCAmelCase = our_model(_snake_case ).logits assert torch.allclose(_snake_case , _snake_case ), "The model logits don't match the original one." UpperCAmelCase = name print(_snake_case ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) UpperCAmelCase = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(F'''Pushed {checkpoint_name}''' ) def _a ( _snake_case , _snake_case = None , _snake_case = True ): """simple docstring""" UpperCAmelCase = """imagenet-1k-id2label.json""" UpperCAmelCase = 1000 UpperCAmelCase = (1, num_labels) UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = num_labels UpperCAmelCase = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(_snake_case ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = partial(_snake_case , num_labels=_snake_case , idalabel=_snake_case , labelaid=_snake_case ) UpperCAmelCase = { """levit-128S""": 128, """levit-128""": 128, """levit-192""": 192, """levit-256""": 256, """levit-384""": 384, } UpperCAmelCase = { """levit-128S""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-128""": ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), """levit-192""": ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-256""": ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), """levit-384""": ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , _snake_case , names_to_config[model_name] , _snake_case , _snake_case ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , _snake_case , _snake_case , _snake_case , _snake_case ) return config, expected_shape if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default=None, type=str, help="""The name of the model you wish to convert, it must be one of the supported Levit* architecture,""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""levit-dump-folder/""", type=Path, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") parser.add_argument( """--no-push_to_hub""", dest="""push_to_hub""", action="""store_false""", help="""Do not push model and image processor to the hub""", ) _UpperCamelCase = parser.parse_args() _UpperCamelCase = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
74
"""simple docstring""" def _a ( _snake_case = 10 , _snake_case = 22 ): """simple docstring""" UpperCAmelCase = range(1 , _snake_case ) UpperCAmelCase = range(1 , _snake_case ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F"""{solution(10, 22) = }""")
74
1
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = 30 UpperCAmelCase = self.seq_length + self.mem_len UpperCAmelCase = 15 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 99 UpperCAmelCase = [10, 50, 80] UpperCAmelCase = 32 UpperCAmelCase = 32 UpperCAmelCase = 4 UpperCAmelCase = 8 UpperCAmelCase = 128 UpperCAmelCase = 2 UpperCAmelCase = 2 UpperCAmelCase = None UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 3 UpperCAmelCase = self.vocab_size - 1 UpperCAmelCase = 0.01 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCamelCase ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLLMHeadModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase , UpperCAmelCase = model([input_ids_a, mems_a] ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLForSequenceClassification(A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids_a} return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = () if is_tf_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,d_embed=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A ) def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: UpperCAmelCase = model_class(A ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: UpperCAmelCase = model.get_output_embeddings() assert isinstance(A ,tf.keras.layers.Layer ) UpperCAmelCase = model.get_bias() assert name is None else: UpperCAmelCase = model.get_output_embeddings() assert x is None UpperCAmelCase = model.get_bias() assert name is None def _UpperCamelCase ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _UpperCamelCase ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFTransfoXLModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""This model doesn't play well with fit() due to not returning a single loss.""" ) def _UpperCamelCase ( self ): pass @require_tf class lowerCamelCase__ ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off UpperCAmelCase = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off UpperCAmelCase = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> UpperCAmelCase = model.generate(A ,max_length=200 ,do_sample=A ) self.assertListEqual(output_ids[0].numpy().tolist() ,A )
74
"""simple docstring""" from __future__ import annotations def _a ( _snake_case ): """simple docstring""" return len(set(_snake_case ) ) == len(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = None def _a ( _snake_case , _snake_case=0.999 , _snake_case="cosine" , ): """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(_snake_case ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_snake_case ): return math.exp(t * -12.0 ) else: raise ValueError(F'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) UpperCAmelCase = [] for i in range(_snake_case ): UpperCAmelCase = i / num_diffusion_timesteps UpperCAmelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(_snake_case ) / alpha_bar_fn(_snake_case ) , _snake_case ) ) return torch.tensor(_snake_case , dtype=torch.floataa ) class lowerCamelCase__ ( snake_case , snake_case ): @register_to_config def __init__( self ,A = 1_000 ,A = "fixed_small_log" ,A = True ,A = 1.0 ,A = "epsilon" ,A = "squaredcos_cap_v2" ,): if beta_schedule != "squaredcos_cap_v2": raise ValueError("""UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'""" ) UpperCAmelCase = betas_for_alpha_bar(A ) UpperCAmelCase = 1.0 - self.betas UpperCAmelCase = torch.cumprod(self.alphas ,dim=0 ) UpperCAmelCase = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase = 1.0 # setable values UpperCAmelCase = None UpperCAmelCase = torch.from_numpy(np.arange(0 ,A )[::-1].copy() ) UpperCAmelCase = variance_type def _UpperCamelCase ( self ,A ,A = None ): return sample def _UpperCamelCase ( self ,A ,A = None ): UpperCAmelCase = num_inference_steps UpperCAmelCase = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase = (np.arange(0 ,A ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase = torch.from_numpy(A ).to(A ) def _UpperCamelCase ( self ,A ,A=None ,A=None ,A=None ): if prev_timestep is None: UpperCAmelCase = t - 1 UpperCAmelCase = self.alphas_cumprod[t] UpperCAmelCase = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase = 1 - alpha_prod_t UpperCAmelCase = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase = self.betas[t] else: UpperCAmelCase = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase = torch.log(torch.clamp(A ,min=1e-2_0 ) ) UpperCAmelCase = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase = variance.log() UpperCAmelCase = beta.log() UpperCAmelCase = (predicted_variance + 1) / 2 UpperCAmelCase = frac * max_log + (1 - frac) * min_log return variance def _UpperCamelCase ( self ,A ,A ,A ,A = None ,A=None ,A = True ,): UpperCAmelCase = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase , UpperCAmelCase = torch.split(A ,sample.shape[1] ,dim=1 ) else: UpperCAmelCase = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase = t - 1 UpperCAmelCase = self.alphas_cumprod[t] UpperCAmelCase = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase = 1 - alpha_prod_t UpperCAmelCase = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase = self.betas[t] UpperCAmelCase = self.alphas[t] else: UpperCAmelCase = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase = model_output else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`''' """ for the UnCLIPScheduler.""" ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase = torch.clamp( A ,-self.config.clip_sample_range ,self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase = 0 if t > 0: UpperCAmelCase = randn_tensor( model_output.shape ,dtype=model_output.dtype ,generator=A ,device=model_output.device ) UpperCAmelCase = self._get_variance( A ,predicted_variance=A ,prev_timestep=A ,) if self.variance_type == "fixed_small_log": UpperCAmelCase = variance elif self.variance_type == "learned_range": UpperCAmelCase = (0.5 * variance).exp() else: raise ValueError( F'''variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`''' """ for the UnCLIPScheduler.""" ) UpperCAmelCase = variance * variance_noise UpperCAmelCase = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=A ,pred_original_sample=A ) def _UpperCamelCase ( self ,A ,A ,A ,): # Make sure alphas_cumprod and timestep have same device and dtype as original_samples UpperCAmelCase = self.alphas_cumprod.to(device=original_samples.device ,dtype=original_samples.dtype ) UpperCAmelCase = timesteps.to(original_samples.device ) UpperCAmelCase = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
74
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( _snake_case = 0.1 ): """simple docstring""" UpperCAmelCase = 3 UpperCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( _snake_case = 0.1 ): """simple docstring""" UpperCAmelCase = 3 UpperCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE = '''CLIPImageProcessor''' SCREAMING_SNAKE_CASE = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self ,A=None ,A=None ,**A ): UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) def __call__( self ,A=None ,A=None ,A=None ,**A ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.batch_decode(*A ,**A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.decode(*A ,**A ) @property def _UpperCamelCase ( self ): UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,A ,) return self.image_processor
74
1
"""simple docstring""" def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = 0 while b > 0: if b & 1: UpperCAmelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
74
"""simple docstring""" from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def _a ( _snake_case = "mumbai" ): """simple docstring""" UpperCAmelCase = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): UpperCAmelCase = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() UpperCAmelCase = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("""Bangalore"""), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
74
1
"""simple docstring""" import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''''' SCREAMING_SNAKE_CASE = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) SCREAMING_SNAKE_CASE = None # compression type in fsspec. ex: "gzip" SCREAMING_SNAKE_CASE = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self ,A = "" ,A = None ,A = None ,**A ): super().__init__(self ,**A ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode UpperCAmelCase = fsspec.open( A ,mode="""rb""" ,protocol=A ,compression=self.compression ,client_kwargs={ """requote_redirect_url""": False, # see https://github.com/huggingface/datasets/pull/5459 """trust_env""": True, # Enable reading proxy env variables. **(target_options or {}).pop("""client_kwargs""" ,{} ), # To avoid issues if it was already passed. } ,**(target_options or {}) ,) UpperCAmelCase = os.path.basename(self.file.path.split("""::""" )[0] ) UpperCAmelCase = ( self.compressed_name[: self.compressed_name.rindex(""".""" )] if """.""" in self.compressed_name else self.compressed_name ) UpperCAmelCase = None @classmethod def _UpperCamelCase ( cls ,A ): # compressed file paths are always relative to the archive root return super()._strip_protocol(A ).lstrip("""/""" ) def _UpperCamelCase ( self ): if self.dir_cache is None: UpperCAmelCase = {**self.file.fs.info(self.file.path ), """name""": self.uncompressed_name} UpperCAmelCase = {f["""name"""]: f} def _UpperCamelCase ( self ,A ): return self.file.open().read() def _UpperCamelCase ( self ,A ,A = "rb" ,A=None ,A=True ,A=None ,**A ,): UpperCAmelCase = self._strip_protocol(A ) if mode != "rb": raise ValueError(F'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''' ) return self.file.open() class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''bz2''' SCREAMING_SNAKE_CASE = '''bz2''' SCREAMING_SNAKE_CASE = '''.bz2''' class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''gzip''' SCREAMING_SNAKE_CASE = '''gzip''' SCREAMING_SNAKE_CASE = '''.gz''' class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''lz4''' SCREAMING_SNAKE_CASE = '''lz4''' SCREAMING_SNAKE_CASE = '''.lz4''' class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''xz''' SCREAMING_SNAKE_CASE = '''xz''' SCREAMING_SNAKE_CASE = '''.xz''' class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''zstd''' SCREAMING_SNAKE_CASE = '''zstd''' SCREAMING_SNAKE_CASE = '''.zst''' def __init__( self ,A ,A = "rb" ,A = None ,A = None ,A = DEFAULT_BLOCK_SIZE ,**A ,): super().__init__( fo=A ,mode=A ,target_protocol=A ,target_options=A ,block_size=A ,**A ,) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 UpperCAmelCase = self.file.__enter__ class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = file_ def __enter__( self ): self._file.__enter__() return self def __exit__( self ,*A ,**A ): self._file.__exit__(*A ,**A ) def __iter__( self ): return iter(self._file ) def _UpperCamelCase ( self ): return next(self._file ) def __getattr__( self ,A ): return getattr(self._file ,A ) def fixed_enter(*A ,**A ): return WrappedFile(_enter(*A ,**A ) ) UpperCAmelCase = fixed_enter
74
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): UpperCAmelCase = ["""a""", """b""", """c"""] # Defaults to last layer if both are None UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,A ,A ) self.assertEqual(A ,["""c"""] ) self.assertEqual(A ,[2] ) # Out indices set to match out features UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(["""a""", """c"""] ,A ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features set to match out indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[0, 2] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features selected from negative indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[-3, -1] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[-3, -1] ) def _UpperCamelCase ( self ): # Stage names must be set with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,A ) # Out features must be a list with self.assertRaises(A ): verify_out_features_out_indices(("""a""", """b""") ,(0, 1) ,["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(A ): verify_out_features_out_indices(A ,0 ,["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(A ,(0, 1) ,["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0,) ,["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 2) ,["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(A ): verify_out_features_out_indices(["""b""", """a"""] ,(0, 1) ,["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] ,(0, 1, -1) ,["""a""", """b""", """c""", """d"""] ) def _UpperCamelCase ( self ): UpperCAmelCase = BackboneMixin() UpperCAmelCase = ["""a""", """b""", """c"""] UpperCAmelCase = ["""a""", """c"""] UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[0, 2] ) # Check out features and indices are updated correctly UpperCAmelCase = ["""a""", """b"""] self.assertEqual(backbone.out_features ,["""a""", """b"""] ) self.assertEqual(backbone.out_indices ,[0, 1] ) UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[-3, -1] )
74
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """weiweishi/roc-bert-base-zh""": """https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json""", } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''roc_bert''' def __init__( self ,A=30_522 ,A=768 ,A=12 ,A=12 ,A=3_072 ,A="gelu" ,A=0.1 ,A=0.1 ,A=512 ,A=2 ,A=0.02 ,A=1e-1_2 ,A=True ,A=0 ,A="absolute" ,A=None ,A=True ,A=True ,A=768 ,A=910 ,A=512 ,A=24_858 ,A=True ,**A ,): UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = type_vocab_size UpperCAmelCase = layer_norm_eps UpperCAmelCase = use_cache UpperCAmelCase = enable_pronunciation UpperCAmelCase = enable_shape UpperCAmelCase = pronunciation_embed_dim UpperCAmelCase = pronunciation_vocab_size UpperCAmelCase = shape_embed_dim UpperCAmelCase = shape_vocab_size UpperCAmelCase = concat_input UpperCAmelCase = position_embedding_type UpperCAmelCase = classifier_dropout super().__init__(pad_token_id=A ,**A )
74
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCamelCase__ : def __init__( self ,A = 6 ): UpperCAmelCase = None UpperCAmelCase = None self.create_linked_list(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = current_node UpperCAmelCase = current_node for _ in range(1 ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = previous_node UpperCAmelCase = current_node UpperCAmelCase = self.front UpperCAmelCase = previous_node def _UpperCamelCase ( self ): return ( self.front == self.rear and self.front is not None and self.front.data is None ) def _UpperCamelCase ( self ): self.check_can_perform_operation() return self.front.data if self.front else None def _UpperCamelCase ( self ,A ): if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase = self.rear.next if self.rear: UpperCAmelCase = data def _UpperCamelCase ( self ): self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase = self.front.data UpperCAmelCase = None return data UpperCAmelCase = self.front UpperCAmelCase = old_front.next UpperCAmelCase = old_front.data UpperCAmelCase = None return data def _UpperCamelCase ( self ): if self.is_empty(): raise Exception("""Empty Queue""" ) def _UpperCamelCase ( self ): if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" def _a ( _snake_case , _snake_case ): """simple docstring""" if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) UpperCAmelCase = str(bin(_snake_case ) )[2:] # remove the leading "0b" UpperCAmelCase = str(bin(_snake_case ) )[2:] # remove the leading "0b" UpperCAmelCase = max(len(_snake_case ) , len(_snake_case ) ) return "0b" + "".join( str(int(char_a == """1""" and char_b == """1""" ) ) for char_a, char_b in zip(a_binary.zfill(_snake_case ) , b_binary.zfill(_snake_case ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("""transformers.models.speecht5""") def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" hf_model.apply_weight_norm() UpperCAmelCase = checkpoint["""input_conv.weight_g"""] UpperCAmelCase = checkpoint["""input_conv.weight_v"""] UpperCAmelCase = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , ): """simple docstring""" if config_path is not None: UpperCAmelCase = SpeechTaHifiGanConfig.from_pretrained(_snake_case ) else: UpperCAmelCase = SpeechTaHifiGanConfig() UpperCAmelCase = SpeechTaHifiGan(_snake_case ) UpperCAmelCase = torch.load(_snake_case ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _snake_case , _snake_case ) UpperCAmelCase = np.load(_snake_case ) UpperCAmelCase = stats[0].reshape(-1 ) UpperCAmelCase = stats[1].reshape(-1 ) UpperCAmelCase = torch.from_numpy(_snake_case ).float() UpperCAmelCase = torch.from_numpy(_snake_case ).float() model.save_pretrained(_snake_case ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
74
1
"""simple docstring""" import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse("""3.8"""): import importlib_metadata else: import importlib.metadata as importlib_metadata def _a ( _snake_case , _snake_case=False ): """simple docstring""" try: UpperCAmelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase = strtobool(_snake_case ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value _UpperCamelCase = parse_flag_from_env("""RUN_SLOW""", default=False) _UpperCamelCase = parse_flag_from_env("""RUN_REMOTE""", default=False) _UpperCamelCase = parse_flag_from_env("""RUN_LOCAL""", default=True) _UpperCamelCase = parse_flag_from_env("""RUN_PACKAGED""", default=True) # Compression _UpperCamelCase = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason="""test requires lz4""") _UpperCamelCase = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason="""test requires py7zr""") _UpperCamelCase = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason="""test requires zstandard""") # Audio _UpperCamelCase = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec("""soundfile""") is None or version.parse(importlib_metadata.version("""soundfile""")) < version.parse("""0.12.0"""), reason="""test requires sndfile>=0.12.1: 'pip install \"soundfile>=0.12.1\"'; """, ) # Beam _UpperCamelCase = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse("""0.3.2"""), reason="""test requires apache-beam and a compatible dill version""", ) # Dill-cloudpickle compatibility _UpperCamelCase = pytest.mark.skipif( config.DILL_VERSION <= version.parse("""0.3.2"""), reason="""test requires dill>0.3.2 for cloudpickle compatibility""", ) # Windows _UpperCamelCase = pytest.mark.skipif( sys.platform == """win32""", reason="""test should not be run on Windows""", ) def _a ( _snake_case ): """simple docstring""" try: import faiss # noqa except ImportError: UpperCAmelCase = unittest.skip("""test requires faiss""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" try: import regex # noqa except ImportError: UpperCAmelCase = unittest.skip("""test requires regex""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" try: import elasticsearch # noqa except ImportError: UpperCAmelCase = unittest.skip("""test requires elasticsearch""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" try: import sqlalchemy # noqa except ImportError: UpperCAmelCase = unittest.skip("""test requires sqlalchemy""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not config.TORCH_AVAILABLE: UpperCAmelCase = unittest.skip("""test requires PyTorch""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not config.TF_AVAILABLE: UpperCAmelCase = unittest.skip("""test requires TensorFlow""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not config.JAX_AVAILABLE: UpperCAmelCase = unittest.skip("""test requires JAX""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not config.PIL_AVAILABLE: UpperCAmelCase = unittest.skip("""test requires Pillow""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip("""test requires transformers""" )(_snake_case ) else: return test_case def _a ( _snake_case ): """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip("""test requires tiktoken""" )(_snake_case ) else: return test_case def _a ( _snake_case ): """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip("""test requires spacy""" )(_snake_case ) else: return test_case def _a ( _snake_case ): """simple docstring""" def _require_spacy_model(_snake_case ): try: import spacy # noqa F401 spacy.load(_snake_case ) except ImportError: return unittest.skip("""test requires spacy""" )(_snake_case ) except OSError: return unittest.skip("""test requires spacy model '{}'""".format(_snake_case ) )(_snake_case ) else: return test_case return _require_spacy_model def _a ( _snake_case ): """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip("""test requires pyspark""" )(_snake_case ) else: return test_case def _a ( _snake_case ): """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip("""test requires joblibspark""" )(_snake_case ) else: return test_case def _a ( _snake_case ): """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase = unittest.skip("""test is slow""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase = unittest.skip("""test is local""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase = unittest.skip("""test is packaged""" )(_snake_case ) return test_case def _a ( _snake_case ): """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase = unittest.skip("""test requires remote""" )(_snake_case ) return test_case def _a ( *_snake_case ): """simple docstring""" def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(_snake_case ) and name.startswith("""test""" ): for decorator in decorators: UpperCAmelCase = decorator(_snake_case ) setattr(cls , _snake_case , _snake_case ) return cls return decorate class lowerCamelCase__ ( snake_case ): pass class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = 2 @contextmanager def _a ( _snake_case=OfflineSimulationMode.CONNECTION_FAILS , _snake_case=1E-16 ): """simple docstring""" UpperCAmelCase = requests.Session().request def timeout_request(_snake_case , _snake_case , _snake_case , **_snake_case ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase = """https://10.255.255.1""" if kwargs.get("""timeout""" ) is None: raise RequestWouldHangIndefinitelyError( F'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) UpperCAmelCase = timeout try: return online_request(_snake_case , _snake_case , **_snake_case ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase = url UpperCAmelCase = e.args[0] UpperCAmelCase = (max_retry_error.args[0].replace("""10.255.255.1""" , F'''OfflineMock[{url}]''' ),) UpperCAmelCase = (max_retry_error,) raise def raise_connection_error(_snake_case , _snake_case , **_snake_case ): raise requests.ConnectionError("""Offline mode is enabled.""" , request=_snake_case ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("""requests.Session.send""" , _snake_case ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("""requests.Session.request""" , _snake_case ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("""datasets.config.HF_DATASETS_OFFLINE""" , _snake_case ): yield else: raise ValueError("""Please use a value from the OfflineSimulationMode enum.""" ) @contextmanager def _a ( *_snake_case , **_snake_case ): """simple docstring""" UpperCAmelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*_snake_case , **_snake_case ) as tmp_dir: try: os.chdir(_snake_case ) yield finally: os.chdir(_snake_case ) @contextmanager def _a ( ): """simple docstring""" import gc gc.collect() UpperCAmelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def _a ( ): """simple docstring""" import gc gc.collect() UpperCAmelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def _a ( _snake_case , _snake_case ): """simple docstring""" return deepcopy(_snake_case ).integers(0 , 100 , 10 ).tolist() == deepcopy(_snake_case ).integers(0 , 100 , 10 ).tolist() def _a ( _snake_case ): """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(_snake_case , *_snake_case , **_snake_case ): try: return func(*_snake_case , **_snake_case ) except HTTPError as err: if str(_snake_case ).startswith("""500""" ) or str(_snake_case ).startswith("""502""" ): pytest.xfail(str(_snake_case ) ) raise err return decorator.decorator(_wrapper , _snake_case ) class lowerCamelCase__ : def __init__( self ,A ,A ,A ): UpperCAmelCase = returncode UpperCAmelCase = stdout UpperCAmelCase = stderr async def _a ( _snake_case , _snake_case ): """simple docstring""" while True: UpperCAmelCase = await stream.readline() if line: callback(_snake_case ) else: break async def _a ( _snake_case , _snake_case=None , _snake_case=None , _snake_case=None , _snake_case=False , _snake_case=False ): """simple docstring""" if echo: print("""\nRunning: """ , """ """.join(_snake_case ) ) UpperCAmelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_snake_case , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_snake_case , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase = [] UpperCAmelCase = [] def tee(_snake_case , _snake_case , _snake_case , _snake_case="" ): UpperCAmelCase = line.decode("""utf-8""" ).rstrip() sink.append(_snake_case ) if not quiet: print(_snake_case , _snake_case , file=_snake_case ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda _snake_case : tee(_snake_case , _snake_case , sys.stdout , label="""stdout:""" ) ), _read_stream(p.stderr , lambda _snake_case : tee(_snake_case , _snake_case , sys.stderr , label="""stderr:""" ) ), ] , timeout=_snake_case , ) return _RunOutput(await p.wait() , _snake_case , _snake_case ) def _a ( _snake_case , _snake_case=None , _snake_case=None , _snake_case=180 , _snake_case=False , _snake_case=True ): """simple docstring""" UpperCAmelCase = asyncio.get_event_loop() UpperCAmelCase = loop.run_until_complete( _stream_subprocess(_snake_case , env=_snake_case , stdin=_snake_case , timeout=_snake_case , quiet=_snake_case , echo=_snake_case ) ) UpperCAmelCase = """ """.join(_snake_case ) if result.returncode > 0: UpperCAmelCase = """\n""".join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(F'''\'{cmd_str}\' produced no output.''' ) return result def _a ( ): """simple docstring""" UpperCAmelCase = os.environ.get("""PYTEST_XDIST_WORKER""" , """gw0""" ) UpperCAmelCase = re.sub(R"""^gw""" , """""" , _snake_case , 0 , re.M ) return int(_snake_case ) def _a ( ): """simple docstring""" UpperCAmelCase = 2_9500 UpperCAmelCase = pytest_xdist_worker_id() return port + uniq_delta
74
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase = concatenate_datasets _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadManager _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
74
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """vinvino02/glpn-kitti""": """https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json""", # See all GLPN models at https://huggingface.co/models?filter=glpn } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''glpn''' def __init__( self ,A=3 ,A=4 ,A=[2, 2, 2, 2] ,A=[8, 4, 2, 1] ,A=[32, 64, 160, 256] ,A=[7, 3, 3, 3] ,A=[4, 2, 2, 2] ,A=[1, 2, 5, 8] ,A=[4, 4, 4, 4] ,A="gelu" ,A=0.0 ,A=0.0 ,A=0.02 ,A=0.1 ,A=1e-6 ,A=64 ,A=10 ,A=-1 ,**A ,): super().__init__(**A ) UpperCAmelCase = num_channels UpperCAmelCase = num_encoder_blocks UpperCAmelCase = depths UpperCAmelCase = sr_ratios UpperCAmelCase = hidden_sizes UpperCAmelCase = patch_sizes UpperCAmelCase = strides UpperCAmelCase = mlp_ratios UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = drop_path_rate UpperCAmelCase = layer_norm_eps UpperCAmelCase = decoder_hidden_size UpperCAmelCase = max_depth UpperCAmelCase = head_in_index
74
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCamelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") _UpperCamelCase = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) _UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the training data.'''} ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the validation data.'''} ) SCREAMING_SNAKE_CASE = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) SCREAMING_SNAKE_CASE = field(default=32 , metadata={'''help''': '''The size of the square patches to use for masking.'''} ) SCREAMING_SNAKE_CASE = field( default=0.6 , metadata={'''help''': '''Percentage of patches to mask.'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def _UpperCamelCase ( self ): UpperCAmelCase = {} if self.train_dir is not None: UpperCAmelCase = self.train_dir if self.validation_dir is not None: UpperCAmelCase = self.validation_dir UpperCAmelCase = data_files if data_files else None @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ''' '''checkpoint identifier on the hub. ''' '''Don\'t set if you want to train a model from scratch.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(snake_case )} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'''} , ) SCREAMING_SNAKE_CASE = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''Name or path of preprocessor config.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each image. If not specified, will use `image_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Stride to use for the encoder.'''} , ) class lowerCamelCase__ : def __init__( self ,A=192 ,A=32 ,A=4 ,A=0.6 ): UpperCAmelCase = input_size UpperCAmelCase = mask_patch_size UpperCAmelCase = model_patch_size UpperCAmelCase = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError("""Input size must be divisible by mask patch size""" ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError("""Mask patch size must be divisible by model patch size""" ) UpperCAmelCase = self.input_size // self.mask_patch_size UpperCAmelCase = self.mask_patch_size // self.model_patch_size UpperCAmelCase = self.rand_size**2 UpperCAmelCase = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self ): UpperCAmelCase = np.random.permutation(self.token_count )[: self.mask_count] UpperCAmelCase = np.zeros(self.token_count ,dtype=A ) UpperCAmelCase = 1 UpperCAmelCase = mask.reshape((self.rand_size, self.rand_size) ) UpperCAmelCase = mask.repeat(self.scale ,axis=0 ).repeat(self.scale ,axis=1 ) return torch.tensor(mask.flatten() ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = torch.stack([example["""pixel_values"""] for example in examples] ) UpperCAmelCase = torch.stack([example["""mask"""] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mim""" , _snake_case , _snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_snake_case ) transformers.utils.logging.set_verbosity(_snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. UpperCAmelCase = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _snake_case ) and data_args.train_val_split > 0.0: UpperCAmelCase = ds["""train"""].train_test_split(data_args.train_val_split ) UpperCAmelCase = split["""train"""] UpperCAmelCase = split["""test"""] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.config_name_or_path , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_snake_case , """decoder_type""" ): UpperCAmelCase = """simmim""" # adapt config UpperCAmelCase = model_args.image_size if model_args.image_size is not None else config.image_size UpperCAmelCase = model_args.patch_size if model_args.patch_size is not None else config.patch_size UpperCAmelCase = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { """image_size""": model_args.image_size, """patch_size""": model_args.patch_size, """encoder_stride""": model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } UpperCAmelCase = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: UpperCAmelCase = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) UpperCAmelCase = AutoModelForMaskedImageModeling.from_config(_snake_case ) if training_args.do_train: UpperCAmelCase = ds["""train"""].column_names else: UpperCAmelCase = ds["""validation"""].column_names if data_args.image_column_name is not None: UpperCAmelCase = data_args.image_column_name elif "image" in column_names: UpperCAmelCase = """image""" elif "img" in column_names: UpperCAmelCase = """img""" else: UpperCAmelCase = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py UpperCAmelCase = Compose( [ Lambda(lambda _snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator UpperCAmelCase = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_snake_case ): UpperCAmelCase = [transforms(_snake_case ) for image in examples[image_column_name]] UpperCAmelCase = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: UpperCAmelCase = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: UpperCAmelCase = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_snake_case ) # Initialize our trainer UpperCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase = last_checkpoint UpperCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase = trainer.evaluate() trainer.log_metrics("""eval""" , _snake_case ) trainer.save_metrics("""eval""" , _snake_case ) # Write model card and (optionally) push to hub UpperCAmelCase = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """masked-image-modeling""", """dataset""": data_args.dataset_name, """tags""": ["""masked-image-modeling"""], } if training_args.push_to_hub: trainer.push_to_hub(**_snake_case ) else: trainer.create_model_card(**_snake_case ) if __name__ == "__main__": main()
74
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCamelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") _UpperCamelCase = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) _UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the training data.'''} ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the validation data.'''} ) SCREAMING_SNAKE_CASE = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) SCREAMING_SNAKE_CASE = field(default=32 , metadata={'''help''': '''The size of the square patches to use for masking.'''} ) SCREAMING_SNAKE_CASE = field( default=0.6 , metadata={'''help''': '''Percentage of patches to mask.'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def _UpperCamelCase ( self ): UpperCAmelCase = {} if self.train_dir is not None: UpperCAmelCase = self.train_dir if self.validation_dir is not None: UpperCAmelCase = self.validation_dir UpperCAmelCase = data_files if data_files else None @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ''' '''checkpoint identifier on the hub. ''' '''Don\'t set if you want to train a model from scratch.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(snake_case )} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'''} , ) SCREAMING_SNAKE_CASE = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''Name or path of preprocessor config.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each image. If not specified, will use `image_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Stride to use for the encoder.'''} , ) class lowerCamelCase__ : def __init__( self ,A=192 ,A=32 ,A=4 ,A=0.6 ): UpperCAmelCase = input_size UpperCAmelCase = mask_patch_size UpperCAmelCase = model_patch_size UpperCAmelCase = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError("""Input size must be divisible by mask patch size""" ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError("""Mask patch size must be divisible by model patch size""" ) UpperCAmelCase = self.input_size // self.mask_patch_size UpperCAmelCase = self.mask_patch_size // self.model_patch_size UpperCAmelCase = self.rand_size**2 UpperCAmelCase = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self ): UpperCAmelCase = np.random.permutation(self.token_count )[: self.mask_count] UpperCAmelCase = np.zeros(self.token_count ,dtype=A ) UpperCAmelCase = 1 UpperCAmelCase = mask.reshape((self.rand_size, self.rand_size) ) UpperCAmelCase = mask.repeat(self.scale ,axis=0 ).repeat(self.scale ,axis=1 ) return torch.tensor(mask.flatten() ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = torch.stack([example["""pixel_values"""] for example in examples] ) UpperCAmelCase = torch.stack([example["""mask"""] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mim""" , _snake_case , _snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_snake_case ) transformers.utils.logging.set_verbosity(_snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. UpperCAmelCase = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _snake_case ) and data_args.train_val_split > 0.0: UpperCAmelCase = ds["""train"""].train_test_split(data_args.train_val_split ) UpperCAmelCase = split["""train"""] UpperCAmelCase = split["""test"""] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.config_name_or_path , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_snake_case , """decoder_type""" ): UpperCAmelCase = """simmim""" # adapt config UpperCAmelCase = model_args.image_size if model_args.image_size is not None else config.image_size UpperCAmelCase = model_args.patch_size if model_args.patch_size is not None else config.patch_size UpperCAmelCase = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { """image_size""": model_args.image_size, """patch_size""": model_args.patch_size, """encoder_stride""": model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } UpperCAmelCase = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: UpperCAmelCase = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) UpperCAmelCase = AutoModelForMaskedImageModeling.from_config(_snake_case ) if training_args.do_train: UpperCAmelCase = ds["""train"""].column_names else: UpperCAmelCase = ds["""validation"""].column_names if data_args.image_column_name is not None: UpperCAmelCase = data_args.image_column_name elif "image" in column_names: UpperCAmelCase = """image""" elif "img" in column_names: UpperCAmelCase = """img""" else: UpperCAmelCase = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py UpperCAmelCase = Compose( [ Lambda(lambda _snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator UpperCAmelCase = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_snake_case ): UpperCAmelCase = [transforms(_snake_case ) for image in examples[image_column_name]] UpperCAmelCase = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: UpperCAmelCase = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: UpperCAmelCase = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_snake_case ) # Initialize our trainer UpperCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase = last_checkpoint UpperCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase = trainer.evaluate() trainer.log_metrics("""eval""" , _snake_case ) trainer.save_metrics("""eval""" , _snake_case ) # Write model card and (optionally) push to hub UpperCAmelCase = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """masked-image-modeling""", """dataset""": data_args.dataset_name, """tags""": ["""masked-image-modeling"""], } if training_args.push_to_hub: trainer.push_to_hub(**_snake_case ) else: trainer.create_model_card(**_snake_case ) if __name__ == "__main__": main()
74
1
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _UpperCamelCase = 16 _UpperCamelCase = 32 def _a ( _snake_case , _snake_case = 16 ): """simple docstring""" UpperCAmelCase = AutoTokenizer.from_pretrained("""bert-base-cased""" ) UpperCAmelCase = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(_snake_case ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=_snake_case , max_length=_snake_case ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCAmelCase = datasets.map( _snake_case , batched=_snake_case , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(_snake_case ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCAmelCase = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCAmelCase = 16 elif accelerator.mixed_precision != "no": UpperCAmelCase = 8 else: UpperCAmelCase = None return tokenizer.pad( _snake_case , padding="""longest""" , max_length=_snake_case , pad_to_multiple_of=_snake_case , return_tensors="""pt""" , ) # Instantiate dataloaders. UpperCAmelCase = DataLoader( tokenized_datasets["""train"""] , shuffle=_snake_case , collate_fn=_snake_case , batch_size=_snake_case ) UpperCAmelCase = DataLoader( tokenized_datasets["""validation"""] , shuffle=_snake_case , collate_fn=_snake_case , batch_size=_snake_case ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _UpperCamelCase = mocked_dataloaders # noqa: F811 def _a ( _snake_case , _snake_case ): """simple docstring""" if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , _snake_case ) == "1": UpperCAmelCase = 2 # New Code # UpperCAmelCase = int(args.gradient_accumulation_steps ) UpperCAmelCase = int(args.local_sgd_steps ) # Initialize accelerator UpperCAmelCase = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_snake_case ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError("""LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)""" ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase = config["""lr"""] UpperCAmelCase = int(config["""num_epochs"""] ) UpperCAmelCase = int(config["""seed"""] ) UpperCAmelCase = int(config["""batch_size"""] ) UpperCAmelCase = evaluate.load("""glue""" , """mrpc""" ) set_seed(_snake_case ) UpperCAmelCase , UpperCAmelCase = get_dataloaders(_snake_case , _snake_case ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=_snake_case ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCAmelCase = model.to(accelerator.device ) # Instantiate optimizer UpperCAmelCase = AdamW(params=model.parameters() , lr=_snake_case ) # Instantiate scheduler UpperCAmelCase = get_linear_schedule_with_warmup( optimizer=_snake_case , num_warmup_steps=100 , num_training_steps=(len(_snake_case ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) # Now we train the model for epoch in range(_snake_case ): model.train() with LocalSGD( accelerator=_snake_case , model=_snake_case , local_sgd_steps=_snake_case , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(_snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_snake_case ): UpperCAmelCase = model(**_snake_case ) UpperCAmelCase = output.loss accelerator.backward(_snake_case ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(_snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase = model(**_snake_case ) UpperCAmelCase = outputs.logits.argmax(dim=-1 ) UpperCAmelCase , UpperCAmelCase = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=_snake_case , references=_snake_case , ) UpperCAmelCase = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , _snake_case ) def _a ( ): """simple docstring""" UpperCAmelCase = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=_snake_case , default=_snake_case , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) # New Code # parser.add_argument( """--gradient_accumulation_steps""" , type=_snake_case , default=1 , help="""The number of minibatches to be ran before gradients are accumulated.""" , ) parser.add_argument( """--local_sgd_steps""" , type=_snake_case , default=8 , help="""Number of local SGD steps or None to disable local SGD""" ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) UpperCAmelCase = parser.parse_args() UpperCAmelCase = {"""lr""": 2E-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(_snake_case , _snake_case ) if __name__ == "__main__": main()
74
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): _UpperCamelCase = True from torch.cuda.amp import autocast _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to log verbose messages or not.'''} , ) SCREAMING_SNAKE_CASE = field( default=2.0 , metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.5 , metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.99_99_95 , metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def _a ( _snake_case , _snake_case ): """simple docstring""" logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase = logging.WARNING if model_args.verbose_logging: UpperCAmelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase = logging.INFO logger.setLevel(_snake_case ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default='''train''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) SCREAMING_SNAKE_CASE = field( default='''validation''' , metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) } , ) SCREAMING_SNAKE_CASE = field( default='''file''' , metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) SCREAMING_SNAKE_CASE = field( default=1 , metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) SCREAMING_SNAKE_CASE = field( default=20.0 , metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = "longest" SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __call__( self ,A ): # reformat list to dict and set to pytorch format UpperCAmelCase = self.feature_extractor.pad( A ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="""pt""" ,) UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""input_values"""].shape[-1] ) UpperCAmelCase = batch["""input_values"""].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""attention_mask"""].sum(-1 ) ).to( torch.long ) UpperCAmelCase = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["""input_values"""].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase = 1 UpperCAmelCase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=A ,min_masks=2 ,) return batch class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=1 ,A=0 ,A=1.0 ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = 0 UpperCAmelCase = max_gumbel_temp UpperCAmelCase = min_gumbel_temp UpperCAmelCase = gumbel_temp_decay def _UpperCamelCase ( self ,A ,A ): model.train() UpperCAmelCase = self._prepare_inputs(A ) if self.use_amp: with autocast(): UpperCAmelCase = self.compute_loss(A ,A ) else: UpperCAmelCase = self.compute_loss(A ,A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase = loss.sum() / (inputs["""mask_time_indices"""]).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() configure_logger(_snake_case , _snake_case ) # Downloading and loading a dataset from the hub. UpperCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="""validation""" , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=_snake_case ) def prepare_dataset(_snake_case ): # check that all files have the correct sampling rate UpperCAmelCase , UpperCAmelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase = datasets.map( _snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["""train"""].column_names ) # filter audio files that are too long UpperCAmelCase = vectorized_datasets.filter( lambda _snake_case : len(data["""speech"""] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(_snake_case ): return feature_extractor(batch["""speech"""] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase = vectorized_datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["""train"""].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( """PreTraining is only supported for ``config.do_stable_layer_norm=True`` and""" """ ``config.feat_extract_norm='layer'""" ) UpperCAmelCase = WavaVecaForPreTraining(_snake_case ) UpperCAmelCase = DataCollatorForWavaVecaPretraining(model=_snake_case , feature_extractor=_snake_case ) UpperCAmelCase = WavaVecaPreTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , train_dataset=vectorized_datasets["""train"""] , eval_dataset=vectorized_datasets["""validation"""] , tokenizer=_snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
74
1
"""simple docstring""" import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = XCLIPTextConfig() # derive patch size from model name UpperCAmelCase = model_name.find("""patch""" ) UpperCAmelCase = int(model_name[start_idx + len("""patch""" ) : start_idx + len("""patch""" ) + 2] ) UpperCAmelCase = XCLIPVisionConfig(patch_size=_snake_case , num_frames=_snake_case ) if "large" in model_name: UpperCAmelCase = 768 UpperCAmelCase = 3072 UpperCAmelCase = 12 UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 16 UpperCAmelCase = 24 UpperCAmelCase = 768 UpperCAmelCase = 3072 if model_name == "xclip-large-patch14-16-frames": UpperCAmelCase = 336 UpperCAmelCase = XCLIPConfig.from_text_vision_configs(_snake_case , _snake_case ) if "large" in model_name: UpperCAmelCase = 768 return config def _a ( _snake_case ): """simple docstring""" if name == "token_embedding.weight": UpperCAmelCase = name.replace("""token_embedding.weight""" , """text_model.embeddings.token_embedding.weight""" ) if name == "positional_embedding": UpperCAmelCase = name.replace("""positional_embedding""" , """text_model.embeddings.position_embedding.weight""" ) if "ln_1" in name: UpperCAmelCase = name.replace("""ln_1""" , """layer_norm1""" ) if "ln_2" in name: UpperCAmelCase = name.replace("""ln_2""" , """layer_norm2""" ) if "c_fc" in name: UpperCAmelCase = name.replace("""c_fc""" , """fc1""" ) if "c_proj" in name: UpperCAmelCase = name.replace("""c_proj""" , """fc2""" ) if name.startswith("""transformer.resblocks""" ): UpperCAmelCase = name.replace("""transformer.resblocks""" , """text_model.encoder.layers""" ) if "attn.out_proj" in name and "message" not in name: UpperCAmelCase = name.replace("""attn.out_proj""" , """self_attn.out_proj""" ) if "ln_final" in name: UpperCAmelCase = name.replace("""ln_final""" , """text_model.final_layer_norm""" ) # visual encoder if name == "visual.class_embedding": UpperCAmelCase = name.replace("""visual.class_embedding""" , """vision_model.embeddings.class_embedding""" ) if name == "visual.positional_embedding": UpperCAmelCase = name.replace("""visual.positional_embedding""" , """vision_model.embeddings.position_embedding.weight""" ) if name.startswith("""visual.transformer.resblocks""" ): UpperCAmelCase = name.replace("""visual.transformer.resblocks""" , """vision_model.encoder.layers""" ) if "visual.conv1" in name: UpperCAmelCase = name.replace("""visual.conv1""" , """vision_model.embeddings.patch_embedding""" ) if "visual.ln_pre" in name: UpperCAmelCase = name.replace("""visual.ln_pre""" , """vision_model.pre_layernorm""" ) if "visual.ln_post" in name: UpperCAmelCase = name.replace("""visual.ln_post""" , """vision_model.post_layernorm""" ) if "visual.proj" in name: UpperCAmelCase = name.replace("""visual.proj""" , """visual_projection.weight""" ) if "text_projection" in name: UpperCAmelCase = name.replace("""text_projection""" , """text_projection.weight""" ) # things on top if "prompts_visual_proj" in name: UpperCAmelCase = name.replace("""prompts_visual_proj""" , """prompts_visual_projection""" ) if "prompts_visual_ln" in name: UpperCAmelCase = name.replace("""prompts_visual_ln""" , """prompts_visual_layernorm""" ) # mit if name == "mit.positional_embedding": UpperCAmelCase = name.replace("""positional""" , """position""" ) if name.startswith("""mit.resblocks""" ): UpperCAmelCase = name.replace("""mit.resblocks""" , """mit.encoder.layers""" ) # prompts generator if name.startswith("""prompts_generator.norm""" ): UpperCAmelCase = name.replace("""prompts_generator.norm""" , """prompts_generator.layernorm""" ) return name def _a ( _snake_case , _snake_case ): """simple docstring""" for key in orig_state_dict.copy().keys(): UpperCAmelCase = orig_state_dict.pop(_snake_case ) if "attn.in_proj" in key: UpperCAmelCase = key.split(""".""" ) if key.startswith("""visual""" ): UpperCAmelCase = key_split[3] UpperCAmelCase = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: UpperCAmelCase = val[ :dim, : ] UpperCAmelCase = val[ dim : dim * 2, : ] UpperCAmelCase = val[ -dim:, : ] else: UpperCAmelCase = val[ :dim ] UpperCAmelCase = val[ dim : dim * 2 ] UpperCAmelCase = val[ -dim: ] else: if "weight" in key: UpperCAmelCase = val[ :dim, : ] UpperCAmelCase = val[ dim : dim * 2, : ] UpperCAmelCase = val[ -dim:, : ] else: UpperCAmelCase = val[:dim] UpperCAmelCase = val[ dim : dim * 2 ] UpperCAmelCase = val[-dim:] elif key.startswith("""mit""" ): UpperCAmelCase = key_split[2] UpperCAmelCase = config.vision_config.mit_hidden_size if "weight" in key: UpperCAmelCase = val[:dim, :] UpperCAmelCase = val[dim : dim * 2, :] UpperCAmelCase = val[-dim:, :] else: UpperCAmelCase = val[:dim] UpperCAmelCase = val[dim : dim * 2] UpperCAmelCase = val[-dim:] else: UpperCAmelCase = key_split[2] UpperCAmelCase = config.text_config.hidden_size if "weight" in key: UpperCAmelCase = val[:dim, :] UpperCAmelCase = val[ dim : dim * 2, : ] UpperCAmelCase = val[-dim:, :] else: UpperCAmelCase = val[:dim] UpperCAmelCase = val[ dim : dim * 2 ] UpperCAmelCase = val[-dim:] else: UpperCAmelCase = rename_key(_snake_case ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: UpperCAmelCase = val.T UpperCAmelCase = val return orig_state_dict def _a ( _snake_case ): """simple docstring""" if num_frames == 8: UpperCAmelCase = """eating_spaghetti_8_frames.npy""" elif num_frames == 16: UpperCAmelCase = """eating_spaghetti.npy""" elif num_frames == 32: UpperCAmelCase = """eating_spaghetti_32_frames.npy""" UpperCAmelCase = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""" , filename=_snake_case , repo_type="""dataset""" , ) UpperCAmelCase = np.load(_snake_case ) return list(_snake_case ) def _a ( _snake_case , _snake_case=None , _snake_case=False ): """simple docstring""" UpperCAmelCase = { # fully supervised kinetics-400 checkpoints """xclip-base-patch32""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth""", """xclip-base-patch32-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth""" ), """xclip-base-patch16""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth""", """xclip-base-patch16-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth""" ), """xclip-large-patch14""": """https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb""", """xclip-large-patch14-16-frames""": """https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f""", # fully supervised kinetics-600 checkpoints """xclip-base-patch16-kinetics-600""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth""" ), """xclip-base-patch16-kinetics-600-16-frames""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth""" ), """xclip-large-patch14-kinetics-600""": """https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be""", # few shot """xclip-base-patch16-hmdb-2-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth""" ), """xclip-base-patch16-hmdb-4-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth""" ), """xclip-base-patch16-hmdb-8-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth""" ), """xclip-base-patch16-hmdb-16-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth""" ), """xclip-base-patch16-ucf-2-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth""" ), """xclip-base-patch16-ucf-4-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth""" ), """xclip-base-patch16-ucf-8-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth""" ), """xclip-base-patch16-ucf-16-shot""": ( """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth""" ), # zero shot """xclip-base-patch16-zero-shot""": """https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth""", } UpperCAmelCase = model_to_url[model_name] UpperCAmelCase = 8 if "16-frames" in model_name: UpperCAmelCase = 16 elif "shot" in model_name: UpperCAmelCase = 32 UpperCAmelCase = get_xclip_config(_snake_case , _snake_case ) UpperCAmelCase = XCLIPModel(_snake_case ) model.eval() if "drive" in checkpoint_url: UpperCAmelCase = """pytorch_model.bin""" gdown.cached_download(_snake_case , _snake_case , quiet=_snake_case ) UpperCAmelCase = torch.load(_snake_case , map_location="""cpu""" )["""model"""] else: UpperCAmelCase = torch.hub.load_state_dict_from_url(_snake_case )["""model"""] UpperCAmelCase = convert_state_dict(_snake_case , _snake_case ) UpperCAmelCase = XCLIPModel(_snake_case ) UpperCAmelCase , UpperCAmelCase = model.load_state_dict(_snake_case , strict=_snake_case ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() UpperCAmelCase = 336 if model_name == """xclip-large-patch14-16-frames""" else 224 UpperCAmelCase = VideoMAEImageProcessor(size=_snake_case ) UpperCAmelCase = CLIPTokenizer.from_pretrained("""openai/clip-vit-base-patch32""" ) UpperCAmelCase = CLIPTokenizerFast.from_pretrained("""openai/clip-vit-base-patch32""" ) UpperCAmelCase = XCLIPProcessor(image_processor=_snake_case , tokenizer=_snake_case ) UpperCAmelCase = prepare_video(_snake_case ) UpperCAmelCase = processor( text=["""playing sports""", """eating spaghetti""", """go shopping"""] , videos=_snake_case , return_tensors="""pt""" , padding=_snake_case ) print("""Shape of pixel values:""" , inputs.pixel_values.shape ) with torch.no_grad(): UpperCAmelCase = model(**_snake_case ) # Verify outputs UpperCAmelCase = outputs.logits_per_video UpperCAmelCase = logits_per_video.softmax(dim=1 ) print("""Probs:""" , _snake_case ) # kinetics-400 if model_name == "xclip-base-patch32": UpperCAmelCase = torch.tensor([[0.0019, 0.9951, 0.0030]] ) elif model_name == "xclip-base-patch32-16-frames": UpperCAmelCase = torch.tensor([[7.0999E-04, 9.9883E-01, 4.5580E-04]] ) elif model_name == "xclip-base-patch16": UpperCAmelCase = torch.tensor([[0.0083, 0.9681, 0.0236]] ) elif model_name == "xclip-base-patch16-16-frames": UpperCAmelCase = torch.tensor([[7.6937E-04, 9.9728E-01, 1.9473E-03]] ) elif model_name == "xclip-large-patch14": UpperCAmelCase = torch.tensor([[0.0062, 0.9864, 0.0075]] ) elif model_name == "xclip-large-patch14-16-frames": UpperCAmelCase = torch.tensor([[3.3877E-04, 9.9937E-01, 2.8888E-04]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": UpperCAmelCase = torch.tensor([[0.0555, 0.8914, 0.0531]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": UpperCAmelCase = torch.tensor([[3.8554E-04, 9.9929E-01, 3.2754E-04]] ) elif model_name == "xclip-large-patch14-kinetics-600": UpperCAmelCase = torch.tensor([[0.0036, 0.9920, 0.0045]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": UpperCAmelCase = torch.tensor([[7.1890E-06, 9.9994E-01, 5.6559E-05]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": UpperCAmelCase = torch.tensor([[1.0320E-05, 9.9993E-01, 6.2435E-05]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": UpperCAmelCase = torch.tensor([[4.1377E-06, 9.9990E-01, 9.8386E-05]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": UpperCAmelCase = torch.tensor([[4.1347E-05, 9.9962E-01, 3.3411E-04]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": UpperCAmelCase = torch.tensor([[8.5857E-05, 9.9928E-01, 6.3291E-04]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": UpperCAmelCase = torch.tensor([[8.5857E-05, 9.9928E-01, 6.3291E-04]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": UpperCAmelCase = torch.tensor([[0.0027, 0.9904, 0.0070]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": UpperCAmelCase = torch.tensor([[9.8219E-04, 9.9593E-01, 3.0863E-03]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": UpperCAmelCase = torch.tensor([[3.5082E-04, 9.9785E-01, 1.7966E-03]] ) else: raise ValueError(F'''Model name {model_name} not supported''' ) assert torch.allclose(_snake_case , _snake_case , atol=1E-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) if push_to_hub: print("""Pushing model, processor and slow tokenizer files to the hub...""" ) model.push_to_hub(_snake_case , organization="""nielsr""" ) processor.push_to_hub(_snake_case , organization="""nielsr""" ) slow_tokenizer.push_to_hub(_snake_case , organization="""nielsr""" ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""xclip-base-patch32""", type=str, help="""Name of the model.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
74
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = 30 UpperCAmelCase = self.seq_length + self.mem_len UpperCAmelCase = 15 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 99 UpperCAmelCase = [10, 50, 80] UpperCAmelCase = 32 UpperCAmelCase = 32 UpperCAmelCase = 4 UpperCAmelCase = 8 UpperCAmelCase = 128 UpperCAmelCase = 2 UpperCAmelCase = 2 UpperCAmelCase = None UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 3 UpperCAmelCase = self.vocab_size - 1 UpperCAmelCase = 0.01 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCamelCase ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLLMHeadModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase , UpperCAmelCase = model([input_ids_a, mems_a] ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLForSequenceClassification(A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids_a} return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = () if is_tf_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,d_embed=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A ) def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: UpperCAmelCase = model_class(A ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: UpperCAmelCase = model.get_output_embeddings() assert isinstance(A ,tf.keras.layers.Layer ) UpperCAmelCase = model.get_bias() assert name is None else: UpperCAmelCase = model.get_output_embeddings() assert x is None UpperCAmelCase = model.get_bias() assert name is None def _UpperCamelCase ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _UpperCamelCase ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFTransfoXLModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""This model doesn't play well with fit() due to not returning a single loss.""" ) def _UpperCamelCase ( self ): pass @require_tf class lowerCamelCase__ ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off UpperCAmelCase = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off UpperCAmelCase = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> UpperCAmelCase = model.generate(A ,max_length=200 ,do_sample=A ) self.assertListEqual(output_ids[0].numpy().tolist() ,A )
74
1
"""simple docstring""" import datasets from .evaluate import evaluate _UpperCamelCase = """\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } """ _UpperCamelCase = """ This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. """ _UpperCamelCase = """ Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair as given in the references (see below) - 'prediction_text': the text of the answer references: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair (see above), - 'answers': a Dict in the SQuAD dataset format { 'text': list of possible texts for the answer, as a list of strings 'answer_start': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: 'exact_match': Exact match (the normalized answer exactly match the gold answer) 'f1': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{'prediction_text': '1976', 'id': '56e10a3be3433e1400422b22'}] >>> references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': '56e10a3be3433e1400422b22'}] >>> squad_metric = datasets.load_metric(\"squad\") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 100.0, 'f1': 100.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase__ ( datasets.Metric ): def _UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) ,codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] ,reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] ,) def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} UpperCAmelCase = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] UpperCAmelCase = evaluate(dataset=A ,predictions=A ) return score
74
"""simple docstring""" from math import sqrt def _a ( _snake_case = 100_0000 ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_snake_case , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
74
1
"""simple docstring""" import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class lowerCamelCase__ : def __init__( self ,A = "cpu" ,A = "openai/clip-vit-large-patch14" ): UpperCAmelCase = device UpperCAmelCase = CLIPTokenizerFast.from_pretrained(A ) UpperCAmelCase = [0.48145466, 0.4578275, 0.40821073] UpperCAmelCase = [0.26862954, 0.26130258, 0.27577711] UpperCAmelCase = torchvision.transforms.Normalize(self.image_mean ,self.image_std ) UpperCAmelCase = torchvision.transforms.Resize(224 ) UpperCAmelCase = torchvision.transforms.CenterCrop(224 ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.resize(A ) UpperCAmelCase = self.center_crop(A ) UpperCAmelCase = self.normalize(A ) return images def __call__( self ,A=None ,A=None ,**A ): UpperCAmelCase = self.tokenizer(text=A ,**A ) UpperCAmelCase = self.preprocess_img(A ) UpperCAmelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class lowerCamelCase__ ( nn.Module ): def __init__( self ,A=10 ,A=0.01 ,A=None ,A=None ,A=None ,A=None ,A=None ,A=None ,A=False ,A=True ,A="image" ,A=True ,A=False ,A=False ,A=False ,): super().__init__() UpperCAmelCase = None UpperCAmelCase = device if device else get_device() if vqgan: UpperCAmelCase = vqgan else: UpperCAmelCase = load_vqgan(self.device ,conf_path=A ,ckpt_path=A ) self.vqgan.eval() if clip: UpperCAmelCase = clip else: UpperCAmelCase = CLIPModel.from_pretrained("""openai/clip-vit-base-patch32""" ) self.clip.to(self.device ) UpperCAmelCase = ProcessorGradientFlow(device=self.device ) UpperCAmelCase = iterations UpperCAmelCase = lr UpperCAmelCase = log UpperCAmelCase = make_grid UpperCAmelCase = return_val UpperCAmelCase = quantize UpperCAmelCase = self.vqgan.decoder.z_shape def _UpperCamelCase ( self ,A=None ,A=None ,A=5 ,A=True ): UpperCAmelCase = [] if output_path is None: UpperCAmelCase = """./animation.gif""" if input_path is None: UpperCAmelCase = self.save_path UpperCAmelCase = sorted(glob(input_path + """/*""" ) ) if not len(A ): raise ValueError( """No images found in save path, aborting (did you pass save_intermediate=True to the generate""" """ function?)""" ) if len(A ) == 1: print("""Only one image found in save path, (did you pass save_intermediate=True to the generate function?)""" ) UpperCAmelCase = total_duration / len(A ) UpperCAmelCase = [frame_duration] * len(A ) if extend_frames: UpperCAmelCase = 1.5 UpperCAmelCase = 3 for file_name in paths: if file_name.endswith(""".png""" ): images.append(imageio.imread(A ) ) imageio.mimsave(A ,A ,duration=A ) print(F'''gif saved to {output_path}''' ) def _UpperCamelCase ( self ,A=None ,A=None ): if not (path or img): raise ValueError("""Input either path or tensor""" ) if img is not None: raise NotImplementedError UpperCAmelCase = preprocess(Image.open(A ) ,target_image_size=256 ).to(self.device ) UpperCAmelCase = preprocess_vqgan(A ) UpperCAmelCase , *UpperCAmelCase = self.vqgan.encode(A ) return z def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.latent.detach().requires_grad_() UpperCAmelCase = base_latent + transform_vector if self.quantize: UpperCAmelCase , *UpperCAmelCase = self.vqgan.quantize(A ) else: UpperCAmelCase = trans_latent return self.vqgan.decode(A ) def _UpperCamelCase ( self ,A ,A ,A=None ): UpperCAmelCase = self.clip_preprocessor(text=A ,images=A ,return_tensors="""pt""" ,padding=A ) UpperCAmelCase = self.clip(**A ) UpperCAmelCase = clip_outputs.logits_per_image if weights is not None: UpperCAmelCase = similarity_logits * weights return similarity_logits.sum() def _UpperCamelCase ( self ,A ,A ,A ): UpperCAmelCase = self._get_clip_similarity(pos_prompts["""prompts"""] ,A ,weights=(1 / pos_prompts["""weights"""]) ) if neg_prompts: UpperCAmelCase = self._get_clip_similarity(neg_prompts["""prompts"""] ,A ,weights=neg_prompts["""weights"""] ) else: UpperCAmelCase = torch.tensor([1] ,device=self.device ) UpperCAmelCase = -torch.log(A ) + torch.log(A ) return loss def _UpperCamelCase ( self ,A ,A ,A ): UpperCAmelCase = torch.randn_like(self.latent ,requires_grad=A ,device=self.device ) UpperCAmelCase = torch.optim.Adam([vector] ,lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() UpperCAmelCase = self._add_vector(A ) UpperCAmelCase = loop_post_process(A ) UpperCAmelCase = self._get_CLIP_loss(A ,A ,A ) print("""CLIP loss""" ,A ) if self.log: wandb.log({"""CLIP Loss""": clip_loss} ) clip_loss.backward(retain_graph=A ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def _UpperCamelCase ( self ,A ,A ,A ): wandb.init(reinit=A ,project="""face-editor""" ) wandb.config.update({"""Positive Prompts""": positive_prompts} ) wandb.config.update({"""Negative Prompts""": negative_prompts} ) wandb.config.update({"""lr""": self.lr, """iterations""": self.iterations} ) if image_path: UpperCAmelCase = Image.open(A ) UpperCAmelCase = image.resize((256, 256) ) wandb.log("""Original Image""" ,wandb.Image(A ) ) def _UpperCamelCase ( self ,A ): if not prompts: return [] UpperCAmelCase = [] UpperCAmelCase = [] if isinstance(A ,A ): UpperCAmelCase = [prompt.strip() for prompt in prompts.split("""|""" )] for prompt in prompts: if isinstance(A ,(tuple, list) ): UpperCAmelCase = prompt[0] UpperCAmelCase = float(prompt[1] ) elif ":" in prompt: UpperCAmelCase , UpperCAmelCase = prompt.split(""":""" ) UpperCAmelCase = float(A ) else: UpperCAmelCase = prompt UpperCAmelCase = 1.0 processed_prompts.append(A ) weights.append(A ) return { "prompts": processed_prompts, "weights": torch.tensor(A ,device=self.device ), } def _UpperCamelCase ( self ,A ,A=None ,A=None ,A=True ,A=False ,A=True ,A=True ,A=None ,): if image_path: UpperCAmelCase = self._get_latent(A ) else: UpperCAmelCase = torch.randn(self.latent_dim ,device=self.device ) if self.log: self._init_logging(A ,A ,A ) assert pos_prompts, "You must provide at least one positive prompt." UpperCAmelCase = self.process_prompts(A ) UpperCAmelCase = self.process_prompts(A ) if save_final and save_path is None: UpperCAmelCase = os.path.join("""./outputs/""" ,"""_""".join(pos_prompts["""prompts"""] ) ) if not os.path.exists(A ): os.makedirs(A ) else: UpperCAmelCase = save_path + """_""" + get_timestamp() os.makedirs(A ) UpperCAmelCase = save_path UpperCAmelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print("""Original Image""" ) show_pil(custom_to_pil(A ) ) UpperCAmelCase = loop_post_process(A ) for iter, transformed_img in enumerate(self._optimize_CLIP(A ,A ,A ) ): if show_intermediate: show_pil(A ) if save_intermediate: transformed_img.save(os.path.join(self.save_path ,F'''iter_{iter:03d}.png''' ) ) if self.log: wandb.log({"""Image""": wandb.Image(A )} ) if show_final: show_pil(A ) if save_final: transformed_img.save(os.path.join(self.save_path ,F'''iter_{iter:03d}_final.png''' ) )
74
"""simple docstring""" import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _UpperCamelCase = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } _UpperCamelCase = AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def _a ( _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = create_model( """HTSAT-tiny""" , """roberta""" , _snake_case , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=_snake_case , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = {} UpperCAmelCase = R""".*sequential.(\d+).*""" UpperCAmelCase = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase = key.replace(_snake_case , _snake_case ) if re.match(_snake_case , _snake_case ): # replace sequential layers with list UpperCAmelCase = re.match(_snake_case , _snake_case ).group(1 ) UpperCAmelCase = key.replace(F'''sequential.{sequential_layer}.''' , F'''layers.{int(_snake_case )//3}.linear.''' ) elif re.match(_snake_case , _snake_case ): UpperCAmelCase = int(re.match(_snake_case , _snake_case ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... UpperCAmelCase = 1 if projecton_layer == 0 else 2 UpperCAmelCase = key.replace(F'''_projection.{projecton_layer}.''' , F'''_projection.linear{transformers_projection_layer}.''' ) if "audio" and "qkv" in key: # split qkv into query key and value UpperCAmelCase = value UpperCAmelCase = mixed_qkv.size(0 ) // 3 UpperCAmelCase = mixed_qkv[:qkv_dim] UpperCAmelCase = mixed_qkv[qkv_dim : qkv_dim * 2] UpperCAmelCase = mixed_qkv[qkv_dim * 2 :] UpperCAmelCase = query_layer UpperCAmelCase = key_layer UpperCAmelCase = value_layer else: UpperCAmelCase = value return model_state_dict def _a ( _snake_case , _snake_case , _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = init_clap(_snake_case , enable_fusion=_snake_case ) clap_model.eval() UpperCAmelCase = clap_model.state_dict() UpperCAmelCase = rename_state_dict(_snake_case ) UpperCAmelCase = ClapConfig() UpperCAmelCase = enable_fusion UpperCAmelCase = ClapModel(_snake_case ) # ignore the spectrogram embedding layer model.load_state_dict(_snake_case , strict=_snake_case ) model.save_pretrained(_snake_case ) transformers_config.save_pretrained(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") _UpperCamelCase = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
74
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """microsoft/wavlm-base""": """https://huggingface.co/microsoft/wavlm-base/resolve/main/config.json""", # See all WavLM models at https://huggingface.co/models?filter=wavlm } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''wavlm''' def __init__( self ,A=32 ,A=768 ,A=12 ,A=12 ,A=3_072 ,A="gelu" ,A=0.1 ,A=0.1 ,A=0.1 ,A=0.0 ,A=0.1 ,A=0.1 ,A=0.02 ,A=1e-5 ,A="group" ,A="gelu" ,A=(512, 512, 512, 512, 512, 512, 512) ,A=(5, 2, 2, 2, 2, 2, 2) ,A=(10, 3, 3, 3, 3, 2, 2) ,A=False ,A=128 ,A=16 ,A=320 ,A=800 ,A=False ,A=True ,A=0.05 ,A=10 ,A=2 ,A=0.0 ,A=10 ,A=320 ,A=2 ,A=0.1 ,A=100 ,A=256 ,A=256 ,A=0.1 ,A="mean" ,A=False ,A=False ,A=256 ,A=(512, 512, 512, 512, 1_500) ,A=(5, 3, 3, 1, 1) ,A=(1, 2, 3, 1, 1) ,A=512 ,A=80 ,A=0 ,A=1 ,A=2 ,A=False ,A=3 ,A=2 ,A=3 ,A=None ,**A ,): super().__init__(**A ,pad_token_id=A ,bos_token_id=A ,eos_token_id=A ) UpperCAmelCase = hidden_size UpperCAmelCase = feat_extract_norm UpperCAmelCase = feat_extract_activation UpperCAmelCase = list(A ) UpperCAmelCase = list(A ) UpperCAmelCase = list(A ) UpperCAmelCase = conv_bias UpperCAmelCase = num_buckets UpperCAmelCase = max_bucket_distance UpperCAmelCase = num_conv_pos_embeddings UpperCAmelCase = num_conv_pos_embedding_groups UpperCAmelCase = len(self.conv_dim ) UpperCAmelCase = num_hidden_layers UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = feat_proj_dropout UpperCAmelCase = final_dropout UpperCAmelCase = layerdrop UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = num_ctc_classes UpperCAmelCase = vocab_size UpperCAmelCase = do_stable_layer_norm UpperCAmelCase = use_weighted_layer_sum UpperCAmelCase = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase = apply_spec_augment UpperCAmelCase = mask_time_prob UpperCAmelCase = mask_time_length UpperCAmelCase = mask_time_min_masks UpperCAmelCase = mask_feature_prob UpperCAmelCase = mask_feature_length # parameters for pretraining with codevector quantized representations UpperCAmelCase = num_codevectors_per_group UpperCAmelCase = num_codevector_groups UpperCAmelCase = contrastive_logits_temperature UpperCAmelCase = num_negatives UpperCAmelCase = codevector_dim UpperCAmelCase = proj_codevector_dim UpperCAmelCase = diversity_loss_weight # ctc loss UpperCAmelCase = ctc_loss_reduction UpperCAmelCase = ctc_zero_infinity # adapter UpperCAmelCase = add_adapter UpperCAmelCase = adapter_kernel_size UpperCAmelCase = adapter_stride UpperCAmelCase = num_adapter_layers UpperCAmelCase = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase = list(A ) UpperCAmelCase = list(A ) UpperCAmelCase = list(A ) UpperCAmelCase = xvector_output_dim @property def _UpperCamelCase ( self ): return functools.reduce(operator.mul ,self.conv_stride ,1 )
74
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _UpperCamelCase = abspath(join(dirname(dirname(__file__)), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_snake_case ) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_terminal_summary_main UpperCAmelCase = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(_snake_case , id=_snake_case )
74
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """alibaba-damo/mgp-str-base""": """https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json""", } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''mgp-str''' def __init__( self ,A=[32, 128] ,A=4 ,A=3 ,A=27 ,A=38 ,A=50_257 ,A=30_522 ,A=768 ,A=12 ,A=12 ,A=4.0 ,A=True ,A=False ,A=1e-5 ,A=0.0 ,A=0.0 ,A=0.0 ,A=False ,A=0.02 ,**A ,): super().__init__(**A ) UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = max_token_length UpperCAmelCase = num_character_labels UpperCAmelCase = num_bpe_labels UpperCAmelCase = num_wordpiece_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = mlp_ratio UpperCAmelCase = distilled UpperCAmelCase = layer_norm_eps UpperCAmelCase = drop_rate UpperCAmelCase = qkv_bias UpperCAmelCase = attn_drop_rate UpperCAmelCase = drop_path_rate UpperCAmelCase = output_aa_attentions UpperCAmelCase = initializer_range
74
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class lowerCamelCase__ : def __init__( self ,A ,A ): if len(A ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) UpperCAmelCase = list(A ) UpperCAmelCase = degree def __add__( self ,A ): if self.degree > polynomial_a.degree: UpperCAmelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,A ) else: UpperCAmelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,A ) def __sub__( self ,A ): return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self ): return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self ,A ): UpperCAmelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): UpperCAmelCase = """""" for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(A ) return polynomial def __repr__( self ): return self.__str__() def _UpperCamelCase ( self ): UpperCAmelCase = [0] * self.degree for i in range(self.degree ): UpperCAmelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,A ) def _UpperCamelCase ( self ,A = 0 ): UpperCAmelCase = [0] * (self.degree + 2) UpperCAmelCase = constant for i in range(self.degree + 1 ): UpperCAmelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,A ) def __eq__( self ,A ): if not isinstance(A ,A ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self ,A ): return not self.__eq__(A )
74
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = SwinvaConfig() UpperCAmelCase = swinva_name.split("""_""" ) UpperCAmelCase = name_split[1] if "to" in name_split[3]: UpperCAmelCase = int(name_split[3][-3:] ) else: UpperCAmelCase = int(name_split[3] ) if "to" in name_split[2]: UpperCAmelCase = int(name_split[2][-2:] ) else: UpperCAmelCase = int(name_split[2][6:] ) if model_size == "tiny": UpperCAmelCase = 96 UpperCAmelCase = (2, 2, 6, 2) UpperCAmelCase = (3, 6, 12, 24) elif model_size == "small": UpperCAmelCase = 96 UpperCAmelCase = (2, 2, 18, 2) UpperCAmelCase = (3, 6, 12, 24) elif model_size == "base": UpperCAmelCase = 128 UpperCAmelCase = (2, 2, 18, 2) UpperCAmelCase = (4, 8, 16, 32) else: UpperCAmelCase = 192 UpperCAmelCase = (2, 2, 18, 2) UpperCAmelCase = (6, 12, 24, 48) if "to" in swinva_name: UpperCAmelCase = (12, 12, 12, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): UpperCAmelCase = 2_1841 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """imagenet-22k-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(_snake_case ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} else: UpperCAmelCase = 1000 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """imagenet-1k-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(_snake_case ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = img_size UpperCAmelCase = num_classes UpperCAmelCase = embed_dim UpperCAmelCase = depths UpperCAmelCase = num_heads UpperCAmelCase = window_size return config def _a ( _snake_case ): """simple docstring""" if "patch_embed.proj" in name: UpperCAmelCase = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: UpperCAmelCase = name.replace("""patch_embed.norm""" , """embeddings.norm""" ) if "layers" in name: UpperCAmelCase = """encoder.""" + name if "attn.proj" in name: UpperCAmelCase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: UpperCAmelCase = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: UpperCAmelCase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: UpperCAmelCase = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: UpperCAmelCase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: UpperCAmelCase = name.replace("""mlp.fc2""" , """output.dense""" ) if "q_bias" in name: UpperCAmelCase = name.replace("""q_bias""" , """query.bias""" ) if "k_bias" in name: UpperCAmelCase = name.replace("""k_bias""" , """key.bias""" ) if "v_bias" in name: UpperCAmelCase = name.replace("""v_bias""" , """value.bias""" ) if "cpb_mlp" in name: UpperCAmelCase = name.replace("""cpb_mlp""" , """continuous_position_bias_mlp""" ) if name == "norm.weight": UpperCAmelCase = """layernorm.weight""" if name == "norm.bias": UpperCAmelCase = """layernorm.bias""" if "head" in name: UpperCAmelCase = name.replace("""head""" , """classifier""" ) else: UpperCAmelCase = """swinv2.""" + name return name def _a ( _snake_case , _snake_case ): """simple docstring""" for key in orig_state_dict.copy().keys(): UpperCAmelCase = orig_state_dict.pop(_snake_case ) if "mask" in key: continue elif "qkv" in key: UpperCAmelCase = key.split(""".""" ) UpperCAmelCase = int(key_split[1] ) UpperCAmelCase = int(key_split[3] ) UpperCAmelCase = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: UpperCAmelCase = val[:dim, :] UpperCAmelCase = val[dim : dim * 2, :] UpperCAmelCase = val[-dim:, :] else: UpperCAmelCase = val[:dim] UpperCAmelCase = val[ dim : dim * 2 ] UpperCAmelCase = val[-dim:] else: UpperCAmelCase = val return orig_state_dict def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() UpperCAmelCase = get_swinva_config(_snake_case ) UpperCAmelCase = SwinvaForImageClassification(_snake_case ) model.eval() UpperCAmelCase = convert_state_dict(timm_model.state_dict() , _snake_case ) model.load_state_dict(_snake_case ) UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = AutoImageProcessor.from_pretrained("""microsoft/{}""".format(swinva_name.replace("""_""" , """-""" ) ) ) UpperCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) UpperCAmelCase = image_processor(images=_snake_case , return_tensors="""pt""" ) UpperCAmelCase = timm_model(inputs["""pixel_values"""] ) UpperCAmelCase = model(**_snake_case ).logits assert torch.allclose(_snake_case , _snake_case , atol=1E-3 ) print(F'''Saving model {swinva_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(_snake_case ) model.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nandwalritik""" , commit_message="""Add model""" , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--swinv2_name""", default="""swinv2_tiny_patch4_window8_256""", type=str, help="""Name of the Swinv2 timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) _UpperCamelCase = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
74
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy as np import tensorflow as tf from transformers import ( TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST, FlaubertConfig, TFFlaubertForMultipleChoice, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForSequenceClassification, TFFlaubertForTokenClassification, TFFlaubertModel, TFFlaubertWithLMHeadModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = 2 UpperCAmelCase = 99 UpperCAmelCase = 0 UpperCAmelCase = 32 UpperCAmelCase = 2 UpperCAmelCase = 4 UpperCAmelCase = 0.1 UpperCAmelCase = 0.1 UpperCAmelCase = 512 UpperCAmelCase = 16 UpperCAmelCase = 2 UpperCAmelCase = 0.02 UpperCAmelCase = 3 UpperCAmelCase = 4 UpperCAmelCase = """last""" UpperCAmelCase = True UpperCAmelCase = None UpperCAmelCase = 0 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ,dtype=tf.floataa ) UpperCAmelCase = None if self.use_input_lengths: UpperCAmelCase = ( ids_tensor([self.batch_size] ,vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.n_langs ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] ,2 ,dtype=tf.floataa ) UpperCAmelCase = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase = FlaubertConfig( vocab_size=self.vocab_size ,n_special=self.n_special ,emb_dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,gelu_activation=self.gelu_activation ,sinusoidal_embeddings=self.sinusoidal_embeddings ,asm=self.asm ,causal=self.causal ,n_langs=self.n_langs ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,summary_type=self.summary_type ,use_proj=self.use_proj ,bos_token_id=self.bos_token_id ,) return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertModel(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) UpperCAmelCase = [input_ids, input_mask] UpperCAmelCase = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertWithLMHeadModel(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForQuestionAnsweringSimple(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForSequenceClassification(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_labels UpperCAmelCase = TFFlaubertForTokenClassification(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_choices UpperCAmelCase = TFFlaubertForMultipleChoice(config=A ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """langs""": token_type_ids, """lengths""": input_lengths, } return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( ( TFFlaubertModel, TFFlaubertWithLMHeadModel, TFFlaubertForSequenceClassification, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForTokenClassification, TFFlaubertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = ( (TFFlaubertWithLMHeadModel,) if is_tf_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFFlaubertModel, '''fill-mask''': TFFlaubertWithLMHeadModel, '''question-answering''': TFFlaubertForQuestionAnsweringSimple, '''text-classification''': TFFlaubertForSequenceClassification, '''token-classification''': TFFlaubertForTokenClassification, '''zero-shot''': TFFlaubertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,emb_dim=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_token_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_multiple_choice(*A ) @slow def _UpperCamelCase ( self ): for model_name in TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFFlaubertModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_tf @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( unittest.TestCase ): @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModel.from_pretrained("""jplu/tf-flaubert-small-cased""" ) UpperCAmelCase = tf.convert_to_tensor( [[0, 158, 735, 2_592, 1_424, 6_727, 82, 1]] ,dtype=tf.intaa ,) # "J'aime flaubert !" UpperCAmelCase = model(A )[0] UpperCAmelCase = tf.TensorShape((1, 8, 512) ) self.assertEqual(output.shape ,A ) # compare the actual values for a slice. UpperCAmelCase = tf.convert_to_tensor( [ [ [-1.8768773, -1.566555, 0.27072418], [-1.6920038, -0.5873505, 1.9329599], [-2.9563985, -1.6993835, 1.7972052], ] ] ,dtype=tf.floataa ,) self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1e-4 ) )
74
1
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _UpperCamelCase = """src/diffusers""" _UpperCamelCase = """.""" # This is to make sure the diffusers module imported is the one in the repo. _UpperCamelCase = importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) _UpperCamelCase = spec.loader.load_module() def _a ( _snake_case , _snake_case ): """simple docstring""" return line.startswith(_snake_case ) or len(_snake_case ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""" , _snake_case ) is not None def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = object_name.split(""".""" ) UpperCAmelCase = 0 # First let's find the module where our object lives. UpperCAmelCase = parts[i] while i < len(_snake_case ) and not os.path.isfile(os.path.join(_snake_case , F'''{module}.py''' ) ): i += 1 if i < len(_snake_case ): UpperCAmelCase = os.path.join(_snake_case , parts[i] ) if i >= len(_snake_case ): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''' ) with open(os.path.join(_snake_case , F'''{module}.py''' ) , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() # Now let's find the class / func in the code! UpperCAmelCase = """""" UpperCAmelCase = 0 for name in parts[i + 1 :]: while ( line_index < len(_snake_case ) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(_snake_case ): raise ValueError(F''' {object_name} does not match any function or class in {module}.''' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). UpperCAmelCase = line_index while line_index < len(_snake_case ) and _should_continue(lines[line_index] , _snake_case ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 UpperCAmelCase = lines[start_index:line_index] return "".join(_snake_case ) _UpperCamelCase = re.compile(R"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") _UpperCamelCase = re.compile(R"""^\s*(\S+)->(\S+)(\s+.*|$)""") _UpperCamelCase = re.compile(R"""<FILL\s+[^>]*>""") def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = code.split("""\n""" ) UpperCAmelCase = 0 while idx < len(_snake_case ) and len(lines[idx] ) == 0: idx += 1 if idx < len(_snake_case ): return re.search(R"""^(\s*)\S""" , lines[idx] ).groups()[0] return "" def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = len(get_indent(_snake_case ) ) > 0 if has_indent: UpperCAmelCase = F'''class Bla:\n{code}''' UpperCAmelCase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=_snake_case ) UpperCAmelCase = black.format_str(_snake_case , mode=_snake_case ) UpperCAmelCase , UpperCAmelCase = style_docstrings_in_code(_snake_case ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _a ( _snake_case , _snake_case=False ): """simple docstring""" with open(_snake_case , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = [] UpperCAmelCase = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(_snake_case ): UpperCAmelCase = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = search.groups() UpperCAmelCase = find_code_in_diffusers(_snake_case ) UpperCAmelCase = get_indent(_snake_case ) UpperCAmelCase = line_index + 1 if indent == theoretical_indent else line_index + 2 UpperCAmelCase = theoretical_indent UpperCAmelCase = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. UpperCAmelCase = True while line_index < len(_snake_case ) and should_continue: line_index += 1 if line_index >= len(_snake_case ): break UpperCAmelCase = lines[line_index] UpperCAmelCase = _should_continue(_snake_case , _snake_case ) and re.search(F'''^{indent}# End copy''' , _snake_case ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 UpperCAmelCase = lines[start_index:line_index] UpperCAmelCase = """""".join(_snake_case ) # Remove any nested `Copied from` comments to avoid circular copies UpperCAmelCase = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(_snake_case ) is None] UpperCAmelCase = """\n""".join(_snake_case ) # Before comparing, use the `replace_pattern` on the original code. if len(_snake_case ) > 0: UpperCAmelCase = replace_pattern.replace("""with""" , """""" ).split(""",""" ) UpperCAmelCase = [_re_replace_pattern.search(_snake_case ) for p in patterns] for pattern in patterns: if pattern is None: continue UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = pattern.groups() UpperCAmelCase = re.sub(_snake_case , _snake_case , _snake_case ) if option.strip() == "all-casing": UpperCAmelCase = re.sub(obja.lower() , obja.lower() , _snake_case ) UpperCAmelCase = re.sub(obja.upper() , obja.upper() , _snake_case ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line UpperCAmelCase = blackify(lines[start_index - 1] + theoretical_code ) UpperCAmelCase = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: UpperCAmelCase = lines[:start_index] + [theoretical_code] + lines[line_index:] UpperCAmelCase = start_index + 1 if overwrite and len(_snake_case ) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''' ) with open(_snake_case , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(_snake_case ) return diffs def _a ( _snake_case = False ): """simple docstring""" UpperCAmelCase = glob.glob(os.path.join(_snake_case , """**/*.py""" ) , recursive=_snake_case ) UpperCAmelCase = [] for filename in all_files: UpperCAmelCase = is_copy_consistent(_snake_case , _snake_case ) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(_snake_case ) > 0: UpperCAmelCase = """\n""".join(_snake_case ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _UpperCamelCase = parser.parse_args() check_copies(args.fix_and_overwrite)
74
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): UpperCAmelCase = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): UpperCAmelCase = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] UpperCAmelCase = key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(_snake_case )-1}''' ) if "norm" in key: UpperCAmelCase = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] UpperCAmelCase = key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(_snake_case )-1}''' ) if "layer_norm1" in key: UpperCAmelCase = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: UpperCAmelCase = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase = key[key.find("""block""" ) + len("""block""" )] UpperCAmelCase = key.replace(F'''block{idx}''' , F'''block.{int(_snake_case )-1}''' ) if "attn.q" in key: UpperCAmelCase = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: UpperCAmelCase = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: UpperCAmelCase = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: UpperCAmelCase = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: UpperCAmelCase = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: UpperCAmelCase = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: UpperCAmelCase = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) UpperCAmelCase = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase = key[key.find("""linear_c""" ) + len("""linear_c""" )] UpperCAmelCase = key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(_snake_case )-1}''' ) if "bot_conv" in key: UpperCAmelCase = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: UpperCAmelCase = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: UpperCAmelCase = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: UpperCAmelCase = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: UpperCAmelCase = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: UpperCAmelCase = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: UpperCAmelCase = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): UpperCAmelCase = key.replace("""module.last_layer_depth""" , """head.head""" ) UpperCAmelCase = value return new_state_dict def _a ( _snake_case , _snake_case ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict UpperCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase = kv_bias[config.hidden_sizes[i] :] def _a ( ): """simple docstring""" UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case=False , _snake_case=None ): """simple docstring""" UpperCAmelCase = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) UpperCAmelCase = GLPNImageProcessor() # prepare image UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=_snake_case , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict UpperCAmelCase = torch.load(_snake_case , map_location=torch.device("""cpu""" ) ) # rename keys UpperCAmelCase = rename_keys(_snake_case ) # key and value matrices need special treatment read_in_k_v(_snake_case , _snake_case ) # create HuggingFace model and load state dict UpperCAmelCase = GLPNForDepthEstimation(_snake_case ) model.load_state_dict(_snake_case ) model.eval() # forward pass UpperCAmelCase = model(_snake_case ) UpperCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: UpperCAmelCase = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: UpperCAmelCase = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) UpperCAmelCase = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _snake_case , atol=1E-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=_snake_case , ) image_processor.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=_snake_case , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCamelCase = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
74
1
"""simple docstring""" from __future__ import annotations def _a ( _snake_case , _snake_case ): """simple docstring""" if nth_term == "": return [""] UpperCAmelCase = int(_snake_case ) UpperCAmelCase = int(_snake_case ) UpperCAmelCase = [] for temp in range(int(_snake_case ) ): series.append(F'''1 / {pow(temp + 1 , int(_snake_case ) )}''' if series else """1""" ) return series if __name__ == "__main__": import doctest doctest.testmod() _UpperCamelCase = int(input("""Enter the last number (nth term) of the P-Series""")) _UpperCamelCase = int(input("""Enter the power for P-Series""")) print("""Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p""") print(p_series(nth_term, power))
74
"""simple docstring""" def _a ( _snake_case ): # noqa: E741 """simple docstring""" UpperCAmelCase = len(_snake_case ) UpperCAmelCase = 0 UpperCAmelCase = [0] * n UpperCAmelCase = [False] * n UpperCAmelCase = [False] * n def dfs(_snake_case , _snake_case , _snake_case , _snake_case ): if parent == root: out_edge_count += 1 UpperCAmelCase = True UpperCAmelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase = dfs(_snake_case , _snake_case , _snake_case , _snake_case ) UpperCAmelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase = True # AP found via cycle if at == low[to]: UpperCAmelCase = True else: UpperCAmelCase = min(low[at] , _snake_case ) return out_edge_count for i in range(_snake_case ): if not visited[i]: UpperCAmelCase = 0 UpperCAmelCase = dfs(_snake_case , _snake_case , -1 , _snake_case ) UpperCAmelCase = out_edge_count > 1 for x in range(len(_snake_case ) ): if is_art[x] is True: print(_snake_case ) # Adjacency list of graph _UpperCamelCase = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
74
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = 42 class lowerCamelCase__ ( snake_case , snake_case ): @register_to_config def __init__( self ,A = 16 ,A = 88 ,A = None ,A = None ,A = 1 ,A = 0.0 ,A = 32 ,A = None ,A = False ,A = None ,A = "geglu" ,A = True ,A = True ,): super().__init__() UpperCAmelCase = num_attention_heads UpperCAmelCase = attention_head_dim UpperCAmelCase = num_attention_heads * attention_head_dim UpperCAmelCase = in_channels UpperCAmelCase = torch.nn.GroupNorm(num_groups=A ,num_channels=A ,eps=1e-6 ,affine=A ) UpperCAmelCase = nn.Linear(A ,A ) # 3. Define transformers blocks UpperCAmelCase = nn.ModuleList( [ BasicTransformerBlock( A ,A ,A ,dropout=A ,cross_attention_dim=A ,activation_fn=A ,attention_bias=A ,double_self_attention=A ,norm_elementwise_affine=A ,) for d in range(A ) ] ) UpperCAmelCase = nn.Linear(A ,A ) def _UpperCamelCase ( self ,A ,A=None ,A=None ,A=None ,A=1 ,A=None ,A = True ,): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = hidden_states.shape UpperCAmelCase = batch_frames // num_frames UpperCAmelCase = hidden_states UpperCAmelCase = hidden_states[None, :].reshape(A ,A ,A ,A ,A ) UpperCAmelCase = hidden_states.permute(0 ,2 ,1 ,3 ,4 ) UpperCAmelCase = self.norm(A ) UpperCAmelCase = hidden_states.permute(0 ,3 ,4 ,2 ,1 ).reshape(batch_size * height * width ,A ,A ) UpperCAmelCase = self.proj_in(A ) # 2. Blocks for block in self.transformer_blocks: UpperCAmelCase = block( A ,encoder_hidden_states=A ,timestep=A ,cross_attention_kwargs=A ,class_labels=A ,) # 3. Output UpperCAmelCase = self.proj_out(A ) UpperCAmelCase = ( hidden_states[None, None, :] .reshape(A ,A ,A ,A ,A ) .permute(0 ,3 ,4 ,1 ,2 ) .contiguous() ) UpperCAmelCase = hidden_states.reshape(A ,A ,A ,A ) UpperCAmelCase = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=A )
74
"""simple docstring""" _UpperCamelCase = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ _UpperCamelCase = [{"""type""": """code""", """content""": INSTALL_CONTENT}] _UpperCamelCase = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
74
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _UpperCamelCase = { """configuration_falcon""": ["""FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FalconConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """FALCON_PRETRAINED_MODEL_ARCHIVE_LIST""", """FalconForCausalLM""", """FalconModel""", """FalconPreTrainedModel""", """FalconForSequenceClassification""", """FalconForTokenClassification""", """FalconForQuestionAnswering""", ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" import argparse import struct import unittest class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = data # Initialize hash values UpperCAmelCase = [ 0x6A_09_E6_67, 0xBB_67_AE_85, 0x3C_6E_F3_72, 0xA5_4F_F5_3A, 0x51_0E_52_7F, 0x9B_05_68_8C, 0x1F_83_D9_AB, 0x5B_E0_CD_19, ] # Initialize round constants UpperCAmelCase = [ 0x42_8A_2F_98, 0x71_37_44_91, 0xB5_C0_FB_CF, 0xE9_B5_DB_A5, 0x39_56_C2_5B, 0x59_F1_11_F1, 0x92_3F_82_A4, 0xAB_1C_5E_D5, 0xD8_07_AA_98, 0x12_83_5B_01, 0x24_31_85_BE, 0x55_0C_7D_C3, 0x72_BE_5D_74, 0x80_DE_B1_FE, 0x9B_DC_06_A7, 0xC1_9B_F1_74, 0xE4_9B_69_C1, 0xEF_BE_47_86, 0x0F_C1_9D_C6, 0x24_0C_A1_CC, 0x2D_E9_2C_6F, 0x4A_74_84_AA, 0x5C_B0_A9_DC, 0x76_F9_88_DA, 0x98_3E_51_52, 0xA8_31_C6_6D, 0xB0_03_27_C8, 0xBF_59_7F_C7, 0xC6_E0_0B_F3, 0xD5_A7_91_47, 0x06_CA_63_51, 0x14_29_29_67, 0x27_B7_0A_85, 0x2E_1B_21_38, 0x4D_2C_6D_FC, 0x53_38_0D_13, 0x65_0A_73_54, 0x76_6A_0A_BB, 0x81_C2_C9_2E, 0x92_72_2C_85, 0xA2_BF_E8_A1, 0xA8_1A_66_4B, 0xC2_4B_8B_70, 0xC7_6C_51_A3, 0xD1_92_E8_19, 0xD6_99_06_24, 0xF4_0E_35_85, 0x10_6A_A0_70, 0x19_A4_C1_16, 0x1E_37_6C_08, 0x27_48_77_4C, 0x34_B0_BC_B5, 0x39_1C_0C_B3, 0x4E_D8_AA_4A, 0x5B_9C_CA_4F, 0x68_2E_6F_F3, 0x74_8F_82_EE, 0x78_A5_63_6F, 0x84_C8_78_14, 0x8C_C7_02_08, 0x90_BE_FF_FA, 0xA4_50_6C_EB, 0xBE_F9_A3_F7, 0xC6_71_78_F2, ] UpperCAmelCase = self.preprocessing(self.data ) self.final_hash() @staticmethod def _UpperCamelCase ( A ): UpperCAmelCase = b"""\x80""" + (b"""\x00""" * (63 - (len(A ) + 8) % 64)) UpperCAmelCase = struct.pack(""">Q""" ,(len(A ) * 8) ) return data + padding + big_endian_integer def _UpperCamelCase ( self ): # Convert into blocks of 64 bytes UpperCAmelCase = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers UpperCAmelCase = list(struct.unpack(""">16L""" ,A ) ) # add 48 0-ed integers words += [0] * 48 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array UpperCAmelCase = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) UpperCAmelCase = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) UpperCAmelCase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_00_00_00_00 # Compression UpperCAmelCase = self.ror(A ,6 ) ^ self.ror(A ,11 ) ^ self.ror(A ,25 ) UpperCAmelCase = (e & f) ^ ((~e & 0xFF_FF_FF_FF) & g) UpperCAmelCase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_00_00_00_00 UpperCAmelCase = self.ror(A ,2 ) ^ self.ror(A ,13 ) ^ self.ror(A ,22 ) UpperCAmelCase = (a & b) ^ (a & c) ^ (b & c) UpperCAmelCase = (sa + maj) % 0x1_00_00_00_00 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = ( g, f, e, ((d + tempa) % 0x1_00_00_00_00), c, b, a, ((tempa + tempa) % 0x1_00_00_00_00), ) UpperCAmelCase = [a, b, c, d, e, f, g, h] # Modify final values UpperCAmelCase = [ ((element + mutated_hash_values[index]) % 0x1_00_00_00_00) for index, element in enumerate(self.hashes ) ] UpperCAmelCase = """""".join([hex(A )[2:].zfill(8 ) for value in self.hashes] ) def _UpperCamelCase ( self ,A ,A ): return 0xFF_FF_FF_FF & (value << (32 - rotations)) | (value >> rotations) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): import hashlib UpperCAmelCase = bytes("""Test String""" ,"""utf-8""" ) self.assertEqual(SHAaaa(A ).hash ,hashlib.shaaaa(A ).hexdigest() ) def _a ( ): """simple docstring""" import doctest doctest.testmod() UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase = parser.parse_args() UpperCAmelCase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase = f.read() else: UpperCAmelCase = bytes(_snake_case , """utf-8""" ) print(SHAaaa(_snake_case ).hash ) if __name__ == "__main__": main()
74
1
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """huggingface/informer-tourism-monthly""": ( """https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json""" ), # See all Informer models at https://huggingface.co/models?filter=informer } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''informer''' SCREAMING_SNAKE_CASE = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self ,A = None ,A = None ,A = "student_t" ,A = "nll" ,A = 1 ,A = None ,A = "mean" ,A = 0 ,A = 0 ,A = 0 ,A = 0 ,A = None ,A = None ,A = 64 ,A = 32 ,A = 32 ,A = 2 ,A = 2 ,A = 2 ,A = 2 ,A = True ,A = "gelu" ,A = 0.05 ,A = 0.1 ,A = 0.1 ,A = 0.1 ,A = 0.1 ,A = 100 ,A = 0.02 ,A=True ,A = "prob" ,A = 5 ,A = True ,**A ,): # time series specific configuration UpperCAmelCase = prediction_length UpperCAmelCase = context_length or prediction_length UpperCAmelCase = distribution_output UpperCAmelCase = loss UpperCAmelCase = input_size UpperCAmelCase = num_time_features UpperCAmelCase = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] UpperCAmelCase = scaling UpperCAmelCase = num_dynamic_real_features UpperCAmelCase = num_static_real_features UpperCAmelCase = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(A ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase = cardinality else: UpperCAmelCase = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(A ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) UpperCAmelCase = embedding_dimension else: UpperCAmelCase = [min(50 ,(cat + 1) // 2 ) for cat in self.cardinality] UpperCAmelCase = num_parallel_samples # Transformer architecture configuration UpperCAmelCase = input_size * len(self.lags_sequence ) + self._number_of_features UpperCAmelCase = d_model UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_attention_heads UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = decoder_layers UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = use_cache # Informer UpperCAmelCase = attention_type UpperCAmelCase = sampling_factor UpperCAmelCase = distil super().__init__(is_encoder_decoder=A ,**A ) @property def _UpperCamelCase ( self ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
74
"""simple docstring""" def _a ( _snake_case = 10 , _snake_case = 22 ): """simple docstring""" UpperCAmelCase = range(1 , _snake_case ) UpperCAmelCase = range(1 , _snake_case ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F"""{solution(10, 22) = }""")
74
1
"""simple docstring""" import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def _a ( _snake_case ): """simple docstring""" assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def _a ( ): """simple docstring""" assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def _a ( ): """simple docstring""" UpperCAmelCase = """mock-s3-bucket""" UpperCAmelCase = F'''s3://{mock_bucket}''' UpperCAmelCase = extract_path_from_uri(_snake_case ) assert dataset_path.startswith("""s3://""" ) is False UpperCAmelCase = """./local/path""" UpperCAmelCase = extract_path_from_uri(_snake_case ) assert dataset_path == new_dataset_path def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = is_remote_filesystem(_snake_case ) assert is_remote is True UpperCAmelCase = fsspec.filesystem("""file""" ) UpperCAmelCase = is_remote_filesystem(_snake_case ) assert is_remote is False @pytest.mark.parametrize("""compression_fs_class""" , _snake_case ) def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_file, """bz2""": bza_file, """lz4""": lza_file} UpperCAmelCase = input_paths[compression_fs_class.protocol] if input_path is None: UpperCAmelCase = F'''for \'{compression_fs_class.protocol}\' compression protocol, ''' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(_snake_case ) UpperCAmelCase = fsspec.filesystem(compression_fs_class.protocol , fo=_snake_case ) assert isinstance(_snake_case , _snake_case ) UpperCAmelCase = os.path.basename(_snake_case ) UpperCAmelCase = expected_filename[: expected_filename.rindex(""".""" )] assert fs.glob("""*""" ) == [expected_filename] with fs.open(_snake_case , """r""" , encoding="""utf-8""" ) as f, open(_snake_case , encoding="""utf-8""" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("""protocol""" , ["""zip""", """gzip"""] ) def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = {"""zip""": zip_jsonl_path, """gzip""": jsonl_gz_path} UpperCAmelCase = compressed_file_paths[protocol] UpperCAmelCase = """dataset.jsonl""" UpperCAmelCase = F'''{protocol}://{member_file_path}::{compressed_file_path}''' UpperCAmelCase , *UpperCAmelCase = fsspec.get_fs_token_paths(_snake_case ) assert fs.isfile(_snake_case ) assert not fs.isfile("""non_existing_""" + member_file_path ) @pytest.mark.integration def _a ( _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = hf_api.dataset_info(_snake_case , token=_snake_case ) UpperCAmelCase = HfFileSystem(repo_info=_snake_case , token=_snake_case ) assert sorted(hffs.glob("""*""" ) ) == [".gitattributes", "data"] assert hffs.isdir("""data""" ) assert hffs.isfile(""".gitattributes""" ) and hffs.isfile("""data/text_data.txt""" ) with open(_snake_case ) as f: assert hffs.open("""data/text_data.txt""" , """r""" ).read() == f.read() def _a ( ): """simple docstring""" UpperCAmelCase = """bz2""" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(_snake_case , _snake_case , clobber=_snake_case ) with pytest.warns(_snake_case ) as warning_info: importlib.reload(datasets.filesystems ) assert len(_snake_case ) == 1 assert ( str(warning_info[0].message ) == F'''A filesystem protocol was already set for {protocol} and will be overwritten.''' )
74
"""simple docstring""" from __future__ import annotations def _a ( _snake_case ): """simple docstring""" return len(set(_snake_case ) ) == len(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast _UpperCamelCase = datasets.utils.logging.get_logger(__name__) @dataclass class lowerCamelCase__ ( datasets.BuilderConfig ): SCREAMING_SNAKE_CASE = 10_000 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None class lowerCamelCase__ ( datasets.ArrowBasedBuilder ): SCREAMING_SNAKE_CASE = ParquetConfig def _UpperCamelCase ( self ): return datasets.DatasetInfo(features=self.config.features ) def _UpperCamelCase ( self ,A ): if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) UpperCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(A ,(str, list, tuple) ): UpperCAmelCase = data_files if isinstance(A ,A ): UpperCAmelCase = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase = [dl_manager.iter_files(A ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN ,gen_kwargs={"""files""": files} )] UpperCAmelCase = [] for split_name, files in data_files.items(): if isinstance(A ,A ): UpperCAmelCase = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase = [dl_manager.iter_files(A ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(A ): with open(A ,"""rb""" ) as f: UpperCAmelCase = datasets.Features.from_arrow_schema(pq.read_schema(A ) ) break splits.append(datasets.SplitGenerator(name=A ,gen_kwargs={"""files""": files} ) ) return splits def _UpperCamelCase ( self ,A ): if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example UpperCAmelCase = table_cast(A ,self.info.features.arrow_schema ) return pa_table def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F'''Tried to load parquet data with columns \'{self.config.columns}\' with mismatching features \'{self.info.features}\'''' ) for file_idx, file in enumerate(itertools.chain.from_iterable(A ) ): with open(A ,"""rb""" ) as f: UpperCAmelCase = pq.ParquetFile(A ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size ,columns=self.config.columns ) ): UpperCAmelCase = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F'''{file_idx}_{batch_idx}''', self._cast_table(A ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(A )}: {e}''' ) raise
74
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( _snake_case = 0.1 ): """simple docstring""" UpperCAmelCase = 3 UpperCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" from __future__ import annotations import math import random from typing import Any class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = [] UpperCAmelCase = 0 UpperCAmelCase = 0 def _UpperCamelCase ( self ): return self.head == self.tail def _UpperCamelCase ( self ,A ): self.data.append(A ) UpperCAmelCase = self.tail + 1 def _UpperCamelCase ( self ): UpperCAmelCase = self.data[self.head] UpperCAmelCase = self.head + 1 return ret def _UpperCamelCase ( self ): return self.tail - self.head def _UpperCamelCase ( self ): print(self.data ) print("""**************""" ) print(self.data[self.head : self.tail] ) class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = data UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = 1 def _UpperCamelCase ( self ): return self.data def _UpperCamelCase ( self ): return self.left def _UpperCamelCase ( self ): return self.right def _UpperCamelCase ( self ): return self.height def _UpperCamelCase ( self ,A ): UpperCAmelCase = data def _UpperCamelCase ( self ,A ): UpperCAmelCase = node def _UpperCamelCase ( self ,A ): UpperCAmelCase = node def _UpperCamelCase ( self ,A ): UpperCAmelCase = height def _a ( _snake_case ): """simple docstring""" if node is None: return 0 return node.get_height() def _a ( _snake_case , _snake_case ): """simple docstring""" if a > b: return a return b def _a ( _snake_case ): """simple docstring""" print("""left rotation node:""" , node.get_data() ) UpperCAmelCase = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(_snake_case ) UpperCAmelCase = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) UpperCAmelCase = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(_snake_case ) return ret def _a ( _snake_case ): """simple docstring""" print("""right rotation node:""" , node.get_data() ) UpperCAmelCase = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(_snake_case ) UpperCAmelCase = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) UpperCAmelCase = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(_snake_case ) return ret def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = node.get_left() assert left_child is not None node.set_left(left_rotation(_snake_case ) ) return right_rotation(_snake_case ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = node.get_right() assert right_child is not None node.set_right(right_rotation(_snake_case ) ) return left_rotation(_snake_case ) def _a ( _snake_case , _snake_case ): """simple docstring""" if node is None: return MyNode(_snake_case ) if data < node.get_data(): node.set_left(insert_node(node.get_left() , _snake_case ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected UpperCAmelCase = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child UpperCAmelCase = right_rotation(_snake_case ) else: UpperCAmelCase = lr_rotation(_snake_case ) else: node.set_right(insert_node(node.get_right() , _snake_case ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: UpperCAmelCase = node.get_right() assert right_child is not None if data < right_child.get_data(): UpperCAmelCase = rl_rotation(_snake_case ) else: UpperCAmelCase = left_rotation(_snake_case ) UpperCAmelCase = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) return node def _a ( _snake_case ): """simple docstring""" while True: UpperCAmelCase = root.get_right() if right_child is None: break UpperCAmelCase = right_child return root.get_data() def _a ( _snake_case ): """simple docstring""" while True: UpperCAmelCase = root.get_left() if left_child is None: break UpperCAmelCase = left_child return root.get_data() def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = root.get_left() UpperCAmelCase = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: UpperCAmelCase = get_left_most(_snake_case ) root.set_data(_snake_case ) root.set_right(del_node(_snake_case , _snake_case ) ) elif left_child is not None: UpperCAmelCase = left_child elif right_child is not None: UpperCAmelCase = right_child else: return None elif root.get_data() > data: if left_child is None: print("""No such data""" ) return root else: root.set_left(del_node(_snake_case , _snake_case ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(_snake_case , _snake_case ) ) if get_height(_snake_case ) - get_height(_snake_case ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): UpperCAmelCase = left_rotation(_snake_case ) else: UpperCAmelCase = rl_rotation(_snake_case ) elif get_height(_snake_case ) - get_height(_snake_case ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): UpperCAmelCase = right_rotation(_snake_case ) else: UpperCAmelCase = lr_rotation(_snake_case ) UpperCAmelCase = my_max(get_height(root.get_right() ) , get_height(root.get_left() ) ) + 1 root.set_height(_snake_case ) return root class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = None def _UpperCamelCase ( self ): return get_height(self.root ) def _UpperCamelCase ( self ,A ): print("""insert:""" + str(A ) ) UpperCAmelCase = insert_node(self.root ,A ) def _UpperCamelCase ( self ,A ): print("""delete:""" + str(A ) ) if self.root is None: print("""Tree is empty!""" ) return UpperCAmelCase = del_node(self.root ,A ) def __str__( self ,): # a level traversale, gives a more intuitive look on the tree UpperCAmelCase = """""" UpperCAmelCase = MyQueue() q.push(self.root ) UpperCAmelCase = self.get_height() if layer == 0: return output UpperCAmelCase = 0 while not q.is_empty(): UpperCAmelCase = q.pop() UpperCAmelCase = """ """ * int(math.pow(2 ,layer - 1 ) ) output += space if node is None: output += "*" q.push(A ) q.push(A ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space UpperCAmelCase = cnt + 1 for i in range(100 ): if cnt == math.pow(2 ,A ) - 1: UpperCAmelCase = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def _a ( ): """simple docstring""" import doctest doctest.testmod() if __name__ == "__main__": _test() _UpperCamelCase = AVLtree() _UpperCamelCase = list(range(10)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
74
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE = '''CLIPImageProcessor''' SCREAMING_SNAKE_CASE = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self ,A=None ,A=None ,**A ): UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) def __call__( self ,A=None ,A=None ,A=None ,**A ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.batch_decode(*A ,**A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.decode(*A ,**A ) @property def _UpperCamelCase ( self ): UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,A ,) return self.image_processor
74
1
"""simple docstring""" from itertools import count def _a ( _snake_case = 50 ): """simple docstring""" UpperCAmelCase = [1] * min_block_length for n in count(_snake_case ): fill_count_functions.append(1 ) for block_length in range(_snake_case , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 100_0000: break return n if __name__ == "__main__": print(F"""{solution() = }""")
74
"""simple docstring""" from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def _a ( _snake_case = "mumbai" ): """simple docstring""" UpperCAmelCase = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): UpperCAmelCase = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() UpperCAmelCase = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("""Bangalore"""), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
74
1
"""simple docstring""" import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate _UpperCamelCase = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataRow("""""", """|""", """|"""), datarow=DataRow("""""", """|""", """|"""), padding=1, with_header_hide=None, ) _UpperCamelCase = [] _UpperCamelCase = [] _UpperCamelCase = {"""type""": """section""", """text""": {"""type""": """plain_text""", """text""": """No failed tests! 🤗""", """emoji""": True}} _UpperCamelCase = [ { """type""": """header""", """text""": { """type""": """plain_text""", """text""": F"""🤗 Accelerate nightly {os.environ.get("TEST_TYPE", "")} test results""", """emoji""": True, }, } ] _UpperCamelCase = 0 for log in Path().glob("""*.log"""): _UpperCamelCase = 0 with open(log, """r""") as f: for line in f: _UpperCamelCase = json.loads(line) if line.get("""nodeid""", """""") != "": _UpperCamelCase = line["""nodeid"""] if line.get("""duration""", None) is not None: _UpperCamelCase = F"""{line["duration"]:.4f}""" if line.get("""outcome""", """""") == "failed": section_num_failed += 1 failed.append([test, duration, log.name.split("""_""")[0]]) total_num_failed += 1 group_info.append([str(log), section_num_failed, failed]) _UpperCamelCase = [] log.unlink() _UpperCamelCase = """""" _UpperCamelCase = [] if total_num_failed > 0: for name, num_failed, failed_tests in group_info: if num_failed > 0: if num_failed == 1: message += F"*{name[1:]}: {num_failed} failed test*\n" else: message += F"*{name[1:]}: {num_failed} failed tests*\n" _UpperCamelCase = [] _UpperCamelCase = {} for test in failed_tests: _UpperCamelCase = test[0].split("""::""") _UpperCamelCase = data[0].split("""/""")[-1] if data[0] not in filesafailed: _UpperCamelCase = [data[1:]] else: filesafailed[data[0]] += [data[1:]] failed_table.append(data) _UpperCamelCase = [test[0] for test in failed_table] _UpperCamelCase = list(set(files)) # Count number of instances in failed_tests _UpperCamelCase = [] for file in individual_files: table.append([file, len(filesafailed[file])]) _UpperCamelCase = tabulate( table, headers=["""Test Location""", """Num Failed"""], tablefmt=hf_table_format, stralign="""right""", ) message += F"\n```\n{failed_table}\n```" all_filesafailed.append(filesafailed) if len(message) > 3000: _UpperCamelCase = """Too many failed tests, please see the full report in the Action results.""" _UpperCamelCase = len(err) + 10 _UpperCamelCase = message[: 3000 - offset] + F"""\n...\n```\n{err}""" print(F"""### {message}""") else: _UpperCamelCase = """No failed tests! 🤗""" print(F"""## {message}""") payload.append(no_error_payload) if os.environ.get("""TEST_TYPE""", """""") != "": from slack_sdk import WebClient _UpperCamelCase = WebClient(token=os.environ["""SLACK_API_TOKEN"""]) if message != "No failed tests! 🤗": _UpperCamelCase = { """type""": """section""", """text""": { """type""": """mrkdwn""", """text""": message, }, } payload.append(md_report) _UpperCamelCase = { """type""": """section""", """text""": { """type""": """mrkdwn""", """text""": """*For more details:*""", }, """accessory""": { """type""": """button""", """text""": { """type""": """plain_text""", """text""": """Check Action results""", """emoji""": True, }, """url""": F"""https://github.com/{os.environ["GITHUB_REPOSITORY"]}/actions/runs/{os.environ["GITHUB_RUN_ID"]}""", }, } payload.append(action_button) _UpperCamelCase = { """type""": """context""", """elements""": [ { """type""": """plain_text""", """text""": F"""Nightly {os.environ.get("TEST_TYPE")} test results for {date.today()}""", } ], } payload.append(date_report) _UpperCamelCase = client.chat_postMessage(channel="""#accelerate-ci-daily""", text=message, blocks=payload) _UpperCamelCase = response.data["""ts"""] for failed_file in all_filesafailed: for test_location, test_failures in failed_file.items(): # Keep only the first instance of the test name _UpperCamelCase = """""" for i, row in enumerate(test_failures): if row[0] != test_class: _UpperCamelCase = row[0] else: _UpperCamelCase = """""" _UpperCamelCase = { """type""": """section""", """text""": { """type""": """mrkdwn""", """text""": F"""Test location: {test_location}\n```\n{tabulate(test_failures, headers=["Class", "Test"], tablefmt=hf_table_format, stralign="right")}\n```""", }, } client.chat_postMessage( channel="""#accelerate-ci-daily""", thread_ts=ts, blocks=[payload], )
74
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): UpperCAmelCase = ["""a""", """b""", """c"""] # Defaults to last layer if both are None UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,A ,A ) self.assertEqual(A ,["""c"""] ) self.assertEqual(A ,[2] ) # Out indices set to match out features UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(["""a""", """c"""] ,A ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features set to match out indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[0, 2] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features selected from negative indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[-3, -1] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[-3, -1] ) def _UpperCamelCase ( self ): # Stage names must be set with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,A ) # Out features must be a list with self.assertRaises(A ): verify_out_features_out_indices(("""a""", """b""") ,(0, 1) ,["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(A ): verify_out_features_out_indices(A ,0 ,["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(A ,(0, 1) ,["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0,) ,["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 2) ,["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(A ): verify_out_features_out_indices(["""b""", """a"""] ,(0, 1) ,["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] ,(0, 1, -1) ,["""a""", """b""", """c""", """d"""] ) def _UpperCamelCase ( self ): UpperCAmelCase = BackboneMixin() UpperCAmelCase = ["""a""", """b""", """c"""] UpperCAmelCase = ["""a""", """c"""] UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[0, 2] ) # Check out features and indices are updated correctly UpperCAmelCase = ["""a""", """b"""] self.assertEqual(backbone.out_features ,["""a""", """b"""] ) self.assertEqual(backbone.out_indices ,[0, 1] ) UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[-3, -1] )
74
1
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE = '''CLIPImageProcessor''' SCREAMING_SNAKE_CASE = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self ,A=None ,A=None ,**A ): UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) def __call__( self ,A=None ,A=None ,A=None ,**A ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.batch_decode(*A ,**A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.decode(*A ,**A ) @property def _UpperCamelCase ( self ): UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,A ,) return self.image_processor
74
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCamelCase__ : def __init__( self ,A = 6 ): UpperCAmelCase = None UpperCAmelCase = None self.create_linked_list(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = current_node UpperCAmelCase = current_node for _ in range(1 ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = previous_node UpperCAmelCase = current_node UpperCAmelCase = self.front UpperCAmelCase = previous_node def _UpperCamelCase ( self ): return ( self.front == self.rear and self.front is not None and self.front.data is None ) def _UpperCamelCase ( self ): self.check_can_perform_operation() return self.front.data if self.front else None def _UpperCamelCase ( self ,A ): if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase = self.rear.next if self.rear: UpperCAmelCase = data def _UpperCamelCase ( self ): self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase = self.front.data UpperCAmelCase = None return data UpperCAmelCase = self.front UpperCAmelCase = old_front.next UpperCAmelCase = old_front.data UpperCAmelCase = None return data def _UpperCamelCase ( self ): if self.is_empty(): raise Exception("""Empty Queue""" ) def _UpperCamelCase ( self ): if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel _UpperCamelCase = False _UpperCamelCase = True _UpperCamelCase = False if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--repo_path""", default=None, type=str, required=True, help="""The config json file corresponding to the architecture.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") _UpperCamelCase = parser.parse_args() _UpperCamelCase = { """image_size""": """sample_size""", """num_res_blocks""": """layers_per_block""", """block_channels""": """block_out_channels""", """down_blocks""": """down_block_types""", """up_blocks""": """up_block_types""", """downscale_freq_shift""": """freq_shift""", """resnet_num_groups""": """norm_num_groups""", """resnet_act_fn""": """act_fn""", """resnet_eps""": """norm_eps""", """num_head_channels""": """attention_head_dim""", } _UpperCamelCase = { """time_steps""": """time_proj""", """mid""": """mid_block""", """downsample_blocks""": """down_blocks""", """upsample_blocks""": """up_blocks""", } _UpperCamelCase = """""" if has_file(args.repo_path, """config.json""") else """unet""" with open(os.path.join(args.repo_path, subfolder, """config.json"""), """r""", encoding="""utf-8""") as reader: _UpperCamelCase = reader.read() _UpperCamelCase = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, """config.json"""): _UpperCamelCase = UNetaDModel(**config) else: _UpperCamelCase = UNetaDConditionModel if """ldm-text2im-large-256""" in args.repo_path else UNetaDModel _UpperCamelCase = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) _UpperCamelCase = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: _UpperCamelCase = config[key] del config[key] _UpperCamelCase = [k.replace("""UNetRes""", """""") for k in config["""down_block_types"""]] _UpperCamelCase = [k.replace("""UNetRes""", """""") for k in config["""up_block_types"""]] if do_only_weights: _UpperCamelCase = torch.load(os.path.join(args.repo_path, subfolder, """diffusion_pytorch_model.bin""")) _UpperCamelCase = {} for param_key, param_value in state_dict.items(): if param_key.endswith(""".op.bias""") or param_key.endswith(""".op.weight"""): continue _UpperCamelCase = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(""".""")[0] == key: _UpperCamelCase = param_value _UpperCamelCase = True if not has_changed: _UpperCamelCase = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
74
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("""transformers.models.speecht5""") def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" hf_model.apply_weight_norm() UpperCAmelCase = checkpoint["""input_conv.weight_g"""] UpperCAmelCase = checkpoint["""input_conv.weight_v"""] UpperCAmelCase = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , ): """simple docstring""" if config_path is not None: UpperCAmelCase = SpeechTaHifiGanConfig.from_pretrained(_snake_case ) else: UpperCAmelCase = SpeechTaHifiGanConfig() UpperCAmelCase = SpeechTaHifiGan(_snake_case ) UpperCAmelCase = torch.load(_snake_case ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _snake_case , _snake_case ) UpperCAmelCase = np.load(_snake_case ) UpperCAmelCase = stats[0].reshape(-1 ) UpperCAmelCase = stats[1].reshape(-1 ) UpperCAmelCase = torch.from_numpy(_snake_case ).float() UpperCAmelCase = torch.from_numpy(_snake_case ).float() model.save_pretrained(_snake_case ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
74
1
"""simple docstring""" from __future__ import annotations def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = [True] * limit UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCAmelCase = i * 2 while index < limit: UpperCAmelCase = False UpperCAmelCase = index + i UpperCAmelCase = [2] for i in range(3 , _snake_case , 2 ): if is_prime[i]: primes.append(_snake_case ) return primes def _a ( _snake_case = 100_0000 ): """simple docstring""" UpperCAmelCase = prime_sieve(_snake_case ) UpperCAmelCase = 0 UpperCAmelCase = 0 for i in range(len(_snake_case ) ): for j in range(i + length , len(_snake_case ) ): UpperCAmelCase = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCAmelCase = j - i UpperCAmelCase = sol return largest if __name__ == "__main__": print(F"""{solution() = }""")
74
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase = concatenate_datasets _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadManager _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
74
1
"""simple docstring""" from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _a ( _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=True ): """simple docstring""" model.train() UpperCAmelCase = model(_snake_case ) UpperCAmelCase = F.mse_loss(_snake_case , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(_snake_case ) def _a ( _snake_case , _snake_case=False ): """simple docstring""" set_seed(42 ) UpperCAmelCase = RegressionModel() UpperCAmelCase = deepcopy(_snake_case ) UpperCAmelCase = RegressionDataset(length=80 ) UpperCAmelCase = DataLoader(_snake_case , batch_size=16 ) model.to(accelerator.device ) if sched: UpperCAmelCase = AdamW(params=model.parameters() , lr=1E-3 ) UpperCAmelCase = AdamW(params=ddp_model.parameters() , lr=1E-3 ) UpperCAmelCase = LambdaLR(_snake_case , lr_lambda=lambda _snake_case : epoch**0.65 ) UpperCAmelCase = LambdaLR(_snake_case , lr_lambda=lambda _snake_case : epoch**0.65 ) # Make a copy of `model` if sched: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = accelerator.prepare(_snake_case , _snake_case , _snake_case , _snake_case ) else: UpperCAmelCase , UpperCAmelCase = accelerator.prepare(_snake_case , _snake_case ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _a ( _snake_case ): """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = get_training_setup(_snake_case ) # Use a single batch UpperCAmelCase , UpperCAmelCase = next(iter(_snake_case ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model UpperCAmelCase , UpperCAmelCase = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase , UpperCAmelCase = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_snake_case , _snake_case , _snake_case , _snake_case ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(_snake_case ): step_model(_snake_case , _snake_case , _snake_case , _snake_case ) else: # Sync grads step_model(_snake_case , _snake_case , _snake_case , _snake_case ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(_snake_case , _snake_case , _snake_case , _snake_case ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) UpperCAmelCase = ddp_input[torch.randperm(len(_snake_case ) )] def _a ( _snake_case ): """simple docstring""" UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = get_training_setup(_snake_case ) # Use a single batch UpperCAmelCase , UpperCAmelCase = next(iter(_snake_case ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model UpperCAmelCase , UpperCAmelCase = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase , UpperCAmelCase = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_snake_case , _snake_case , _snake_case , _snake_case ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(_snake_case ): step_model(_snake_case , _snake_case , _snake_case , _snake_case ) else: # Sync grads step_model(_snake_case , _snake_case , _snake_case , _snake_case ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) UpperCAmelCase = ddp_input[torch.randperm(len(_snake_case ) )] def _a ( _snake_case=False , _snake_case=False ): """simple docstring""" UpperCAmelCase = Accelerator( split_batches=_snake_case , dispatch_batches=_snake_case , gradient_accumulation_steps=2 ) # Test that context manager behaves properly UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = get_training_setup(_snake_case ) for iteration, batch in enumerate(_snake_case ): UpperCAmelCase , UpperCAmelCase = batch.values() # Gather the distributed inputs and targs for the base model UpperCAmelCase , UpperCAmelCase = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase , UpperCAmelCase = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) # Do "gradient accumulation" (noop) with accelerator.accumulate(_snake_case ): step_model(_snake_case , _snake_case , _snake_case , _snake_case ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(_snake_case ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) UpperCAmelCase = ddp_input[torch.randperm(len(_snake_case ) )] GradientState._reset_state() def _a ( _snake_case=False , _snake_case=False ): """simple docstring""" UpperCAmelCase = Accelerator( split_batches=_snake_case , dispatch_batches=_snake_case , gradient_accumulation_steps=2 ) # Test that context manager behaves properly UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = get_training_setup(_snake_case , _snake_case ) for iteration, batch in enumerate(_snake_case ): UpperCAmelCase , UpperCAmelCase = batch.values() # Gather the distributed inputs and targs for the base model UpperCAmelCase , UpperCAmelCase = accelerator.gather((ddp_input, ddp_target) ) UpperCAmelCase , UpperCAmelCase = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(_snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(_snake_case )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(_snake_case ): step_model(_snake_case , _snake_case , _snake_case , _snake_case ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n''' UpperCAmelCase = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(_snake_case )) if accelerator.num_processes > 1: check_model_parameters(_snake_case , _snake_case , _snake_case , _snake_case ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def _a ( ): """simple docstring""" UpperCAmelCase = Accelerator() UpperCAmelCase = RegressionDataset(length=80 ) UpperCAmelCase = DataLoader(_snake_case , batch_size=16 ) UpperCAmelCase = RegressionDataset(length=96 ) UpperCAmelCase = DataLoader(_snake_case , batch_size=16 ) UpperCAmelCase , UpperCAmelCase = accelerator.prepare(_snake_case , _snake_case ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(_snake_case ): assert id(accelerator.gradient_state.active_dataloader ) == id(_snake_case ) if iteration < len(_snake_case ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(_snake_case ): assert id(accelerator.gradient_state.active_dataloader ) == id(_snake_case ) if batch_num < len(_snake_case ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _a ( ): """simple docstring""" UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.state if state.local_process_index == 0: print("""**Test `accumulate` gradient accumulation with dataloader break**""" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("""**Test NOOP `no_sync` context manager**""" ) test_noop_sync(_snake_case ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("""**Test Distributed `no_sync` context manager**""" ) test_distributed_sync(_snake_case ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation, """ , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation(_snake_case , _snake_case ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("""<""" , """2.0""" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , """`split_batches=False`, `dispatch_batches=False`**""" , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( """**Test `accumulate` gradient accumulation with optimizer and scheduler, """ , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation_with_opt_and_scheduler(_snake_case , _snake_case ) def _a ( _snake_case ): """simple docstring""" main() if __name__ == "__main__": main()
74
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class lowerCamelCase__ : def __init__( self ,A ,A ,A ): if dst_width < 0 or dst_height < 0: raise ValueError("""Destination width/height should be > 0""" ) UpperCAmelCase = img UpperCAmelCase = img.shape[1] UpperCAmelCase = img.shape[0] UpperCAmelCase = dst_width UpperCAmelCase = dst_height UpperCAmelCase = self.src_w / self.dst_w UpperCAmelCase = self.src_h / self.dst_h UpperCAmelCase = UpperCAmelCase = ( np.ones((self.dst_h, self.dst_w, 3) ,np.uinta ) * 255 ) def _UpperCamelCase ( self ): for i in range(self.dst_h ): for j in range(self.dst_w ): UpperCAmelCase = self.img[self.get_y(A )][self.get_x(A )] def _UpperCamelCase ( self ,A ): return int(self.ratio_x * x ) def _UpperCamelCase ( self ,A ): return int(self.ratio_y * y ) if __name__ == "__main__": _UpperCamelCase , _UpperCamelCase = 800, 600 _UpperCamelCase = imread("""image_data/lena.jpg""", 1) _UpperCamelCase = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F"""Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}""", n.output ) waitKey(0) destroyAllWindows()
74
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCamelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") _UpperCamelCase = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) _UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the training data.'''} ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the validation data.'''} ) SCREAMING_SNAKE_CASE = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) SCREAMING_SNAKE_CASE = field(default=32 , metadata={'''help''': '''The size of the square patches to use for masking.'''} ) SCREAMING_SNAKE_CASE = field( default=0.6 , metadata={'''help''': '''Percentage of patches to mask.'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def _UpperCamelCase ( self ): UpperCAmelCase = {} if self.train_dir is not None: UpperCAmelCase = self.train_dir if self.validation_dir is not None: UpperCAmelCase = self.validation_dir UpperCAmelCase = data_files if data_files else None @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ''' '''checkpoint identifier on the hub. ''' '''Don\'t set if you want to train a model from scratch.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(snake_case )} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'''} , ) SCREAMING_SNAKE_CASE = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''Name or path of preprocessor config.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each image. If not specified, will use `image_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Stride to use for the encoder.'''} , ) class lowerCamelCase__ : def __init__( self ,A=192 ,A=32 ,A=4 ,A=0.6 ): UpperCAmelCase = input_size UpperCAmelCase = mask_patch_size UpperCAmelCase = model_patch_size UpperCAmelCase = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError("""Input size must be divisible by mask patch size""" ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError("""Mask patch size must be divisible by model patch size""" ) UpperCAmelCase = self.input_size // self.mask_patch_size UpperCAmelCase = self.mask_patch_size // self.model_patch_size UpperCAmelCase = self.rand_size**2 UpperCAmelCase = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self ): UpperCAmelCase = np.random.permutation(self.token_count )[: self.mask_count] UpperCAmelCase = np.zeros(self.token_count ,dtype=A ) UpperCAmelCase = 1 UpperCAmelCase = mask.reshape((self.rand_size, self.rand_size) ) UpperCAmelCase = mask.repeat(self.scale ,axis=0 ).repeat(self.scale ,axis=1 ) return torch.tensor(mask.flatten() ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = torch.stack([example["""pixel_values"""] for example in examples] ) UpperCAmelCase = torch.stack([example["""mask"""] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mim""" , _snake_case , _snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_snake_case ) transformers.utils.logging.set_verbosity(_snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. UpperCAmelCase = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _snake_case ) and data_args.train_val_split > 0.0: UpperCAmelCase = ds["""train"""].train_test_split(data_args.train_val_split ) UpperCAmelCase = split["""train"""] UpperCAmelCase = split["""test"""] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.config_name_or_path , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_snake_case , """decoder_type""" ): UpperCAmelCase = """simmim""" # adapt config UpperCAmelCase = model_args.image_size if model_args.image_size is not None else config.image_size UpperCAmelCase = model_args.patch_size if model_args.patch_size is not None else config.patch_size UpperCAmelCase = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { """image_size""": model_args.image_size, """patch_size""": model_args.patch_size, """encoder_stride""": model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } UpperCAmelCase = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: UpperCAmelCase = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) UpperCAmelCase = AutoModelForMaskedImageModeling.from_config(_snake_case ) if training_args.do_train: UpperCAmelCase = ds["""train"""].column_names else: UpperCAmelCase = ds["""validation"""].column_names if data_args.image_column_name is not None: UpperCAmelCase = data_args.image_column_name elif "image" in column_names: UpperCAmelCase = """image""" elif "img" in column_names: UpperCAmelCase = """img""" else: UpperCAmelCase = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py UpperCAmelCase = Compose( [ Lambda(lambda _snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator UpperCAmelCase = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_snake_case ): UpperCAmelCase = [transforms(_snake_case ) for image in examples[image_column_name]] UpperCAmelCase = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: UpperCAmelCase = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: UpperCAmelCase = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_snake_case ) # Initialize our trainer UpperCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase = last_checkpoint UpperCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase = trainer.evaluate() trainer.log_metrics("""eval""" , _snake_case ) trainer.save_metrics("""eval""" , _snake_case ) # Write model card and (optionally) push to hub UpperCAmelCase = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """masked-image-modeling""", """dataset""": data_args.dataset_name, """tags""": ["""masked-image-modeling"""], } if training_args.push_to_hub: trainer.push_to_hub(**_snake_case ) else: trainer.create_model_card(**_snake_case ) if __name__ == "__main__": main()
74
1
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] UpperCAmelCase = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } UpperCAmelCase = F'''{src_lang}-{tgt_lang}''' UpperCAmelCase = F''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(_snake_case , exist_ok=_snake_case ) UpperCAmelCase = os.path.join(_snake_case , """README.md""" ) print(F'''Generating {path}''' ) with open(_snake_case , """w""" , encoding="""utf-8""" ) as f: f.write(_snake_case ) # make sure we are under the root of the project _UpperCamelCase = Path(__file__).resolve().parent.parent.parent _UpperCamelCase = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = model_name.split("""-""") _UpperCamelCase = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
74
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): _UpperCamelCase = True from torch.cuda.amp import autocast _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to log verbose messages or not.'''} , ) SCREAMING_SNAKE_CASE = field( default=2.0 , metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.5 , metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.99_99_95 , metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def _a ( _snake_case , _snake_case ): """simple docstring""" logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase = logging.WARNING if model_args.verbose_logging: UpperCAmelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase = logging.INFO logger.setLevel(_snake_case ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default='''train''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) SCREAMING_SNAKE_CASE = field( default='''validation''' , metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) } , ) SCREAMING_SNAKE_CASE = field( default='''file''' , metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) SCREAMING_SNAKE_CASE = field( default=1 , metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) SCREAMING_SNAKE_CASE = field( default=20.0 , metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = "longest" SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __call__( self ,A ): # reformat list to dict and set to pytorch format UpperCAmelCase = self.feature_extractor.pad( A ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="""pt""" ,) UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""input_values"""].shape[-1] ) UpperCAmelCase = batch["""input_values"""].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""attention_mask"""].sum(-1 ) ).to( torch.long ) UpperCAmelCase = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["""input_values"""].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase = 1 UpperCAmelCase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=A ,min_masks=2 ,) return batch class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=1 ,A=0 ,A=1.0 ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = 0 UpperCAmelCase = max_gumbel_temp UpperCAmelCase = min_gumbel_temp UpperCAmelCase = gumbel_temp_decay def _UpperCamelCase ( self ,A ,A ): model.train() UpperCAmelCase = self._prepare_inputs(A ) if self.use_amp: with autocast(): UpperCAmelCase = self.compute_loss(A ,A ) else: UpperCAmelCase = self.compute_loss(A ,A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase = loss.sum() / (inputs["""mask_time_indices"""]).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() configure_logger(_snake_case , _snake_case ) # Downloading and loading a dataset from the hub. UpperCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="""validation""" , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=_snake_case ) def prepare_dataset(_snake_case ): # check that all files have the correct sampling rate UpperCAmelCase , UpperCAmelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase = datasets.map( _snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["""train"""].column_names ) # filter audio files that are too long UpperCAmelCase = vectorized_datasets.filter( lambda _snake_case : len(data["""speech"""] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(_snake_case ): return feature_extractor(batch["""speech"""] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase = vectorized_datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["""train"""].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( """PreTraining is only supported for ``config.do_stable_layer_norm=True`` and""" """ ``config.feat_extract_norm='layer'""" ) UpperCAmelCase = WavaVecaForPreTraining(_snake_case ) UpperCAmelCase = DataCollatorForWavaVecaPretraining(model=_snake_case , feature_extractor=_snake_case ) UpperCAmelCase = WavaVecaPreTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , train_dataset=vectorized_datasets["""train"""] , eval_dataset=vectorized_datasets["""validation"""] , tokenizer=_snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
74
1
"""simple docstring""" import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) _UpperCamelCase = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation="""relu""") ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation="""relu""")) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=128, activation="""relu""")) classifier.add(layers.Dense(units=1, activation="""sigmoid""")) # Compiling the CNN classifier.compile( optimizer="""adam""", loss="""binary_crossentropy""", metrics=["""accuracy"""] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') _UpperCamelCase = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) _UpperCamelCase = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255) _UpperCamelCase = train_datagen.flow_from_directory( """dataset/training_set""", target_size=(64, 64), batch_size=32, class_mode="""binary""" ) _UpperCamelCase = test_datagen.flow_from_directory( """dataset/test_set""", target_size=(64, 64), batch_size=32, class_mode="""binary""" ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save("""cnn.h5""") # Part 3 - Making new predictions _UpperCamelCase = tf.keras.preprocessing.image.load_img( """dataset/single_prediction/image.png""", target_size=(64, 64) ) _UpperCamelCase = tf.keras.preprocessing.image.img_to_array(test_image) _UpperCamelCase = np.expand_dims(test_image, axis=0) _UpperCamelCase = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: _UpperCamelCase = """Normal""" if result[0][0] == 1: _UpperCamelCase = """Abnormality detected"""
74
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = 30 UpperCAmelCase = self.seq_length + self.mem_len UpperCAmelCase = 15 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 99 UpperCAmelCase = [10, 50, 80] UpperCAmelCase = 32 UpperCAmelCase = 32 UpperCAmelCase = 4 UpperCAmelCase = 8 UpperCAmelCase = 128 UpperCAmelCase = 2 UpperCAmelCase = 2 UpperCAmelCase = None UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 3 UpperCAmelCase = self.vocab_size - 1 UpperCAmelCase = 0.01 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCamelCase ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLLMHeadModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase , UpperCAmelCase = model([input_ids_a, mems_a] ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLForSequenceClassification(A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids_a} return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = () if is_tf_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,d_embed=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A ) def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: UpperCAmelCase = model_class(A ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: UpperCAmelCase = model.get_output_embeddings() assert isinstance(A ,tf.keras.layers.Layer ) UpperCAmelCase = model.get_bias() assert name is None else: UpperCAmelCase = model.get_output_embeddings() assert x is None UpperCAmelCase = model.get_bias() assert name is None def _UpperCamelCase ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _UpperCamelCase ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFTransfoXLModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""This model doesn't play well with fit() due to not returning a single loss.""" ) def _UpperCamelCase ( self ): pass @require_tf class lowerCamelCase__ ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off UpperCAmelCase = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off UpperCAmelCase = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> UpperCAmelCase = model.generate(A ,max_length=200 ,do_sample=A ) self.assertListEqual(output_ids[0].numpy().tolist() ,A )
74
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """facebook/xlm-roberta-xl""": """https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json""", """facebook/xlm-roberta-xxl""": """https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json""", # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''xlm-roberta-xl''' def __init__( self ,A=250_880 ,A=2_560 ,A=36 ,A=32 ,A=10_240 ,A="gelu" ,A=0.1 ,A=0.1 ,A=514 ,A=1 ,A=0.02 ,A=1e-0_5 ,A=1 ,A=0 ,A=2 ,A="absolute" ,A=True ,A=None ,**A ,): super().__init__(pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,**A ) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = position_embedding_type UpperCAmelCase = use_cache UpperCAmelCase = classifier_dropout class lowerCamelCase__ ( snake_case ): @property def _UpperCamelCase ( self ): if self.task == "multiple-choice": UpperCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: UpperCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
74
"""simple docstring""" from math import sqrt def _a ( _snake_case = 100_0000 ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_snake_case , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
74
1
"""simple docstring""" import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class lowerCamelCase__ : @staticmethod def _UpperCamelCase ( *A ,**A ): pass @is_pipeline_test @require_torch @require_vision class lowerCamelCase__ ( unittest.TestCase ): SCREAMING_SNAKE_CASE = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def _UpperCamelCase ( self ,A ,A ,A ): UpperCAmelCase = pipeline("""visual-question-answering""" ,model="""hf-internal-testing/tiny-vilt-random-vqa""" ) UpperCAmelCase = [ { """image""": Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ), """question""": """How many cats are there?""", }, { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """question""": """How many cats are there?""", }, ] return vqa_pipeline, examples def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = vqa_pipeline(A ,top_k=1 ) self.assertEqual( A ,[ [{"""score""": ANY(A ), """answer""": ANY(A )}], [{"""score""": ANY(A ), """answer""": ANY(A )}], ] ,) @require_torch def _UpperCamelCase ( self ): UpperCAmelCase = pipeline("""visual-question-answering""" ,model="""hf-internal-testing/tiny-vilt-random-vqa""" ) UpperCAmelCase = """./tests/fixtures/tests_samples/COCO/000000039769.png""" UpperCAmelCase = """How many cats are there?""" UpperCAmelCase = vqa_pipeline(image=A ,question="""How many cats are there?""" ,top_k=2 ) self.assertEqual( A ,[{"""score""": ANY(A ), """answer""": ANY(A )}, {"""score""": ANY(A ), """answer""": ANY(A )}] ) UpperCAmelCase = vqa_pipeline({"""image""": image, """question""": question} ,top_k=2 ) self.assertEqual( A ,[{"""score""": ANY(A ), """answer""": ANY(A )}, {"""score""": ANY(A ), """answer""": ANY(A )}] ) @slow @require_torch def _UpperCamelCase ( self ): UpperCAmelCase = pipeline("""visual-question-answering""" ,model="""dandelin/vilt-b32-finetuned-vqa""" ) UpperCAmelCase = """./tests/fixtures/tests_samples/COCO/000000039769.png""" UpperCAmelCase = """How many cats are there?""" UpperCAmelCase = vqa_pipeline(image=A ,question=A ,top_k=2 ) self.assertEqual( nested_simplify(A ,decimals=4 ) ,[{"""score""": 0.8799, """answer""": """2"""}, {"""score""": 0.296, """answer""": """1"""}] ) UpperCAmelCase = vqa_pipeline({"""image""": image, """question""": question} ,top_k=2 ) self.assertEqual( nested_simplify(A ,decimals=4 ) ,[{"""score""": 0.8799, """answer""": """2"""}, {"""score""": 0.296, """answer""": """1"""}] ) UpperCAmelCase = vqa_pipeline( [{"""image""": image, """question""": question}, {"""image""": image, """question""": question}] ,top_k=2 ) self.assertEqual( nested_simplify(A ,decimals=4 ) ,[[{"""score""": 0.8799, """answer""": """2"""}, {"""score""": 0.296, """answer""": """1"""}]] * 2 ,) @require_tf @unittest.skip("""Visual question answering not implemented in TF""" ) def _UpperCamelCase ( self ): pass
74
"""simple docstring""" import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _UpperCamelCase = { """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } _UpperCamelCase = AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def _a ( _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = create_model( """HTSAT-tiny""" , """roberta""" , _snake_case , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=_snake_case , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = {} UpperCAmelCase = R""".*sequential.(\d+).*""" UpperCAmelCase = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: UpperCAmelCase = key.replace(_snake_case , _snake_case ) if re.match(_snake_case , _snake_case ): # replace sequential layers with list UpperCAmelCase = re.match(_snake_case , _snake_case ).group(1 ) UpperCAmelCase = key.replace(F'''sequential.{sequential_layer}.''' , F'''layers.{int(_snake_case )//3}.linear.''' ) elif re.match(_snake_case , _snake_case ): UpperCAmelCase = int(re.match(_snake_case , _snake_case ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... UpperCAmelCase = 1 if projecton_layer == 0 else 2 UpperCAmelCase = key.replace(F'''_projection.{projecton_layer}.''' , F'''_projection.linear{transformers_projection_layer}.''' ) if "audio" and "qkv" in key: # split qkv into query key and value UpperCAmelCase = value UpperCAmelCase = mixed_qkv.size(0 ) // 3 UpperCAmelCase = mixed_qkv[:qkv_dim] UpperCAmelCase = mixed_qkv[qkv_dim : qkv_dim * 2] UpperCAmelCase = mixed_qkv[qkv_dim * 2 :] UpperCAmelCase = query_layer UpperCAmelCase = key_layer UpperCAmelCase = value_layer else: UpperCAmelCase = value return model_state_dict def _a ( _snake_case , _snake_case , _snake_case , _snake_case=False ): """simple docstring""" UpperCAmelCase , UpperCAmelCase = init_clap(_snake_case , enable_fusion=_snake_case ) clap_model.eval() UpperCAmelCase = clap_model.state_dict() UpperCAmelCase = rename_state_dict(_snake_case ) UpperCAmelCase = ClapConfig() UpperCAmelCase = enable_fusion UpperCAmelCase = ClapModel(_snake_case ) # ignore the spectrogram embedding layer model.load_state_dict(_snake_case , strict=_snake_case ) model.save_pretrained(_snake_case ) transformers_config.save_pretrained(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") _UpperCamelCase = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
74
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { """microsoft/swin-tiny-patch4-window7-224""": ( """https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json""" ), # See all Swin models at https://huggingface.co/models?filter=swin } class lowerCamelCase__ ( snake_case , snake_case ): SCREAMING_SNAKE_CASE = '''swin''' SCREAMING_SNAKE_CASE = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self ,A=224 ,A=4 ,A=3 ,A=96 ,A=[2, 2, 6, 2] ,A=[3, 6, 12, 24] ,A=7 ,A=4.0 ,A=True ,A=0.0 ,A=0.0 ,A=0.1 ,A="gelu" ,A=False ,A=0.02 ,A=1e-5 ,A=32 ,A=None ,A=None ,**A ,): super().__init__(**A ) UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = embed_dim UpperCAmelCase = depths UpperCAmelCase = len(A ) UpperCAmelCase = num_heads UpperCAmelCase = window_size UpperCAmelCase = mlp_ratio UpperCAmelCase = qkv_bias UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = drop_path_rate UpperCAmelCase = hidden_act UpperCAmelCase = use_absolute_embeddings UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCAmelCase = int(embed_dim * 2 ** (len(A ) - 1) ) UpperCAmelCase = ["""stem"""] + [F'''stage{idx}''' for idx in range(1 ,len(A ) + 1 )] UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices( out_features=A ,out_indices=A ,stage_names=self.stage_names ) class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = version.parse('''1.11''' ) @property def _UpperCamelCase ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _UpperCamelCase ( self ): return 1e-4
74
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _UpperCamelCase = abspath(join(dirname(dirname(__file__)), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_snake_case ) def _a ( _snake_case ): """simple docstring""" from diffusers.utils.testing_utils import pytest_terminal_summary_main UpperCAmelCase = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(_snake_case , id=_snake_case )
74
1
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = [] if len(_snake_case ) == 1: return [nums.copy()] for _ in range(len(_snake_case ) ): UpperCAmelCase = nums.pop(0 ) UpperCAmelCase = permute(_snake_case ) for perm in permutations: perm.append(_snake_case ) result.extend(_snake_case ) nums.append(_snake_case ) return result def _a ( _snake_case ): """simple docstring""" def backtrack(_snake_case ): if start == len(_snake_case ) - 1: output.append(nums[:] ) else: for i in range(_snake_case , len(_snake_case ) ): UpperCAmelCase , UpperCAmelCase = nums[i], nums[start] backtrack(start + 1 ) UpperCAmelCase , UpperCAmelCase = nums[i], nums[start] # backtrack UpperCAmelCase = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function _UpperCamelCase = permutea([1, 2, 3]) print(res) doctest.testmod()
74
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class lowerCamelCase__ : def __init__( self ,A ,A ): if len(A ) != degree + 1: raise ValueError( """The number of coefficients should be equal to the degree + 1.""" ) UpperCAmelCase = list(A ) UpperCAmelCase = degree def __add__( self ,A ): if self.degree > polynomial_a.degree: UpperCAmelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,A ) else: UpperCAmelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,A ) def __sub__( self ,A ): return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self ): return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self ,A ): UpperCAmelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): UpperCAmelCase = """""" for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(A ) return polynomial def __repr__( self ): return self.__str__() def _UpperCamelCase ( self ): UpperCAmelCase = [0] * self.degree for i in range(self.degree ): UpperCAmelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,A ) def _UpperCamelCase ( self ,A = 0 ): UpperCAmelCase = [0] * (self.degree + 2) UpperCAmelCase = constant for i in range(self.degree + 1 ): UpperCAmelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,A ) def __eq__( self ,A ): if not isinstance(A ,A ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self ,A ): return not self.__eq__(A )
74
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {} class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = '''llama''' SCREAMING_SNAKE_CASE = ['''past_key_values'''] def __init__( self ,A=32_000 ,A=4_096 ,A=11_008 ,A=32 ,A=32 ,A=None ,A="silu" ,A=2_048 ,A=0.02 ,A=1e-6 ,A=True ,A=0 ,A=1 ,A=2 ,A=1 ,A=False ,A=None ,**A ,): UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = hidden_size UpperCAmelCase = intermediate_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads # for backward compatibility if num_key_value_heads is None: UpperCAmelCase = num_attention_heads UpperCAmelCase = num_key_value_heads UpperCAmelCase = hidden_act UpperCAmelCase = initializer_range UpperCAmelCase = rms_norm_eps UpperCAmelCase = pretraining_tp UpperCAmelCase = use_cache UpperCAmelCase = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A ,bos_token_id=A ,eos_token_id=A ,tie_word_embeddings=A ,**A ,) def _UpperCamelCase ( self ): if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ F'''got {self.rope_scaling}''' ) UpperCAmelCase = self.rope_scaling.get("""type""" ,A ) UpperCAmelCase = self.rope_scaling.get("""factor""" ,A ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(A ,A ) or rope_scaling_factor <= 1.0: raise ValueError(F'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
74
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy as np import tensorflow as tf from transformers import ( TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST, FlaubertConfig, TFFlaubertForMultipleChoice, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForSequenceClassification, TFFlaubertForTokenClassification, TFFlaubertModel, TFFlaubertWithLMHeadModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = 2 UpperCAmelCase = 99 UpperCAmelCase = 0 UpperCAmelCase = 32 UpperCAmelCase = 2 UpperCAmelCase = 4 UpperCAmelCase = 0.1 UpperCAmelCase = 0.1 UpperCAmelCase = 512 UpperCAmelCase = 16 UpperCAmelCase = 2 UpperCAmelCase = 0.02 UpperCAmelCase = 3 UpperCAmelCase = 4 UpperCAmelCase = """last""" UpperCAmelCase = True UpperCAmelCase = None UpperCAmelCase = 0 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ,dtype=tf.floataa ) UpperCAmelCase = None if self.use_input_lengths: UpperCAmelCase = ( ids_tensor([self.batch_size] ,vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.n_langs ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] ,2 ,dtype=tf.floataa ) UpperCAmelCase = ids_tensor([self.batch_size] ,self.num_choices ) UpperCAmelCase = FlaubertConfig( vocab_size=self.vocab_size ,n_special=self.n_special ,emb_dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,gelu_activation=self.gelu_activation ,sinusoidal_embeddings=self.sinusoidal_embeddings ,asm=self.asm ,causal=self.causal ,n_langs=self.n_langs ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,summary_type=self.summary_type ,use_proj=self.use_proj ,bos_token_id=self.bos_token_id ,) return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertModel(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) UpperCAmelCase = [input_ids, input_mask] UpperCAmelCase = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertWithLMHeadModel(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths, """langs""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForQuestionAnsweringSimple(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = TFFlaubertForSequenceClassification(A ) UpperCAmelCase = {"""input_ids""": input_ids, """lengths""": input_lengths} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_labels UpperCAmelCase = TFFlaubertForTokenClassification(config=A ) UpperCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self ,A ,A ,A ,A ,A ,A ,A ,A ,A ,): UpperCAmelCase = self.num_choices UpperCAmelCase = TFFlaubertForMultipleChoice(config=A ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = tf.tile(tf.expand_dims(A ,1 ) ,(1, self.num_choices, 1) ) UpperCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """langs""": token_type_ids, """lengths""": input_lengths, } return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( ( TFFlaubertModel, TFFlaubertWithLMHeadModel, TFFlaubertForSequenceClassification, TFFlaubertForQuestionAnsweringSimple, TFFlaubertForTokenClassification, TFFlaubertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = ( (TFFlaubertWithLMHeadModel,) if is_tf_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFFlaubertModel, '''fill-mask''': TFFlaubertWithLMHeadModel, '''question-answering''': TFFlaubertForQuestionAnsweringSimple, '''text-classification''': TFFlaubertForSequenceClassification, '''token-classification''': TFFlaubertForTokenClassification, '''zero-shot''': TFFlaubertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,emb_dim=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_token_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_for_multiple_choice(*A ) @slow def _UpperCamelCase ( self ): for model_name in TF_FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFFlaubertModel.from_pretrained(A ) self.assertIsNotNone(A ) @require_tf @require_sentencepiece @require_tokenizers class lowerCamelCase__ ( unittest.TestCase ): @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFFlaubertModel.from_pretrained("""jplu/tf-flaubert-small-cased""" ) UpperCAmelCase = tf.convert_to_tensor( [[0, 158, 735, 2_592, 1_424, 6_727, 82, 1]] ,dtype=tf.intaa ,) # "J'aime flaubert !" UpperCAmelCase = model(A )[0] UpperCAmelCase = tf.TensorShape((1, 8, 512) ) self.assertEqual(output.shape ,A ) # compare the actual values for a slice. UpperCAmelCase = tf.convert_to_tensor( [ [ [-1.8768773, -1.566555, 0.27072418], [-1.6920038, -0.5873505, 1.9329599], [-2.9563985, -1.6993835, 1.7972052], ] ] ,dtype=tf.floataa ,) self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1e-4 ) )
74
1
"""simple docstring""" from __future__ import annotations import math _UpperCamelCase = """2020.9.26""" _UpperCamelCase = """xcodz-dot, cclaus, dhruvmanila""" def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" if not all(isinstance(_snake_case , (float, int) ) for val in locals().values() ): UpperCAmelCase = F'''Input values must either be float or int: {list(locals().values() )}''' raise TypeError(_snake_case ) UpperCAmelCase = ((x * distance) / (z + distance)) * scale UpperCAmelCase = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise TypeError("""Axis must be a str""" ) UpperCAmelCase = locals() del input_variables["axis"] if not all(isinstance(_snake_case , (float, int) ) for val in input_variables.values() ): UpperCAmelCase = ( """Input values except axis must either be float or int: """ F'''{list(input_variables.values() )}''' ) raise TypeError(_snake_case ) UpperCAmelCase = (angle % 360) / 450 * 180 / math.pi if axis == "z": UpperCAmelCase = x * math.cos(_snake_case ) - y * math.sin(_snake_case ) UpperCAmelCase = y * math.cos(_snake_case ) + x * math.sin(_snake_case ) UpperCAmelCase = z elif axis == "x": UpperCAmelCase = y * math.cos(_snake_case ) - z * math.sin(_snake_case ) UpperCAmelCase = z * math.cos(_snake_case ) + y * math.sin(_snake_case ) UpperCAmelCase = x elif axis == "y": UpperCAmelCase = x * math.cos(_snake_case ) - z * math.sin(_snake_case ) UpperCAmelCase = z * math.cos(_snake_case ) + x * math.sin(_snake_case ) UpperCAmelCase = y else: raise ValueError("""not a valid axis, choose one of 'x', 'y', 'z'""" ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(F"""{convert_to_ad(1.0, 2.0, 3.0, 10.0, 10.0) = }""") print(F"""{rotate(1.0, 2.0, 3.0, "y", 90.0) = }""")
74
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): UpperCAmelCase = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): UpperCAmelCase = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] UpperCAmelCase = key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(_snake_case )-1}''' ) if "norm" in key: UpperCAmelCase = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] UpperCAmelCase = key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(_snake_case )-1}''' ) if "layer_norm1" in key: UpperCAmelCase = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: UpperCAmelCase = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase = key[key.find("""block""" ) + len("""block""" )] UpperCAmelCase = key.replace(F'''block{idx}''' , F'''block.{int(_snake_case )-1}''' ) if "attn.q" in key: UpperCAmelCase = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: UpperCAmelCase = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: UpperCAmelCase = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: UpperCAmelCase = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: UpperCAmelCase = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: UpperCAmelCase = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: UpperCAmelCase = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) UpperCAmelCase = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase = key[key.find("""linear_c""" ) + len("""linear_c""" )] UpperCAmelCase = key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(_snake_case )-1}''' ) if "bot_conv" in key: UpperCAmelCase = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: UpperCAmelCase = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: UpperCAmelCase = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: UpperCAmelCase = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: UpperCAmelCase = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: UpperCAmelCase = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: UpperCAmelCase = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): UpperCAmelCase = key.replace("""module.last_layer_depth""" , """head.head""" ) UpperCAmelCase = value return new_state_dict def _a ( _snake_case , _snake_case ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict UpperCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase = kv_bias[config.hidden_sizes[i] :] def _a ( ): """simple docstring""" UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case=False , _snake_case=None ): """simple docstring""" UpperCAmelCase = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) UpperCAmelCase = GLPNImageProcessor() # prepare image UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=_snake_case , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict UpperCAmelCase = torch.load(_snake_case , map_location=torch.device("""cpu""" ) ) # rename keys UpperCAmelCase = rename_keys(_snake_case ) # key and value matrices need special treatment read_in_k_v(_snake_case , _snake_case ) # create HuggingFace model and load state dict UpperCAmelCase = GLPNForDepthEstimation(_snake_case ) model.load_state_dict(_snake_case ) model.eval() # forward pass UpperCAmelCase = model(_snake_case ) UpperCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: UpperCAmelCase = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: UpperCAmelCase = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) UpperCAmelCase = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _snake_case , atol=1E-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=_snake_case , ) image_processor.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=_snake_case , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCamelCase = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
74
1
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline _UpperCamelCase = """path-to-your-trained-model""" _UpperCamelCase = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") _UpperCamelCase = """A photo of sks dog in a bucket""" _UpperCamelCase = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
74
"""simple docstring""" def _a ( _snake_case ): # noqa: E741 """simple docstring""" UpperCAmelCase = len(_snake_case ) UpperCAmelCase = 0 UpperCAmelCase = [0] * n UpperCAmelCase = [False] * n UpperCAmelCase = [False] * n def dfs(_snake_case , _snake_case , _snake_case , _snake_case ): if parent == root: out_edge_count += 1 UpperCAmelCase = True UpperCAmelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase = dfs(_snake_case , _snake_case , _snake_case , _snake_case ) UpperCAmelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase = True # AP found via cycle if at == low[to]: UpperCAmelCase = True else: UpperCAmelCase = min(low[at] , _snake_case ) return out_edge_count for i in range(_snake_case ): if not visited[i]: UpperCAmelCase = 0 UpperCAmelCase = dfs(_snake_case , _snake_case , -1 , _snake_case ) UpperCAmelCase = out_edge_count > 1 for x in range(len(_snake_case ) ): if is_art[x] is True: print(_snake_case ) # Adjacency list of graph _UpperCamelCase = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
74
1
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def _a ( _snake_case="ro" , _snake_case="en" , _snake_case="wmt16" , _snake_case=None ): """simple docstring""" try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("""run pip install datasets""" ) UpperCAmelCase = F'''{src_lang}-{tgt_lang}''' print(F'''Converting {dataset}-{pair}''' ) UpperCAmelCase = datasets.load_dataset(_snake_case , _snake_case ) if save_dir is None: UpperCAmelCase = F'''{dataset}-{pair}''' UpperCAmelCase = Path(_snake_case ) save_dir.mkdir(exist_ok=_snake_case ) for split in ds.keys(): print(F'''Splitting {split} with {ds[split].num_rows} records''' ) # to save to val.source, val.target like summary datasets UpperCAmelCase = """val""" if split == """validation""" else split UpperCAmelCase = save_dir.joinpath(F'''{fn}.source''' ) UpperCAmelCase = save_dir.joinpath(F'''{fn}.target''' ) UpperCAmelCase = src_path.open("""w+""" ) UpperCAmelCase = tgt_path.open("""w+""" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): UpperCAmelCase = x["""translation"""] src_fp.write(ex[src_lang] + """\n""" ) tgt_fp.write(ex[tgt_lang] + """\n""" ) print(F'''Saved {dataset} dataset to {save_dir}''' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
74
"""simple docstring""" _UpperCamelCase = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ _UpperCamelCase = [{"""type""": """code""", """content""": INSTALL_CONTENT}] _UpperCamelCase = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
74
1
"""simple docstring""" import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class lowerCamelCase__ : def __init__( self ,A ,A=13 ,A=7 ,A=True ,A=True ,A=True ,A=True ,A=99 ,A=32 ,A=5 ,A=4 ,A=4 ,A="gelu" ,A=0.0 ,A=0.1 ,A=True ,A=512 ,A=16 ,A=2 ,A=0.02 ,A=3 ,A=4 ,A=None ,): UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_multiple_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout UpperCAmelCase = attention_dropout UpperCAmelCase = weight_tying UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) UpperCAmelCase = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self ): return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_multiple_size=self.intermediate_multiple_size ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,weight_tying=self.weight_tying ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A ,initializer_range=self.initializer_range ,) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase = True return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self ,A ,A ,A ): UpperCAmelCase = GPTNeoXJapaneseModel(config=A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ): UpperCAmelCase = True UpperCAmelCase = GPTNeoXJapaneseModel(A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = GPTNeoXJapaneseForCausalLM(config=A ) model.to(A ) model.eval() UpperCAmelCase = model(A ,attention_mask=A ,labels=A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self ,A ,A ,A ): UpperCAmelCase = True UpperCAmelCase = GPTNeoXJapaneseForCausalLM(config=A ) model.to(A ) model.eval() # first forward pass UpperCAmelCase = model(A ,attention_mask=A ,use_cache=A ) UpperCAmelCase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase = ids_tensor((self.batch_size, 3) ,config.vocab_size ) UpperCAmelCase = ids_tensor((self.batch_size, 3) ,vocab_size=2 ) # append to next input_ids and UpperCAmelCase = torch.cat([input_ids, next_tokens] ,dim=-1 ) UpperCAmelCase = torch.cat([input_mask, next_mask] ,dim=-1 ) UpperCAmelCase = model(A ,attention_mask=A ,output_hidden_states=A ) UpperCAmelCase = output_from_no_past["""hidden_states"""][0] UpperCAmelCase = model( A ,attention_mask=A ,past_key_values=A ,output_hidden_states=A ,)["""hidden_states"""][0] # select random slice UpperCAmelCase = ids_tensor((1,) ,output_from_past.shape[-1] ).item() UpperCAmelCase = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A ,A ,atol=1e-3 ) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () SCREAMING_SNAKE_CASE = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () SCREAMING_SNAKE_CASE = ( {'''feature-extraction''': GPTNeoXJapaneseModel, '''text-generation''': GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ): UpperCAmelCase = GPTNeoXJapaneseModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,hidden_size=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(A ,A ,A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(A ,A ,A ) def _UpperCamelCase ( self ): # This regression test was failing with PyTorch < 1.3 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_decoder() UpperCAmelCase = None self.model_tester.create_and_check_model_as_decoder(A ,A ,A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(A ,A ,A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*A ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = """abeja/gpt-neox-japanese-2.7b""" UpperCAmelCase = ["""データサイエンティストとは、""", """100年後に必要とされる会社は、""", """フルリモートの環境で働くために必要なことは、""", """国境の長いトンネルを抜けると""", """美味しい日本食といえば、"""] UpperCAmelCase = [ """データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。""", """100年後に必要とされる会社は、「人」が中心の会社です。""", """フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。""", """国境の長いトンネルを抜けると、そこは雪国だった。""", """美味しい日本食といえば、やっぱりお寿司ですよね。""", ] UpperCAmelCase = GPTNeoXJapaneseTokenizer.from_pretrained(A ) UpperCAmelCase = GPTNeoXJapaneseForCausalLM.from_pretrained(A ) UpperCAmelCase = [] for prompt in prompts: UpperCAmelCase = tokenizer(A ,return_tensors="""pt""" ).input_ids UpperCAmelCase = model.generate(A ,max_length=50 ) UpperCAmelCase = tokenizer.batch_decode(A ,skip_special_tokens=A ) predicted_outputs += generated_string self.assertListEqual(A ,A )
74
"""simple docstring""" import argparse import struct import unittest class lowerCamelCase__ : def __init__( self ,A ): UpperCAmelCase = data # Initialize hash values UpperCAmelCase = [ 0x6A_09_E6_67, 0xBB_67_AE_85, 0x3C_6E_F3_72, 0xA5_4F_F5_3A, 0x51_0E_52_7F, 0x9B_05_68_8C, 0x1F_83_D9_AB, 0x5B_E0_CD_19, ] # Initialize round constants UpperCAmelCase = [ 0x42_8A_2F_98, 0x71_37_44_91, 0xB5_C0_FB_CF, 0xE9_B5_DB_A5, 0x39_56_C2_5B, 0x59_F1_11_F1, 0x92_3F_82_A4, 0xAB_1C_5E_D5, 0xD8_07_AA_98, 0x12_83_5B_01, 0x24_31_85_BE, 0x55_0C_7D_C3, 0x72_BE_5D_74, 0x80_DE_B1_FE, 0x9B_DC_06_A7, 0xC1_9B_F1_74, 0xE4_9B_69_C1, 0xEF_BE_47_86, 0x0F_C1_9D_C6, 0x24_0C_A1_CC, 0x2D_E9_2C_6F, 0x4A_74_84_AA, 0x5C_B0_A9_DC, 0x76_F9_88_DA, 0x98_3E_51_52, 0xA8_31_C6_6D, 0xB0_03_27_C8, 0xBF_59_7F_C7, 0xC6_E0_0B_F3, 0xD5_A7_91_47, 0x06_CA_63_51, 0x14_29_29_67, 0x27_B7_0A_85, 0x2E_1B_21_38, 0x4D_2C_6D_FC, 0x53_38_0D_13, 0x65_0A_73_54, 0x76_6A_0A_BB, 0x81_C2_C9_2E, 0x92_72_2C_85, 0xA2_BF_E8_A1, 0xA8_1A_66_4B, 0xC2_4B_8B_70, 0xC7_6C_51_A3, 0xD1_92_E8_19, 0xD6_99_06_24, 0xF4_0E_35_85, 0x10_6A_A0_70, 0x19_A4_C1_16, 0x1E_37_6C_08, 0x27_48_77_4C, 0x34_B0_BC_B5, 0x39_1C_0C_B3, 0x4E_D8_AA_4A, 0x5B_9C_CA_4F, 0x68_2E_6F_F3, 0x74_8F_82_EE, 0x78_A5_63_6F, 0x84_C8_78_14, 0x8C_C7_02_08, 0x90_BE_FF_FA, 0xA4_50_6C_EB, 0xBE_F9_A3_F7, 0xC6_71_78_F2, ] UpperCAmelCase = self.preprocessing(self.data ) self.final_hash() @staticmethod def _UpperCamelCase ( A ): UpperCAmelCase = b"""\x80""" + (b"""\x00""" * (63 - (len(A ) + 8) % 64)) UpperCAmelCase = struct.pack(""">Q""" ,(len(A ) * 8) ) return data + padding + big_endian_integer def _UpperCamelCase ( self ): # Convert into blocks of 64 bytes UpperCAmelCase = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers UpperCAmelCase = list(struct.unpack(""">16L""" ,A ) ) # add 48 0-ed integers words += [0] * 48 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array UpperCAmelCase = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) UpperCAmelCase = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) UpperCAmelCase = ( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_00_00_00_00 # Compression UpperCAmelCase = self.ror(A ,6 ) ^ self.ror(A ,11 ) ^ self.ror(A ,25 ) UpperCAmelCase = (e & f) ^ ((~e & 0xFF_FF_FF_FF) & g) UpperCAmelCase = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_00_00_00_00 UpperCAmelCase = self.ror(A ,2 ) ^ self.ror(A ,13 ) ^ self.ror(A ,22 ) UpperCAmelCase = (a & b) ^ (a & c) ^ (b & c) UpperCAmelCase = (sa + maj) % 0x1_00_00_00_00 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = ( g, f, e, ((d + tempa) % 0x1_00_00_00_00), c, b, a, ((tempa + tempa) % 0x1_00_00_00_00), ) UpperCAmelCase = [a, b, c, d, e, f, g, h] # Modify final values UpperCAmelCase = [ ((element + mutated_hash_values[index]) % 0x1_00_00_00_00) for index, element in enumerate(self.hashes ) ] UpperCAmelCase = """""".join([hex(A )[2:].zfill(8 ) for value in self.hashes] ) def _UpperCamelCase ( self ,A ,A ): return 0xFF_FF_FF_FF & (value << (32 - rotations)) | (value >> rotations) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): import hashlib UpperCAmelCase = bytes("""Test String""" ,"""utf-8""" ) self.assertEqual(SHAaaa(A ).hash ,hashlib.shaaaa(A ).hexdigest() ) def _a ( ): """simple docstring""" import doctest doctest.testmod() UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) UpperCAmelCase = parser.parse_args() UpperCAmelCase = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: UpperCAmelCase = f.read() else: UpperCAmelCase = bytes(_snake_case , """utf-8""" ) print(SHAaaa(_snake_case ).hash ) if __name__ == "__main__": main()
74
1
"""simple docstring""" import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(""">=""", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType _UpperCamelCase = get_logger(__name__) def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=0 ): """simple docstring""" os.makedirs(_snake_case , exist_ok=_snake_case ) with FSDP.state_dict_type( _snake_case , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): UpperCAmelCase = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: UpperCAmelCase = F'''{MODEL_NAME}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}.bin''' UpperCAmelCase = os.path.join(_snake_case , _snake_case ) if accelerator.process_index == 0: logger.info(F'''Saving model to {output_model_file}''' ) torch.save(_snake_case , _snake_case ) logger.info(F'''Model saved to {output_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: UpperCAmelCase = ( F'''{MODEL_NAME}_rank{accelerator.process_index}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin''' ) UpperCAmelCase = os.path.join(_snake_case , _snake_case ) logger.info(F'''Saving model to {output_model_file}''' ) torch.save(_snake_case , _snake_case ) logger.info(F'''Model saved to {output_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: UpperCAmelCase = os.path.join(_snake_case , F'''{MODEL_NAME}_{model_index}''' ) os.makedirs(_snake_case , exist_ok=_snake_case ) logger.info(F'''Saving model to {ckpt_dir}''' ) UpperCAmelCase = {"""model""": state_dict} dist_cp.save_state_dict( state_dict=_snake_case , storage_writer=dist_cp.FileSystemWriter(_snake_case ) , planner=DefaultSavePlanner() , ) logger.info(F'''Model saved to {ckpt_dir}''' ) def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=0 ): """simple docstring""" accelerator.wait_for_everyone() with FSDP.state_dict_type( _snake_case , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(_snake_case ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( """Set the `sync_module_states` flag to `True` so that model states are synced across processes when """ """initializing FSDP object""" ) return UpperCAmelCase = F'''{MODEL_NAME}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}.bin''' UpperCAmelCase = os.path.join(_snake_case , _snake_case ) logger.info(F'''Loading model from {input_model_file}''' ) UpperCAmelCase = torch.load(_snake_case ) logger.info(F'''Model loaded from {input_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: UpperCAmelCase = ( F'''{MODEL_NAME}_rank{accelerator.process_index}.bin''' if model_index == 0 else F'''{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin''' ) UpperCAmelCase = os.path.join(_snake_case , _snake_case ) logger.info(F'''Loading model from {input_model_file}''' ) UpperCAmelCase = torch.load(_snake_case ) logger.info(F'''Model loaded from {input_model_file}''' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: UpperCAmelCase = ( os.path.join(_snake_case , F'''{MODEL_NAME}_{model_index}''' ) if F'''{MODEL_NAME}''' not in input_dir else input_dir ) logger.info(F'''Loading model from {ckpt_dir}''' ) UpperCAmelCase = {"""model""": model.state_dict()} dist_cp.load_state_dict( state_dict=_snake_case , storage_reader=dist_cp.FileSystemReader(_snake_case ) , planner=DefaultLoadPlanner() , ) UpperCAmelCase = state_dict["""model"""] logger.info(F'''Model loaded from {ckpt_dir}''' ) model.load_state_dict(_snake_case ) def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=0 ): """simple docstring""" os.makedirs(_snake_case , exist_ok=_snake_case ) with FSDP.state_dict_type( _snake_case , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): UpperCAmelCase = FSDP.optim_state_dict(_snake_case , _snake_case ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: UpperCAmelCase = ( F'''{OPTIMIZER_NAME}.bin''' if optimizer_index == 0 else F'''{OPTIMIZER_NAME}_{optimizer_index}.bin''' ) UpperCAmelCase = os.path.join(_snake_case , _snake_case ) logger.info(F'''Saving Optimizer state to {output_optimizer_file}''' ) torch.save(_snake_case , _snake_case ) logger.info(F'''Optimizer state saved in {output_optimizer_file}''' ) else: UpperCAmelCase = os.path.join(_snake_case , F'''{OPTIMIZER_NAME}_{optimizer_index}''' ) os.makedirs(_snake_case , exist_ok=_snake_case ) logger.info(F'''Saving Optimizer state to {ckpt_dir}''' ) dist_cp.save_state_dict( state_dict={"""optimizer""": optim_state} , storage_writer=dist_cp.FileSystemWriter(_snake_case ) , planner=DefaultSavePlanner() , ) logger.info(F'''Optimizer state saved in {ckpt_dir}''' ) def _a ( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case=0 ): """simple docstring""" accelerator.wait_for_everyone() with FSDP.state_dict_type( _snake_case , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: UpperCAmelCase = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: UpperCAmelCase = ( F'''{OPTIMIZER_NAME}.bin''' if optimizer_index == 0 else F'''{OPTIMIZER_NAME}_{optimizer_index}.bin''' ) UpperCAmelCase = os.path.join(_snake_case , _snake_case ) logger.info(F'''Loading Optimizer state from {input_optimizer_file}''' ) UpperCAmelCase = torch.load(_snake_case ) logger.info(F'''Optimizer state loaded from {input_optimizer_file}''' ) else: UpperCAmelCase = ( os.path.join(_snake_case , F'''{OPTIMIZER_NAME}_{optimizer_index}''' ) if F'''{OPTIMIZER_NAME}''' not in input_dir else input_dir ) logger.info(F'''Loading Optimizer from {ckpt_dir}''' ) UpperCAmelCase = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key="""optimizer""" , storage_reader=dist_cp.FileSystemReader(_snake_case ) , ) UpperCAmelCase = optim_state["""optimizer"""] logger.info(F'''Optimizer loaded from {ckpt_dir}''' ) UpperCAmelCase = FSDP.optim_state_dict_to_load(_snake_case , _snake_case , _snake_case ) optimizer.load_state_dict(_snake_case )
74
"""simple docstring""" def _a ( _snake_case = 10 , _snake_case = 22 ): """simple docstring""" UpperCAmelCase = range(1 , _snake_case ) UpperCAmelCase = range(1 , _snake_case ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(F"""{solution(10, 22) = }""")
74
1
"""simple docstring""" import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = False, False, False @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = None # Automatically constructed SCREAMING_SNAKE_CASE = "dict" SCREAMING_SNAKE_CASE = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) SCREAMING_SNAKE_CASE = field(default='''Audio''' , init=snake_case , repr=snake_case ) def __call__( self ): return self.pa_type def _UpperCamelCase ( self ,A ): try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("""To support encoding audio data, please install 'soundfile'.""" ) from err if isinstance(A ,A ): return {"bytes": None, "path": value} elif isinstance(A ,A ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes UpperCAmelCase = BytesIO() sf.write(A ,value["""array"""] ,value["""sampling_rate"""] ,format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("""pcm""" ): # "PCM" only has raw audio bytes if value.get("""sampling_rate""" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("""To use PCM files, please specify a 'sampling_rate' in Audio object""" ) if value.get("""bytes""" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) UpperCAmelCase = np.frombuffer(value["""bytes"""] ,dtype=np.intaa ).astype(np.floataa ) / 32_767 else: UpperCAmelCase = np.memmap(value["""path"""] ,dtype="""h""" ,mode="""r""" ).astype(np.floataa ) / 32_767 UpperCAmelCase = BytesIO(bytes() ) sf.write(A ,A ,value["""sampling_rate"""] ,format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( F'''An audio sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def _UpperCamelCase ( self ,A ,A = None ): if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Audio(decode=True) instead.""" ) UpperCAmelCase , UpperCAmelCase = (value["""path"""], BytesIO(value["""bytes"""] )) if value["""bytes"""] is not None else (value["""path"""], None) if path is None and file is None: raise ValueError(F'''An audio sample should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("""To support decoding audio files, please install 'librosa' and 'soundfile'.""" ) from err UpperCAmelCase = xsplitext(A )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( """Decoding 'opus' files requires system library 'libsndfile'>=1.0.31, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( """Decoding 'mp3' files requires system library 'libsndfile'>=1.1.0, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) if file is None: UpperCAmelCase = token_per_repo_id or {} UpperCAmelCase = path.split("""::""" )[-1] try: UpperCAmelCase = string_to_dict(A ,config.HUB_DATASETS_URL )["""repo_id"""] UpperCAmelCase = token_per_repo_id[repo_id] except (ValueError, KeyError): UpperCAmelCase = None with xopen(A ,"""rb""" ,use_auth_token=A ) as f: UpperCAmelCase , UpperCAmelCase = sf.read(A ) else: UpperCAmelCase , UpperCAmelCase = sf.read(A ) UpperCAmelCase = array.T if self.mono: UpperCAmelCase = librosa.to_mono(A ) if self.sampling_rate and self.sampling_rate != sampling_rate: UpperCAmelCase = librosa.resample(A ,orig_sr=A ,target_sr=self.sampling_rate ) UpperCAmelCase = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def _UpperCamelCase ( self ): from .features import Value if self.decode: raise ValueError("""Cannot flatten a decoded Audio feature.""" ) return { "bytes": Value("""binary""" ), "path": Value("""string""" ), } def _UpperCamelCase ( self ,A ): if pa.types.is_string(storage.type ): UpperCAmelCase = pa.array([None] * len(A ) ,type=pa.binary() ) UpperCAmelCase = pa.StructArray.from_arrays([bytes_array, storage] ,["""bytes""", """path"""] ,mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): UpperCAmelCase = pa.array([None] * len(A ) ,type=pa.string() ) UpperCAmelCase = pa.StructArray.from_arrays([storage, path_array] ,["""bytes""", """path"""] ,mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("""array""" ): UpperCAmelCase = pa.array([Audio().encode_example(A ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: UpperCAmelCase = storage.field("""bytes""" ) else: UpperCAmelCase = pa.array([None] * len(A ) ,type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: UpperCAmelCase = storage.field("""path""" ) else: UpperCAmelCase = pa.array([None] * len(A ) ,type=pa.string() ) UpperCAmelCase = pa.StructArray.from_arrays([bytes_array, path_array] ,["""bytes""", """path"""] ,mask=storage.is_null() ) return array_cast(A ,self.pa_type ) def _UpperCamelCase ( self ,A ): @no_op_if_value_is_null def path_to_bytes(A ): with xopen(A ,"""rb""" ) as f: UpperCAmelCase = f.read() return bytes_ UpperCAmelCase = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] ,type=pa.binary() ,) UpperCAmelCase = pa.array( [os.path.basename(A ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] ,type=pa.string() ,) UpperCAmelCase = pa.StructArray.from_arrays([bytes_array, path_array] ,["""bytes""", """path"""] ,mask=bytes_array.is_null() ) return array_cast(A ,self.pa_type )
74
"""simple docstring""" from __future__ import annotations def _a ( _snake_case ): """simple docstring""" return len(set(_snake_case ) ) == len(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" from __future__ import annotations def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" if (voltage, current, resistance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if resistance < 0: raise ValueError("""Resistance cannot be negative""" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" import math def _a ( _snake_case ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( _snake_case = 0.1 ): """simple docstring""" UpperCAmelCase = 3 UpperCAmelCase = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" if isinstance(_snake_case , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(_snake_case , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(_snake_case ): return [[videos]] raise ValueError(F'''Could not make batched video from {videos}''' ) class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''pixel_values'''] def __init__( self ,A = True ,A = None ,A = PILImageResampling.BILINEAR ,A = True ,A = None ,A = True ,A = 1 / 255 ,A = True ,A = True ,A = None ,A = None ,**A ,): super().__init__(**A ) UpperCAmelCase = size if size is not None else {"""shortest_edge""": 256} UpperCAmelCase = get_size_dict(A ,default_to_square=A ) UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} UpperCAmelCase = get_size_dict(A ,param_name="""crop_size""" ) UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = resample UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = offset UpperCAmelCase = do_normalize UpperCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def _UpperCamelCase ( self ,A ,A ,A = PILImageResampling.BILINEAR ,A = None ,**A ,): UpperCAmelCase = get_size_dict(A ,default_to_square=A ) if "shortest_edge" in size: UpperCAmelCase = get_resize_output_image_size(A ,size["""shortest_edge"""] ,default_to_square=A ) elif "height" in size and "width" in size: UpperCAmelCase = (size["""height"""], size["""width"""]) else: raise ValueError(F'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(A ,size=A ,resample=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A = None ,**A ,): UpperCAmelCase = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(F'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(A ,size=(size["""height"""], size["""width"""]) ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A = True ,A = None ,**A ,): UpperCAmelCase = image.astype(np.floataa ) if offset: UpperCAmelCase = image - (scale / 2) return rescale(A ,scale=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A ,A = None ,**A ,): return normalize(A ,mean=A ,std=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = ChannelDimension.FIRST ,): if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) if offset and not do_rescale: raise ValueError("""For offset, do_rescale must also be set to True.""" ) # All transformations expect numpy arrays. UpperCAmelCase = to_numpy_array(A ) if do_resize: UpperCAmelCase = self.resize(image=A ,size=A ,resample=A ) if do_center_crop: UpperCAmelCase = self.center_crop(A ,size=A ) if do_rescale: UpperCAmelCase = self.rescale(image=A ,scale=A ,offset=A ) if do_normalize: UpperCAmelCase = self.normalize(image=A ,mean=A ,std=A ) UpperCAmelCase = to_channel_dimension_format(A ,A ) return image def _UpperCamelCase ( self ,A ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = ChannelDimension.FIRST ,**A ,): UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = offset if offset is not None else self.offset UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(A ,default_to_square=A ) UpperCAmelCase = crop_size if crop_size is not None else self.crop_size UpperCAmelCase = get_size_dict(A ,param_name="""crop_size""" ) if not valid_images(A ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) UpperCAmelCase = make_batched(A ) UpperCAmelCase = [ [ self._preprocess_image( image=A ,do_resize=A ,size=A ,resample=A ,do_center_crop=A ,crop_size=A ,do_rescale=A ,rescale_factor=A ,offset=A ,do_normalize=A ,image_mean=A ,image_std=A ,data_format=A ,) for img in video ] for video in videos ] UpperCAmelCase = {"""pixel_values""": videos} return BatchFeature(data=A ,tensor_type=A )
74
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE = '''CLIPImageProcessor''' SCREAMING_SNAKE_CASE = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self ,A=None ,A=None ,**A ): UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,A ,) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A ,A ) def __call__( self ,A=None ,A=None ,A=None ,**A ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(A ,return_tensors=A ,**A ) if images is not None: UpperCAmelCase = self.image_processor(A ,return_tensors=A ,**A ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) ,tensor_type=A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.batch_decode(*A ,**A ) def _UpperCamelCase ( self ,*A ,**A ): return self.tokenizer.decode(*A ,**A ) @property def _UpperCamelCase ( self ): UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,A ,) return self.image_processor_class @property def _UpperCamelCase ( self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,A ,) return self.image_processor
74
1
"""simple docstring""" from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING _UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(snake_case ) class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,**A ): super().__init__(*A ,**A ) requires_backends(self ,"""vision""" ) self.check_model_type(A ) def __call__( self ,A ,**A ): return super().__call__(A ,**A ) def _UpperCamelCase ( self ,**A ): return {}, {}, {} def _UpperCamelCase ( self ,A ): UpperCAmelCase = load_image(A ) UpperCAmelCase = image.size UpperCAmelCase = self.image_processor(images=A ,return_tensors=self.framework ) return model_inputs def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.model(**A ) return model_outputs def _UpperCamelCase ( self ,A ): UpperCAmelCase = model_outputs.predicted_depth UpperCAmelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) ,size=self.image_size[::-1] ,mode="""bicubic""" ,align_corners=A ) UpperCAmelCase = prediction.squeeze().cpu().numpy() UpperCAmelCase = (output * 255 / np.max(A )).astype("""uint8""" ) UpperCAmelCase = Image.fromarray(A ) UpperCAmelCase = {} UpperCAmelCase = predicted_depth UpperCAmelCase = depth return output_dict
74
"""simple docstring""" from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase = """https://www.indeed.co.in/jobs?q=mobile+app+development&l=""" def _a ( _snake_case = "mumbai" ): """simple docstring""" UpperCAmelCase = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): UpperCAmelCase = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() UpperCAmelCase = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("""Bangalore"""), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
74
1
"""simple docstring""" from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable _UpperCamelCase = {"""configuration_dpt""": ["""DPT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DPTConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""DPTFeatureExtractor"""] _UpperCamelCase = ["""DPTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """DPT_PRETRAINED_MODEL_ARCHIVE_LIST""", """DPTForDepthEstimation""", """DPTForSemanticSegmentation""", """DPTModel""", """DPTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dpt import DPT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPTConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_dpt import DPTFeatureExtractor from .image_processing_dpt import DPTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dpt import ( DPT_PRETRAINED_MODEL_ARCHIVE_LIST, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel, DPTPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ): UpperCAmelCase = ["""a""", """b""", """c"""] # Defaults to last layer if both are None UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,A ,A ) self.assertEqual(A ,["""c"""] ) self.assertEqual(A ,[2] ) # Out indices set to match out features UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(["""a""", """c"""] ,A ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features set to match out indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[0, 2] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[0, 2] ) # Out features selected from negative indices UpperCAmelCase , UpperCAmelCase = get_aligned_output_features_output_indices(A ,[-3, -1] ,A ) self.assertEqual(A ,["""a""", """c"""] ) self.assertEqual(A ,[-3, -1] ) def _UpperCamelCase ( self ): # Stage names must be set with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,A ) # Out features must be a list with self.assertRaises(A ): verify_out_features_out_indices(("""a""", """b""") ,(0, 1) ,["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 1) ,["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(A ): verify_out_features_out_indices(A ,0 ,["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(A ): verify_out_features_out_indices(A ,(0, 1) ,["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0,) ,["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(A ): verify_out_features_out_indices(["""a""", """b"""] ,(0, 2) ,["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(A ): verify_out_features_out_indices(["""b""", """a"""] ,(0, 1) ,["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] ,(0, 1, -1) ,["""a""", """b""", """c""", """d"""] ) def _UpperCamelCase ( self ): UpperCAmelCase = BackboneMixin() UpperCAmelCase = ["""a""", """b""", """c"""] UpperCAmelCase = ["""a""", """c"""] UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[0, 2] ) # Check out features and indices are updated correctly UpperCAmelCase = ["""a""", """b"""] self.assertEqual(backbone.out_features ,["""a""", """b"""] ) self.assertEqual(backbone.out_indices ,[0, 1] ) UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features ,["""a""", """c"""] ) self.assertEqual(backbone.out_indices ,[-3, -1] )
74
1
"""simple docstring""" from collections.abc import Iterable from typing import Any class lowerCamelCase__ : def __init__( self ,A = None ): UpperCAmelCase = value UpperCAmelCase = None # Added in order to delete a node easier UpperCAmelCase = None UpperCAmelCase = None def __repr__( self ): from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F'''{self.value}''': (self.left, self.right)} ,indent=1 ) class lowerCamelCase__ : def __init__( self ,A = None ): UpperCAmelCase = root def __str__( self ): return str(self.root ) def _UpperCamelCase ( self ,A ,A ): if new_children is not None: # reset its kids UpperCAmelCase = node.parent if node.parent is not None: # reset its parent if self.is_right(A ): # If it is the right children UpperCAmelCase = new_children else: UpperCAmelCase = new_children else: UpperCAmelCase = new_children def _UpperCamelCase ( self ,A ): if node.parent and node.parent.right: return node == node.parent.right return False def _UpperCamelCase ( self ): return self.root is None def _UpperCamelCase ( self ,A ): UpperCAmelCase = Node(A ) # create a new Node if self.empty(): # if Tree is empty UpperCAmelCase = new_node # set its root else: # Tree is not empty UpperCAmelCase = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: UpperCAmelCase = new_node # We insert the new node in a leaf break else: UpperCAmelCase = parent_node.left else: if parent_node.right is None: UpperCAmelCase = new_node break else: UpperCAmelCase = parent_node.right UpperCAmelCase = parent_node def _UpperCamelCase ( self ,*A ): for value in values: self.__insert(A ) def _UpperCamelCase ( self ,A ): if self.empty(): raise IndexError("""Warning: Tree is empty! please use another.""" ) else: UpperCAmelCase = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: UpperCAmelCase = node.left if value < node.value else node.right return node def _UpperCamelCase ( self ,A = None ): if node is None: if self.root is None: return None UpperCAmelCase = self.root if not self.empty(): while node.right is not None: UpperCAmelCase = node.right return node def _UpperCamelCase ( self ,A = None ): if node is None: UpperCAmelCase = self.root if self.root is None: return None if not self.empty(): UpperCAmelCase = self.root while node.left is not None: UpperCAmelCase = node.left return node def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.search(A ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(A ,A ) elif node.left is None: # Has only right children self.__reassign_nodes(A ,node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(A ,node.left ) else: UpperCAmelCase = self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore UpperCAmelCase = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _UpperCamelCase ( self ,A ): if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _UpperCamelCase ( self ,A=None ): if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _UpperCamelCase ( self ,A ,A ): if node: self.inorder(A ,node.left ) arr.append(node.value ) self.inorder(A ,node.right ) def _UpperCamelCase ( self ,A ,A ): UpperCAmelCase = [] self.inorder(A ,A ) # append all values to list using inorder traversal return arr[k - 1] def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = [] if curr_node is not None: UpperCAmelCase = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def _a ( ): """simple docstring""" UpperCAmelCase = (8, 3, 6, 1, 10, 14, 13, 4, 7) UpperCAmelCase = BinarySearchTree() for i in testlist: t.insert(_snake_case ) # Prints all the elements of the list in order traversal print(_snake_case ) if t.search(6 ) is not None: print("""The value 6 exists""" ) else: print("""The value 6 doesn't exist""" ) if t.search(-1 ) is not None: print("""The value -1 exists""" ) else: print("""The value -1 doesn't exist""" ) if not t.empty(): print("""Max Value: """ , t.get_max().value ) # type: ignore print("""Min Value: """ , t.get_min().value ) # type: ignore for i in testlist: t.remove(_snake_case ) print(_snake_case ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
74
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCamelCase__ : def __init__( self ,A = 6 ): UpperCAmelCase = None UpperCAmelCase = None self.create_linked_list(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = current_node UpperCAmelCase = current_node for _ in range(1 ,A ): UpperCAmelCase = Node() UpperCAmelCase = current_node UpperCAmelCase = previous_node UpperCAmelCase = current_node UpperCAmelCase = self.front UpperCAmelCase = previous_node def _UpperCamelCase ( self ): return ( self.front == self.rear and self.front is not None and self.front.data is None ) def _UpperCamelCase ( self ): self.check_can_perform_operation() return self.front.data if self.front else None def _UpperCamelCase ( self ,A ): if self.rear is None: return self.check_is_full() if not self.is_empty(): UpperCAmelCase = self.rear.next if self.rear: UpperCAmelCase = data def _UpperCamelCase ( self ): self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: UpperCAmelCase = self.front.data UpperCAmelCase = None return data UpperCAmelCase = self.front UpperCAmelCase = old_front.next UpperCAmelCase = old_front.data UpperCAmelCase = None return data def _UpperCamelCase ( self ): if self.is_empty(): raise Exception("""Empty Queue""" ) def _UpperCamelCase ( self ): if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""" ) class lowerCamelCase__ : def __init__( self ): UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = len(_snake_case ) UpperCAmelCase = len(matrix[0] ) UpperCAmelCase = min(_snake_case , _snake_case ) for row in range(_snake_case ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , _snake_case ): UpperCAmelCase = matrix[col][row] / matrix[row][row] for i in range(_snake_case , _snake_case ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows UpperCAmelCase = True for i in range(row + 1 , _snake_case ): if matrix[i][row] != 0: UpperCAmelCase , UpperCAmelCase = matrix[i], matrix[row] UpperCAmelCase = False break if reduce: rank -= 1 for i in range(_snake_case ): UpperCAmelCase = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger("""transformers.models.speecht5""") def _a ( _snake_case , _snake_case , _snake_case ): """simple docstring""" hf_model.apply_weight_norm() UpperCAmelCase = checkpoint["""input_conv.weight_g"""] UpperCAmelCase = checkpoint["""input_conv.weight_v"""] UpperCAmelCase = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case , _snake_case=None , _snake_case=None , ): """simple docstring""" if config_path is not None: UpperCAmelCase = SpeechTaHifiGanConfig.from_pretrained(_snake_case ) else: UpperCAmelCase = SpeechTaHifiGanConfig() UpperCAmelCase = SpeechTaHifiGan(_snake_case ) UpperCAmelCase = torch.load(_snake_case ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _snake_case , _snake_case ) UpperCAmelCase = np.load(_snake_case ) UpperCAmelCase = stats[0].reshape(-1 ) UpperCAmelCase = stats[1].reshape(-1 ) UpperCAmelCase = torch.from_numpy(_snake_case ).float() UpperCAmelCase = torch.from_numpy(_snake_case ).float() model.save_pretrained(_snake_case ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_snake_case ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCamelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
74
1
"""simple docstring""" import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput _UpperCamelCase = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=None ,A=None ,A=None ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = eval_examples UpperCAmelCase = post_process_function UpperCAmelCase = quant_trainer_args UpperCAmelCase = 128 # default number of calibration samples def _UpperCamelCase ( self ,A=None ): if calib_dataset is None and self.calib_dataset is None: raise ValueError("""Trainer: calibration requires an calib_dataset.""" ) UpperCAmelCase = calib_dataset if calib_dataset is not None else self.calib_dataset UpperCAmelCase = self._remove_unused_columns(A ,description="""Calibration""" ) return DataLoader( A ,batch_size=self.args.eval_batch_size ,collate_fn=self.data_collator ,drop_last=self.args.dataloader_drop_last ,num_workers=self.args.dataloader_num_workers ,pin_memory=self.args.dataloader_pin_memory ,shuffle=A ,) def _UpperCamelCase ( self ,A=None ): UpperCAmelCase = self.train_dataset if calib_dataset is None else calib_dataset UpperCAmelCase = self.get_calib_dataloader(A ) UpperCAmelCase = self.model quant_trainer.configure_model(A ,self.quant_trainer_args ,calib=A ) model.eval() quant_trainer.enable_calibration(A ) logger.info("""***** Running calibration *****""" ) logger.info(F''' Num examples = {self.calib_num}''' ) logger.info(F''' Batch size = {calib_dataloader.batch_size}''' ) for step, inputs in enumerate(A ): # Prediction step UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.prediction_step(A ,A ,prediction_loss_only=A ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(A ,self.quant_trainer_args ) UpperCAmelCase = model def _UpperCamelCase ( self ,A=None ,A=None ,A=None ,A = "eval" ): UpperCAmelCase = self.eval_dataset if eval_dataset is None else eval_dataset UpperCAmelCase = self.get_eval_dataloader(A ) UpperCAmelCase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase = self.compute_metrics UpperCAmelCase = None UpperCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: UpperCAmelCase = eval_loop( A ,description="""Evaluation""" ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A ,) finally: UpperCAmelCase = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: UpperCAmelCase = self.post_process_function(A ,A ,output.predictions ) UpperCAmelCase = self.compute_metrics(A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): UpperCAmelCase = metrics.pop(A ) self.log(A ) else: UpperCAmelCase = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) UpperCAmelCase = self.callback_handler.on_evaluate(self.args ,self.state ,self.control ,A ) return metrics def _UpperCamelCase ( self ,A ,A ,A=None ,A = "test" ): UpperCAmelCase = self.get_test_dataloader(A ) # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase = self.compute_metrics UpperCAmelCase = None UpperCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: UpperCAmelCase = eval_loop( A ,description="""Prediction""" ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A ,) finally: UpperCAmelCase = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output UpperCAmelCase = self.post_process_function(A ,A ,output.predictions ,"""predict""" ) UpperCAmelCase = self.compute_metrics(A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): UpperCAmelCase = metrics.pop(A ) return PredictionOutput(predictions=predictions.predictions ,label_ids=predictions.label_ids ,metrics=A ) def _UpperCamelCase ( self ,A="./" ): UpperCAmelCase = self.eval_dataset UpperCAmelCase = self.get_eval_dataloader(A ) UpperCAmelCase = next(iter(A ) ) # saving device - to make it consistent UpperCAmelCase = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) # convert to tuple UpperCAmelCase = tuple(v.to(A ) for k, v in batch.items() ) logger.info("""Converting model to be onnx compatible""" ) from pytorch_quantization.nn import TensorQuantizer UpperCAmelCase = True UpperCAmelCase = self.model.to(A ) model.eval() model.float() UpperCAmelCase = model.module if hasattr(A ,"""module""" ) else model quant_trainer.configure_model(A ,self.quant_trainer_args ) UpperCAmelCase = os.path.join(A ,"""model.onnx""" ) logger.info(F'''exporting model to {output_model_file}''' ) UpperCAmelCase = {0: """batch_size""", 1: """seq_len"""} torch.onnx.export( A ,A ,A ,export_params=A ,opset_version=13 ,do_constant_folding=A ,input_names=["""input_ids""", """attention_mask""", """token_type_ids"""] ,output_names=["""output_start_logits""", """output_end_logits"""] ,dynamic_axes={ """input_ids""": axes, """attention_mask""": axes, """token_type_ids""": axes, """output_start_logits""": axes, """output_end_logits""": axes, } ,verbose=A ,) logger.info("""onnx export finished""" )
74
"""simple docstring""" # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _UpperCamelCase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _UpperCamelCase = concatenate_datasets _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadManager _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadConfig _UpperCamelCase = DownloadMode _UpperCamelCase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
74
1
"""simple docstring""" def _a ( _snake_case , _snake_case ): """simple docstring""" UpperCAmelCase = len(_snake_case ) + 1 UpperCAmelCase = len(_snake_case ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. UpperCAmelCase = [[0 for i in range(_snake_case )] for j in range(_snake_case )] # since string of zero length match pattern of zero length UpperCAmelCase = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , _snake_case ): UpperCAmelCase = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , _snake_case ): UpperCAmelCase = dp[0][j - 2] if pattern[j - 1] == """*""" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , _snake_case ): for j in range(1 , _snake_case ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": UpperCAmelCase = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: UpperCAmelCase = 1 elif pattern[j - 2] in (input_string[i - 1], "."): UpperCAmelCase = dp[i - 1][j] else: UpperCAmelCase = 0 else: UpperCAmelCase = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") _UpperCamelCase = """aab""" _UpperCamelCase = """c*a*b""" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F"""{input_string} matches the given pattern {pattern}""") else: print(F"""{input_string} does not match with the given pattern {pattern}""")
74
"""simple docstring""" def _a ( _snake_case ): """simple docstring""" if not isinstance(_snake_case , _snake_case ): raise ValueError("""Input must be an integer""" ) if input_num <= 0: raise ValueError("""Input must be positive""" ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
74
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCamelCase = logging.get_logger(__name__) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): UpperCAmelCase = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): UpperCAmelCase = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] UpperCAmelCase = key.replace(F'''patch_embed{idx}''' , F'''patch_embeddings.{int(_snake_case )-1}''' ) if "norm" in key: UpperCAmelCase = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] UpperCAmelCase = key.replace(F'''layer_norm{idx}''' , F'''layer_norm.{int(_snake_case )-1}''' ) if "layer_norm1" in key: UpperCAmelCase = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: UpperCAmelCase = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase = key[key.find("""block""" ) + len("""block""" )] UpperCAmelCase = key.replace(F'''block{idx}''' , F'''block.{int(_snake_case )-1}''' ) if "attn.q" in key: UpperCAmelCase = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: UpperCAmelCase = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: UpperCAmelCase = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: UpperCAmelCase = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: UpperCAmelCase = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: UpperCAmelCase = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: UpperCAmelCase = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) UpperCAmelCase = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase = key[key.find("""linear_c""" ) + len("""linear_c""" )] UpperCAmelCase = key.replace(F'''linear_c{idx}''' , F'''linear_c.{int(_snake_case )-1}''' ) if "bot_conv" in key: UpperCAmelCase = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: UpperCAmelCase = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: UpperCAmelCase = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: UpperCAmelCase = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: UpperCAmelCase = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: UpperCAmelCase = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: UpperCAmelCase = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): UpperCAmelCase = key.replace("""module.last_layer_depth""" , """head.head""" ) UpperCAmelCase = value return new_state_dict def _a ( _snake_case , _snake_case ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.weight''' ) UpperCAmelCase = state_dict.pop(F'''glpn.encoder.block.{i}.{j}.attention.self.kv.bias''' ) # next, add keys and values (in that order) to the state dict UpperCAmelCase = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase = kv_bias[config.hidden_sizes[i] :] def _a ( ): """simple docstring""" UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image @torch.no_grad() def _a ( _snake_case , _snake_case , _snake_case=False , _snake_case=None ): """simple docstring""" UpperCAmelCase = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) UpperCAmelCase = GLPNImageProcessor() # prepare image UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=_snake_case , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict UpperCAmelCase = torch.load(_snake_case , map_location=torch.device("""cpu""" ) ) # rename keys UpperCAmelCase = rename_keys(_snake_case ) # key and value matrices need special treatment read_in_k_v(_snake_case , _snake_case ) # create HuggingFace model and load state dict UpperCAmelCase = GLPNForDepthEstimation(_snake_case ) model.load_state_dict(_snake_case ) model.eval() # forward pass UpperCAmelCase = model(_snake_case ) UpperCAmelCase = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: UpperCAmelCase = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: UpperCAmelCase = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(F'''Unknown model name: {model_name}''' ) UpperCAmelCase = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _snake_case , atol=1E-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=_snake_case , ) image_processor.push_to_hub( repo_path_or_name=Path(_snake_case , _snake_case ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=_snake_case , ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) parser.add_argument( """--model_name""", default="""glpn-kitti""", type=str, help="""Name of the model in case you're pushing to the hub.""", ) _UpperCamelCase = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
74
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCamelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") _UpperCamelCase = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) _UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default='''cifar10''' , metadata={'''help''': '''Name of a dataset from the datasets package'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The column name of the images in the files. If not set, will try to use \'image\' or \'img\'.'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the training data.'''} ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''A folder containing the validation data.'''} ) SCREAMING_SNAKE_CASE = field( default=0.15 , metadata={'''help''': '''Percent to split off of train for validation.'''} ) SCREAMING_SNAKE_CASE = field(default=32 , metadata={'''help''': '''The size of the square patches to use for masking.'''} ) SCREAMING_SNAKE_CASE = field( default=0.6 , metadata={'''help''': '''Percentage of patches to mask.'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } , ) def _UpperCamelCase ( self ): UpperCAmelCase = {} if self.train_dir is not None: UpperCAmelCase = self.train_dir if self.validation_dir is not None: UpperCAmelCase = self.validation_dir UpperCAmelCase = data_files if data_files else None @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a ''' '''checkpoint identifier on the hub. ''' '''Don\'t set if you want to train a model from scratch.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''If training from scratch, pass a model type from the list: ''' + ''', '''.join(snake_case )} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store (cache) the pretrained models/datasets downloaded from the hub'''} , ) SCREAMING_SNAKE_CASE = field( default='''main''' , metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} , ) SCREAMING_SNAKE_CASE = field(default=snake_case , metadata={'''help''': '''Name or path of preprocessor config.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each image. If not specified, will use `image_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={ '''help''': ( '''The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.''' ) } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Stride to use for the encoder.'''} , ) class lowerCamelCase__ : def __init__( self ,A=192 ,A=32 ,A=4 ,A=0.6 ): UpperCAmelCase = input_size UpperCAmelCase = mask_patch_size UpperCAmelCase = model_patch_size UpperCAmelCase = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError("""Input size must be divisible by mask patch size""" ) if self.mask_patch_size % self.model_patch_size != 0: raise ValueError("""Mask patch size must be divisible by model patch size""" ) UpperCAmelCase = self.input_size // self.mask_patch_size UpperCAmelCase = self.mask_patch_size // self.model_patch_size UpperCAmelCase = self.rand_size**2 UpperCAmelCase = int(np.ceil(self.token_count * self.mask_ratio ) ) def __call__( self ): UpperCAmelCase = np.random.permutation(self.token_count )[: self.mask_count] UpperCAmelCase = np.zeros(self.token_count ,dtype=A ) UpperCAmelCase = 1 UpperCAmelCase = mask.reshape((self.rand_size, self.rand_size) ) UpperCAmelCase = mask.repeat(self.scale ,axis=0 ).repeat(self.scale ,axis=1 ) return torch.tensor(mask.flatten() ) def _a ( _snake_case ): """simple docstring""" UpperCAmelCase = torch.stack([example["""pixel_values"""] for example in examples] ) UpperCAmelCase = torch.stack([example["""mask"""] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mim""" , _snake_case , _snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_snake_case ) transformers.utils.logging.set_verbosity(_snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. UpperCAmelCase = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _snake_case ) and data_args.train_val_split > 0.0: UpperCAmelCase = ds["""train"""].train_test_split(data_args.train_val_split ) UpperCAmelCase = split["""train"""] UpperCAmelCase = split["""test"""] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCAmelCase = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.config_name_or_path , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoConfig.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = CONFIG_MAPPING[model_args.model_type]() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_snake_case , """decoder_type""" ): UpperCAmelCase = """simmim""" # adapt config UpperCAmelCase = model_args.image_size if model_args.image_size is not None else config.image_size UpperCAmelCase = model_args.patch_size if model_args.patch_size is not None else config.patch_size UpperCAmelCase = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { """image_size""": model_args.image_size, """patch_size""": model_args.patch_size, """encoder_stride""": model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_snake_case ) elif model_args.model_name_or_path: UpperCAmelCase = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_snake_case ) else: UpperCAmelCase = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } UpperCAmelCase = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: UpperCAmelCase = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) UpperCAmelCase = AutoModelForMaskedImageModeling.from_config(_snake_case ) if training_args.do_train: UpperCAmelCase = ds["""train"""].column_names else: UpperCAmelCase = ds["""validation"""].column_names if data_args.image_column_name is not None: UpperCAmelCase = data_args.image_column_name elif "image" in column_names: UpperCAmelCase = """image""" elif "img" in column_names: UpperCAmelCase = """img""" else: UpperCAmelCase = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py UpperCAmelCase = Compose( [ Lambda(lambda _snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator UpperCAmelCase = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_snake_case ): UpperCAmelCase = [transforms(_snake_case ) for image in examples[image_column_name]] UpperCAmelCase = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: UpperCAmelCase = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: UpperCAmelCase = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_snake_case ) # Initialize our trainer UpperCAmelCase = Trainer( model=_snake_case , args=_snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: UpperCAmelCase = last_checkpoint UpperCAmelCase = trainer.train(resume_from_checkpoint=_snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: UpperCAmelCase = trainer.evaluate() trainer.log_metrics("""eval""" , _snake_case ) trainer.save_metrics("""eval""" , _snake_case ) # Write model card and (optionally) push to hub UpperCAmelCase = { """finetuned_from""": model_args.model_name_or_path, """tasks""": """masked-image-modeling""", """dataset""": data_args.dataset_name, """tags""": ["""masked-image-modeling"""], } if training_args.push_to_hub: trainer.push_to_hub(**_snake_case ) else: trainer.create_model_card(**_snake_case ) if __name__ == "__main__": main()
74
1
"""simple docstring""" import string import numpy def _a ( _snake_case , _snake_case ): """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , _snake_case ) class lowerCamelCase__ : SCREAMING_SNAKE_CASE = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) SCREAMING_SNAKE_CASE = numpy.vectorize(lambda snake_case : x % 36 ) SCREAMING_SNAKE_CASE = numpy.vectorize(snake_case ) def __init__( self ,A ): UpperCAmelCase = self.modulus(A ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key UpperCAmelCase = encrypt_key.shape[0] def _UpperCamelCase ( self ,A ): return self.key_string.index(A ) def _UpperCamelCase ( self ,A ): return self.key_string[round(A )] def _UpperCamelCase ( self ): UpperCAmelCase = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: UpperCAmelCase = det % len(self.key_string ) UpperCAmelCase = len(self.key_string ) if greatest_common_divisor(A ,len(self.key_string ) ) != 1: UpperCAmelCase = ( F'''determinant modular {req_l} of encryption key({det}) ''' F'''is not co prime w.r.t {req_l}.\nTry another key.''' ) raise ValueError(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = [char for char in text.upper() if char in self.key_string] UpperCAmelCase = chars[-1] while len(A ) % self.break_key != 0: chars.append(A ) return "".join(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.process_text(text.upper() ) UpperCAmelCase = """""" for i in range(0 ,len(A ) - self.break_key + 1 ,self.break_key ): UpperCAmelCase = text[i : i + self.break_key] UpperCAmelCase = [self.replace_letters(A ) for char in batch] UpperCAmelCase = numpy.array([vec] ).T UpperCAmelCase = self.modulus(self.encrypt_key.dot(A ) ).T.tolist()[ 0 ] UpperCAmelCase = """""".join( self.replace_digits(A ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def _UpperCamelCase ( self ): UpperCAmelCase = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: UpperCAmelCase = det % len(self.key_string ) UpperCAmelCase = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: UpperCAmelCase = i break UpperCAmelCase = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(A ) ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.make_decrypt_key() UpperCAmelCase = self.process_text(text.upper() ) UpperCAmelCase = """""" for i in range(0 ,len(A ) - self.break_key + 1 ,self.break_key ): UpperCAmelCase = text[i : i + self.break_key] UpperCAmelCase = [self.replace_letters(A ) for char in batch] UpperCAmelCase = numpy.array([vec] ).T UpperCAmelCase = self.modulus(decrypt_key.dot(A ) ).T.tolist()[0] UpperCAmelCase = """""".join( self.replace_digits(A ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def _a ( ): """simple docstring""" UpperCAmelCase = int(input("""Enter the order of the encryption key: """ ) ) UpperCAmelCase = [] print("""Enter each row of the encryption key with space separated integers""" ) for _ in range(_snake_case ): UpperCAmelCase = [int(_snake_case ) for x in input().split()] hill_matrix.append(_snake_case ) UpperCAmelCase = HillCipher(numpy.array(_snake_case ) ) print("""Would you like to encrypt or decrypt some text? (1 or 2)""" ) UpperCAmelCase = input("""\n1. Encrypt\n2. Decrypt\n""" ) if option == "1": UpperCAmelCase = input("""What text would you like to encrypt?: """ ) print("""Your encrypted text is:""" ) print(hc.encrypt(_snake_case ) ) elif option == "2": UpperCAmelCase = input("""What text would you like to decrypt?: """ ) print("""Your decrypted text is:""" ) print(hc.decrypt(_snake_case ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
74
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("""1.6"""): _UpperCamelCase = True from torch.cuda.amp import autocast _UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Whether to log verbose messages or not.'''} , ) SCREAMING_SNAKE_CASE = field( default=2.0 , metadata={'''help''': '''Maximum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.5 , metadata={'''help''': '''Minimum temperature for gumbel softmax.'''} ) SCREAMING_SNAKE_CASE = field( default=0.99_99_95 , metadata={'''help''': '''Decay of gumbel temperature during training.'''} ) def _a ( _snake_case , _snake_case ): """simple docstring""" logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase = logging.WARNING if model_args.verbose_logging: UpperCAmelCase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase = logging.INFO logger.setLevel(_snake_case ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) SCREAMING_SNAKE_CASE = field( default='''train''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) SCREAMING_SNAKE_CASE = field( default='''validation''' , metadata={ '''help''': ( '''The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'''' ) } , ) SCREAMING_SNAKE_CASE = field( default='''file''' , metadata={'''help''': '''Column in the dataset that contains speech file path. Defaults to \'file\''''} , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) SCREAMING_SNAKE_CASE = field( default=1 , metadata={ '''help''': '''The percentage of the train set used as validation set in case there\'s no validation split''' } , ) SCREAMING_SNAKE_CASE = field( default=snake_case , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) SCREAMING_SNAKE_CASE = field( default=20.0 , metadata={'''help''': '''Filter audio files that are longer than `max_duration_in_seconds` seconds'''} ) @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = "longest" SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None def __call__( self ,A ): # reformat list to dict and set to pytorch format UpperCAmelCase = self.feature_extractor.pad( A ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="""pt""" ,) UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""input_values"""].shape[-1] ) UpperCAmelCase = batch["""input_values"""].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase = self.model._get_feat_extract_output_lengths(batch["""attention_mask"""].sum(-1 ) ).to( torch.long ) UpperCAmelCase = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["""input_values"""].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase = 1 UpperCAmelCase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=A ,min_masks=2 ,) return batch class lowerCamelCase__ ( snake_case ): def __init__( self ,*A ,A=1 ,A=0 ,A=1.0 ,**A ): super().__init__(*A ,**A ) UpperCAmelCase = 0 UpperCAmelCase = max_gumbel_temp UpperCAmelCase = min_gumbel_temp UpperCAmelCase = gumbel_temp_decay def _UpperCamelCase ( self ,A ,A ): model.train() UpperCAmelCase = self._prepare_inputs(A ) if self.use_amp: with autocast(): UpperCAmelCase = self.compute_loss(A ,A ) else: UpperCAmelCase = self.compute_loss(A ,A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase = loss.sum() / (inputs["""mask_time_indices"""]).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A ).backward() elif self.use_apex: with amp.scale_loss(A ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def _a ( ): """simple docstring""" UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = parser.parse_args_into_dataclasses() configure_logger(_snake_case , _snake_case ) # Downloading and loading a dataset from the hub. UpperCAmelCase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase = DatasetDict() UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="""validation""" , cache_dir=model_args.cache_dir , ) UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=_snake_case ) def prepare_dataset(_snake_case ): # check that all files have the correct sampling rate UpperCAmelCase , UpperCAmelCase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase = datasets.map( _snake_case , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["""train"""].column_names ) # filter audio files that are too long UpperCAmelCase = vectorized_datasets.filter( lambda _snake_case : len(data["""speech"""] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(_snake_case ): return feature_extractor(batch["""speech"""] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase = vectorized_datasets.map( _snake_case , batched=_snake_case , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["""train"""].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( """PreTraining is only supported for ``config.do_stable_layer_norm=True`` and""" """ ``config.feat_extract_norm='layer'""" ) UpperCAmelCase = WavaVecaForPreTraining(_snake_case ) UpperCAmelCase = DataCollatorForWavaVecaPretraining(model=_snake_case , feature_extractor=_snake_case ) UpperCAmelCase = WavaVecaPreTrainer( model=_snake_case , data_collator=_snake_case , args=_snake_case , train_dataset=vectorized_datasets["""train"""] , eval_dataset=vectorized_datasets["""validation"""] , tokenizer=_snake_case , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
74
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
74
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCamelCase__ : def __init__( self ,A ,): UpperCAmelCase = parent UpperCAmelCase = 13 UpperCAmelCase = 7 UpperCAmelCase = 30 UpperCAmelCase = self.seq_length + self.mem_len UpperCAmelCase = 15 UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = 99 UpperCAmelCase = [10, 50, 80] UpperCAmelCase = 32 UpperCAmelCase = 32 UpperCAmelCase = 4 UpperCAmelCase = 8 UpperCAmelCase = 128 UpperCAmelCase = 2 UpperCAmelCase = 2 UpperCAmelCase = None UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 3 UpperCAmelCase = self.vocab_size - 1 UpperCAmelCase = 0.01 def _UpperCamelCase ( self ): UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCamelCase ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLLMHeadModel(A ) UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() UpperCAmelCase , UpperCAmelCase = model([input_ids_a, mems_a] ).to_tuple() UpperCAmelCase = {"""input_ids""": input_ids_a, """mems""": mems_a, """labels""": lm_labels} UpperCAmelCase , UpperCAmelCase = model(A ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def _UpperCamelCase ( self ,A ,A ,A ,A ): UpperCAmelCase = TFTransfoXLForSequenceClassification(A ) UpperCAmelCase = model(A ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ): UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids_a} return config, inputs_dict @require_tf class lowerCamelCase__ ( snake_case , snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE = () if is_tf_available() else () SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _UpperCamelCase ( self ,A ,A ,A ,A ,A ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLModelTester(self ) UpperCAmelCase = ConfigTester(self ,config_class=A ,d_embed=37 ) def _UpperCamelCase ( self ): self.config_tester.run_common_tests() def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A ) def _UpperCamelCase ( self ): self.model_tester.set_seed() UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A ) def _UpperCamelCase ( self ): UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A ) def _UpperCamelCase ( self ): UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: UpperCAmelCase = model_class(A ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: UpperCAmelCase = model.get_output_embeddings() assert isinstance(A ,tf.keras.layers.Layer ) UpperCAmelCase = model.get_bias() assert name is None else: UpperCAmelCase = model.get_output_embeddings() assert x is None UpperCAmelCase = model.get_bias() assert name is None def _UpperCamelCase ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def _UpperCamelCase ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFTransfoXLModel.from_pretrained(A ) self.assertIsNotNone(A ) @unittest.skip(reason="""This model doesn't play well with fit() due to not returning a single loss.""" ) def _UpperCamelCase ( self ): pass @require_tf class lowerCamelCase__ ( unittest.TestCase ): @unittest.skip("""Skip test until #12651 is resolved.""" ) @slow def _UpperCamelCase ( self ): UpperCAmelCase = TFTransfoXLLMHeadModel.from_pretrained("""transfo-xl-wt103""" ) # fmt: off UpperCAmelCase = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off UpperCAmelCase = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> UpperCAmelCase = model.generate(A ,max_length=200 ,do_sample=A ) self.assertListEqual(output_ids[0].numpy().tolist() ,A )
74
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _UpperCamelCase = logging.get_logger(__name__) class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = ['''pixel_values'''] def __init__( self ,A = True ,A = None ,A = None ,A = PILImageResampling.BILINEAR ,A = True ,A = 1 / 255 ,A = True ,A = None ,A = None ,**A ,): super().__init__(**A ) UpperCAmelCase = size if size is not None else {"""shortest_edge""": 384} UpperCAmelCase = get_size_dict(A ,default_to_square=A ) UpperCAmelCase = do_resize UpperCAmelCase = size # Default value set here for backwards compatibility where the value in config is None UpperCAmelCase = crop_pct if crop_pct is not None else 224 / 256 UpperCAmelCase = resample UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def _UpperCamelCase ( self ,A ,A ,A ,A = PILImageResampling.BICUBIC ,A = None ,**A ,): UpperCAmelCase = get_size_dict(A ,default_to_square=A ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) UpperCAmelCase = size["""shortest_edge"""] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct UpperCAmelCase = int(shortest_edge / crop_pct ) UpperCAmelCase = get_resize_output_image_size(A ,size=A ,default_to_square=A ) UpperCAmelCase = resize(image=A ,size=A ,resample=A ,data_format=A ,**A ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=A ,size=(shortest_edge, shortest_edge) ,data_format=A ,**A ) else: # warping (no cropping) when evaluated at 384 or larger return resize( A ,size=(shortest_edge, shortest_edge) ,resample=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A = None ,**A ,): return rescale(A ,scale=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A ,A ,A = None ,**A ,): return normalize(A ,mean=A ,std=A ,data_format=A ,**A ) def _UpperCamelCase ( self ,A ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = None ,A = ChannelDimension.FIRST ,**A ,): UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = crop_pct if crop_pct is not None else self.crop_pct UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(A ,default_to_square=A ) UpperCAmelCase = make_list_of_images(A ) if not valid_images(A ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase = [to_numpy_array(A ) for image in images] if do_resize: UpperCAmelCase = [self.resize(image=A ,size=A ,crop_pct=A ,resample=A ) for image in images] if do_rescale: UpperCAmelCase = [self.rescale(image=A ,scale=A ) for image in images] if do_normalize: UpperCAmelCase = [self.normalize(image=A ,mean=A ,std=A ) for image in images] UpperCAmelCase = [to_channel_dimension_format(A ,A ) for image in images] UpperCAmelCase = {"""pixel_values""": images} return BatchFeature(data=A ,tensor_type=A )
74
"""simple docstring""" from math import sqrt def _a ( _snake_case = 100_0000 ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_snake_case , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
74
1