code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCamelCase = '''src/diffusers''' UpperCamelCase = '''.''' # This is to make sure the diffusers module imported is the one in the repo. UpperCamelCase = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) UpperCamelCase = spec.loader.load_module() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> int: """simple docstring""" return line.startswith(snake_case__ ) or len(snake_case__ ) <= 1 or re.search(r"""^\s*\)(\s*->.*:|:)\s*$""" ,snake_case__ ) is not None def __lowerCamelCase ( snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = object_name.split(""".""" ) _SCREAMING_SNAKE_CASE = 0 # First let's find the module where our object lives. _SCREAMING_SNAKE_CASE = parts[i] while i < len(snake_case__ ) and not os.path.isfile(os.path.join(snake_case__ ,F'{module}.py' ) ): i += 1 if i < len(snake_case__ ): _SCREAMING_SNAKE_CASE = os.path.join(snake_case__ ,parts[i] ) if i >= len(snake_case__ ): raise ValueError(F'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(snake_case__ ,F'{module}.py' ) ,"""r""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: _SCREAMING_SNAKE_CASE = f.readlines() # Now let's find the class / func in the code! _SCREAMING_SNAKE_CASE = """""" _SCREAMING_SNAKE_CASE = 0 for name in parts[i + 1 :]: while ( line_index < len(snake_case__ ) and re.search(rF'^{indent}(class|def)\s+{name}(\(|\:)' ,lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(snake_case__ ): raise ValueError(F' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). _SCREAMING_SNAKE_CASE = line_index while line_index < len(snake_case__ ) and _should_continue(lines[line_index] ,snake_case__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 _SCREAMING_SNAKE_CASE = lines[start_index:line_index] return "".join(snake_case__ ) UpperCamelCase = re.compile(R'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') UpperCamelCase = re.compile(R'''^\s*(\S+)->(\S+)(\s+.*|$)''') UpperCamelCase = re.compile(R'''<FILL\s+[^>]*>''') def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = code.split("""\n""" ) _SCREAMING_SNAKE_CASE = 0 while idx < len(snake_case__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(snake_case__ ): return re.search(r"""^(\s*)\S""" ,lines[idx] ).groups()[0] return "" def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = len(get_indent(snake_case__ ) ) > 0 if has_indent: _SCREAMING_SNAKE_CASE = F'class Bla:\n{code}' _SCREAMING_SNAKE_CASE = black.Mode(target_versions={black.TargetVersion.PYaa} ,line_length=1_19 ,preview=snake_case__ ) _SCREAMING_SNAKE_CASE = black.format_str(snake_case__ ,mode=snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = style_docstrings_in_code(snake_case__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def __lowerCamelCase ( snake_case__ ,snake_case__=False ) -> List[str]: """simple docstring""" with open(snake_case__ ,"""r""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: _SCREAMING_SNAKE_CASE = f.readlines() _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(snake_case__ ): _SCREAMING_SNAKE_CASE = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = search.groups() _SCREAMING_SNAKE_CASE = find_code_in_diffusers(snake_case__ ) _SCREAMING_SNAKE_CASE = get_indent(snake_case__ ) _SCREAMING_SNAKE_CASE = line_index + 1 if indent == theoretical_indent else line_index + 2 _SCREAMING_SNAKE_CASE = theoretical_indent _SCREAMING_SNAKE_CASE = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. _SCREAMING_SNAKE_CASE = True while line_index < len(snake_case__ ) and should_continue: line_index += 1 if line_index >= len(snake_case__ ): break _SCREAMING_SNAKE_CASE = lines[line_index] _SCREAMING_SNAKE_CASE = _should_continue(snake_case__ ,snake_case__ ) and re.search(F'^{indent}# End copy' ,snake_case__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 _SCREAMING_SNAKE_CASE = lines[start_index:line_index] _SCREAMING_SNAKE_CASE = """""".join(snake_case__ ) # Remove any nested `Copied from` comments to avoid circular copies _SCREAMING_SNAKE_CASE = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(snake_case__ ) is None] _SCREAMING_SNAKE_CASE = """\n""".join(snake_case__ ) # Before comparing, use the `replace_pattern` on the original code. if len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = replace_pattern.replace("""with""" ,"""""" ).split(""",""" ) _SCREAMING_SNAKE_CASE = [_re_replace_pattern.search(snake_case__ ) for p in patterns] for pattern in patterns: if pattern is None: continue _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = pattern.groups() _SCREAMING_SNAKE_CASE = re.sub(snake_case__ ,snake_case__ ,snake_case__ ) if option.strip() == "all-casing": _SCREAMING_SNAKE_CASE = re.sub(obja.lower() ,obja.lower() ,snake_case__ ) _SCREAMING_SNAKE_CASE = re.sub(obja.upper() ,obja.upper() ,snake_case__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line _SCREAMING_SNAKE_CASE = blackify(lines[start_index - 1] + theoretical_code ) _SCREAMING_SNAKE_CASE = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: _SCREAMING_SNAKE_CASE = lines[:start_index] + [theoretical_code] + lines[line_index:] _SCREAMING_SNAKE_CASE = start_index + 1 if overwrite and len(snake_case__ ) > 0: # Warn the user a file has been modified. print(F'Detected changes, rewriting {filename}.' ) with open(snake_case__ ,"""w""" ,encoding="""utf-8""" ,newline="""\n""" ) as f: f.writelines(snake_case__ ) return diffs def __lowerCamelCase ( snake_case__ = False ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = glob.glob(os.path.join(snake_case__ ,"""**/*.py""" ) ,recursive=snake_case__ ) _SCREAMING_SNAKE_CASE = [] for filename in all_files: _SCREAMING_SNAKE_CASE = is_copy_consistent(snake_case__ ,snake_case__ ) diffs += [F'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = """\n""".join(snake_case__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') UpperCamelCase = parser.parse_args() check_copies(args.fix_and_overwrite)
306
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase ) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) self.check_model_type(UpperCAmelCase_ ) def UpperCamelCase ( self: str , **UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} # preprocess args if "points_per_batch" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self: Optional[Any] , UpperCAmelCase_: Tuple , *UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=None , UpperCAmelCase_: Tuple=None , **UpperCAmelCase_: Any ): '''simple docstring''' return super().__call__(UpperCAmelCase_ , *UpperCAmelCase_ , num_workers=UpperCAmelCase_ , batch_size=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict=64 , UpperCAmelCase_: int = 0 , UpperCAmelCase_: float = 512 / 1_500 , UpperCAmelCase_: Optional[int] = 32 , UpperCAmelCase_: Optional[int] = 1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor.size["""longest_edge"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.generate_crop_boxes( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": _SCREAMING_SNAKE_CASE = self.get_inference_context() with inference_context(): _SCREAMING_SNAKE_CASE = self._ensure_tensor_on_device(UpperCAmelCase_ , device=self.device ) _SCREAMING_SNAKE_CASE = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) _SCREAMING_SNAKE_CASE = image_embeddings _SCREAMING_SNAKE_CASE = grid_points.shape[1] _SCREAMING_SNAKE_CASE = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = grid_points[:, i : i + points_per_batch, :, :] _SCREAMING_SNAKE_CASE = input_labels[:, i : i + points_per_batch] _SCREAMING_SNAKE_CASE = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=0.88 , UpperCAmelCase_: Dict=0.95 , UpperCAmelCase_: Tuple=0 , UpperCAmelCase_: str=1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_inputs.pop("""input_boxes""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""is_last""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""original_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = model_inputs.pop("""reshaped_input_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks _SCREAMING_SNAKE_CASE = model_outputs["""pred_masks"""] _SCREAMING_SNAKE_CASE = self.image_processor.post_process_masks( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , binarize=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model_outputs["""iou_scores"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCamelCase ( self: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=False , UpperCAmelCase_: Any=0.7 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.post_process_for_mask_generation( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = defaultdict(UpperCAmelCase_ ) for output in model_outputs: for k, v in output.items(): extra[k].append(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {} if output_rle_mask: _SCREAMING_SNAKE_CASE = rle_mask if output_bboxes_mask: _SCREAMING_SNAKE_CASE = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
306
1
import heapq import sys import numpy as np UpperCamelCase = tuple[int, int] class __UpperCAmelCase : def __init__( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = set() def UpperCamelCase ( self: int ): '''simple docstring''' if not self.empty(): return self.elements[0][0] else: return float("""inf""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' return len(self.elements ) == 0 def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(UpperCAmelCase_ ) else: # update # print("update", item) _SCREAMING_SNAKE_CASE = [] ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[int] ): '''simple docstring''' if item in self.set: self.set.remove(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [] ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def UpperCamelCase ( self: int ): '''simple docstring''' return self.elements[0][1] def UpperCamelCase ( self: Any ): '''simple docstring''' ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = heapq.heappop(self.elements ) self.set.remove(UpperCAmelCase_ ) return (priority, item) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.array(snake_case__ ) _SCREAMING_SNAKE_CASE = np.array(snake_case__ ) return np.linalg.norm(a - b ) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" return consistent_heuristic(snake_case__ ,snake_case__ ) // t def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Optional[Any]: """simple docstring""" return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = g_function[start] + Wa * heuristics[i](snake_case__ ,snake_case__ ) return ans def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = np.chararray((n, n) ) for i in range(snake_case__ ): for j in range(snake_case__ ): _SCREAMING_SNAKE_CASE = """*""" for i in range(snake_case__ ): for j in range(snake_case__ ): if (j, (n - 1) - i) in blocks: _SCREAMING_SNAKE_CASE = """#""" _SCREAMING_SNAKE_CASE = """-""" _SCREAMING_SNAKE_CASE = back_pointer[goal] while x != start: ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = x # print(x) _SCREAMING_SNAKE_CASE = """-""" _SCREAMING_SNAKE_CASE = back_pointer[x] _SCREAMING_SNAKE_CASE = """-""" for i in range(snake_case__ ): for j in range(snake_case__ ): if (i, j) == (0, n - 1): print(grid[i][j] ,end=""" """ ) print("""<-- End position""" ,end=""" """ ) else: print(grid[i][j] ,end=""" """ ) print() print("""^""" ) print("""Start position""" ) print() print("""# is an obstacle""" ) print("""- is the path taken by algorithm""" ) print("""PATH TAKEN BY THE ALGORITHM IS:-""" ) _SCREAMING_SNAKE_CASE = back_pointer[goal] while x != start: print(snake_case__ ,end=""" """ ) _SCREAMING_SNAKE_CASE = back_pointer[x] print(snake_case__ ) sys.exit() def __lowerCamelCase ( snake_case__ ) -> Optional[int]: """simple docstring""" if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,) -> Tuple: """simple docstring""" for itera in range(snake_case__ ): open_list[itera].remove_element(snake_case__ ) # print("s", s) # print("j", j) ((_SCREAMING_SNAKE_CASE) , (_SCREAMING_SNAKE_CASE)) = s _SCREAMING_SNAKE_CASE = (x - 1, y) _SCREAMING_SNAKE_CASE = (x + 1, y) _SCREAMING_SNAKE_CASE = (x, y + 1) _SCREAMING_SNAKE_CASE = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(snake_case__ ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(snake_case__ ) _SCREAMING_SNAKE_CASE = -1 _SCREAMING_SNAKE_CASE = float("""inf""" ) if valid(snake_case__ ) and g_function[neighbours] > g_function[s] + 1: _SCREAMING_SNAKE_CASE = g_function[s] + 1 _SCREAMING_SNAKE_CASE = s if neighbours not in close_list_anchor: open_list[0].put(snake_case__ ,key(snake_case__ ,0 ,snake_case__ ,snake_case__ ) ) if neighbours not in close_list_inad: for var in range(1 ,snake_case__ ): if key(snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) <= Wa * key( snake_case__ ,0 ,snake_case__ ,snake_case__ ): open_list[j].put( snake_case__ ,key(snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) ) def __lowerCamelCase ( ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = [] for x in range(1 ,5 ): for y in range(1 ,6 ): some_list.append((x, y) ) for x in range(15 ,20 ): some_list.append((x, 17) ) for x in range(10 ,19 ): for y in range(1 ,15 ): some_list.append((x, y) ) # L block for x in range(1 ,4 ): for y in range(12 ,19 ): some_list.append((x, y) ) for x in range(3 ,13 ): for y in range(16 ,19 ): some_list.append((x, y) ) return some_list UpperCamelCase = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} UpperCamelCase = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] UpperCamelCase = make_common_ground() UpperCamelCase = blocks_blk # hyper parameters UpperCamelCase = 1 UpperCamelCase = 1 UpperCamelCase = 20 UpperCamelCase = 3 # one consistent and two other inconsistent # start and end destination UpperCamelCase = (0, 0) UpperCamelCase = (n - 1, n - 1) UpperCamelCase = 1 def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = {start: 0, goal: float("""inf""" )} _SCREAMING_SNAKE_CASE = {start: -1, goal: -1} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = set() for i in range(snake_case__ ): open_list.append(PriorityQueue() ) open_list[i].put(snake_case__ ,key(snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) ) _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] while open_list[0].minkey() < float("""inf""" ): for i in range(1 ,snake_case__ ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("""inf""" ): do_something(snake_case__ ,snake_case__ ,snake_case__ ) else: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = open_list[i].top_show() visited.add(snake_case__ ) expand_state( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,) close_list_inad.append(snake_case__ ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("""inf""" ): do_something(snake_case__ ,snake_case__ ,snake_case__ ) else: _SCREAMING_SNAKE_CASE = open_list[0].top_show() visited.add(snake_case__ ) expand_state( snake_case__ ,0 ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,) close_list_anchor.append(snake_case__ ) print("""No path found to goal""" ) print() for i in range(n - 1 ,-1 ,-1 ): for j in range(snake_case__ ): if (j, i) in blocks: print("""#""" ,end=""" """ ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("""*""" ,end=""" """ ) else: print("""-""" ,end=""" """ ) else: print("""*""" ,end=""" """ ) if (j, i) == (n - 1, n - 1): print("""<-- End position""" ,end=""" """ ) print() print("""^""" ) print("""Start position""" ) print() print("""# is an obstacle""" ) print("""- is the path taken by algorithm""" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
306
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging UpperCamelCase = logging.get_logger(__name__) logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" if "xprophetnet" in prophetnet_checkpoint_path: _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) else: _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) _SCREAMING_SNAKE_CASE = ["""key_proj""", """value_proj""", """query_proj"""] _SCREAMING_SNAKE_CASE = { """self_attn""": """ngram_self_attn""", """cross_attn""": """encoder_attn""", """cross_attn_layer_norm""": """encoder_attn_layer_norm""", """feed_forward_layer_norm""": """final_layer_norm""", """feed_forward""": """""", """intermediate""": """fc1""", """output""": """fc2""", """key_proj""": """k_proj""", """query_proj""": """q_proj""", """value_proj""": """v_proj""", """word_embeddings""": """embed_tokens""", """embeddings_layer_norm""": """emb_layer_norm""", """relative_pos_embeddings""": """relative_linear""", """ngram_embeddings""": """ngram_input_embed""", """position_embeddings""": """embed_positions""", } for key in loading_info["missing_keys"]: _SCREAMING_SNAKE_CASE = key.split(""".""" ) if attributes[0] == "lm_head": _SCREAMING_SNAKE_CASE = prophet _SCREAMING_SNAKE_CASE = prophet_old else: _SCREAMING_SNAKE_CASE = prophet.prophetnet _SCREAMING_SNAKE_CASE = prophet_old.model _SCREAMING_SNAKE_CASE = False for attribute in attributes: if attribute in mapping: _SCREAMING_SNAKE_CASE = mapping[attribute] if not hasattr(snake_case__ ,snake_case__ ) and len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = attribute elif hasattr(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.weight logger.info(F'{attribute} is initialized.' ) _SCREAMING_SNAKE_CASE = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.bias logger.info(F'{attribute} is initialized' ) _SCREAMING_SNAKE_CASE = True break elif attribute in special_keys and hasattr(snake_case__ ,"""in_proj_weight""" ): _SCREAMING_SNAKE_CASE = old_model.in_proj_weight.shape[0] // 3 _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _SCREAMING_SNAKE_CASE = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_12, "We want 512 position_embeddings." _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.embed_positions.weight[:5_12, :] ) _SCREAMING_SNAKE_CASE = True break if attribute.isdigit(): _SCREAMING_SNAKE_CASE = model[int(snake_case__ )] _SCREAMING_SNAKE_CASE = old_model[int(snake_case__ )] else: _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if old_attribute == "": _SCREAMING_SNAKE_CASE = old_model else: if not hasattr(snake_case__ ,snake_case__ ): raise ValueError(F'{old_model} does not have {old_attribute}' ) _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if not is_key_init: raise ValueError(F'{key} was not correctly initialized!' ) print(F'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--prophetnet_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
306
1
def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> int: """simple docstring""" return int((input_a, input_a).count(1 ) != 0 ) def __lowerCamelCase ( ) -> None: """simple docstring""" assert or_gate(0 ,0 ) == 0 assert or_gate(0 ,1 ) == 1 assert or_gate(1 ,0 ) == 1 assert or_gate(1 ,1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
306
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: _SCREAMING_SNAKE_CASE = i + 1 else: _SCREAMING_SNAKE_CASE = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"{two_pointer([2, 7, 11, 15], 9) = }")
306
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = 3 _SCREAMING_SNAKE_CASE = (32, 32) _SCREAMING_SNAKE_CASE = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(UpperCAmelCase_ ) return image @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 32, 64) , layers_per_block=2 , sample_size=32 , in_channels=7 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=8 , use_linear_projection=UpperCAmelCase_ , only_cross_attention=(True, True, False) , num_class_embeds=100 , ) return model @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) return CLIPTextModel(UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.dummy_cond_unet_upscale _SCREAMING_SNAKE_CASE = DDPMScheduler() _SCREAMING_SNAKE_CASE = DDIMScheduler(prediction_type="""v_prediction""" ) _SCREAMING_SNAKE_CASE = self.dummy_vae _SCREAMING_SNAKE_CASE = self.dummy_text_encoder _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(UpperCAmelCase_ ) ).convert("""RGB""" ).resize((64, 64) ) # make sure here that pndm scheduler skips prk _SCREAMING_SNAKE_CASE = StableDiffusionUpscalePipeline( unet=UpperCAmelCase_ , low_res_scheduler=UpperCAmelCase_ , scheduler=UpperCAmelCase_ , vae=UpperCAmelCase_ , text_encoder=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ , max_noise_level=350 , ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger""" _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = sd_pipe( [prompt] , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = output.images _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = sd_pipe( [prompt] , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , return_dict=UpperCAmelCase_ , )[0] _SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] _SCREAMING_SNAKE_CASE = image_from_tuple[0, -3:, -3:, -1] _SCREAMING_SNAKE_CASE = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) _SCREAMING_SNAKE_CASE = np.array([0.31_13, 0.39_10, 0.42_72, 0.48_59, 0.50_61, 0.46_52, 0.53_62, 0.57_15, 0.56_61] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.dummy_cond_unet_upscale _SCREAMING_SNAKE_CASE = DDPMScheduler() _SCREAMING_SNAKE_CASE = DDIMScheduler(prediction_type="""v_prediction""" ) _SCREAMING_SNAKE_CASE = self.dummy_vae _SCREAMING_SNAKE_CASE = self.dummy_text_encoder _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(UpperCAmelCase_ ) ).convert("""RGB""" ).resize((64, 64) ) # make sure here that pndm scheduler skips prk _SCREAMING_SNAKE_CASE = StableDiffusionUpscalePipeline( unet=UpperCAmelCase_ , low_res_scheduler=UpperCAmelCase_ , scheduler=UpperCAmelCase_ , vae=UpperCAmelCase_ , text_encoder=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ , max_noise_level=350 , ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger""" _SCREAMING_SNAKE_CASE = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = output.images assert image.shape[0] == 2 _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = sd_pipe( [prompt] , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=20 , num_inference_steps=2 , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.dummy_cond_unet_upscale _SCREAMING_SNAKE_CASE = DDPMScheduler() _SCREAMING_SNAKE_CASE = DDIMScheduler(prediction_type="""v_prediction""" ) _SCREAMING_SNAKE_CASE = self.dummy_vae _SCREAMING_SNAKE_CASE = self.dummy_text_encoder _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(UpperCAmelCase_ ) ).convert("""RGB""" ).resize((64, 64) ) # put models in fp16, except vae as it overflows in fp16 _SCREAMING_SNAKE_CASE = unet.half() _SCREAMING_SNAKE_CASE = text_encoder.half() # make sure here that pndm scheduler skips prk _SCREAMING_SNAKE_CASE = StableDiffusionUpscalePipeline( unet=UpperCAmelCase_ , low_res_scheduler=UpperCAmelCase_ , scheduler=UpperCAmelCase_ , vae=UpperCAmelCase_ , text_encoder=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ , max_noise_level=350 , ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """A painting of a squirrel eating a burger""" _SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = sd_pipe( [prompt] , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""np""" , ).images _SCREAMING_SNAKE_CASE = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: Dict ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-upscale/low_res_cat.png""" ) _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale""" """/upsampled_cat.npy""" ) _SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-x4-upscaler""" _SCREAMING_SNAKE_CASE = StableDiffusionUpscalePipeline.from_pretrained(UpperCAmelCase_ ) pipe.to(UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) pipe.enable_attention_slicing() _SCREAMING_SNAKE_CASE = """a cat sitting on a park bench""" _SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-3 def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-upscale/low_res_cat.png""" ) _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale""" """/upsampled_cat_fp16.npy""" ) _SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-x4-upscaler""" _SCREAMING_SNAKE_CASE = StableDiffusionUpscalePipeline.from_pretrained( UpperCAmelCase_ , torch_dtype=torch.floataa , ) pipe.to(UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) pipe.enable_attention_slicing() _SCREAMING_SNAKE_CASE = """a cat sitting on a park bench""" _SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCamelCase ( self: str ): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-upscale/low_res_cat.png""" ) _SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-x4-upscaler""" _SCREAMING_SNAKE_CASE = StableDiffusionUpscalePipeline.from_pretrained( UpperCAmelCase_ , torch_dtype=torch.floataa , ) pipe.to(UpperCAmelCase_ ) pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _SCREAMING_SNAKE_CASE = """a cat sitting on a park bench""" _SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe( prompt=UpperCAmelCase_ , image=UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=5 , output_type="""np""" , ) _SCREAMING_SNAKE_CASE = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 10**9
306
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig UpperCamelCase = logging.get_logger(__name__) # General docstring UpperCamelCase = '''MobileNetV1Config''' # Base docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = [1, 1_024, 7, 7] # Image classification docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = '''tabby, tabby cat''' UpperCamelCase = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = {} if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = model.mobilenet_va else: _SCREAMING_SNAKE_CASE = model _SCREAMING_SNAKE_CASE = """MobilenetV1/Conv2d_0/""" _SCREAMING_SNAKE_CASE = backbone.conv_stem.convolution.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.bias _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_mean _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_var for i in range(13 ): _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = i * 2 _SCREAMING_SNAKE_CASE = backbone.layer[pt_index] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_depthwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var _SCREAMING_SNAKE_CASE = backbone.layer[pt_index + 1] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_pointwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = """MobilenetV1/Logits/Conv2d_1c_1x1/""" _SCREAMING_SNAKE_CASE = model.classifier.weight _SCREAMING_SNAKE_CASE = model.classifier.bias return tf_to_pt_map def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> List[str]: """simple docstring""" try: import numpy as np import tensorflow as tf except ImportError: logger.error( """Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see """ """https://www.tensorflow.org/install/ for installation instructions.""" ) raise # Load weights from TF model _SCREAMING_SNAKE_CASE = tf.train.list_variables(snake_case__ ) _SCREAMING_SNAKE_CASE = {} for name, shape in init_vars: logger.info(F'Loading TF weight {name} with shape {shape}' ) _SCREAMING_SNAKE_CASE = tf.train.load_variable(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = array # Build TF to PyTorch weights loading map _SCREAMING_SNAKE_CASE = _build_tf_to_pytorch_map(snake_case__ ,snake_case__ ,snake_case__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F'Importing {name}' ) if name not in tf_weights: logger.info(F'{name} not in tf pre-trained weights, skipping' ) continue _SCREAMING_SNAKE_CASE = tf_weights[name] if "depthwise_weights" in name: logger.info("""Transposing depthwise""" ) _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(2, 3, 0, 1) ) elif "weights" in name: logger.info("""Transposing""" ) if len(pointer.shape ) == 2: # copying into linear layer _SCREAMING_SNAKE_CASE = array.squeeze().transpose() else: _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'Pointer shape {pointer.shape} and array shape {array.shape} mismatched' ) logger.info(F'Initialize PyTorch weight {name} {array.shape}' ) _SCREAMING_SNAKE_CASE = torch.from_numpy(snake_case__ ) tf_weights.pop(snake_case__ ,snake_case__ ) tf_weights.pop(name + """/RMSProp""" ,snake_case__ ) tf_weights.pop(name + """/RMSProp_1""" ,snake_case__ ) tf_weights.pop(name + """/ExponentialMovingAverage""" ,snake_case__ ) logger.info(F'Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}' ) return model def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> torch.Tensor: """simple docstring""" _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = features.shape[-2:] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.stride _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.kernel_size if in_height % stride_height == 0: _SCREAMING_SNAKE_CASE = max(kernel_height - stride_height ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_height - (in_height % stride_height) ,0 ) if in_width % stride_width == 0: _SCREAMING_SNAKE_CASE = max(kernel_width - stride_width ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_width - (in_width % stride_width) ,0 ) _SCREAMING_SNAKE_CASE = pad_along_width // 2 _SCREAMING_SNAKE_CASE = pad_along_width - pad_left _SCREAMING_SNAKE_CASE = pad_along_height // 2 _SCREAMING_SNAKE_CASE = pad_along_height - pad_top _SCREAMING_SNAKE_CASE = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(snake_case__ ,snake_case__ ,"""constant""" ,0.0 ) class __UpperCAmelCase (nn.Module ): def __init__( self: Optional[Any] , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: bool = False , UpperCAmelCase_: Optional[bool] = True , UpperCAmelCase_: Optional[bool or str] = True , ): '''simple docstring''' super().__init__() _SCREAMING_SNAKE_CASE = config if in_channels % groups != 0: raise ValueError(F'Input channels ({in_channels}) are not divisible by {groups} groups.' ) if out_channels % groups != 0: raise ValueError(F'Output channels ({out_channels}) are not divisible by {groups} groups.' ) _SCREAMING_SNAKE_CASE = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) _SCREAMING_SNAKE_CASE = nn.Convad( in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=UpperCAmelCase_ , groups=UpperCAmelCase_ , bias=UpperCAmelCase_ , padding_mode="""zeros""" , ) if use_normalization: _SCREAMING_SNAKE_CASE = nn.BatchNormad( num_features=UpperCAmelCase_ , eps=config.layer_norm_eps , momentum=0.99_97 , affine=UpperCAmelCase_ , track_running_stats=UpperCAmelCase_ , ) else: _SCREAMING_SNAKE_CASE = None if use_activation: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[use_activation] elif isinstance(config.hidden_act , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: _SCREAMING_SNAKE_CASE = config.hidden_act else: _SCREAMING_SNAKE_CASE = None def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' if self.config.tf_padding: _SCREAMING_SNAKE_CASE = apply_tf_padding(UpperCAmelCase_ , self.convolution ) _SCREAMING_SNAKE_CASE = self.convolution(UpperCAmelCase_ ) if self.normalization is not None: _SCREAMING_SNAKE_CASE = self.normalization(UpperCAmelCase_ ) if self.activation is not None: _SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase_ ) return features class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Dict = MobileNetVaConfig __snake_case : Any = load_tf_weights_in_mobilenet_va __snake_case : Any = "mobilenet_v1" __snake_case : List[Any] = "pixel_values" __snake_case : Any = False def UpperCamelCase ( self: str , UpperCAmelCase_: Union[nn.Linear, nn.Convad] ): '''simple docstring''' if isinstance(UpperCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCAmelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) UpperCamelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' UpperCamelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: bool = True ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config _SCREAMING_SNAKE_CASE = 32 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) _SCREAMING_SNAKE_CASE = MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=config.num_channels , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=2 , ) _SCREAMING_SNAKE_CASE = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] _SCREAMING_SNAKE_CASE = nn.ModuleList() for i in range(13 ): _SCREAMING_SNAKE_CASE = out_channels if strides[i] == 2 or i == 0: depth *= 2 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=strides[i] , groups=UpperCAmelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=1 , ) ) _SCREAMING_SNAKE_CASE = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCamelCase ( self: Dict , UpperCAmelCase_: Tuple ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase ( self: int , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) _SCREAMING_SNAKE_CASE = self.conv_stem(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): _SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase_ ) if output_hidden_states: _SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) _SCREAMING_SNAKE_CASE = hidden_states if self.pooler is not None: _SCREAMING_SNAKE_CASE = torch.flatten(self.pooler(UpperCAmelCase_ ) , start_dim=1 ) else: _SCREAMING_SNAKE_CASE = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase_ , pooler_output=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Dict , UpperCAmelCase_: MobileNetVaConfig ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config.num_labels _SCREAMING_SNAKE_CASE = MobileNetVaModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head _SCREAMING_SNAKE_CASE = nn.Dropout(config.classifier_dropout_prob , inplace=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.Linear(UpperCAmelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict _SCREAMING_SNAKE_CASE = self.mobilenet_va(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.pooler_output if return_dict else outputs[1] _SCREAMING_SNAKE_CASE = self.classifier(self.dropout(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _SCREAMING_SNAKE_CASE = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _SCREAMING_SNAKE_CASE = """single_label_classification""" else: _SCREAMING_SNAKE_CASE = """multi_label_classification""" if self.config.problem_type == "regression": _SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: _SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) elif self.config.problem_type == "single_label_classification": _SCREAMING_SNAKE_CASE = CrossEntropyLoss() _SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: _SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states , )
306
1
import os import pytest from transformers.dynamic_module_utils import get_imports UpperCamelCase = ''' import os ''' UpperCamelCase = ''' def foo(): import os return False ''' UpperCamelCase = ''' def foo(): def bar(): if True: import os return False return bar() ''' UpperCamelCase = ''' import os try: import bar except ImportError: raise ValueError() ''' UpperCamelCase = ''' import os def foo(): try: import bar except ImportError: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar except (ImportError, AttributeError): raise ValueError() ''' UpperCamelCase = ''' import os try: import bar except ImportError as e: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar except: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar import baz except ImportError: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar import baz except ImportError: x = 1 raise ValueError() ''' UpperCamelCase = [ TOP_LEVEL_IMPORT, IMPORT_IN_FUNCTION, DEEPLY_NESTED_IMPORT, TOP_LEVEL_TRY_IMPORT, GENERIC_EXCEPT_IMPORT, MULTILINE_TRY_IMPORT, MULTILINE_BOTH_IMPORT, MULTIPLE_EXCEPTS_IMPORT, EXCEPT_AS_IMPORT, TRY_IMPORT_IN_FUNCTION, ] @pytest.mark.parametrize("""case""" ,snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = os.path.join(snake_case__ ,"""test_file.py""" ) with open(snake_case__ ,"""w""" ) as _tmp_file: _tmp_file.write(snake_case__ ) _SCREAMING_SNAKE_CASE = get_imports(snake_case__ ) assert parsed_imports == ["os"]
306
def __lowerCamelCase ( snake_case__ ) -> list: """simple docstring""" def merge(snake_case__ ,snake_case__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(snake_case__ ) <= 1: return collection _SCREAMING_SNAKE_CASE = len(snake_case__ ) // 2 return merge(merge_sort(collection[:mid] ) ,merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
306
1
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __UpperCAmelCase : def __init__( self: Tuple , UpperCAmelCase_: Dict , UpperCAmelCase_: Optional[Any]=13 , UpperCAmelCase_: Any=64 , UpperCAmelCase_: List[str]=2 , UpperCAmelCase_: List[Any]=3 , UpperCAmelCase_: Optional[int]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Union[str, Any]=32 , UpperCAmelCase_: str=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: List[str]="gelu" , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Tuple=0.1 , UpperCAmelCase_: List[Any]=10 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Any=[1, 16, 4, 4] , UpperCAmelCase_: Tuple=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = patch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = scope _SCREAMING_SNAKE_CASE = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size _SCREAMING_SNAKE_CASE = (self.image_size // 32) ** 2 _SCREAMING_SNAKE_CASE = num_patches + 1 def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = { """global_padding""": """same""", """layer_type""": """bottleneck""", """depths""": [3, 4, 9], """out_features""": ["""stage1""", """stage2""", """stage3"""], """embedding_dynamic_padding""": True, """hidden_sizes""": [4, 8, 16, 32], """num_groups""": 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase_ , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=UpperCAmelCase_ , ) def UpperCamelCase ( self: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTHybridModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: Dict , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.type_sequence_label_size _SCREAMING_SNAKE_CASE = ViTHybridForImageClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[str] = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () __snake_case : Any = ( {"feature-extraction": ViTHybridModel, "image-classification": ViTHybridForImageClassification} if is_torch_available() else {} ) __snake_case : List[str] = False __snake_case : Any = False __snake_case : int = False def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTHybridModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , has_text_modality=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' pass def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _SCREAMING_SNAKE_CASE = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase_ , nn.Linear ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] _SCREAMING_SNAKE_CASE = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() _SCREAMING_SNAKE_CASE = _config_zero_init(UpperCAmelCase_ ) for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(config=UpperCAmelCase_ ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": _SCREAMING_SNAKE_CASE = [F'{name}.{key}' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = ViTHybridModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def __lowerCamelCase ( ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCAmelCase (unittest.TestCase ): @cached_property def UpperCamelCase ( self: Any ): '''simple docstring''' return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.default_image_processor _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ ) # verify the logits _SCREAMING_SNAKE_CASE = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor([-1.90_90, -0.49_93, -0.23_89] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow @require_accelerate def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTHybridImageProcessor.from_pretrained("""google/vit-hybrid-base-bit-384""" ) _SCREAMING_SNAKE_CASE = ViTHybridForImageClassification.from_pretrained("""google/vit-hybrid-base-bit-384""" , device_map="""auto""" ) _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.logits # model predicts one of the 1000 ImageNet classes _SCREAMING_SNAKE_CASE = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , """tabby, tabby cat""" )
306
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length, 2) ,snake_case__ ) else: _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length) ,snake_case__ ) for i, tensor in enumerate(snake_case__ ): if padding_side == "right": if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] return out_tensor.tolist() def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = ord(snake_case__ ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True _SCREAMING_SNAKE_CASE = unicodedata.category(snake_case__ ) if cat.startswith("""P""" ): return True return False @dataclass class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : PreTrainedTokenizerBase __snake_case : Union[bool, str, PaddingStrategy] = True __snake_case : Optional[int] = None __snake_case : Optional[int] = None __snake_case : int = -100 __snake_case : str = "pt" def UpperCamelCase ( self: str , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' import torch _SCREAMING_SNAKE_CASE = """label""" if """label""" in features[0].keys() else """labels""" _SCREAMING_SNAKE_CASE = [feature[label_name] for feature in features] if label_name in features[0].keys() else None _SCREAMING_SNAKE_CASE = self.tokenizer.pad( UpperCAmelCase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" if labels is None else None , ) if labels is None: return batch _SCREAMING_SNAKE_CASE = torch.tensor(batch["""entity_ids"""] ).shape[1] _SCREAMING_SNAKE_CASE = self.tokenizer.padding_side if padding_side == "right": _SCREAMING_SNAKE_CASE = [ list(UpperCAmelCase_ ) + [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) for label in labels ] else: _SCREAMING_SNAKE_CASE = [ [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) + list(UpperCAmelCase_ ) for label in labels ] _SCREAMING_SNAKE_CASE = [feature["""ner_tags"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , -1 , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [feature["""original_entity_spans"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , (-1, -1) , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {k: torch.tensor(UpperCAmelCase_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
306
1
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class __UpperCAmelCase : def __init__( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: int=None , UpperCAmelCase_: Any="resnet50" , UpperCAmelCase_: str=3 , UpperCAmelCase_: Union[str, Any]=32 , UpperCAmelCase_: str=3 , UpperCAmelCase_: Any=True , UpperCAmelCase_: int=True , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = out_indices if out_indices is not None else [4] _SCREAMING_SNAKE_CASE = stage_names _SCREAMING_SNAKE_CASE = out_features _SCREAMING_SNAKE_CASE = backbone _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = use_pretrained_backbone _SCREAMING_SNAKE_CASE = is_training def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TimmBackbone(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch @require_timm class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = (TimmBackbone,) if is_torch_available() else () __snake_case : Tuple = {"feature-extraction": TimmBackbone} if is_torch_available() else {} __snake_case : Union[str, Any] = False __snake_case : int = False __snake_case : Union[str, Any] = False __snake_case : Optional[Any] = False def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TimmBackboneModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , has_text_modality=UpperCAmelCase_ ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """resnet18""" _SCREAMING_SNAKE_CASE = """microsoft/resnet-18""" _SCREAMING_SNAKE_CASE = AutoBackbone.from_pretrained(UpperCAmelCase_ , use_timm_backbone=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = AutoBackbone.from_pretrained(UpperCAmelCase_ ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) _SCREAMING_SNAKE_CASE = AutoBackbone.from_pretrained(UpperCAmelCase_ , use_timm_backbone=UpperCAmelCase_ , out_indices=[1, 2, 3] ) _SCREAMING_SNAKE_CASE = AutoBackbone.from_pretrained(UpperCAmelCase_ , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip("""TimmBackbone doesn't support feed forward chunking""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @unittest.skip("""TimmBackbone doesn't have num_hidden_layers attribute""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @unittest.skip("""TimmBackbone initialization is managed on the timm side""" ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' pass @unittest.skip("""TimmBackbone models doesn't have inputs_embeds""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @unittest.skip("""TimmBackbone models doesn't have inputs_embeds""" ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' pass @unittest.skip("""TimmBackbone model cannot be created without specifying a backbone checkpoint""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @unittest.skip("""model weights aren't tied in TimmBackbone.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip("""model weights aren't tied in TimmBackbone.""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip("""TimmBackbone doesn't have hidden size info in its configuration.""" ) def UpperCamelCase ( self: str ): '''simple docstring''' pass @unittest.skip("""TimmBackbone doesn't support output_attentions.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass @unittest.skip("""Safetensors is not supported by timm.""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] _SCREAMING_SNAKE_CASE = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = self.has_attentions # no need to test all models as different heads yield the same functionality _SCREAMING_SNAKE_CASE = self.all_model_classes[0] _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self._prepare_for_class(UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs[0][-1] # Encoder-/Decoder-only models _SCREAMING_SNAKE_CASE = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: _SCREAMING_SNAKE_CASE = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=UpperCAmelCase_ ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None _SCREAMING_SNAKE_CASE = copy.deepcopy(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights _SCREAMING_SNAKE_CASE = copy.deepcopy(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ )
306
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = weights[0][0][0] _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[0] ) _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # lsh weights + output _SCREAMING_SNAKE_CASE = weights[0][1] if len(snake_case__ ) < 4: set_layer_weights_in_torch_lsh(snake_case__ ,torch_block.attention ,snake_case__ ) else: set_layer_weights_in_torch_local(snake_case__ ,torch_block.attention ,snake_case__ ) # intermediate weighs _SCREAMING_SNAKE_CASE = weights[2][0][1][2] # Chunked Feed Forward if len(snake_case__ ) == 4: _SCREAMING_SNAKE_CASE = intermediate_weights[2] # layernorm 2 _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # intermediate dense _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) # intermediate out _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = torch_model.reformer # word embeds _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings ,torch.tensor(snake_case__ ) ,) if isinstance(weights[3] ,snake_case__ ): _SCREAMING_SNAKE_CASE = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _SCREAMING_SNAKE_CASE = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(torch.tensor(snake_case__ ) ) _SCREAMING_SNAKE_CASE = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( snake_case__ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _SCREAMING_SNAKE_CASE = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(snake_case__ ,snake_case__ ,snake_case__ ) # output layer norm _SCREAMING_SNAKE_CASE = np.asarray(weights[7][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # output embeddings _SCREAMING_SNAKE_CASE = np.asarray(weights[9][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ReformerConfig.from_json_file(snake_case__ ) print(F'Building PyTorch model from configuration: {config}' ) _SCREAMING_SNAKE_CASE = ReformerModelWithLMHead(snake_case__ ) with open(snake_case__ ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = pickle.load(snake_case__ )["""weights"""] set_model_weights_in_torch(snake_case__ ,snake_case__ ,config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() ,snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
306
1
def __lowerCamelCase ( snake_case__ = 1_00 ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = n * (n + 1) * (2 * n + 1) / 6 _SCREAMING_SNAKE_CASE = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f"{solution() = }")
306
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = TextToVideoSDPipeline __snake_case : Optional[int] = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __snake_case : Optional[int] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase ( self: int ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) _SCREAMING_SNAKE_CASE = CLIPTextModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict=0 ): '''simple docstring''' if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.get_dummy_inputs(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """np""" _SCREAMING_SNAKE_CASE = sd_pipe(**UpperCAmelCase_ ).frames _SCREAMING_SNAKE_CASE = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _SCREAMING_SNAKE_CASE = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=1E-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=25 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
306
1
from __future__ import annotations from collections import Counter from random import random class __UpperCAmelCase : def __init__( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} def UpperCamelCase ( self: Dict , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: float ): '''simple docstring''' if nodea not in self.connections: self.add_node(UpperCAmelCase_ ) if nodea not in self.connections: self.add_node(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = probability def UpperCamelCase ( self: int ): '''simple docstring''' return list(self.connections ) def UpperCamelCase ( self: int , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> dict[str, int]: """simple docstring""" _SCREAMING_SNAKE_CASE = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(snake_case__ ,snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = Counter(graph.get_nodes() ) _SCREAMING_SNAKE_CASE = start for _ in range(snake_case__ ): _SCREAMING_SNAKE_CASE = graph.transition(snake_case__ ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
306
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class __UpperCAmelCase : def __init__( self: Union[str, Any] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int=13 , UpperCAmelCase_: Optional[int]=7 , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=False , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Optional[int]=33 , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: List[Any]=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Any=37 , UpperCAmelCase_: Optional[Any]="gelu" , UpperCAmelCase_: Dict=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: int=16 , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: Tuple=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: str=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = EsmForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = False __snake_case : Dict = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __snake_case : List[Any] = () __snake_case : Dict = ( { "feature-extraction": EsmModel, "fill-mask": EsmForMaskedLM, "text-classification": EsmForSequenceClassification, "token-classification": EsmForTokenClassification, "zero-shot": EsmForSequenceClassification, } if is_torch_available() else {} ) __snake_case : int = True def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: int ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _SCREAMING_SNAKE_CASE = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _SCREAMING_SNAKE_CASE = create_position_ids_from_input_ids(UpperCAmelCase_ , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.empty(2 , 4 , 30 ) _SCREAMING_SNAKE_CASE = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _SCREAMING_SNAKE_CASE = torch.as_tensor([expected_single_positions, expected_single_positions] ) _SCREAMING_SNAKE_CASE = embeddings.create_position_ids_from_inputs_embeds(UpperCAmelCase_ ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Dict ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Any ): '''simple docstring''' pass @require_torch class __UpperCAmelCase (_UpperCAmelCase ): @slow def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[[8.92_15, -10.58_98, -6.46_71], [-6.39_67, -13.91_14, -1.12_12], [-7.78_12, -13.95_16, -3.74_06]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] # compare the actual values for a slice. _SCREAMING_SNAKE_CASE = torch.tensor( [[[0.14_44, 0.54_13, 0.32_48], [0.30_34, 0.00_53, 0.31_08], [0.32_28, -0.24_99, 0.34_15]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
1
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Any = MvpTokenizer __snake_case : Optional[Any] = MvpTokenizerFast __snake_case : Dict = True __snake_case : Any = filter_roberta_detectors def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' super().setUp() _SCREAMING_SNAKE_CASE = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] _SCREAMING_SNAKE_CASE = dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_ ) ) ) ) _SCREAMING_SNAKE_CASE = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] _SCREAMING_SNAKE_CASE = {"""unk_token""": """<unk>"""} _SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) _SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(UpperCAmelCase_ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(UpperCAmelCase_ ) ) def UpperCamelCase ( self: Optional[int] , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def UpperCamelCase ( self: str , UpperCAmelCase_: List[str] ): '''simple docstring''' return "lower newer", "lower newer" @cached_property def UpperCamelCase ( self: int ): '''simple docstring''' return MvpTokenizer.from_pretrained("""RUCAIBox/mvp""" ) @cached_property def UpperCamelCase ( self: str ): '''simple docstring''' return MvpTokenizerFast.from_pretrained("""RUCAIBox/mvp""" ) @require_torch def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] _SCREAMING_SNAKE_CASE = [0, 250, 251, 17_818, 13, 39_186, 1_938, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _SCREAMING_SNAKE_CASE = tokenizer(UpperCAmelCase_ , max_length=len(UpperCAmelCase_ ) , padding=UpperCAmelCase_ , return_tensors="""pt""" ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) _SCREAMING_SNAKE_CASE = batch.input_ids.tolist()[0] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Test that special tokens are reset @require_torch def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _SCREAMING_SNAKE_CASE = tokenizer(UpperCAmelCase_ , padding=UpperCAmelCase_ , return_tensors="""pt""" ) # check if input_ids are returned and no labels self.assertIn("""input_ids""" , UpperCAmelCase_ ) self.assertIn("""attention_mask""" , UpperCAmelCase_ ) self.assertNotIn("""labels""" , UpperCAmelCase_ ) self.assertNotIn("""decoder_attention_mask""" , UpperCAmelCase_ ) @require_torch def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [ """Summary of the text.""", """Another summary.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _SCREAMING_SNAKE_CASE = tokenizer(text_target=UpperCAmelCase_ , max_length=32 , padding="""max_length""" , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) @require_torch def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _SCREAMING_SNAKE_CASE = tokenizer( ["""I am a small frog""" * 1_024, """I am a small frog"""] , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , return_tensors="""pt""" ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(batch.input_ids.shape , (2, 1_024) ) @require_torch def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ["""A long paragraph for summarization."""] _SCREAMING_SNAKE_CASE = [ """Summary of the text.""", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: _SCREAMING_SNAKE_CASE = tokenizer(UpperCAmelCase_ , text_target=UpperCAmelCase_ , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = inputs["""input_ids"""] _SCREAMING_SNAKE_CASE = inputs["""labels"""] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' pass def UpperCamelCase ( self: str ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _SCREAMING_SNAKE_CASE = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """A, <mask> AllenNLP sentence.""" _SCREAMING_SNAKE_CASE = tokenizer_r.encode_plus(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tokenizer_p.encode_plus(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) _SCREAMING_SNAKE_CASE = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) _SCREAMING_SNAKE_CASE = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual( UpperCAmelCase_ , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( UpperCAmelCase_ , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] )
306
import random def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = num - 1 _SCREAMING_SNAKE_CASE = 0 while s % 2 == 0: _SCREAMING_SNAKE_CASE = s // 2 t += 1 for _ in range(5 ): _SCREAMING_SNAKE_CASE = random.randrange(2 ,num - 1 ) _SCREAMING_SNAKE_CASE = pow(snake_case__ ,snake_case__ ,snake_case__ ) if v != 1: _SCREAMING_SNAKE_CASE = 0 while v != (num - 1): if i == t - 1: return False else: _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = (v**2) % num return True def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" if num < 2: return False _SCREAMING_SNAKE_CASE = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 1_01, 1_03, 1_07, 1_09, 1_13, 1_27, 1_31, 1_37, 1_39, 1_49, 1_51, 1_57, 1_63, 1_67, 1_73, 1_79, 1_81, 1_91, 1_93, 1_97, 1_99, 2_11, 2_23, 2_27, 2_29, 2_33, 2_39, 2_41, 2_51, 2_57, 2_63, 2_69, 2_71, 2_77, 2_81, 2_83, 2_93, 3_07, 3_11, 3_13, 3_17, 3_31, 3_37, 3_47, 3_49, 3_53, 3_59, 3_67, 3_73, 3_79, 3_83, 3_89, 3_97, 4_01, 4_09, 4_19, 4_21, 4_31, 4_33, 4_39, 4_43, 4_49, 4_57, 4_61, 4_63, 4_67, 4_79, 4_87, 4_91, 4_99, 5_03, 5_09, 5_21, 5_23, 5_41, 5_47, 5_57, 5_63, 5_69, 5_71, 5_77, 5_87, 5_93, 5_99, 6_01, 6_07, 6_13, 6_17, 6_19, 6_31, 6_41, 6_43, 6_47, 6_53, 6_59, 6_61, 6_73, 6_77, 6_83, 6_91, 7_01, 7_09, 7_19, 7_27, 7_33, 7_39, 7_43, 7_51, 7_57, 7_61, 7_69, 7_73, 7_87, 7_97, 8_09, 8_11, 8_21, 8_23, 8_27, 8_29, 8_39, 8_53, 8_57, 8_59, 8_63, 8_77, 8_81, 8_83, 8_87, 9_07, 9_11, 9_19, 9_29, 9_37, 9_41, 9_47, 9_53, 9_67, 9_71, 9_77, 9_83, 9_91, 9_97, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(snake_case__ ) def __lowerCamelCase ( snake_case__ = 10_24 ) -> int: """simple docstring""" while True: _SCREAMING_SNAKE_CASE = random.randrange(2 ** (keysize - 1) ,2 ** (keysize) ) if is_prime_low_num(snake_case__ ): return num if __name__ == "__main__": UpperCamelCase = generate_large_prime() print(('''Prime number:''', num)) print(('''is_prime_low_num:''', is_prime_low_num(num)))
306
1
def __lowerCamelCase ( snake_case__ ) -> list: """simple docstring""" def merge(snake_case__ ,snake_case__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(snake_case__ ) <= 1: return collection _SCREAMING_SNAKE_CASE = len(snake_case__ ) // 2 return merge(merge_sort(collection[:mid] ) ,merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
306
def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _SCREAMING_SNAKE_CASE = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
306
1
from manim import * class __UpperCAmelCase (_UpperCAmelCase ): def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = Rectangle(height=0.5 , width=0.5 ) _SCREAMING_SNAKE_CASE = Rectangle(height=0.25 , width=0.25 ) _SCREAMING_SNAKE_CASE = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("""CPU""" , font_size=24 ) _SCREAMING_SNAKE_CASE = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(4 )] _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("""GPU""" , font_size=24 ) _SCREAMING_SNAKE_CASE = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) gpu.move_to([-1, -1, 0] ) self.add(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("""Model""" , font_size=24 ) _SCREAMING_SNAKE_CASE = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) model.move_to([3, -1.0, 0] ) self.add(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(UpperCAmelCase_ ): rect.set_stroke(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(UpperCAmelCase_ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=UpperCAmelCase_ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=UpperCAmelCase_ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=UpperCAmelCase_ , buff=0.0 ) self.add(UpperCAmelCase_ ) model_cpu_arr.append(UpperCAmelCase_ ) self.add(*UpperCAmelCase_ , *UpperCAmelCase_ , *UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("""Loaded Checkpoint""" , font_size=24 ) _SCREAMING_SNAKE_CASE = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) checkpoint.move_to([3, 0.5, 0] ) self.add(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = fill.copy().set_fill(UpperCAmelCase_ , opacity=0.7 ) target.move_to(UpperCAmelCase_ ) ckpt_arr.append(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(UpperCAmelCase_ ) self.add(*UpperCAmelCase_ , *UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _SCREAMING_SNAKE_CASE = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = MarkupText( F'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' , font_size=18 , ) blue_text.next_to(UpperCAmelCase_ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = MarkupText( F'Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.' , font_size=24 , ) step_a.move_to([2, 2, 0] ) _SCREAMING_SNAKE_CASE = [meta_mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = [meta_mem.copy() for i in range(6 )] _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(*UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = VGroup(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0 ) _SCREAMING_SNAKE_CASE = Text("""Disk""" , font_size=24 ) _SCREAMING_SNAKE_CASE = Group(UpperCAmelCase_ , UpperCAmelCase_ ).arrange(UpperCAmelCase_ , buff=0.5 , aligned_edge=UpperCAmelCase_ ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(UpperCAmelCase_ , run_time=3 ) , Write(UpperCAmelCase_ , run_time=1 ) , Create(UpperCAmelCase_ , run_time=1 ) ) _SCREAMING_SNAKE_CASE = [] for i, rect in enumerate(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(UpperCAmelCase_ , run_time=1.5 ) ) self.play(*UpperCAmelCase_ ) self.play(FadeOut(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = MarkupText(F'Then, the checkpoint is removed from memory\nthrough garbage collection.' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(UpperCAmelCase_ , run_time=3 ) ) self.play( FadeOut(UpperCAmelCase_ , UpperCAmelCase_ , *UpperCAmelCase_ , *UpperCAmelCase_ ) , ) self.wait()
306
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
1
from __future__ import annotations import math def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list: """simple docstring""" if len(snake_case__ ) != 2 or len(a[0] ) != 2 or len(snake_case__ ) != 2 or len(b[0] ) != 2: raise Exception("""Matrices are not 2x2""" ) _SCREAMING_SNAKE_CASE = [ [a[0][0] * b[0][0] + a[0][1] * b[1][0], a[0][0] * b[0][1] + a[0][1] * b[1][1]], [a[1][0] * b[0][0] + a[1][1] * b[1][0], a[1][0] * b[0][1] + a[1][1] * b[1][1]], ] return new_matrix def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> int: """simple docstring""" return [ [matrix_a[row][col] + matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(snake_case__ ) ) ] def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Any: """simple docstring""" return [ [matrix_a[row][col] - matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(snake_case__ ) ) ] def __lowerCamelCase ( snake_case__ ) -> tuple[list, list, list, list]: """simple docstring""" if len(snake_case__ ) % 2 != 0 or len(a[0] ) % 2 != 0: raise Exception("""Odd matrices are not supported!""" ) _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = matrix_length // 2 _SCREAMING_SNAKE_CASE = [[a[i][j] for j in range(snake_case__ ,snake_case__ )] for i in range(snake_case__ )] _SCREAMING_SNAKE_CASE = [ [a[i][j] for j in range(snake_case__ ,snake_case__ )] for i in range(snake_case__ ,snake_case__ ) ] _SCREAMING_SNAKE_CASE = [[a[i][j] for j in range(snake_case__ )] for i in range(snake_case__ )] _SCREAMING_SNAKE_CASE = [[a[i][j] for j in range(snake_case__ )] for i in range(snake_case__ ,snake_case__ )] return top_left, top_right, bot_left, bot_right def __lowerCamelCase ( snake_case__ ) -> tuple[int, int]: """simple docstring""" return len(snake_case__ ), len(matrix[0] ) def __lowerCamelCase ( snake_case__ ) -> None: """simple docstring""" print("""\n""".join(str(snake_case__ ) for line in matrix ) ) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list: """simple docstring""" if matrix_dimensions(snake_case__ ) == (2, 2): return default_matrix_multiplication(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = split_matrix(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = split_matrix(snake_case__ ) _SCREAMING_SNAKE_CASE = actual_strassen(snake_case__ ,matrix_subtraction(snake_case__ ,snake_case__ ) ) _SCREAMING_SNAKE_CASE = actual_strassen(matrix_addition(snake_case__ ,snake_case__ ) ,snake_case__ ) _SCREAMING_SNAKE_CASE = actual_strassen(matrix_addition(snake_case__ ,snake_case__ ) ,snake_case__ ) _SCREAMING_SNAKE_CASE = actual_strassen(snake_case__ ,matrix_subtraction(snake_case__ ,snake_case__ ) ) _SCREAMING_SNAKE_CASE = actual_strassen(matrix_addition(snake_case__ ,snake_case__ ) ,matrix_addition(snake_case__ ,snake_case__ ) ) _SCREAMING_SNAKE_CASE = actual_strassen(matrix_subtraction(snake_case__ ,snake_case__ ) ,matrix_addition(snake_case__ ,snake_case__ ) ) _SCREAMING_SNAKE_CASE = actual_strassen(matrix_subtraction(snake_case__ ,snake_case__ ) ,matrix_addition(snake_case__ ,snake_case__ ) ) _SCREAMING_SNAKE_CASE = matrix_addition(matrix_subtraction(matrix_addition(snake_case__ ,snake_case__ ) ,snake_case__ ) ,snake_case__ ) _SCREAMING_SNAKE_CASE = matrix_addition(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = matrix_addition(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = matrix_subtraction(matrix_subtraction(matrix_addition(snake_case__ ,snake_case__ ) ,snake_case__ ) ,snake_case__ ) # construct the new matrix from our 4 quadrants _SCREAMING_SNAKE_CASE = [] for i in range(len(snake_case__ ) ): new_matrix.append(top_left[i] + top_right[i] ) for i in range(len(snake_case__ ) ): new_matrix.append(bot_left[i] + bot_right[i] ) return new_matrix def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list: """simple docstring""" if matrix_dimensions(snake_case__ )[1] != matrix_dimensions(snake_case__ )[0]: _SCREAMING_SNAKE_CASE = ( """Unable to multiply these matrices, please check the dimensions.\n""" F'Matrix A: {matrixa}\n' F'Matrix B: {matrixa}' ) raise Exception(snake_case__ ) _SCREAMING_SNAKE_CASE = matrix_dimensions(snake_case__ ) _SCREAMING_SNAKE_CASE = matrix_dimensions(snake_case__ ) if dimensiona[0] == dimensiona[1] and dimensiona[0] == dimensiona[1]: return [matrixa, matrixa] _SCREAMING_SNAKE_CASE = max(*snake_case__ ,*snake_case__ ) _SCREAMING_SNAKE_CASE = int(math.pow(2 ,math.ceil(math.loga(snake_case__ ) ) ) ) _SCREAMING_SNAKE_CASE = matrixa _SCREAMING_SNAKE_CASE = matrixa # Adding zeros to the matrices so that the arrays dimensions are the same and also # power of 2 for i in range(0 ,snake_case__ ): if i < dimensiona[0]: for _ in range(dimensiona[1] ,snake_case__ ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) if i < dimensiona[0]: for _ in range(dimensiona[1] ,snake_case__ ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) _SCREAMING_SNAKE_CASE = actual_strassen(snake_case__ ,snake_case__ ) # Removing the additional zeros for i in range(0 ,snake_case__ ): if i < dimensiona[0]: for _ in range(dimensiona[1] ,snake_case__ ): final_matrix[i].pop() else: final_matrix.pop() return final_matrix if __name__ == "__main__": UpperCamelCase = [ [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 2, 3, 1], ] UpperCamelCase = [[0, 2, 1, 1], [16, 2, 3, 3], [2, 2, 7, 7], [13, 11, 22, 4]] print(strassen(matrixa, matrixa))
306
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: UpperCamelCase = None UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1_024, '''facebook/nllb-200-distilled-600M''': 1_024, } # fmt: off UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : List[str] = VOCAB_FILES_NAMES __snake_case : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Tuple = ["input_ids", "attention_mask"] __snake_case : Dict = NllbTokenizer __snake_case : List[int] = [] __snake_case : List[int] = [] def __init__( self: Tuple , UpperCAmelCase_: str=None , UpperCAmelCase_: List[str]=None , UpperCAmelCase_: Tuple="<s>" , UpperCAmelCase_: str="</s>" , UpperCAmelCase_: Union[str, Any]="</s>" , UpperCAmelCase_: int="<s>" , UpperCAmelCase_: Union[str, Any]="<unk>" , UpperCAmelCase_: Union[str, Any]="<pad>" , UpperCAmelCase_: str="<mask>" , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int=None , UpperCAmelCase_: str=False , **UpperCAmelCase_: int , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token _SCREAMING_SNAKE_CASE = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , src_lang=UpperCAmelCase_ , tgt_lang=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , legacy_behaviour=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = vocab_file _SCREAMING_SNAKE_CASE = False if not self.vocab_file else True _SCREAMING_SNAKE_CASE = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"""additional_special_tokens""": _additional_special_tokens} ) _SCREAMING_SNAKE_CASE = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase_ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } _SCREAMING_SNAKE_CASE = src_lang if src_lang is not None else """eng_Latn""" _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(self._src_lang ) _SCREAMING_SNAKE_CASE = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCamelCase ( self: int ): '''simple docstring''' return self._src_lang @src_lang.setter def UpperCamelCase ( self: int , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase ( self: List[str] , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase ( self: Tuple , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] , UpperCAmelCase_: Optional[str] , **UpperCAmelCase_: Any ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = self(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tgt_lang_id return inputs def UpperCamelCase ( self: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str = "eng_Latn" , UpperCAmelCase_: Optional[List[str]] = None , UpperCAmelCase_: str = "fra_Latn" , **UpperCAmelCase_: List[str] , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCAmelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory.' ) return _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) return (out_vocab_file,)
306
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : str = "dpr" def __init__( self: Tuple , UpperCAmelCase_: List[Any]=30_522 , UpperCAmelCase_: int=768 , UpperCAmelCase_: Dict=12 , UpperCAmelCase_: List[str]=12 , UpperCAmelCase_: int=3_072 , UpperCAmelCase_: Union[str, Any]="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: Union[str, Any]=0.1 , UpperCAmelCase_: Tuple=512 , UpperCAmelCase_: List[str]=2 , UpperCAmelCase_: str=0.02 , UpperCAmelCase_: int=1E-12 , UpperCAmelCase_: List[str]=0 , UpperCAmelCase_: Tuple="absolute" , UpperCAmelCase_: int = 0 , **UpperCAmelCase_: Union[str, Any] , ): '''simple docstring''' super().__init__(pad_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = layer_norm_eps _SCREAMING_SNAKE_CASE = projection_dim _SCREAMING_SNAKE_CASE = position_embedding_type
306
def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> list: """simple docstring""" _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = [[0] * n for i in range(snake_case__ )] for i in range(snake_case__ ): _SCREAMING_SNAKE_CASE = y_points[i] for i in range(2 ,snake_case__ ): for j in range(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
306
1
from math import factorial, pi def __lowerCamelCase ( snake_case__ ,snake_case__ = 30 ) -> float: """simple docstring""" if not isinstance(snake_case__ ,(int, float) ): raise ValueError("""maclaurin_sin() requires either an int or float for theta""" ) if not isinstance(snake_case__ ,snake_case__ ) or accuracy <= 0: raise ValueError("""maclaurin_sin() requires a positive int for accuracy""" ) _SCREAMING_SNAKE_CASE = float(snake_case__ ) _SCREAMING_SNAKE_CASE = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(snake_case__ ) ) def __lowerCamelCase ( snake_case__ ,snake_case__ = 30 ) -> float: """simple docstring""" if not isinstance(snake_case__ ,(int, float) ): raise ValueError("""maclaurin_cos() requires either an int or float for theta""" ) if not isinstance(snake_case__ ,snake_case__ ) or accuracy <= 0: raise ValueError("""maclaurin_cos() requires a positive int for accuracy""" ) _SCREAMING_SNAKE_CASE = float(snake_case__ ) _SCREAMING_SNAKE_CASE = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(snake_case__ ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(10)) print(maclaurin_sin(-10)) print(maclaurin_sin(10, 15)) print(maclaurin_sin(-10, 15)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(10, 15)) print(maclaurin_cos(-10, 15))
306
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCamelCase = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
1
import flax.linen as nn import jax import jax.numpy as jnp class __UpperCAmelCase (nn.Module ): __snake_case : int __snake_case : jnp.dtype = jnp.floataa def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self: List[str] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = hidden_states.shape _SCREAMING_SNAKE_CASE = jax.image.resize( UpperCAmelCase_ , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) _SCREAMING_SNAKE_CASE = self.conv(UpperCAmelCase_ ) return hidden_states class __UpperCAmelCase (nn.Module ): __snake_case : int __snake_case : jnp.dtype = jnp.floataa def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self: Optional[int] , UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.conv(UpperCAmelCase_ ) return hidden_states class __UpperCAmelCase (nn.Module ): __snake_case : int __snake_case : int = None __snake_case : float = 0.0 __snake_case : bool = None __snake_case : jnp.dtype = jnp.floataa def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.in_channels if self.out_channels is None else self.out_channels _SCREAMING_SNAKE_CASE = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) _SCREAMING_SNAKE_CASE = nn.Conv( UpperCAmelCase_ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _SCREAMING_SNAKE_CASE = nn.Dense(UpperCAmelCase_ , dtype=self.dtype ) _SCREAMING_SNAKE_CASE = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) _SCREAMING_SNAKE_CASE = nn.Dropout(self.dropout_prob ) _SCREAMING_SNAKE_CASE = nn.Conv( UpperCAmelCase_ , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _SCREAMING_SNAKE_CASE = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut _SCREAMING_SNAKE_CASE = None if use_nin_shortcut: _SCREAMING_SNAKE_CASE = nn.Conv( UpperCAmelCase_ , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self: List[Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any]=True ): '''simple docstring''' _SCREAMING_SNAKE_CASE = hidden_states _SCREAMING_SNAKE_CASE = self.norma(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.swish(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.conva(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.time_emb_proj(nn.swish(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = jnp.expand_dims(jnp.expand_dims(UpperCAmelCase_ , 1 ) , 1 ) _SCREAMING_SNAKE_CASE = hidden_states + temb _SCREAMING_SNAKE_CASE = self.norma(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.swish(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.dropout(UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.conva(UpperCAmelCase_ ) if self.conv_shortcut is not None: _SCREAMING_SNAKE_CASE = self.conv_shortcut(UpperCAmelCase_ ) return hidden_states + residual
306
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class __UpperCAmelCase : __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : int __snake_case : int __snake_case : float __snake_case : float __snake_case : Tuple[int] def UpperCamelCase ( self: str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = torch.arange(self.height * self.width ) _SCREAMING_SNAKE_CASE = torch.stack( [ pixel_indices % self.width, torch.div(UpperCAmelCase_ , self.width , rounding_mode="""trunc""" ), ] , axis=1 , ) return coords @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE = self.shape _SCREAMING_SNAKE_CASE = int(np.prod(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = self.get_image_coords() _SCREAMING_SNAKE_CASE = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) _SCREAMING_SNAKE_CASE = self.get_camera_rays(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = rays.view(UpperCAmelCase_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCamelCase ( self: Any , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] _SCREAMING_SNAKE_CASE = coords.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = self.resolution() _SCREAMING_SNAKE_CASE = self.fov() _SCREAMING_SNAKE_CASE = (flat.float() / (res - 1)) * 2 - 1 _SCREAMING_SNAKE_CASE = fracs * torch.tan(fov / 2 ) _SCREAMING_SNAKE_CASE = fracs.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = ( self.z.view(UpperCAmelCase_ , 1 , 3 ) + self.x.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, 1:] ) _SCREAMING_SNAKE_CASE = directions / directions.norm(dim=-1 , keepdim=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.stack( [ torch.broadcast_to(self.origin.view(UpperCAmelCase_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(UpperCAmelCase_ , *UpperCAmelCase_ , 2 , 3 ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=UpperCAmelCase_ , height=UpperCAmelCase_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __lowerCamelCase ( snake_case__ ) -> DifferentiableProjectiveCamera: """simple docstring""" _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for theta in np.linspace(0 ,2 * np.pi ,num=20 ): _SCREAMING_SNAKE_CASE = np.array([np.sin(snake_case__ ), np.cos(snake_case__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) _SCREAMING_SNAKE_CASE = -z * 4 _SCREAMING_SNAKE_CASE = np.array([np.cos(snake_case__ ), -np.sin(snake_case__ ), 0.0] ) _SCREAMING_SNAKE_CASE = np.cross(snake_case__ ,snake_case__ ) origins.append(snake_case__ ) xs.append(snake_case__ ) ys.append(snake_case__ ) zs.append(snake_case__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,x=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,y=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,z=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,width=snake_case__ ,height=snake_case__ ,x_fov=0.7 ,y_fov=0.7 ,shape=(1, len(snake_case__ )) ,)
306
1
from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ): __snake_case : int = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] = None , UpperCAmelCase_: int = 50_257 , UpperCAmelCase_: int = 1_024 , UpperCAmelCase_: int = 768 , UpperCAmelCase_: int = 12 , UpperCAmelCase_: int = 12 , UpperCAmelCase_: Optional[int] = None , UpperCAmelCase_: str = "gelu_new" , UpperCAmelCase_: float = 0.1 , UpperCAmelCase_: float = 0.1 , UpperCAmelCase_: float = 0.1 , UpperCAmelCase_: float = 1E-5 , UpperCAmelCase_: float = 0.02 , UpperCAmelCase_: bool = True , UpperCAmelCase_: bool = True , UpperCAmelCase_: bool = False , UpperCAmelCase_: bool = False , ): '''simple docstring''' super().__init__() _SCREAMING_SNAKE_CASE = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and' F' `n_embd`: {n_embd} are not equal.' ) _SCREAMING_SNAKE_CASE = prefix_inner_dim _SCREAMING_SNAKE_CASE = prefix_hidden_dim _SCREAMING_SNAKE_CASE = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _SCREAMING_SNAKE_CASE = ( nn.Linear(self.prefix_hidden_dim , UpperCAmelCase_ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _SCREAMING_SNAKE_CASE = GPTaConfig( vocab_size=UpperCAmelCase_ , n_positions=UpperCAmelCase_ , n_embd=UpperCAmelCase_ , n_layer=UpperCAmelCase_ , n_head=UpperCAmelCase_ , n_inner=UpperCAmelCase_ , activation_function=UpperCAmelCase_ , resid_pdrop=UpperCAmelCase_ , embd_pdrop=UpperCAmelCase_ , attn_pdrop=UpperCAmelCase_ , layer_norm_epsilon=UpperCAmelCase_ , initializer_range=UpperCAmelCase_ , scale_attn_weights=UpperCAmelCase_ , use_cache=UpperCAmelCase_ , scale_attn_by_inverse_layer_idx=UpperCAmelCase_ , reorder_and_upcast_attn=UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = GPTaLMHeadModel(UpperCAmelCase_ ) def UpperCamelCase ( self: str , UpperCAmelCase_: torch.Tensor , UpperCAmelCase_: torch.Tensor , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[torch.Tensor] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.transformer.transformer.wte(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.encode_prefix(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.decode_prefix(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _SCREAMING_SNAKE_CASE = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _SCREAMING_SNAKE_CASE = torch.cat((dummy_token, input_ids) , dim=1 ) _SCREAMING_SNAKE_CASE = self.transformer(inputs_embeds=UpperCAmelCase_ , labels=UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def UpperCamelCase ( self: int , UpperCAmelCase_: int , UpperCAmelCase_: torch.device ): '''simple docstring''' return torch.zeros(UpperCAmelCase_ , self.prefix_length , dtype=torch.intaa , device=UpperCAmelCase_ ) def UpperCamelCase ( self: int , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' return self.encode_prefix(UpperCAmelCase_ ) @torch.no_grad() def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = torch.split(UpperCAmelCase_ , 1 , dim=0 ) _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for feature in features: _SCREAMING_SNAKE_CASE = self.decode_prefix(feature.to(UpperCAmelCase_ ) ) # back to the clip feature # Only support beam search for now _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.generate_beam( input_embeds=UpperCAmelCase_ , device=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _SCREAMING_SNAKE_CASE = torch.stack(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.stack(UpperCAmelCase_ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def UpperCamelCase ( self: int , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: List[Any]=None , UpperCAmelCase_: int = 5 , UpperCAmelCase_: int = 67 , UpperCAmelCase_: float = 1.0 , UpperCAmelCase_: Optional[int] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = eos_token_id _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = torch.ones(UpperCAmelCase_ , device=UpperCAmelCase_ , dtype=torch.int ) _SCREAMING_SNAKE_CASE = torch.zeros(UpperCAmelCase_ , device=UpperCAmelCase_ , dtype=torch.bool ) if input_embeds is not None: _SCREAMING_SNAKE_CASE = input_embeds else: _SCREAMING_SNAKE_CASE = self.transformer.transformer.wte(UpperCAmelCase_ ) for i in range(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = self.transformer(inputs_embeds=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.logits _SCREAMING_SNAKE_CASE = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _SCREAMING_SNAKE_CASE = logits.softmax(-1 ).log() if scores is None: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = logits.topk(UpperCAmelCase_ , -1 ) _SCREAMING_SNAKE_CASE = generated.expand(UpperCAmelCase_ , *generated.shape[1:] ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _SCREAMING_SNAKE_CASE = next_tokens else: _SCREAMING_SNAKE_CASE = tokens.expand(UpperCAmelCase_ , *tokens.shape[1:] ) _SCREAMING_SNAKE_CASE = torch.cat((tokens, next_tokens) , dim=1 ) else: _SCREAMING_SNAKE_CASE = -float(np.inf ) _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _SCREAMING_SNAKE_CASE = scores_sum / seq_lengths[:, None] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = scores_sum_average.view(-1 ).topk(UpperCAmelCase_ , -1 ) _SCREAMING_SNAKE_CASE = next_tokens // scores_sum.shape[1] _SCREAMING_SNAKE_CASE = seq_lengths[next_tokens_source] _SCREAMING_SNAKE_CASE = next_tokens % scores_sum.shape[1] _SCREAMING_SNAKE_CASE = next_tokens.unsqueeze(1 ) _SCREAMING_SNAKE_CASE = tokens[next_tokens_source] _SCREAMING_SNAKE_CASE = torch.cat((tokens, next_tokens) , dim=1 ) _SCREAMING_SNAKE_CASE = generated[next_tokens_source] _SCREAMING_SNAKE_CASE = scores_sum_average * seq_lengths _SCREAMING_SNAKE_CASE = is_stopped[next_tokens_source] _SCREAMING_SNAKE_CASE = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _SCREAMING_SNAKE_CASE = torch.cat((generated, next_token_embed) , dim=1 ) _SCREAMING_SNAKE_CASE = is_stopped + next_tokens.eq(UpperCAmelCase_ ).squeeze() if is_stopped.all(): break _SCREAMING_SNAKE_CASE = scores / seq_lengths _SCREAMING_SNAKE_CASE = scores.argsort(descending=UpperCAmelCase_ ) # tokens tensors are already padded to max_seq_length _SCREAMING_SNAKE_CASE = [tokens[i] for i in order] _SCREAMING_SNAKE_CASE = torch.stack(UpperCAmelCase_ , dim=0 ) _SCREAMING_SNAKE_CASE = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
306
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __UpperCAmelCase (unittest.TestCase ): def __init__( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any]=13 , UpperCAmelCase_: List[str]=7 , UpperCAmelCase_: Tuple=True , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: str=99 , UpperCAmelCase_: List[Any]=32 , UpperCAmelCase_: Dict=5 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: Optional[int]="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: List[Any]=512 , UpperCAmelCase_: Any=16 , UpperCAmelCase_: Dict=2 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Union[str, Any]=4 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_attention_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_choices def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_attention_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=UpperCAmelCase_ , ) return config, input_ids, attention_mask def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Optional[int] = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase ( self: List[str] ): '''simple docstring''' for model_class_name in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCAmelCase_ ) @require_flax class __UpperCAmelCase (unittest.TestCase ): @slow def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 ) )
306
1
from PIL import Image def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Image: """simple docstring""" def brightness(snake_case__ ) -> float: return 1_28 + level + (c - 1_28) if not -255.0 <= level <= 255.0: raise ValueError("""level must be between -255.0 (black) and 255.0 (white)""" ) return img.point(snake_case__ ) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save('''image_data/lena_brightness.png''', format='''png''')
306
import argparse import hashlib # hashlib is only used inside the Test class import struct class __UpperCAmelCase : def __init__( self: List[str] , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = data _SCREAMING_SNAKE_CASE = [0x67_452_301, 0xef_cda_b89, 0x98_bad_cfe, 0x10_325_476, 0xc3_d2e_1f0] @staticmethod def UpperCamelCase ( UpperCAmelCase_: int , UpperCAmelCase_: List[str] ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0xff_fff_fff def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _SCREAMING_SNAKE_CASE = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = list(struct.unpack(""">16L""" , UpperCAmelCase_ ) ) + [0] * 64 for i in range(16 , 80 ): _SCREAMING_SNAKE_CASE = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.padding() _SCREAMING_SNAKE_CASE = self.split_blocks() for block in self.blocks: _SCREAMING_SNAKE_CASE = self.expand_block(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.h for i in range(0 , 80 ): if 0 <= i < 20: _SCREAMING_SNAKE_CASE = (b & c) | ((~b) & d) _SCREAMING_SNAKE_CASE = 0x5a_827_999 elif 20 <= i < 40: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0x6e_d9e_ba1 elif 40 <= i < 60: _SCREAMING_SNAKE_CASE = (b & c) | (b & d) | (c & d) _SCREAMING_SNAKE_CASE = 0x8f_1bb_cdc elif 60 <= i < 80: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0xca_62c_1d6 _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ( self.rotate(UpperCAmelCase_ , 5 ) + f + e + k + expanded_block[i] & 0xff_fff_fff, a, self.rotate(UpperCAmelCase_ , 30 ), c, d, ) _SCREAMING_SNAKE_CASE = ( self.h[0] + a & 0xff_fff_fff, self.h[1] + b & 0xff_fff_fff, self.h[2] + c & 0xff_fff_fff, self.h[3] + d & 0xff_fff_fff, self.h[4] + e & 0xff_fff_fff, ) return ("{:08x}" * 5).format(*self.h ) def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = b"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def __lowerCamelCase ( ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" ,dest="""input_string""" ,default="""Hello World!! Welcome to Cryptography""" ,help="""Hash the string""" ,) parser.add_argument("""--file""" ,dest="""input_file""" ,help="""Hash contents of a file""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = f.read() else: _SCREAMING_SNAKE_CASE = bytes(snake_case__ ,"""utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
306
1
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __UpperCAmelCase : def __init__( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int]=13 , UpperCAmelCase_: str=7 , UpperCAmelCase_: int=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Dict=True , UpperCAmelCase_: Any=True , UpperCAmelCase_: Tuple=99 , UpperCAmelCase_: Optional[Any]=32 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Tuple=37 , UpperCAmelCase_: Union[str, Any]="gelu" , UpperCAmelCase_: List[str]=0.1 , UpperCAmelCase_: int=0.1 , UpperCAmelCase_: str=512 , UpperCAmelCase_: Union[str, Any]=16 , UpperCAmelCase_: List[Any]=2 , UpperCAmelCase_: str=0.02 , UpperCAmelCase_: int=False , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: Optional[Any]="None" , UpperCAmelCase_: Optional[int]=3 , UpperCAmelCase_: Any=4 , UpperCAmelCase_: Optional[int]=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = relative_attention _SCREAMING_SNAKE_CASE = position_biased_input _SCREAMING_SNAKE_CASE = pos_att_type _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=UpperCAmelCase_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _SCREAMING_SNAKE_CASE = [input_ids, input_mask] _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForMaskedLM(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Any , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForSequenceClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForTokenClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: Any , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForQuestionAnswering(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) __snake_case : Union[str, Any] = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) __snake_case : Dict = False __snake_case : Optional[Any] = False def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(UpperCAmelCase_ ) @require_tf class __UpperCAmelCase (unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = tf.constant( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 )
306
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Tuple = VOCAB_FILES_NAMES __snake_case : Optional[int] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Optional[Any] = ["input_ids", "attention_mask"] __snake_case : Optional[int] = None def __init__( self: Dict , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int="<unk>" , UpperCAmelCase_: List[str]="<s>" , UpperCAmelCase_: Tuple="</s>" , UpperCAmelCase_: List[Any]="<pad>" , UpperCAmelCase_: Dict=False , UpperCAmelCase_: Dict=False , **UpperCAmelCase_: Dict , ): '''simple docstring''' super().__init__( UpperCAmelCase_ , UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , add_prefix_space=UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCAmelCase_ ) != add_prefix_space: _SCREAMING_SNAKE_CASE = getattr(UpperCAmelCase_ , pre_tok_state.pop("""type""" ) ) _SCREAMING_SNAKE_CASE = add_prefix_space _SCREAMING_SNAKE_CASE = pre_tok_class(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = add_prefix_space def UpperCamelCase ( self: List[str] , *UpperCAmelCase_: Any , **UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Union[str, Any] , *UpperCAmelCase_: Dict , **UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_ ) return tuple(UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: "Conversation" ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) + [self.eos_token_id] ) if len(UpperCAmelCase_ ) > self.model_max_length: _SCREAMING_SNAKE_CASE = input_ids[-self.model_max_length :] return input_ids
306
1
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int | float: """simple docstring""" if len(snake_case__ ) == 0: raise ValueError("""find_max() arg is an empty sequence""" ) if ( left >= len(snake_case__ ) or left < -len(snake_case__ ) or right >= len(snake_case__ ) or right < -len(snake_case__ ) ): raise IndexError("""list index out of range""" ) if left == right: return nums[left] _SCREAMING_SNAKE_CASE = (left + right) >> 1 # the middle _SCREAMING_SNAKE_CASE = find_max(snake_case__ ,snake_case__ ,snake_case__ ) # find max in range[left, mid] _SCREAMING_SNAKE_CASE = find_max(snake_case__ ,mid + 1 ,snake_case__ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
306
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __UpperCAmelCase : def __init__( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int]=13 , UpperCAmelCase_: str=7 , UpperCAmelCase_: int=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Dict=True , UpperCAmelCase_: Any=True , UpperCAmelCase_: Tuple=99 , UpperCAmelCase_: Optional[Any]=32 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Tuple=37 , UpperCAmelCase_: Union[str, Any]="gelu" , UpperCAmelCase_: List[str]=0.1 , UpperCAmelCase_: int=0.1 , UpperCAmelCase_: str=512 , UpperCAmelCase_: Union[str, Any]=16 , UpperCAmelCase_: List[Any]=2 , UpperCAmelCase_: str=0.02 , UpperCAmelCase_: int=False , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: Optional[Any]="None" , UpperCAmelCase_: Optional[int]=3 , UpperCAmelCase_: Any=4 , UpperCAmelCase_: Optional[int]=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = relative_attention _SCREAMING_SNAKE_CASE = position_biased_input _SCREAMING_SNAKE_CASE = pos_att_type _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=UpperCAmelCase_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _SCREAMING_SNAKE_CASE = [input_ids, input_mask] _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForMaskedLM(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Any , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForSequenceClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForTokenClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: Any , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForQuestionAnswering(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) __snake_case : Union[str, Any] = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) __snake_case : Dict = False __snake_case : Optional[Any] = False def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(UpperCAmelCase_ ) @require_tf class __UpperCAmelCase (unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = tf.constant( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 )
306
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCamelCase = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __lowerCamelCase ( snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = gather(snake_case__ ) assert gathered_tensor.tolist() == list(range(1 ,state.num_processes**2 + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = [state.process_index] _SCREAMING_SNAKE_CASE = gather_object(snake_case__ ) assert len(snake_case__ ) == state.num_processes, F'{gathered_obj}, {len(snake_case__ )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = broadcast(snake_case__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 ,state.num_processes + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Tuple: """simple docstring""" if state.is_main_process: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes + 1 ).to(state.device ) else: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes ).to(state.device ) _SCREAMING_SNAKE_CASE = pad_across_processes(snake_case__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 ,state.num_processes ) ) + [0] def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""sum""" ) _SCREAMING_SNAKE_CASE = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""mean""" ) _SCREAMING_SNAKE_CASE = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> str: """simple docstring""" main() def __lowerCamelCase ( ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = PartialState() state.print(F'State: {state}' ) state.print("""testing gather""" ) test_gather(snake_case__ ) state.print("""testing gather_object""" ) test_gather_object(snake_case__ ) state.print("""testing broadcast""" ) test_broadcast(snake_case__ ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(snake_case__ ) state.print("""testing reduce_sum""" ) test_reduce_sum(snake_case__ ) state.print("""testing reduce_mean""" ) test_reduce_mean(snake_case__ ) if __name__ == "__main__": main()
306
1
def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> float: """simple docstring""" _SCREAMING_SNAKE_CASE = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def __lowerCamelCase ( ) -> str: """simple docstring""" print(sum_of_series(1 ,1 ,10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
306
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __lowerCamelCase ( ) -> tuple[list[int], int]: """simple docstring""" _SCREAMING_SNAKE_CASE = [randint(-10_00 ,10_00 ) for i in range(10 )] _SCREAMING_SNAKE_CASE = randint(-50_00 ,50_00 ) return (arr, r) UpperCamelCase = make_dataset() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(snake_case__ ,3 ): if sum(snake_case__ ) == target: return tuple(sorted(snake_case__ ) ) return (0, 0, 0) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, int, int]: """simple docstring""" arr.sort() _SCREAMING_SNAKE_CASE = len(snake_case__ ) for i in range(n - 1 ): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __lowerCamelCase ( ) -> tuple[float, float]: """simple docstring""" _SCREAMING_SNAKE_CASE = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _SCREAMING_SNAKE_CASE = """ triplet_sum1(*dataset) """ _SCREAMING_SNAKE_CASE = """ triplet_sum2(*dataset) """ _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) return (min(snake_case__ ), min(snake_case__ )) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
306
1
def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def __lowerCamelCase ( snake_case__ = 1_00 ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = 2 for i in range(2 ,max_n + 1 ): _SCREAMING_SNAKE_CASE = pre_numerator _SCREAMING_SNAKE_CASE = 2 * i // 3 if i % 3 == 0 else 1 _SCREAMING_SNAKE_CASE = cur_numerator _SCREAMING_SNAKE_CASE = e_cont * pre_numerator + temp return sum_digits(snake_case__ ) if __name__ == "__main__": print(f"{solution() = }")
306
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase ) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) self.check_model_type(UpperCAmelCase_ ) def UpperCamelCase ( self: str , **UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} # preprocess args if "points_per_batch" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self: Optional[Any] , UpperCAmelCase_: Tuple , *UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=None , UpperCAmelCase_: Tuple=None , **UpperCAmelCase_: Any ): '''simple docstring''' return super().__call__(UpperCAmelCase_ , *UpperCAmelCase_ , num_workers=UpperCAmelCase_ , batch_size=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict=64 , UpperCAmelCase_: int = 0 , UpperCAmelCase_: float = 512 / 1_500 , UpperCAmelCase_: Optional[int] = 32 , UpperCAmelCase_: Optional[int] = 1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor.size["""longest_edge"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.generate_crop_boxes( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": _SCREAMING_SNAKE_CASE = self.get_inference_context() with inference_context(): _SCREAMING_SNAKE_CASE = self._ensure_tensor_on_device(UpperCAmelCase_ , device=self.device ) _SCREAMING_SNAKE_CASE = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) _SCREAMING_SNAKE_CASE = image_embeddings _SCREAMING_SNAKE_CASE = grid_points.shape[1] _SCREAMING_SNAKE_CASE = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = grid_points[:, i : i + points_per_batch, :, :] _SCREAMING_SNAKE_CASE = input_labels[:, i : i + points_per_batch] _SCREAMING_SNAKE_CASE = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=0.88 , UpperCAmelCase_: Dict=0.95 , UpperCAmelCase_: Tuple=0 , UpperCAmelCase_: str=1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_inputs.pop("""input_boxes""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""is_last""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""original_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = model_inputs.pop("""reshaped_input_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks _SCREAMING_SNAKE_CASE = model_outputs["""pred_masks"""] _SCREAMING_SNAKE_CASE = self.image_processor.post_process_masks( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , binarize=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model_outputs["""iou_scores"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCamelCase ( self: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=False , UpperCAmelCase_: Any=0.7 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.post_process_for_mask_generation( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = defaultdict(UpperCAmelCase_ ) for output in model_outputs: for k, v in output.items(): extra[k].append(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {} if output_rle_mask: _SCREAMING_SNAKE_CASE = rle_mask if output_bboxes_mask: _SCREAMING_SNAKE_CASE = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
306
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: UpperCamelCase = None UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = '''▁''' UpperCamelCase = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''}, '''tokenizer_file''': { '''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json''' }, } UpperCamelCase = { '''google/pegasus-xsum''': 512, } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Union[str, Any] = VOCAB_FILES_NAMES __snake_case : Any = PRETRAINED_VOCAB_FILES_MAP __snake_case : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case : str = PegasusTokenizer __snake_case : Optional[Any] = ["input_ids", "attention_mask"] def __init__( self: str , UpperCAmelCase_: Any=None , UpperCAmelCase_: Any=None , UpperCAmelCase_: List[Any]="<pad>" , UpperCAmelCase_: str="</s>" , UpperCAmelCase_: Dict="<unk>" , UpperCAmelCase_: Any="<mask_2>" , UpperCAmelCase_: List[Any]="<mask_1>" , UpperCAmelCase_: Dict=None , UpperCAmelCase_: Any=103 , **UpperCAmelCase_: str , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = offset if additional_special_tokens is not None: if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): raise TypeError( F'additional_special_tokens should be of type {type(UpperCAmelCase_ )}, but is' F' {type(UpperCAmelCase_ )}' ) _SCREAMING_SNAKE_CASE = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'<unk_{i}>' for i in range(len(UpperCAmelCase_ ) , self.offset - 1 ) ] if len(set(UpperCAmelCase_ ) ) != len(UpperCAmelCase_ ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" F' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) _SCREAMING_SNAKE_CASE = additional_special_tokens_extended else: _SCREAMING_SNAKE_CASE = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'<unk_{i}>' for i in range(2 , self.offset )] super().__init__( UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , mask_token_sent=UpperCAmelCase_ , offset=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = vocab_file _SCREAMING_SNAKE_CASE = False if not self.vocab_file else True def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( """There should be 3 special tokens: mask_token, pad_token, and eos_token +""" F' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}' ) return [1 if x in all_special_ids else 0 for x in seq] def UpperCamelCase ( self: int , UpperCAmelCase_: List , UpperCAmelCase_: Optional[List] = None , UpperCAmelCase_: bool = False ): '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(UpperCAmelCase_ ) elif token_ids_a is None: return self._special_token_mask(UpperCAmelCase_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: List[Any]=None ): '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def UpperCamelCase ( self: Any , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCAmelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) return (out_vocab_file,)
306
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging UpperCamelCase = logging.get_logger(__name__) logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" if "xprophetnet" in prophetnet_checkpoint_path: _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) else: _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) _SCREAMING_SNAKE_CASE = ["""key_proj""", """value_proj""", """query_proj"""] _SCREAMING_SNAKE_CASE = { """self_attn""": """ngram_self_attn""", """cross_attn""": """encoder_attn""", """cross_attn_layer_norm""": """encoder_attn_layer_norm""", """feed_forward_layer_norm""": """final_layer_norm""", """feed_forward""": """""", """intermediate""": """fc1""", """output""": """fc2""", """key_proj""": """k_proj""", """query_proj""": """q_proj""", """value_proj""": """v_proj""", """word_embeddings""": """embed_tokens""", """embeddings_layer_norm""": """emb_layer_norm""", """relative_pos_embeddings""": """relative_linear""", """ngram_embeddings""": """ngram_input_embed""", """position_embeddings""": """embed_positions""", } for key in loading_info["missing_keys"]: _SCREAMING_SNAKE_CASE = key.split(""".""" ) if attributes[0] == "lm_head": _SCREAMING_SNAKE_CASE = prophet _SCREAMING_SNAKE_CASE = prophet_old else: _SCREAMING_SNAKE_CASE = prophet.prophetnet _SCREAMING_SNAKE_CASE = prophet_old.model _SCREAMING_SNAKE_CASE = False for attribute in attributes: if attribute in mapping: _SCREAMING_SNAKE_CASE = mapping[attribute] if not hasattr(snake_case__ ,snake_case__ ) and len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = attribute elif hasattr(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.weight logger.info(F'{attribute} is initialized.' ) _SCREAMING_SNAKE_CASE = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.bias logger.info(F'{attribute} is initialized' ) _SCREAMING_SNAKE_CASE = True break elif attribute in special_keys and hasattr(snake_case__ ,"""in_proj_weight""" ): _SCREAMING_SNAKE_CASE = old_model.in_proj_weight.shape[0] // 3 _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _SCREAMING_SNAKE_CASE = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_12, "We want 512 position_embeddings." _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.embed_positions.weight[:5_12, :] ) _SCREAMING_SNAKE_CASE = True break if attribute.isdigit(): _SCREAMING_SNAKE_CASE = model[int(snake_case__ )] _SCREAMING_SNAKE_CASE = old_model[int(snake_case__ )] else: _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if old_attribute == "": _SCREAMING_SNAKE_CASE = old_model else: if not hasattr(snake_case__ ,snake_case__ ): raise ValueError(F'{old_model} does not have {old_attribute}' ) _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if not is_key_init: raise ValueError(F'{key} was not correctly initialized!' ) print(F'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--prophetnet_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
306
1
import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format='''%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s''', datefmt='''%Y-%m-%d %H:%M:%S''', level=os.environ.get('''LOGLEVEL''', '''INFO''').upper(), stream=sys.stdout, ) UpperCamelCase = logging.getLogger(__name__) UpperCamelCase = {'''facebook/bart-base''': BartForConditionalGeneration} UpperCamelCase = {'''facebook/bart-base''': BartTokenizer} def __lowerCamelCase ( ) -> List[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Export Bart model + Beam Search to ONNX graph.""" ) parser.add_argument( """--validation_file""" ,type=snake_case__ ,default=snake_case__ ,help="""A csv or a json file containing the validation data.""" ) parser.add_argument( """--max_length""" ,type=snake_case__ ,default=5 ,help="""The maximum total input sequence length after tokenization.""" ,) parser.add_argument( """--num_beams""" ,type=snake_case__ ,default=snake_case__ ,help=( """Number of beams to use for evaluation. This argument will be """ """passed to ``model.generate``, which is used during ``evaluate`` and ``predict``.""" ) ,) parser.add_argument( """--model_name_or_path""" ,type=snake_case__ ,help="""Path to pretrained model or model identifier from huggingface.co/models.""" ,required=snake_case__ ,) parser.add_argument( """--config_name""" ,type=snake_case__ ,default=snake_case__ ,help="""Pretrained config name or path if not the same as model_name""" ,) parser.add_argument( """--device""" ,type=snake_case__ ,default="""cpu""" ,help="""Device where the model will be run""" ,) parser.add_argument("""--output_file_path""" ,type=snake_case__ ,default=snake_case__ ,help="""Where to store the final ONNX file.""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() return args def __lowerCamelCase ( snake_case__ ,snake_case__="cpu" ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = model_dict[model_name].from_pretrained(snake_case__ ).to(snake_case__ ) _SCREAMING_SNAKE_CASE = tokenizer_dict[model_name].from_pretrained(snake_case__ ) if model_name in ["facebook/bart-base"]: _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = 0 return huggingface_model, tokenizer def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> Any: """simple docstring""" model.eval() _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = torch.jit.script(BARTBeamSearchGenerator(snake_case__ ) ) with torch.no_grad(): _SCREAMING_SNAKE_CASE = """My friends are cool but they eat too many carbs.""" _SCREAMING_SNAKE_CASE = tokenizer([ARTICLE_TO_SUMMARIZE] ,max_length=10_24 ,return_tensors="""pt""" ).to(model.device ) _SCREAMING_SNAKE_CASE = model.generate( inputs["""input_ids"""] ,attention_mask=inputs["""attention_mask"""] ,num_beams=snake_case__ ,max_length=snake_case__ ,early_stopping=snake_case__ ,decoder_start_token_id=model.config.decoder_start_token_id ,) torch.onnx.export( snake_case__ ,( inputs["""input_ids"""], inputs["""attention_mask"""], num_beams, max_length, model.config.decoder_start_token_id, ) ,snake_case__ ,opset_version=14 ,input_names=["""input_ids""", """attention_mask""", """num_beams""", """max_length""", """decoder_start_token_id"""] ,output_names=["""output_ids"""] ,dynamic_axes={ """input_ids""": {0: """batch""", 1: """seq"""}, """output_ids""": {0: """batch""", 1: """seq_out"""}, } ,example_outputs=snake_case__ ,) logger.info("""Model exported to {}""".format(snake_case__ ) ) _SCREAMING_SNAKE_CASE = remove_dup_initializers(os.path.abspath(snake_case__ ) ) logger.info("""Deduplicated and optimized model written to {}""".format(snake_case__ ) ) _SCREAMING_SNAKE_CASE = onnxruntime.InferenceSession(snake_case__ ) _SCREAMING_SNAKE_CASE = ort_sess.run( snake_case__ ,{ """input_ids""": inputs["""input_ids"""].cpu().numpy(), """attention_mask""": inputs["""attention_mask"""].cpu().numpy(), """num_beams""": np.array(snake_case__ ), """max_length""": np.array(snake_case__ ), """decoder_start_token_id""": np.array(model.config.decoder_start_token_id ), } ,) np.testing.assert_allclose(summary_ids.cpu().numpy() ,ort_out[0] ,rtol=1e-3 ,atol=1e-3 ) logger.info("""Model outputs from torch and ONNX Runtime are similar.""" ) logger.info("""Success.""" ) def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = parse_args() _SCREAMING_SNAKE_CASE = 5 _SCREAMING_SNAKE_CASE = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" ,datefmt="""%m/%d/%Y %H:%M:%S""" ,level=logging.INFO ,) logger.setLevel(logging.INFO ) transformers.utils.logging.set_verbosity_error() _SCREAMING_SNAKE_CASE = torch.device(args.device ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = load_model_tokenizer(args.model_name_or_path ,snake_case__ ) if model.config.decoder_start_token_id is None: raise ValueError("""Make sure that `config.decoder_start_token_id` is correctly defined""" ) model.to(snake_case__ ) if args.max_length: _SCREAMING_SNAKE_CASE = args.max_length if args.num_beams: _SCREAMING_SNAKE_CASE = args.num_beams if args.output_file_path: _SCREAMING_SNAKE_CASE = args.output_file_path else: _SCREAMING_SNAKE_CASE = """BART.onnx""" logger.info("""Exporting model to ONNX""" ) export_and_validate_model(snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) if __name__ == "__main__": main()
306
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: _SCREAMING_SNAKE_CASE = i + 1 else: _SCREAMING_SNAKE_CASE = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"{two_pointer([2, 7, 11, 15], 9) = }")
306
1
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar UpperCamelCase = TypeVar('''T''') def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" return (position - 1) // 2 def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" return (2 * position) + 1 def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" return (2 * position) + 2 class __UpperCAmelCase (Generic[T] ): def __init__( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = 0 def __len__( self: List[Any] ): '''simple docstring''' return self.elements def __repr__( self: List[Any] ): '''simple docstring''' return str(self.heap ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return self.elements == 0 def UpperCamelCase ( self: Any , UpperCAmelCase_: T , UpperCAmelCase_: int ): '''simple docstring''' self.heap.append((elem, weight) ) _SCREAMING_SNAKE_CASE = self.elements self.elements += 1 self._bubble_up(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[0] self._bubble_down(UpperCAmelCase_ ) return elem def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: T , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.position_map[elem] _SCREAMING_SNAKE_CASE = (elem, weight) if position > 0: _SCREAMING_SNAKE_CASE = get_parent_position(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[parent_position] if parent_weight > weight: self._bubble_up(UpperCAmelCase_ ) else: self._bubble_down(UpperCAmelCase_ ) else: self._bubble_down(UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: T ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.position_map[elem] if curr_pos == 0: return None _SCREAMING_SNAKE_CASE = get_parent_position(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[curr_pos] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[parent_position] if parent_weight > weight: self._swap_nodes(UpperCAmelCase_ , UpperCAmelCase_ ) return self._bubble_up(UpperCAmelCase_ ) return None def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: T ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.position_map[elem] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[curr_pos] _SCREAMING_SNAKE_CASE = get_child_left_position(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = get_child_right_position(UpperCAmelCase_ ) if child_left_position < self.elements and child_right_position < self.elements: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[child_left_position] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(UpperCAmelCase_ , UpperCAmelCase_ ) return self._bubble_down(UpperCAmelCase_ ) if child_left_position < self.elements: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(UpperCAmelCase_ , UpperCAmelCase_ ) return self._bubble_down(UpperCAmelCase_ ) else: return None if child_right_position < self.elements: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(UpperCAmelCase_ , UpperCAmelCase_ ) return self._bubble_down(UpperCAmelCase_ ) return None def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.heap[nodea_pos][0] _SCREAMING_SNAKE_CASE = self.heap[nodea_pos][0] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ( self.heap[nodea_pos], self.heap[nodea_pos], ) _SCREAMING_SNAKE_CASE = nodea_pos _SCREAMING_SNAKE_CASE = nodea_pos class __UpperCAmelCase (Generic[T] ): def __init__( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = 0 def __repr__( self: Union[str, Any] ): '''simple docstring''' return str(self.connections ) def __len__( self: Tuple ): '''simple docstring''' return self.nodes def UpperCamelCase ( self: Tuple , UpperCAmelCase_: T ): '''simple docstring''' if node not in self.connections: _SCREAMING_SNAKE_CASE = {} self.nodes += 1 def UpperCamelCase ( self: Any , UpperCAmelCase_: T , UpperCAmelCase_: T , UpperCAmelCase_: int ): '''simple docstring''' self.add_node(UpperCAmelCase_ ) self.add_node(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = weight _SCREAMING_SNAKE_CASE = weight def __lowerCamelCase ( snake_case__ ,) -> tuple[dict[T, int], dict[T, T | None]]: """simple docstring""" _SCREAMING_SNAKE_CASE = {node: maxsize for node in graph.connections} _SCREAMING_SNAKE_CASE = {node: None for node in graph.connections} _SCREAMING_SNAKE_CASE = MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(snake_case__ ,snake_case__ ) if priority_queue.is_empty(): return dist, parent # initialization _SCREAMING_SNAKE_CASE = priority_queue.extract_min() _SCREAMING_SNAKE_CASE = 0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: _SCREAMING_SNAKE_CASE = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(snake_case__ ,dist[neighbour] ) _SCREAMING_SNAKE_CASE = node # running prim's algorithm while not priority_queue.is_empty(): _SCREAMING_SNAKE_CASE = priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: _SCREAMING_SNAKE_CASE = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(snake_case__ ,dist[neighbour] ) _SCREAMING_SNAKE_CASE = node return dist, parent
306
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig UpperCamelCase = logging.get_logger(__name__) # General docstring UpperCamelCase = '''MobileNetV1Config''' # Base docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = [1, 1_024, 7, 7] # Image classification docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = '''tabby, tabby cat''' UpperCamelCase = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = {} if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = model.mobilenet_va else: _SCREAMING_SNAKE_CASE = model _SCREAMING_SNAKE_CASE = """MobilenetV1/Conv2d_0/""" _SCREAMING_SNAKE_CASE = backbone.conv_stem.convolution.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.bias _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_mean _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_var for i in range(13 ): _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = i * 2 _SCREAMING_SNAKE_CASE = backbone.layer[pt_index] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_depthwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var _SCREAMING_SNAKE_CASE = backbone.layer[pt_index + 1] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_pointwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = """MobilenetV1/Logits/Conv2d_1c_1x1/""" _SCREAMING_SNAKE_CASE = model.classifier.weight _SCREAMING_SNAKE_CASE = model.classifier.bias return tf_to_pt_map def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> List[str]: """simple docstring""" try: import numpy as np import tensorflow as tf except ImportError: logger.error( """Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see """ """https://www.tensorflow.org/install/ for installation instructions.""" ) raise # Load weights from TF model _SCREAMING_SNAKE_CASE = tf.train.list_variables(snake_case__ ) _SCREAMING_SNAKE_CASE = {} for name, shape in init_vars: logger.info(F'Loading TF weight {name} with shape {shape}' ) _SCREAMING_SNAKE_CASE = tf.train.load_variable(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = array # Build TF to PyTorch weights loading map _SCREAMING_SNAKE_CASE = _build_tf_to_pytorch_map(snake_case__ ,snake_case__ ,snake_case__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F'Importing {name}' ) if name not in tf_weights: logger.info(F'{name} not in tf pre-trained weights, skipping' ) continue _SCREAMING_SNAKE_CASE = tf_weights[name] if "depthwise_weights" in name: logger.info("""Transposing depthwise""" ) _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(2, 3, 0, 1) ) elif "weights" in name: logger.info("""Transposing""" ) if len(pointer.shape ) == 2: # copying into linear layer _SCREAMING_SNAKE_CASE = array.squeeze().transpose() else: _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'Pointer shape {pointer.shape} and array shape {array.shape} mismatched' ) logger.info(F'Initialize PyTorch weight {name} {array.shape}' ) _SCREAMING_SNAKE_CASE = torch.from_numpy(snake_case__ ) tf_weights.pop(snake_case__ ,snake_case__ ) tf_weights.pop(name + """/RMSProp""" ,snake_case__ ) tf_weights.pop(name + """/RMSProp_1""" ,snake_case__ ) tf_weights.pop(name + """/ExponentialMovingAverage""" ,snake_case__ ) logger.info(F'Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}' ) return model def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> torch.Tensor: """simple docstring""" _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = features.shape[-2:] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.stride _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.kernel_size if in_height % stride_height == 0: _SCREAMING_SNAKE_CASE = max(kernel_height - stride_height ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_height - (in_height % stride_height) ,0 ) if in_width % stride_width == 0: _SCREAMING_SNAKE_CASE = max(kernel_width - stride_width ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_width - (in_width % stride_width) ,0 ) _SCREAMING_SNAKE_CASE = pad_along_width // 2 _SCREAMING_SNAKE_CASE = pad_along_width - pad_left _SCREAMING_SNAKE_CASE = pad_along_height // 2 _SCREAMING_SNAKE_CASE = pad_along_height - pad_top _SCREAMING_SNAKE_CASE = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(snake_case__ ,snake_case__ ,"""constant""" ,0.0 ) class __UpperCAmelCase (nn.Module ): def __init__( self: Optional[Any] , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: bool = False , UpperCAmelCase_: Optional[bool] = True , UpperCAmelCase_: Optional[bool or str] = True , ): '''simple docstring''' super().__init__() _SCREAMING_SNAKE_CASE = config if in_channels % groups != 0: raise ValueError(F'Input channels ({in_channels}) are not divisible by {groups} groups.' ) if out_channels % groups != 0: raise ValueError(F'Output channels ({out_channels}) are not divisible by {groups} groups.' ) _SCREAMING_SNAKE_CASE = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) _SCREAMING_SNAKE_CASE = nn.Convad( in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=UpperCAmelCase_ , groups=UpperCAmelCase_ , bias=UpperCAmelCase_ , padding_mode="""zeros""" , ) if use_normalization: _SCREAMING_SNAKE_CASE = nn.BatchNormad( num_features=UpperCAmelCase_ , eps=config.layer_norm_eps , momentum=0.99_97 , affine=UpperCAmelCase_ , track_running_stats=UpperCAmelCase_ , ) else: _SCREAMING_SNAKE_CASE = None if use_activation: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[use_activation] elif isinstance(config.hidden_act , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: _SCREAMING_SNAKE_CASE = config.hidden_act else: _SCREAMING_SNAKE_CASE = None def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' if self.config.tf_padding: _SCREAMING_SNAKE_CASE = apply_tf_padding(UpperCAmelCase_ , self.convolution ) _SCREAMING_SNAKE_CASE = self.convolution(UpperCAmelCase_ ) if self.normalization is not None: _SCREAMING_SNAKE_CASE = self.normalization(UpperCAmelCase_ ) if self.activation is not None: _SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase_ ) return features class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Dict = MobileNetVaConfig __snake_case : Any = load_tf_weights_in_mobilenet_va __snake_case : Any = "mobilenet_v1" __snake_case : List[Any] = "pixel_values" __snake_case : Any = False def UpperCamelCase ( self: str , UpperCAmelCase_: Union[nn.Linear, nn.Convad] ): '''simple docstring''' if isinstance(UpperCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCAmelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) UpperCamelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' UpperCamelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: bool = True ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config _SCREAMING_SNAKE_CASE = 32 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) _SCREAMING_SNAKE_CASE = MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=config.num_channels , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=2 , ) _SCREAMING_SNAKE_CASE = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] _SCREAMING_SNAKE_CASE = nn.ModuleList() for i in range(13 ): _SCREAMING_SNAKE_CASE = out_channels if strides[i] == 2 or i == 0: depth *= 2 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=strides[i] , groups=UpperCAmelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=1 , ) ) _SCREAMING_SNAKE_CASE = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCamelCase ( self: Dict , UpperCAmelCase_: Tuple ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase ( self: int , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) _SCREAMING_SNAKE_CASE = self.conv_stem(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): _SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase_ ) if output_hidden_states: _SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) _SCREAMING_SNAKE_CASE = hidden_states if self.pooler is not None: _SCREAMING_SNAKE_CASE = torch.flatten(self.pooler(UpperCAmelCase_ ) , start_dim=1 ) else: _SCREAMING_SNAKE_CASE = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase_ , pooler_output=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Dict , UpperCAmelCase_: MobileNetVaConfig ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config.num_labels _SCREAMING_SNAKE_CASE = MobileNetVaModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head _SCREAMING_SNAKE_CASE = nn.Dropout(config.classifier_dropout_prob , inplace=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.Linear(UpperCAmelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict _SCREAMING_SNAKE_CASE = self.mobilenet_va(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.pooler_output if return_dict else outputs[1] _SCREAMING_SNAKE_CASE = self.classifier(self.dropout(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _SCREAMING_SNAKE_CASE = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _SCREAMING_SNAKE_CASE = """single_label_classification""" else: _SCREAMING_SNAKE_CASE = """multi_label_classification""" if self.config.problem_type == "regression": _SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: _SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) elif self.config.problem_type == "single_label_classification": _SCREAMING_SNAKE_CASE = CrossEntropyLoss() _SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: _SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states , )
306
1
import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase = '''0.12''' # assumed parallelism: 8 @require_flax @is_staging_test class __UpperCAmelCase (unittest.TestCase ): @classmethod def UpperCamelCase ( cls: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TOKEN HfFolder.save_token(UpperCAmelCase_ ) @classmethod def UpperCamelCase ( cls: List[Any] ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id="""test-model-flax""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-model-flax-org""" ) except HTTPError: pass def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) _SCREAMING_SNAKE_CASE = FlaxBertModel(UpperCAmelCase_ ) model.push_to_hub("""test-model-flax""" , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(F'{USER}/test-model-flax' ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(model.params ) ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _SCREAMING_SNAKE_CASE = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(UpperCAmelCase_ , 1E-3 , msg=F'{key} not identical' ) # Reset repo delete_repo(token=self._token , repo_id="""test-model-flax""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(UpperCAmelCase_ , repo_id="""test-model-flax""" , push_to_hub=UpperCAmelCase_ , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(F'{USER}/test-model-flax' ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(model.params ) ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _SCREAMING_SNAKE_CASE = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(UpperCAmelCase_ , 1E-3 , msg=F'{key} not identical' ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) _SCREAMING_SNAKE_CASE = FlaxBertModel(UpperCAmelCase_ ) model.push_to_hub("""valid_org/test-model-flax-org""" , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained("""valid_org/test-model-flax-org""" ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(model.params ) ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _SCREAMING_SNAKE_CASE = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(UpperCAmelCase_ , 1E-3 , msg=F'{key} not identical' ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-model-flax-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( UpperCAmelCase_ , repo_id="""valid_org/test-model-flax-org""" , push_to_hub=UpperCAmelCase_ , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained("""valid_org/test-model-flax-org""" ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(model.params ) ) _SCREAMING_SNAKE_CASE = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): _SCREAMING_SNAKE_CASE = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(UpperCAmelCase_ , 1E-3 , msg=F'{key} not identical' ) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = flatten_dict(modela.params ) _SCREAMING_SNAKE_CASE = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: _SCREAMING_SNAKE_CASE = False return models_are_equal @require_flax class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = BertConfig.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) _SCREAMING_SNAKE_CASE = FlaxBertModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """bert""" with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) ) with self.assertRaises(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ , subfolder=UpperCAmelCase_ ) self.assertTrue(check_models_equal(UpperCAmelCase_ , UpperCAmelCase_ ) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = BertConfig.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) _SCREAMING_SNAKE_CASE = FlaxBertModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """bert""" with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(UpperCAmelCase_ , UpperCAmelCase_ ) , max_shard_size="""10KB""" ) with self.assertRaises(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ , subfolder=UpperCAmelCase_ ) self.assertTrue(check_models_equal(UpperCAmelCase_ , UpperCAmelCase_ ) ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """bert""" _SCREAMING_SNAKE_CASE = """hf-internal-testing/tiny-random-bert-subfolder""" with self.assertRaises(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ , subfolder=UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """bert""" _SCREAMING_SNAKE_CASE = """hf-internal-testing/tiny-random-bert-sharded-subfolder""" with self.assertRaises(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = FlaxBertModel.from_pretrained(UpperCAmelCase_ , subfolder=UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ )
306
def __lowerCamelCase ( snake_case__ ) -> list: """simple docstring""" def merge(snake_case__ ,snake_case__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(snake_case__ ) <= 1: return collection _SCREAMING_SNAKE_CASE = len(snake_case__ ) // 2 return merge(merge_sort(collection[:mid] ) ,merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
306
1
from ...processing_utils import ProcessorMixin class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Tuple = ["image_processor", "feature_extractor"] __snake_case : Dict = "TvltImageProcessor" __snake_case : Tuple = "TvltFeatureExtractor" def __init__( self: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] ): '''simple docstring''' super().__init__(image_processor=UpperCAmelCase_ , feature_extractor=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = image_processor _SCREAMING_SNAKE_CASE = feature_extractor def __call__( self: List[Any] , UpperCAmelCase_: List[str]=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: List[Any]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: Dict=False , UpperCAmelCase_: Optional[int]=False , *UpperCAmelCase_: Union[str, Any] , **UpperCAmelCase_: str , ): '''simple docstring''' if images is None and audio is None: raise ValueError("""You need to specify either an `images` or `audio` input to process.""" ) _SCREAMING_SNAKE_CASE = None if images is not None: _SCREAMING_SNAKE_CASE = self.image_processor(UpperCAmelCase_ , mask_pixel=UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) if images_mixed is not None: _SCREAMING_SNAKE_CASE = self.image_processor(UpperCAmelCase_ , is_mixed=UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) if audio is not None: _SCREAMING_SNAKE_CASE = self.feature_extractor( UpperCAmelCase_ , *UpperCAmelCase_ , sampling_rate=UpperCAmelCase_ , mask_audio=UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {} if audio is not None: output_dict.update(UpperCAmelCase_ ) if images is not None: output_dict.update(UpperCAmelCase_ ) if images_mixed_dict is not None: output_dict.update(UpperCAmelCase_ ) return output_dict @property def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.image_processor.model_input_names _SCREAMING_SNAKE_CASE = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
306
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length, 2) ,snake_case__ ) else: _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length) ,snake_case__ ) for i, tensor in enumerate(snake_case__ ): if padding_side == "right": if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] return out_tensor.tolist() def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = ord(snake_case__ ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True _SCREAMING_SNAKE_CASE = unicodedata.category(snake_case__ ) if cat.startswith("""P""" ): return True return False @dataclass class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : PreTrainedTokenizerBase __snake_case : Union[bool, str, PaddingStrategy] = True __snake_case : Optional[int] = None __snake_case : Optional[int] = None __snake_case : int = -100 __snake_case : str = "pt" def UpperCamelCase ( self: str , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' import torch _SCREAMING_SNAKE_CASE = """label""" if """label""" in features[0].keys() else """labels""" _SCREAMING_SNAKE_CASE = [feature[label_name] for feature in features] if label_name in features[0].keys() else None _SCREAMING_SNAKE_CASE = self.tokenizer.pad( UpperCAmelCase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" if labels is None else None , ) if labels is None: return batch _SCREAMING_SNAKE_CASE = torch.tensor(batch["""entity_ids"""] ).shape[1] _SCREAMING_SNAKE_CASE = self.tokenizer.padding_side if padding_side == "right": _SCREAMING_SNAKE_CASE = [ list(UpperCAmelCase_ ) + [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) for label in labels ] else: _SCREAMING_SNAKE_CASE = [ [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) + list(UpperCAmelCase_ ) for label in labels ] _SCREAMING_SNAKE_CASE = [feature["""ner_tags"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , -1 , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [feature["""original_entity_spans"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , (-1, -1) , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {k: torch.tensor(UpperCAmelCase_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
306
1
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
306
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = weights[0][0][0] _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[0] ) _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # lsh weights + output _SCREAMING_SNAKE_CASE = weights[0][1] if len(snake_case__ ) < 4: set_layer_weights_in_torch_lsh(snake_case__ ,torch_block.attention ,snake_case__ ) else: set_layer_weights_in_torch_local(snake_case__ ,torch_block.attention ,snake_case__ ) # intermediate weighs _SCREAMING_SNAKE_CASE = weights[2][0][1][2] # Chunked Feed Forward if len(snake_case__ ) == 4: _SCREAMING_SNAKE_CASE = intermediate_weights[2] # layernorm 2 _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # intermediate dense _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) # intermediate out _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = torch_model.reformer # word embeds _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings ,torch.tensor(snake_case__ ) ,) if isinstance(weights[3] ,snake_case__ ): _SCREAMING_SNAKE_CASE = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _SCREAMING_SNAKE_CASE = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(torch.tensor(snake_case__ ) ) _SCREAMING_SNAKE_CASE = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( snake_case__ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _SCREAMING_SNAKE_CASE = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(snake_case__ ,snake_case__ ,snake_case__ ) # output layer norm _SCREAMING_SNAKE_CASE = np.asarray(weights[7][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # output embeddings _SCREAMING_SNAKE_CASE = np.asarray(weights[9][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ReformerConfig.from_json_file(snake_case__ ) print(F'Building PyTorch model from configuration: {config}' ) _SCREAMING_SNAKE_CASE = ReformerModelWithLMHead(snake_case__ ) with open(snake_case__ ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = pickle.load(snake_case__ )["""weights"""] set_model_weights_in_torch(snake_case__ ,snake_case__ ,config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() ,snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
306
1
def __lowerCamelCase ( snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = len(snake_case__ ) for i in range(snake_case__ ): for j in range(i + 1 ,snake_case__ ): if numbers[j] < numbers[i]: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = numbers[j], numbers[i] return numbers if __name__ == "__main__": UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
306
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = TextToVideoSDPipeline __snake_case : Optional[int] = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __snake_case : Optional[int] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase ( self: int ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) _SCREAMING_SNAKE_CASE = CLIPTextModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict=0 ): '''simple docstring''' if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.get_dummy_inputs(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """np""" _SCREAMING_SNAKE_CASE = sd_pipe(**UpperCAmelCase_ ).frames _SCREAMING_SNAKE_CASE = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _SCREAMING_SNAKE_CASE = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=1E-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=25 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
306
1
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProcessor, GPTaTokenizer, PreTrainedTokenizerFast @require_vision class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = tempfile.mkdtemp() _SCREAMING_SNAKE_CASE = BlipImageProcessor() _SCREAMING_SNAKE_CASE = GPTaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-GPT2Model""" ) _SCREAMING_SNAKE_CASE = BlipaProcessor(UpperCAmelCase_ , UpperCAmelCase_ ) processor.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self: Dict , **UpperCAmelCase_: List[str] ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ).tokenizer def UpperCamelCase ( self: Optional[Any] , **UpperCAmelCase_: List[str] ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ).image_processor def UpperCamelCase ( self: Any ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _SCREAMING_SNAKE_CASE = [Image.fromarray(np.moveaxis(UpperCAmelCase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = BlipaProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _SCREAMING_SNAKE_CASE = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _SCREAMING_SNAKE_CASE = self.get_image_processor(do_normalize=UpperCAmelCase_ , padding_value=1.0 ) _SCREAMING_SNAKE_CASE = BlipaProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase_ ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.get_image_processor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = BlipaProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.prepare_image_inputs() _SCREAMING_SNAKE_CASE = image_processor(UpperCAmelCase_ , return_tensors="""np""" ) _SCREAMING_SNAKE_CASE = processor(images=UpperCAmelCase_ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.get_image_processor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = BlipaProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """lower newer""" _SCREAMING_SNAKE_CASE = processor(text=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tokenizer(UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.get_image_processor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = BlipaProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """lower newer""" _SCREAMING_SNAKE_CASE = self.prepare_image_inputs() _SCREAMING_SNAKE_CASE = processor(text=UpperCAmelCase_ , images=UpperCAmelCase_ ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """input_ids""", """attention_mask"""] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase_ ): processor() def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.get_image_processor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = BlipaProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _SCREAMING_SNAKE_CASE = processor.batch_decode(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tokenizer.batch_decode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.get_image_processor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = BlipaProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """lower newer""" _SCREAMING_SNAKE_CASE = self.prepare_image_inputs() _SCREAMING_SNAKE_CASE = processor(text=UpperCAmelCase_ , images=UpperCAmelCase_ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """input_ids""", """attention_mask"""] )
306
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class __UpperCAmelCase : def __init__( self: Union[str, Any] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int=13 , UpperCAmelCase_: Optional[int]=7 , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=False , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Optional[int]=33 , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: List[Any]=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Any=37 , UpperCAmelCase_: Optional[Any]="gelu" , UpperCAmelCase_: Dict=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: int=16 , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: Tuple=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: str=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = EsmForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = False __snake_case : Dict = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __snake_case : List[Any] = () __snake_case : Dict = ( { "feature-extraction": EsmModel, "fill-mask": EsmForMaskedLM, "text-classification": EsmForSequenceClassification, "token-classification": EsmForTokenClassification, "zero-shot": EsmForSequenceClassification, } if is_torch_available() else {} ) __snake_case : int = True def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: int ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _SCREAMING_SNAKE_CASE = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _SCREAMING_SNAKE_CASE = create_position_ids_from_input_ids(UpperCAmelCase_ , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.empty(2 , 4 , 30 ) _SCREAMING_SNAKE_CASE = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _SCREAMING_SNAKE_CASE = torch.as_tensor([expected_single_positions, expected_single_positions] ) _SCREAMING_SNAKE_CASE = embeddings.create_position_ids_from_inputs_embeds(UpperCAmelCase_ ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Dict ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Any ): '''simple docstring''' pass @require_torch class __UpperCAmelCase (_UpperCAmelCase ): @slow def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[[8.92_15, -10.58_98, -6.46_71], [-6.39_67, -13.91_14, -1.12_12], [-7.78_12, -13.95_16, -3.74_06]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] # compare the actual values for a slice. _SCREAMING_SNAKE_CASE = torch.tensor( [[[0.14_44, 0.54_13, 0.32_48], [0.30_34, 0.00_53, 0.31_08], [0.32_28, -0.24_99, 0.34_15]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
1
def __lowerCamelCase ( snake_case__ = 60_08_51_47_51_43 ) -> int: """simple docstring""" try: _SCREAMING_SNAKE_CASE = int(snake_case__ ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = 2 while i * i <= n: while n % i == 0: _SCREAMING_SNAKE_CASE = i n //= i i += 1 if n > 1: _SCREAMING_SNAKE_CASE = n return int(snake_case__ ) if __name__ == "__main__": print(f"{solution() = }")
306
import random def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = num - 1 _SCREAMING_SNAKE_CASE = 0 while s % 2 == 0: _SCREAMING_SNAKE_CASE = s // 2 t += 1 for _ in range(5 ): _SCREAMING_SNAKE_CASE = random.randrange(2 ,num - 1 ) _SCREAMING_SNAKE_CASE = pow(snake_case__ ,snake_case__ ,snake_case__ ) if v != 1: _SCREAMING_SNAKE_CASE = 0 while v != (num - 1): if i == t - 1: return False else: _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = (v**2) % num return True def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" if num < 2: return False _SCREAMING_SNAKE_CASE = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 1_01, 1_03, 1_07, 1_09, 1_13, 1_27, 1_31, 1_37, 1_39, 1_49, 1_51, 1_57, 1_63, 1_67, 1_73, 1_79, 1_81, 1_91, 1_93, 1_97, 1_99, 2_11, 2_23, 2_27, 2_29, 2_33, 2_39, 2_41, 2_51, 2_57, 2_63, 2_69, 2_71, 2_77, 2_81, 2_83, 2_93, 3_07, 3_11, 3_13, 3_17, 3_31, 3_37, 3_47, 3_49, 3_53, 3_59, 3_67, 3_73, 3_79, 3_83, 3_89, 3_97, 4_01, 4_09, 4_19, 4_21, 4_31, 4_33, 4_39, 4_43, 4_49, 4_57, 4_61, 4_63, 4_67, 4_79, 4_87, 4_91, 4_99, 5_03, 5_09, 5_21, 5_23, 5_41, 5_47, 5_57, 5_63, 5_69, 5_71, 5_77, 5_87, 5_93, 5_99, 6_01, 6_07, 6_13, 6_17, 6_19, 6_31, 6_41, 6_43, 6_47, 6_53, 6_59, 6_61, 6_73, 6_77, 6_83, 6_91, 7_01, 7_09, 7_19, 7_27, 7_33, 7_39, 7_43, 7_51, 7_57, 7_61, 7_69, 7_73, 7_87, 7_97, 8_09, 8_11, 8_21, 8_23, 8_27, 8_29, 8_39, 8_53, 8_57, 8_59, 8_63, 8_77, 8_81, 8_83, 8_87, 9_07, 9_11, 9_19, 9_29, 9_37, 9_41, 9_47, 9_53, 9_67, 9_71, 9_77, 9_83, 9_91, 9_97, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(snake_case__ ) def __lowerCamelCase ( snake_case__ = 10_24 ) -> int: """simple docstring""" while True: _SCREAMING_SNAKE_CASE = random.randrange(2 ** (keysize - 1) ,2 ** (keysize) ) if is_prime_low_num(snake_case__ ): return num if __name__ == "__main__": UpperCamelCase = generate_large_prime() print(('''Prime number:''', num)) print(('''is_prime_low_num:''', is_prime_low_num(num)))
306
1
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = TextToVideoSDPipeline __snake_case : Optional[int] = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __snake_case : Optional[int] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase ( self: int ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) _SCREAMING_SNAKE_CASE = CLIPTextModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict=0 ): '''simple docstring''' if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.get_dummy_inputs(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """np""" _SCREAMING_SNAKE_CASE = sd_pipe(**UpperCAmelCase_ ).frames _SCREAMING_SNAKE_CASE = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _SCREAMING_SNAKE_CASE = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=1E-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=25 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
306
def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _SCREAMING_SNAKE_CASE = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
306
1
def __lowerCamelCase ( snake_case__ ) -> None: """simple docstring""" _SCREAMING_SNAKE_CASE = generate_pascal_triangle(snake_case__ ) for row_idx in range(snake_case__ ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=""" """ ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] ,end=""" """ ) else: print(triangle[row_idx][col_idx] ,end="""""" ) print() def __lowerCamelCase ( snake_case__ ) -> list[list[int]]: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) _SCREAMING_SNAKE_CASE = [] for current_row_idx in range(snake_case__ ): _SCREAMING_SNAKE_CASE = populate_current_row(snake_case__ ,snake_case__ ) triangle.append(snake_case__ ) return triangle def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 1, 1 for current_col_idx in range(1 ,snake_case__ ): calculate_current_element( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) return current_row def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,) -> None: """simple docstring""" _SCREAMING_SNAKE_CASE = triangle[current_row_idx - 1][current_col_idx - 1] _SCREAMING_SNAKE_CASE = triangle[current_row_idx - 1][current_col_idx] _SCREAMING_SNAKE_CASE = above_to_left_elt + above_to_right_elt def __lowerCamelCase ( snake_case__ ) -> list[list[int]]: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) _SCREAMING_SNAKE_CASE = [[1]] for row_index in range(1 ,snake_case__ ): _SCREAMING_SNAKE_CASE = [0] + result[-1] + [0] _SCREAMING_SNAKE_CASE = row_index + 1 # Calculate the number of distinct elements in a row _SCREAMING_SNAKE_CASE = sum(divmod(snake_case__ ,2 ) ) _SCREAMING_SNAKE_CASE = [ temp_row[i - 1] + temp_row[i] for i in range(1 ,distinct_elements + 1 ) ] _SCREAMING_SNAKE_CASE = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() _SCREAMING_SNAKE_CASE = row_first_half + row_second_half result.append(snake_case__ ) return result def __lowerCamelCase ( ) -> None: """simple docstring""" from collections.abc import Callable from timeit import timeit def benchmark_a_function(snake_case__ ,snake_case__ ) -> None: _SCREAMING_SNAKE_CASE = F'{func.__name__}({value})' _SCREAMING_SNAKE_CASE = timeit(F'__main__.{call}' ,setup="""import __main__""" ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F'{call:38} -- {timing:.4f} seconds' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(snake_case__ ,snake_case__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
306
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
1
from ..utils import DummyObject, requires_backends class __UpperCAmelCase (metaclass=_UpperCAmelCase ): __snake_case : Union[str, Any] = ["speech"] def __init__( self: Dict , *UpperCAmelCase_: Optional[int] , **UpperCAmelCase_: Dict ): '''simple docstring''' requires_backends(self , ["""speech"""] ) class __UpperCAmelCase (metaclass=_UpperCAmelCase ): __snake_case : Optional[Any] = ["speech"] def __init__( self: int , *UpperCAmelCase_: List[str] , **UpperCAmelCase_: List[str] ): '''simple docstring''' requires_backends(self , ["""speech"""] )
306
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: UpperCamelCase = None UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1_024, '''facebook/nllb-200-distilled-600M''': 1_024, } # fmt: off UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : List[str] = VOCAB_FILES_NAMES __snake_case : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Tuple = ["input_ids", "attention_mask"] __snake_case : Dict = NllbTokenizer __snake_case : List[int] = [] __snake_case : List[int] = [] def __init__( self: Tuple , UpperCAmelCase_: str=None , UpperCAmelCase_: List[str]=None , UpperCAmelCase_: Tuple="<s>" , UpperCAmelCase_: str="</s>" , UpperCAmelCase_: Union[str, Any]="</s>" , UpperCAmelCase_: int="<s>" , UpperCAmelCase_: Union[str, Any]="<unk>" , UpperCAmelCase_: Union[str, Any]="<pad>" , UpperCAmelCase_: str="<mask>" , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int=None , UpperCAmelCase_: str=False , **UpperCAmelCase_: int , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token _SCREAMING_SNAKE_CASE = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , src_lang=UpperCAmelCase_ , tgt_lang=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , legacy_behaviour=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = vocab_file _SCREAMING_SNAKE_CASE = False if not self.vocab_file else True _SCREAMING_SNAKE_CASE = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"""additional_special_tokens""": _additional_special_tokens} ) _SCREAMING_SNAKE_CASE = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase_ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } _SCREAMING_SNAKE_CASE = src_lang if src_lang is not None else """eng_Latn""" _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(self._src_lang ) _SCREAMING_SNAKE_CASE = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCamelCase ( self: int ): '''simple docstring''' return self._src_lang @src_lang.setter def UpperCamelCase ( self: int , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase ( self: List[str] , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase ( self: Tuple , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] , UpperCAmelCase_: Optional[str] , **UpperCAmelCase_: Any ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = self(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tgt_lang_id return inputs def UpperCamelCase ( self: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str = "eng_Latn" , UpperCAmelCase_: Optional[List[str]] = None , UpperCAmelCase_: str = "fra_Latn" , **UpperCAmelCase_: List[str] , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCAmelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory.' ) return _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) return (out_vocab_file,)
306
1
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __UpperCAmelCase : def __init__( self: str , UpperCAmelCase_: Any , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: int=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: List[str]=7 , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: List[str]=99 , UpperCAmelCase_: Optional[Any]=36 , UpperCAmelCase_: Optional[int]=3 , UpperCAmelCase_: Dict=4 , UpperCAmelCase_: List[str]=37 , UpperCAmelCase_: Optional[int]="gelu" , UpperCAmelCase_: List[str]=0.1 , UpperCAmelCase_: str=0.1 , UpperCAmelCase_: str=512 , UpperCAmelCase_: Any=16 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Optional[Any]=6 , UpperCAmelCase_: Any=6 , UpperCAmelCase_: str=3 , UpperCAmelCase_: Optional[int]=4 , UpperCAmelCase_: str=None , UpperCAmelCase_: Any=1_000 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = patch_size _SCREAMING_SNAKE_CASE = text_seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = coordinate_size _SCREAMING_SNAKE_CASE = shape_size _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope _SCREAMING_SNAKE_CASE = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _SCREAMING_SNAKE_CASE = text_seq_length _SCREAMING_SNAKE_CASE = (image_size // patch_size) ** 2 + 1 _SCREAMING_SNAKE_CASE = self.text_seq_length + self.image_seq_length def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _SCREAMING_SNAKE_CASE = bbox[i, j, 3] _SCREAMING_SNAKE_CASE = bbox[i, j, 1] _SCREAMING_SNAKE_CASE = t if bbox[i, j, 2] < bbox[i, j, 0]: _SCREAMING_SNAKE_CASE = bbox[i, j, 2] _SCREAMING_SNAKE_CASE = bbox[i, j, 0] _SCREAMING_SNAKE_CASE = t _SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.text_seq_length] ) _SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def UpperCamelCase ( self: Any , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: str , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = LayoutLMvaModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() # text + image _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , pixel_values=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model( UpperCAmelCase_ , bbox=UpperCAmelCase_ , pixel_values=UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , bbox=UpperCAmelCase_ , pixel_values=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , bbox=UpperCAmelCase_ , pixel_values=UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _SCREAMING_SNAKE_CASE = model(pixel_values=UpperCAmelCase_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Dict , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Dict , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = LayoutLMvaForSequenceClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model( UpperCAmelCase_ , bbox=UpperCAmelCase_ , pixel_values=UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: List[str] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = LayoutLMvaForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model( UpperCAmelCase_ , bbox=UpperCAmelCase_ , pixel_values=UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Any , UpperCAmelCase_: int , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: Tuple , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = LayoutLMvaForQuestionAnswering(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model( UpperCAmelCase_ , bbox=UpperCAmelCase_ , pixel_values=UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , start_positions=UpperCAmelCase_ , end_positions=UpperCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = False __snake_case : Union[str, Any] = False __snake_case : Tuple = False __snake_case : Tuple = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) __snake_case : List[str] = ( {"document-question-answering": LayoutLMvaForQuestionAnswering, "feature-extraction": LayoutLMvaModel} if is_torch_available() else {} ) def UpperCamelCase ( self: str , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Any ): '''simple docstring''' return True def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = LayoutLMvaModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Any=False ): '''simple docstring''' _SCREAMING_SNAKE_CASE = copy.deepcopy(UpperCAmelCase_ ) if model_class in get_values(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(UpperCAmelCase_ , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=UpperCAmelCase_ ) elif model_class in get_values(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCAmelCase_ ) elif model_class in [ *get_values(UpperCAmelCase_ ), ]: _SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCAmelCase_ ) elif model_class in [ *get_values(UpperCAmelCase_ ), ]: _SCREAMING_SNAKE_CASE = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=UpperCAmelCase_ , ) return inputs_dict def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = LayoutLMvaModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class __UpperCAmelCase (unittest.TestCase ): @cached_property def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' return LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase_ ) if is_vision_available() else None @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.default_image_processor _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ).pixel_values.to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor([[1, 2]] ) _SCREAMING_SNAKE_CASE = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass _SCREAMING_SNAKE_CASE = model( input_ids=input_ids.to(UpperCAmelCase_ ) , bbox=bbox.to(UpperCAmelCase_ ) , pixel_values=pixel_values.to(UpperCAmelCase_ ) , ) # verify the logits _SCREAMING_SNAKE_CASE = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[-0.05_29, 0.36_18, 0.16_32], [-0.15_87, -0.16_67, -0.04_00], [-0.15_57, -0.16_71, -0.05_05]] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> list: """simple docstring""" _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = [[0] * n for i in range(snake_case__ )] for i in range(snake_case__ ): _SCREAMING_SNAKE_CASE = y_points[i] for i in range(2 ,snake_case__ ): for j in range(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
306
1
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __UpperCAmelCase (unittest.TestCase ): def __init__( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any]=13 , UpperCAmelCase_: List[str]=7 , UpperCAmelCase_: Tuple=True , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: str=99 , UpperCAmelCase_: List[Any]=32 , UpperCAmelCase_: Dict=5 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: Optional[int]="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: List[Any]=512 , UpperCAmelCase_: Any=16 , UpperCAmelCase_: Dict=2 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Union[str, Any]=4 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_attention_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_choices def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_attention_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=UpperCAmelCase_ , ) return config, input_ids, attention_mask def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Optional[int] = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase ( self: List[str] ): '''simple docstring''' for model_class_name in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCAmelCase_ ) @require_flax class __UpperCAmelCase (unittest.TestCase ): @slow def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 ) )
306
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCamelCase = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCamelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Dict = ["pixel_values"] def __init__( self: Union[str, Any] , UpperCAmelCase_: bool = True , UpperCAmelCase_: Dict[str, int] = None , UpperCAmelCase_: PILImageResampling = PILImageResampling.BICUBIC , UpperCAmelCase_: bool = True , UpperCAmelCase_: Dict[str, int] = None , UpperCAmelCase_: bool = True , UpperCAmelCase_: Union[int, float] = 1 / 255 , UpperCAmelCase_: bool = True , UpperCAmelCase_: Optional[Union[float, List[float]]] = None , UpperCAmelCase_: Optional[Union[float, List[float]]] = None , UpperCAmelCase_: bool = True , **UpperCAmelCase_: Optional[Any] , ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = size if size is not None else {"""shortest_edge""": 224} _SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase_ , default_to_square=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase_ , default_to_square=UpperCAmelCase_ , param_name="""crop_size""" ) _SCREAMING_SNAKE_CASE = do_resize _SCREAMING_SNAKE_CASE = size _SCREAMING_SNAKE_CASE = resample _SCREAMING_SNAKE_CASE = do_center_crop _SCREAMING_SNAKE_CASE = crop_size _SCREAMING_SNAKE_CASE = do_rescale _SCREAMING_SNAKE_CASE = rescale_factor _SCREAMING_SNAKE_CASE = do_normalize _SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else OPENAI_CLIP_MEAN _SCREAMING_SNAKE_CASE = image_std if image_std is not None else OPENAI_CLIP_STD _SCREAMING_SNAKE_CASE = do_convert_rgb def UpperCamelCase ( self: Tuple , UpperCAmelCase_: np.ndarray , UpperCAmelCase_: Dict[str, int] , UpperCAmelCase_: PILImageResampling = PILImageResampling.BICUBIC , UpperCAmelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase_: int , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase_ , default_to_square=UpperCAmelCase_ ) if "shortest_edge" not in size: raise ValueError(F'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}' ) _SCREAMING_SNAKE_CASE = get_resize_output_image_size(UpperCAmelCase_ , size=size["""shortest_edge"""] , default_to_square=UpperCAmelCase_ ) return resize(UpperCAmelCase_ , size=UpperCAmelCase_ , resample=UpperCAmelCase_ , data_format=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: np.ndarray , UpperCAmelCase_: Dict[str, int] , UpperCAmelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase_: Tuple , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase_ ) if "height" not in size or "width" not in size: raise ValueError(F'The `size` parameter must contain the keys (height, width). Got {size.keys()}' ) return center_crop(UpperCAmelCase_ , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: np.ndarray , UpperCAmelCase_: Union[int, float] , UpperCAmelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase_: Tuple , ): '''simple docstring''' return rescale(UpperCAmelCase_ , scale=UpperCAmelCase_ , data_format=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: np.ndarray , UpperCAmelCase_: Union[float, List[float]] , UpperCAmelCase_: Union[float, List[float]] , UpperCAmelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase_: List[str] , ): '''simple docstring''' return normalize(UpperCAmelCase_ , mean=UpperCAmelCase_ , std=UpperCAmelCase_ , data_format=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: int , UpperCAmelCase_: ImageInput , UpperCAmelCase_: bool = None , UpperCAmelCase_: Dict[str, int] = None , UpperCAmelCase_: PILImageResampling = None , UpperCAmelCase_: bool = None , UpperCAmelCase_: int = None , UpperCAmelCase_: bool = None , UpperCAmelCase_: float = None , UpperCAmelCase_: bool = None , UpperCAmelCase_: Optional[Union[float, List[float]]] = None , UpperCAmelCase_: Optional[Union[float, List[float]]] = None , UpperCAmelCase_: bool = None , UpperCAmelCase_: Optional[Union[str, TensorType]] = None , UpperCAmelCase_: Optional[ChannelDimension] = ChannelDimension.FIRST , **UpperCAmelCase_: List[Any] , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize _SCREAMING_SNAKE_CASE = size if size is not None else self.size _SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase_ , param_name="""size""" , default_to_square=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample _SCREAMING_SNAKE_CASE = do_center_crop if do_center_crop is not None else self.do_center_crop _SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else self.crop_size _SCREAMING_SNAKE_CASE = get_size_dict(UpperCAmelCase_ , param_name="""crop_size""" , default_to_square=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale _SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor _SCREAMING_SNAKE_CASE = do_normalize if do_normalize is not None else self.do_normalize _SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else self.image_mean _SCREAMING_SNAKE_CASE = image_std if image_std is not None else self.image_std _SCREAMING_SNAKE_CASE = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb _SCREAMING_SNAKE_CASE = make_list_of_images(UpperCAmelCase_ ) if not valid_images(UpperCAmelCase_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: _SCREAMING_SNAKE_CASE = [convert_to_rgb(UpperCAmelCase_ ) for image in images] # All transformations expect numpy arrays. _SCREAMING_SNAKE_CASE = [to_numpy_array(UpperCAmelCase_ ) for image in images] if do_resize: _SCREAMING_SNAKE_CASE = [self.resize(image=UpperCAmelCase_ , size=UpperCAmelCase_ , resample=UpperCAmelCase_ ) for image in images] if do_center_crop: _SCREAMING_SNAKE_CASE = [self.center_crop(image=UpperCAmelCase_ , size=UpperCAmelCase_ ) for image in images] if do_rescale: _SCREAMING_SNAKE_CASE = [self.rescale(image=UpperCAmelCase_ , scale=UpperCAmelCase_ ) for image in images] if do_normalize: _SCREAMING_SNAKE_CASE = [self.normalize(image=UpperCAmelCase_ , mean=UpperCAmelCase_ , std=UpperCAmelCase_ ) for image in images] _SCREAMING_SNAKE_CASE = [to_channel_dimension_format(UpperCAmelCase_ , UpperCAmelCase_ ) for image in images] _SCREAMING_SNAKE_CASE = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase_ , tensor_type=UpperCAmelCase_ )
306
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class __UpperCAmelCase : __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : int __snake_case : int __snake_case : float __snake_case : float __snake_case : Tuple[int] def UpperCamelCase ( self: str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = torch.arange(self.height * self.width ) _SCREAMING_SNAKE_CASE = torch.stack( [ pixel_indices % self.width, torch.div(UpperCAmelCase_ , self.width , rounding_mode="""trunc""" ), ] , axis=1 , ) return coords @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE = self.shape _SCREAMING_SNAKE_CASE = int(np.prod(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = self.get_image_coords() _SCREAMING_SNAKE_CASE = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) _SCREAMING_SNAKE_CASE = self.get_camera_rays(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = rays.view(UpperCAmelCase_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCamelCase ( self: Any , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] _SCREAMING_SNAKE_CASE = coords.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = self.resolution() _SCREAMING_SNAKE_CASE = self.fov() _SCREAMING_SNAKE_CASE = (flat.float() / (res - 1)) * 2 - 1 _SCREAMING_SNAKE_CASE = fracs * torch.tan(fov / 2 ) _SCREAMING_SNAKE_CASE = fracs.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = ( self.z.view(UpperCAmelCase_ , 1 , 3 ) + self.x.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, 1:] ) _SCREAMING_SNAKE_CASE = directions / directions.norm(dim=-1 , keepdim=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.stack( [ torch.broadcast_to(self.origin.view(UpperCAmelCase_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(UpperCAmelCase_ , *UpperCAmelCase_ , 2 , 3 ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=UpperCAmelCase_ , height=UpperCAmelCase_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __lowerCamelCase ( snake_case__ ) -> DifferentiableProjectiveCamera: """simple docstring""" _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for theta in np.linspace(0 ,2 * np.pi ,num=20 ): _SCREAMING_SNAKE_CASE = np.array([np.sin(snake_case__ ), np.cos(snake_case__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) _SCREAMING_SNAKE_CASE = -z * 4 _SCREAMING_SNAKE_CASE = np.array([np.cos(snake_case__ ), -np.sin(snake_case__ ), 0.0] ) _SCREAMING_SNAKE_CASE = np.cross(snake_case__ ,snake_case__ ) origins.append(snake_case__ ) xs.append(snake_case__ ) ys.append(snake_case__ ) zs.append(snake_case__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,x=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,y=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,z=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,width=snake_case__ ,height=snake_case__ ,x_fov=0.7 ,y_fov=0.7 ,shape=(1, len(snake_case__ )) ,)
306
1
from __future__ import annotations from collections.abc import Sequence from typing import Literal def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> str | Literal[False]: """simple docstring""" _SCREAMING_SNAKE_CASE = list(snake_case__ ) _SCREAMING_SNAKE_CASE = list(snake_case__ ) _SCREAMING_SNAKE_CASE = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count += 1 _SCREAMING_SNAKE_CASE = """_""" if count > 1: return False else: return "".join(snake_case__ ) def __lowerCamelCase ( snake_case__ ) -> list[str]: """simple docstring""" _SCREAMING_SNAKE_CASE = [] while True: _SCREAMING_SNAKE_CASE = ["""$"""] * len(snake_case__ ) _SCREAMING_SNAKE_CASE = [] for i in range(len(snake_case__ ) ): for j in range(i + 1 ,len(snake_case__ ) ): _SCREAMING_SNAKE_CASE = compare_string(binary[i] ,binary[j] ) if k is False: _SCREAMING_SNAKE_CASE = """*""" _SCREAMING_SNAKE_CASE = """*""" temp.append("""X""" ) for i in range(len(snake_case__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case__ ) == 0: return pi _SCREAMING_SNAKE_CASE = list(set(snake_case__ ) ) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[str]: """simple docstring""" _SCREAMING_SNAKE_CASE = [] for minterm in minterms: _SCREAMING_SNAKE_CASE = """""" for _ in range(snake_case__ ): _SCREAMING_SNAKE_CASE = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case__ ) return temp def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = list(snake_case__ ) _SCREAMING_SNAKE_CASE = list(snake_case__ ) _SCREAMING_SNAKE_CASE = 0 for i in range(len(snake_case__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[str]: """simple docstring""" _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [0] * len(snake_case__ ) for i in range(len(chart[0] ) ): _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = -1 for j in range(len(snake_case__ ) ): if chart[j][i] == 1: count += 1 _SCREAMING_SNAKE_CASE = j if count == 1: _SCREAMING_SNAKE_CASE = 1 for i in range(len(snake_case__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case__ ) ): _SCREAMING_SNAKE_CASE = 0 temp.append(prime_implicants[i] ) while True: _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = -1 _SCREAMING_SNAKE_CASE = 0 for i in range(len(snake_case__ ) ): _SCREAMING_SNAKE_CASE = chart[i].count(1 ) if count_n > max_n: _SCREAMING_SNAKE_CASE = count_n _SCREAMING_SNAKE_CASE = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case__ ) ): _SCREAMING_SNAKE_CASE = 0 def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[list[int]]: """simple docstring""" _SCREAMING_SNAKE_CASE = [[0 for x in range(len(snake_case__ ) )] for x in range(len(snake_case__ ) )] for i in range(len(snake_case__ ) ): _SCREAMING_SNAKE_CASE = prime_implicants[i].count("""_""" ) for j in range(len(snake_case__ ) ): if is_for_table(prime_implicants[i] ,binary[j] ,snake_case__ ): _SCREAMING_SNAKE_CASE = 1 return chart def __lowerCamelCase ( ) -> None: """simple docstring""" _SCREAMING_SNAKE_CASE = int(input("""Enter the no. of variables\n""" ) ) _SCREAMING_SNAKE_CASE = [ float(snake_case__ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] _SCREAMING_SNAKE_CASE = decimal_to_binary(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = check(snake_case__ ) print("""Prime Implicants are:""" ) print(snake_case__ ) _SCREAMING_SNAKE_CASE = prime_implicant_chart(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = selection(snake_case__ ,snake_case__ ) print("""Essential Prime Implicants are:""" ) print(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
306
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __UpperCAmelCase (unittest.TestCase ): def __init__( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any]=13 , UpperCAmelCase_: List[str]=7 , UpperCAmelCase_: Tuple=True , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: str=99 , UpperCAmelCase_: List[Any]=32 , UpperCAmelCase_: Dict=5 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: Optional[int]="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: List[Any]=512 , UpperCAmelCase_: Any=16 , UpperCAmelCase_: Dict=2 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Union[str, Any]=4 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_attention_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_choices def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_attention_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=UpperCAmelCase_ , ) return config, input_ids, attention_mask def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Optional[int] = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase ( self: List[str] ): '''simple docstring''' for model_class_name in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCAmelCase_ ) @require_flax class __UpperCAmelCase (unittest.TestCase ): @slow def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 ) )
306
1
import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_flax_available, is_tf_available, is_torch_available, logging from ...utils.generic import _is_jax, _is_numpy UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { '''artists_file''': '''artists.json''', '''lyrics_file''': '''lyrics.json''', '''genres_file''': '''genres.json''', } UpperCamelCase = { '''artists_file''': { '''jukebox''': '''https://huggingface.co/ArthurZ/jukebox/blob/main/artists.json''', }, '''genres_file''': { '''jukebox''': '''https://huggingface.co/ArthurZ/jukebox/blob/main/genres.json''', }, '''lyrics_file''': { '''jukebox''': '''https://huggingface.co/ArthurZ/jukebox/blob/main/lyrics.json''', }, } UpperCamelCase = { '''jukebox''': 512, } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Any = VOCAB_FILES_NAMES __snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Dict = PRETRAINED_LYRIC_TOKENS_SIZES __snake_case : Dict = ["input_ids", "attention_mask"] def __init__( self: List[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Any , UpperCAmelCase_: Tuple , UpperCAmelCase_: Any=["v3", "v2", "v2"] , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: Any=5 , UpperCAmelCase_: Optional[Any]="<|endoftext|>" , **UpperCAmelCase_: Optional[int] , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else unk_token super().__init__( unk_token=UpperCAmelCase_ , n_genres=UpperCAmelCase_ , version=UpperCAmelCase_ , max_n_lyric_tokens=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = version _SCREAMING_SNAKE_CASE = max_n_lyric_tokens _SCREAMING_SNAKE_CASE = n_genres with open(UpperCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _SCREAMING_SNAKE_CASE = json.load(UpperCAmelCase_ ) with open(UpperCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _SCREAMING_SNAKE_CASE = json.load(UpperCAmelCase_ ) with open(UpperCAmelCase_ , encoding="""utf-8""" ) as vocab_handle: _SCREAMING_SNAKE_CASE = json.load(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = R"""[^A-Za-z0-9.,:;!?\-'\"()\[\] \t\n]+""" # In v2, we had a n_vocab=80 and in v3 we missed + and so n_vocab=79 of characters. if len(self.lyrics_encoder ) == 79: _SCREAMING_SNAKE_CASE = oov.replace(R"""\-'""" , R"""\-+'""" ) _SCREAMING_SNAKE_CASE = regex.compile(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {v: k for k, v in self.artists_encoder.items()} _SCREAMING_SNAKE_CASE = {v: k for k, v in self.genres_encoder.items()} _SCREAMING_SNAKE_CASE = {v: k for k, v in self.lyrics_encoder.items()} @property def UpperCamelCase ( self: str ): '''simple docstring''' return len(self.artists_encoder ) + len(self.genres_encoder ) + len(self.lyrics_encoder ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return dict(self.artists_encoder , self.genres_encoder , self.lyrics_encoder ) def UpperCamelCase ( self: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: Any , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [self.artists_encoder.get(UpperCAmelCase_ , 0 ) for artist in list_artists] for genres in range(len(UpperCAmelCase_ ) ): _SCREAMING_SNAKE_CASE = [self.genres_encoder.get(UpperCAmelCase_ , 0 ) for genre in list_genres[genres]] _SCREAMING_SNAKE_CASE = list_genres[genres] + [-1] * (self.n_genres - len(list_genres[genres] )) _SCREAMING_SNAKE_CASE = [[self.lyrics_encoder.get(UpperCAmelCase_ , 0 ) for character in list_lyrics[0]], [], []] return artists_id, list_genres, lyric_ids def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int ): '''simple docstring''' return list(UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str] , **UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.prepare_for_tokenization(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self._tokenize(UpperCAmelCase_ ) return artist, genre, lyrics def UpperCamelCase ( self: Any , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: bool = False ): '''simple docstring''' for idx in range(len(self.version ) ): if self.version[idx] == "v3": _SCREAMING_SNAKE_CASE = artists[idx].lower() _SCREAMING_SNAKE_CASE = [genres[idx].lower()] else: _SCREAMING_SNAKE_CASE = self._normalize(artists[idx] ) + """.v2""" _SCREAMING_SNAKE_CASE = [ self._normalize(UpperCAmelCase_ ) + """.v2""" for genre in genres[idx].split("""_""" ) ] # split is for the full dictionary with combined genres if self.version[0] == "v2": _SCREAMING_SNAKE_CASE = regex.compile(R"""[^A-Za-z0-9.,:;!?\-'\"()\[\] \t\n]+""" ) _SCREAMING_SNAKE_CASE = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789.,:;!?-+'\"()[] \t\n""" _SCREAMING_SNAKE_CASE = {vocab[index]: index + 1 for index in range(len(UpperCAmelCase_ ) )} _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = len(UpperCAmelCase_ ) + 1 _SCREAMING_SNAKE_CASE = self.vocab _SCREAMING_SNAKE_CASE = {v: k for k, v in self.vocab.items()} _SCREAMING_SNAKE_CASE = """""" else: _SCREAMING_SNAKE_CASE = regex.compile(R"""[^A-Za-z0-9.,:;!?\-+'\"()\[\] \t\n]+""" ) _SCREAMING_SNAKE_CASE = self._run_strip_accents(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = lyrics.replace("""\\""" , """\n""" ) _SCREAMING_SNAKE_CASE = self.out_of_vocab.sub("""""" , UpperCAmelCase_ ), [], [] return artists, genres, lyrics def UpperCamelCase ( self: List[str] , UpperCAmelCase_: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = unicodedata.normalize("""NFD""" , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [] for char in text: _SCREAMING_SNAKE_CASE = unicodedata.category(UpperCAmelCase_ ) if cat == "Mn": continue output.append(UpperCAmelCase_ ) return "".join(UpperCAmelCase_ ) def UpperCamelCase ( self: int , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ( [chr(UpperCAmelCase_ ) for i in range(ord("""a""" ) , ord("""z""" ) + 1 )] + [chr(UpperCAmelCase_ ) for i in range(ord("""A""" ) , ord("""Z""" ) + 1 )] + [chr(UpperCAmelCase_ ) for i in range(ord("""0""" ) , ord("""9""" ) + 1 )] + ["""."""] ) _SCREAMING_SNAKE_CASE = frozenset(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = re.compile(R"""_+""" ) _SCREAMING_SNAKE_CASE = """""".join([c if c in accepted else """_""" for c in text.lower()] ) _SCREAMING_SNAKE_CASE = pattern.sub("""_""" , UpperCAmelCase_ ).strip("""_""" ) return text def UpperCamelCase ( self: Any , UpperCAmelCase_: List[str] ): '''simple docstring''' return " ".join(UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[Union[str, TensorType]] = None , UpperCAmelCase_: bool = False ): '''simple docstring''' if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = TensorType(UpperCAmelCase_ ) # Get a function reference for the correct framework if tensor_type == TensorType.TENSORFLOW: if not is_tf_available(): raise ImportError( """Unable to convert output to TensorFlow tensors format, TensorFlow is not installed.""" ) import tensorflow as tf _SCREAMING_SNAKE_CASE = tf.constant _SCREAMING_SNAKE_CASE = tf.is_tensor elif tensor_type == TensorType.PYTORCH: if not is_torch_available(): raise ImportError("""Unable to convert output to PyTorch tensors format, PyTorch is not installed.""" ) import torch _SCREAMING_SNAKE_CASE = torch.tensor _SCREAMING_SNAKE_CASE = torch.is_tensor elif tensor_type == TensorType.JAX: if not is_flax_available(): raise ImportError("""Unable to convert output to JAX tensors format, JAX is not installed.""" ) import jax.numpy as jnp # noqa: F811 _SCREAMING_SNAKE_CASE = jnp.array _SCREAMING_SNAKE_CASE = _is_jax else: _SCREAMING_SNAKE_CASE = np.asarray _SCREAMING_SNAKE_CASE = _is_numpy # Do the tensor conversion in batch try: if prepend_batch_axis: _SCREAMING_SNAKE_CASE = [inputs] if not is_tensor(UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = as_tensor(UpperCAmelCase_ ) except: # noqa E722 raise ValueError( """Unable to create tensor, you should probably activate truncation and/or padding """ """with 'padding=True' 'truncation=True' to have batched tensors with the same length.""" ) return inputs def __call__( self: Dict , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: str="" , UpperCAmelCase_: Optional[Any]="pt" ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [0, 0, 0] _SCREAMING_SNAKE_CASE = [artist] * len(self.version ) _SCREAMING_SNAKE_CASE = [genres] * len(self.version ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.tokenize(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self._convert_token_to_id(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [-INFINITY] * len(full_tokens[-1] ) _SCREAMING_SNAKE_CASE = [ self.convert_to_tensors( [input_ids + [artists_id[i]] + genres_ids[i] + full_tokens[i]] , tensor_type=UpperCAmelCase_ ) for i in range(len(self.version ) ) ] return BatchEncoding({"""input_ids""": input_ids, """attention_masks""": attention_masks} ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' if not os.path.isdir(UpperCAmelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""artists_file"""] ) with open(UpperCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.artists_encoder , ensure_ascii=UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""genres_file"""] ) with open(UpperCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.genres_encoder , ensure_ascii=UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""lyrics_file"""] ) with open(UpperCAmelCase_ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.lyrics_encoder , ensure_ascii=UpperCAmelCase_ ) ) return (artists_file, genres_file, lyrics_file) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.artists_decoder.get(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [self.genres_decoder.get(UpperCAmelCase_ ) for genre in genres_index] _SCREAMING_SNAKE_CASE = [self.lyrics_decoder.get(UpperCAmelCase_ ) for character in lyric_index] return artist, genres, lyrics
306
import argparse import hashlib # hashlib is only used inside the Test class import struct class __UpperCAmelCase : def __init__( self: List[str] , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = data _SCREAMING_SNAKE_CASE = [0x67_452_301, 0xef_cda_b89, 0x98_bad_cfe, 0x10_325_476, 0xc3_d2e_1f0] @staticmethod def UpperCamelCase ( UpperCAmelCase_: int , UpperCAmelCase_: List[str] ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0xff_fff_fff def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _SCREAMING_SNAKE_CASE = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = list(struct.unpack(""">16L""" , UpperCAmelCase_ ) ) + [0] * 64 for i in range(16 , 80 ): _SCREAMING_SNAKE_CASE = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.padding() _SCREAMING_SNAKE_CASE = self.split_blocks() for block in self.blocks: _SCREAMING_SNAKE_CASE = self.expand_block(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.h for i in range(0 , 80 ): if 0 <= i < 20: _SCREAMING_SNAKE_CASE = (b & c) | ((~b) & d) _SCREAMING_SNAKE_CASE = 0x5a_827_999 elif 20 <= i < 40: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0x6e_d9e_ba1 elif 40 <= i < 60: _SCREAMING_SNAKE_CASE = (b & c) | (b & d) | (c & d) _SCREAMING_SNAKE_CASE = 0x8f_1bb_cdc elif 60 <= i < 80: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0xca_62c_1d6 _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ( self.rotate(UpperCAmelCase_ , 5 ) + f + e + k + expanded_block[i] & 0xff_fff_fff, a, self.rotate(UpperCAmelCase_ , 30 ), c, d, ) _SCREAMING_SNAKE_CASE = ( self.h[0] + a & 0xff_fff_fff, self.h[1] + b & 0xff_fff_fff, self.h[2] + c & 0xff_fff_fff, self.h[3] + d & 0xff_fff_fff, self.h[4] + e & 0xff_fff_fff, ) return ("{:08x}" * 5).format(*self.h ) def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = b"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def __lowerCamelCase ( ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" ,dest="""input_string""" ,default="""Hello World!! Welcome to Cryptography""" ,help="""Hash the string""" ,) parser.add_argument("""--file""" ,dest="""input_file""" ,help="""Hash contents of a file""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = f.read() else: _SCREAMING_SNAKE_CASE = bytes(snake_case__ ,"""utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
306
1
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) UpperCamelCase = logging.getLogger(__name__) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> List[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.argmax(snake_case__ ,axis=1 ) return np.sum(outputs == labels ) def __lowerCamelCase ( snake_case__ ) -> str: """simple docstring""" with open(snake_case__ ,encoding="""utf_8""" ) as f: _SCREAMING_SNAKE_CASE = csv.reader(snake_case__ ) _SCREAMING_SNAKE_CASE = [] next(snake_case__ ) # skip the first line for line in tqdm(snake_case__ ): output.append((""" """.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = [] for dataset in encoded_datasets: _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = np.zeros((n_batch, 2, input_len) ,dtype=np.intaa ) _SCREAMING_SNAKE_CASE = np.zeros((n_batch, 2) ,dtype=np.intaa ) _SCREAMING_SNAKE_CASE = np.full((n_batch, 2, input_len) ,fill_value=-1_00 ,dtype=np.intaa ) _SCREAMING_SNAKE_CASE = np.zeros((n_batch,) ,dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(snake_case__ ): _SCREAMING_SNAKE_CASE = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] _SCREAMING_SNAKE_CASE = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] _SCREAMING_SNAKE_CASE = with_conta _SCREAMING_SNAKE_CASE = with_conta _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 _SCREAMING_SNAKE_CASE = with_conta _SCREAMING_SNAKE_CASE = with_conta _SCREAMING_SNAKE_CASE = mc_label _SCREAMING_SNAKE_CASE = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(snake_case__ ) for t in all_inputs ) ) return tensor_datasets def __lowerCamelCase ( ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument("""--model_name""" ,type=snake_case__ ,default="""openai-gpt""" ,help="""pretrained model name""" ) parser.add_argument("""--do_train""" ,action="""store_true""" ,help="""Whether to run training.""" ) parser.add_argument("""--do_eval""" ,action="""store_true""" ,help="""Whether to run eval on the dev set.""" ) parser.add_argument( """--output_dir""" ,default=snake_case__ ,type=snake_case__ ,required=snake_case__ ,help="""The output directory where the model predictions and checkpoints will be written.""" ,) parser.add_argument("""--train_dataset""" ,type=snake_case__ ,default="""""" ) parser.add_argument("""--eval_dataset""" ,type=snake_case__ ,default="""""" ) parser.add_argument("""--seed""" ,type=snake_case__ ,default=42 ) parser.add_argument("""--num_train_epochs""" ,type=snake_case__ ,default=3 ) parser.add_argument("""--train_batch_size""" ,type=snake_case__ ,default=8 ) parser.add_argument("""--eval_batch_size""" ,type=snake_case__ ,default=16 ) parser.add_argument("""--adam_epsilon""" ,default=1e-8 ,type=snake_case__ ,help="""Epsilon for Adam optimizer.""" ) parser.add_argument("""--max_grad_norm""" ,type=snake_case__ ,default=1 ) parser.add_argument( """--max_steps""" ,default=-1 ,type=snake_case__ ,help=( """If > 0: set total number of training steps to perform. Override num_train_epochs.""" ) ,) parser.add_argument( """--gradient_accumulation_steps""" ,type=snake_case__ ,default=1 ,help="""Number of updates steps to accumulate before performing a backward/update pass.""" ,) parser.add_argument("""--learning_rate""" ,type=snake_case__ ,default=6.2_5e-5 ) parser.add_argument("""--warmup_steps""" ,default=0 ,type=snake_case__ ,help="""Linear warmup over warmup_steps.""" ) parser.add_argument("""--lr_schedule""" ,type=snake_case__ ,default="""warmup_linear""" ) parser.add_argument("""--weight_decay""" ,type=snake_case__ ,default=0.01 ) parser.add_argument("""--lm_coef""" ,type=snake_case__ ,default=0.9 ) parser.add_argument("""--n_valid""" ,type=snake_case__ ,default=3_74 ) parser.add_argument("""--server_ip""" ,type=snake_case__ ,default="""""" ,help="""Can be used for distant debugging.""" ) parser.add_argument("""--server_port""" ,type=snake_case__ ,default="""""" ,help="""Can be used for distant debugging.""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() print(snake_case__ ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("""Waiting for debugger attach""" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) ,redirect_output=snake_case__ ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) _SCREAMING_SNAKE_CASE = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) _SCREAMING_SNAKE_CASE = torch.cuda.device_count() logger.info("""device: {}, n_gpu {}""".format(snake_case__ ,snake_case__ ) ) if not args.do_train and not args.do_eval: raise ValueError("""At least one of `do_train` or `do_eval` must be True.""" ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset _SCREAMING_SNAKE_CASE = ["""_start_""", """_delimiter_""", """_classify_"""] _SCREAMING_SNAKE_CASE = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(snake_case__ ) _SCREAMING_SNAKE_CASE = tokenizer.convert_tokens_to_ids(snake_case__ ) _SCREAMING_SNAKE_CASE = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(snake_case__ ) ) model.to(snake_case__ ) # Load and encode the datasets def tokenize_and_encode(snake_case__ ): if isinstance(snake_case__ ,snake_case__ ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(snake_case__ ) ) elif isinstance(snake_case__ ,snake_case__ ): return obj return [tokenize_and_encode(snake_case__ ) for o in obj] logger.info("""Encoding dataset...""" ) _SCREAMING_SNAKE_CASE = load_rocstories_dataset(args.train_dataset ) _SCREAMING_SNAKE_CASE = load_rocstories_dataset(args.eval_dataset ) _SCREAMING_SNAKE_CASE = (train_dataset, eval_dataset) _SCREAMING_SNAKE_CASE = tokenize_and_encode(snake_case__ ) # Compute the max input length for the Transformer _SCREAMING_SNAKE_CASE = model.config.n_positions // 2 - 2 _SCREAMING_SNAKE_CASE = max( len(story[:max_length] ) + max(len(conta[:max_length] ) ,len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) _SCREAMING_SNAKE_CASE = min(snake_case__ ,model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders _SCREAMING_SNAKE_CASE = pre_process_datasets(snake_case__ ,snake_case__ ,snake_case__ ,*snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = tensor_datasets[0], tensor_datasets[1] _SCREAMING_SNAKE_CASE = TensorDataset(*snake_case__ ) _SCREAMING_SNAKE_CASE = RandomSampler(snake_case__ ) _SCREAMING_SNAKE_CASE = DataLoader(snake_case__ ,sampler=snake_case__ ,batch_size=args.train_batch_size ) _SCREAMING_SNAKE_CASE = TensorDataset(*snake_case__ ) _SCREAMING_SNAKE_CASE = SequentialSampler(snake_case__ ) _SCREAMING_SNAKE_CASE = DataLoader(snake_case__ ,sampler=snake_case__ ,batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: _SCREAMING_SNAKE_CASE = args.max_steps _SCREAMING_SNAKE_CASE = args.max_steps // (len(snake_case__ ) // args.gradient_accumulation_steps) + 1 else: _SCREAMING_SNAKE_CASE = len(snake_case__ ) // args.gradient_accumulation_steps * args.num_train_epochs _SCREAMING_SNAKE_CASE = list(model.named_parameters() ) _SCREAMING_SNAKE_CASE = ["""bias""", """LayerNorm.bias""", """LayerNorm.weight"""] _SCREAMING_SNAKE_CASE = [ { """params""": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], """weight_decay""": args.weight_decay, }, {"""params""": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], """weight_decay""": 0.0}, ] _SCREAMING_SNAKE_CASE = AdamW(snake_case__ ,lr=args.learning_rate ,eps=args.adam_epsilon ) _SCREAMING_SNAKE_CASE = get_linear_schedule_with_warmup( snake_case__ ,num_warmup_steps=args.warmup_steps ,num_training_steps=snake_case__ ) if args.do_train: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) ,desc="""Epoch""" ): _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = tqdm(snake_case__ ,desc="""Training""" ) for step, batch in enumerate(snake_case__ ): _SCREAMING_SNAKE_CASE = tuple(t.to(snake_case__ ) for t in batch ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = batch _SCREAMING_SNAKE_CASE = model(snake_case__ ,mc_token_ids=snake_case__ ,lm_labels=snake_case__ ,mc_labels=snake_case__ ) _SCREAMING_SNAKE_CASE = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() _SCREAMING_SNAKE_CASE = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 _SCREAMING_SNAKE_CASE = """Training loss: {:.2e} lr: {:.2e}""".format(snake_case__ ,scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer _SCREAMING_SNAKE_CASE = model.module if hasattr(snake_case__ ,"""module""" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir ,snake_case__ ) _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir ,snake_case__ ) torch.save(model_to_save.state_dict() ,snake_case__ ) model_to_save.config.to_json_file(snake_case__ ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned _SCREAMING_SNAKE_CASE = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) _SCREAMING_SNAKE_CASE = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(snake_case__ ) if args.do_eval: model.eval() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0, 0 _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0, 0 for batch in tqdm(snake_case__ ,desc="""Evaluating""" ): _SCREAMING_SNAKE_CASE = tuple(t.to(snake_case__ ) for t in batch ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = batch with torch.no_grad(): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = model( snake_case__ ,mc_token_ids=snake_case__ ,lm_labels=snake_case__ ,mc_labels=snake_case__ ) _SCREAMING_SNAKE_CASE = mc_logits.detach().cpu().numpy() _SCREAMING_SNAKE_CASE = mc_labels.to("""cpu""" ).numpy() _SCREAMING_SNAKE_CASE = accuracy(snake_case__ ,snake_case__ ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 _SCREAMING_SNAKE_CASE = eval_loss / nb_eval_steps _SCREAMING_SNAKE_CASE = eval_accuracy / nb_eval_examples _SCREAMING_SNAKE_CASE = tr_loss / nb_tr_steps if args.do_train else None _SCREAMING_SNAKE_CASE = {"""eval_loss""": eval_loss, """eval_accuracy""": eval_accuracy, """train_loss""": train_loss} _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir ,"""eval_results.txt""" ) with open(snake_case__ ,"""w""" ) as writer: logger.info("""***** Eval results *****""" ) for key in sorted(result.keys() ): logger.info(""" %s = %s""" ,snake_case__ ,str(result[key] ) ) writer.write("""%s = %s\n""" % (key, str(result[key] )) ) if __name__ == "__main__": main()
306
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Tuple = VOCAB_FILES_NAMES __snake_case : Optional[int] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Optional[Any] = ["input_ids", "attention_mask"] __snake_case : Optional[int] = None def __init__( self: Dict , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int="<unk>" , UpperCAmelCase_: List[str]="<s>" , UpperCAmelCase_: Tuple="</s>" , UpperCAmelCase_: List[Any]="<pad>" , UpperCAmelCase_: Dict=False , UpperCAmelCase_: Dict=False , **UpperCAmelCase_: Dict , ): '''simple docstring''' super().__init__( UpperCAmelCase_ , UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , add_prefix_space=UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCAmelCase_ ) != add_prefix_space: _SCREAMING_SNAKE_CASE = getattr(UpperCAmelCase_ , pre_tok_state.pop("""type""" ) ) _SCREAMING_SNAKE_CASE = add_prefix_space _SCREAMING_SNAKE_CASE = pre_tok_class(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = add_prefix_space def UpperCamelCase ( self: List[str] , *UpperCAmelCase_: Any , **UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Union[str, Any] , *UpperCAmelCase_: Dict , **UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_ ) return tuple(UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: "Conversation" ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) + [self.eos_token_id] ) if len(UpperCAmelCase_ ) > self.model_max_length: _SCREAMING_SNAKE_CASE = input_ids[-self.model_max_length :] return input_ids
306
1
def __lowerCamelCase ( snake_case__ = 10_00 ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = 3 _SCREAMING_SNAKE_CASE = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 15 == 0: result -= a a += 1 return result if __name__ == "__main__": print(f"{solution() = }")
306
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __UpperCAmelCase : def __init__( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int]=13 , UpperCAmelCase_: str=7 , UpperCAmelCase_: int=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Dict=True , UpperCAmelCase_: Any=True , UpperCAmelCase_: Tuple=99 , UpperCAmelCase_: Optional[Any]=32 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Tuple=37 , UpperCAmelCase_: Union[str, Any]="gelu" , UpperCAmelCase_: List[str]=0.1 , UpperCAmelCase_: int=0.1 , UpperCAmelCase_: str=512 , UpperCAmelCase_: Union[str, Any]=16 , UpperCAmelCase_: List[Any]=2 , UpperCAmelCase_: str=0.02 , UpperCAmelCase_: int=False , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: Optional[Any]="None" , UpperCAmelCase_: Optional[int]=3 , UpperCAmelCase_: Any=4 , UpperCAmelCase_: Optional[int]=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = relative_attention _SCREAMING_SNAKE_CASE = position_biased_input _SCREAMING_SNAKE_CASE = pos_att_type _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=UpperCAmelCase_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _SCREAMING_SNAKE_CASE = [input_ids, input_mask] _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForMaskedLM(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Any , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForSequenceClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForTokenClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: Any , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForQuestionAnswering(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) __snake_case : Union[str, Any] = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) __snake_case : Dict = False __snake_case : Optional[Any] = False def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(UpperCAmelCase_ ) @require_tf class __UpperCAmelCase (unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = tf.constant( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 )
306
1
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def __lowerCamelCase ( *snake_case__ ) -> List[str]: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = list(snake_case__ ) for i in range(len(snake_case__ ) ): _SCREAMING_SNAKE_CASE = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = [ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(snake_case__ ,snake_case__ ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def __lowerCamelCase ( snake_case__ = None ,snake_case__ = 1_28 ) -> Optional[Any]: """simple docstring""" if function is None: return functools.partial(snake_case__ ,starting_batch_size=snake_case__ ) _SCREAMING_SNAKE_CASE = starting_batch_size def decorator(*snake_case__ ,**snake_case__ ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() _SCREAMING_SNAKE_CASE = list(inspect.signature(snake_case__ ).parameters.keys() ) # Guard against user error if len(snake_case__ ) < (len(snake_case__ ) + 1): _SCREAMING_SNAKE_CASE = """, """.join([F'{arg}={value}' for arg, value in zip(params[1:] ,args[1:] )] ) raise TypeError( F'Batch size was passed into `{function.__name__}` as the first argument when called.' F'Remove this as the decorator already does so: `{function.__name__}({arg_str})`' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(snake_case__ ,*snake_case__ ,**snake_case__ ) except Exception as e: if should_reduce_batch_size(snake_case__ ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
306
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __lowerCamelCase ( snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = gather(snake_case__ ) assert gathered_tensor.tolist() == list(range(1 ,state.num_processes**2 + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = [state.process_index] _SCREAMING_SNAKE_CASE = gather_object(snake_case__ ) assert len(snake_case__ ) == state.num_processes, F'{gathered_obj}, {len(snake_case__ )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = broadcast(snake_case__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 ,state.num_processes + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Tuple: """simple docstring""" if state.is_main_process: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes + 1 ).to(state.device ) else: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes ).to(state.device ) _SCREAMING_SNAKE_CASE = pad_across_processes(snake_case__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 ,state.num_processes ) ) + [0] def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""sum""" ) _SCREAMING_SNAKE_CASE = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""mean""" ) _SCREAMING_SNAKE_CASE = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> str: """simple docstring""" main() def __lowerCamelCase ( ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = PartialState() state.print(F'State: {state}' ) state.print("""testing gather""" ) test_gather(snake_case__ ) state.print("""testing gather_object""" ) test_gather_object(snake_case__ ) state.print("""testing broadcast""" ) test_broadcast(snake_case__ ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(snake_case__ ) state.print("""testing reduce_sum""" ) test_reduce_sum(snake_case__ ) state.print("""testing reduce_mean""" ) test_reduce_mean(snake_case__ ) if __name__ == "__main__": main()
306
1
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase ) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Tuple , *UpperCAmelCase_: str , **UpperCAmelCase_: Optional[int] ): '''simple docstring''' super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ ) self.check_model_type(UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Tuple=None , UpperCAmelCase_: Dict=None , UpperCAmelCase_: Optional[Any]=None , **UpperCAmelCase_: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = {}, {} if padding is not None: _SCREAMING_SNAKE_CASE = padding if truncation is not None: _SCREAMING_SNAKE_CASE = truncation if top_k is not None: _SCREAMING_SNAKE_CASE = top_k return preprocess_params, {}, postprocess_params def __call__( self: List[str] , UpperCAmelCase_: Union["Image.Image", str] , UpperCAmelCase_: str = None , **UpperCAmelCase_: str ): '''simple docstring''' if isinstance(UpperCAmelCase_ , (Image.Image, str) ) and isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = {"""image""": image, """question""": question} else: _SCREAMING_SNAKE_CASE = image _SCREAMING_SNAKE_CASE = super().__call__(UpperCAmelCase_ , **UpperCAmelCase_ ) return results def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: List[Any]=False , UpperCAmelCase_: Optional[Any]=False ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(inputs["""image"""] ) _SCREAMING_SNAKE_CASE = self.tokenizer( inputs["""question"""] , return_tensors=self.framework , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor(images=UpperCAmelCase_ , return_tensors=self.framework ) model_inputs.update(UpperCAmelCase_ ) return model_inputs def UpperCamelCase ( self: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) return model_outputs def UpperCamelCase ( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: Any=5 ): '''simple docstring''' if top_k > self.model.config.num_labels: _SCREAMING_SNAKE_CASE = self.model.config.num_labels if self.framework == "pt": _SCREAMING_SNAKE_CASE = model_outputs.logits.sigmoid()[0] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = probs.topk(UpperCAmelCase_ ) else: raise ValueError(F'Unsupported framework: {self.framework}' ) _SCREAMING_SNAKE_CASE = scores.tolist() _SCREAMING_SNAKE_CASE = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(UpperCAmelCase_ , UpperCAmelCase_ )]
306
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __lowerCamelCase ( ) -> tuple[list[int], int]: """simple docstring""" _SCREAMING_SNAKE_CASE = [randint(-10_00 ,10_00 ) for i in range(10 )] _SCREAMING_SNAKE_CASE = randint(-50_00 ,50_00 ) return (arr, r) UpperCamelCase = make_dataset() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(snake_case__ ,3 ): if sum(snake_case__ ) == target: return tuple(sorted(snake_case__ ) ) return (0, 0, 0) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, int, int]: """simple docstring""" arr.sort() _SCREAMING_SNAKE_CASE = len(snake_case__ ) for i in range(n - 1 ): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __lowerCamelCase ( ) -> tuple[float, float]: """simple docstring""" _SCREAMING_SNAKE_CASE = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _SCREAMING_SNAKE_CASE = """ triplet_sum1(*dataset) """ _SCREAMING_SNAKE_CASE = """ triplet_sum2(*dataset) """ _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) return (min(snake_case__ ), min(snake_case__ )) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
306
1
import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __UpperCAmelCase (unittest.TestCase ): def __init__( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Any=7 , UpperCAmelCase_: Optional[Any]=3 , UpperCAmelCase_: Union[str, Any]=18 , UpperCAmelCase_: str=30 , UpperCAmelCase_: Optional[int]=400 , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: str=None , UpperCAmelCase_: int=True , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = size if size is not None else {"""height""": 18, """width""": 18} _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = min_resolution _SCREAMING_SNAKE_CASE = max_resolution _SCREAMING_SNAKE_CASE = do_resize _SCREAMING_SNAKE_CASE = size _SCREAMING_SNAKE_CASE = apply_ocr def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Any = LayoutLMvaImageProcessor if is_pytesseract_available() else None def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = LayoutLMvaImageProcessingTester(self ) @property def UpperCamelCase ( self: int ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase_ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCAmelCase_ , """size""" ) ) self.assertTrue(hasattr(UpperCAmelCase_ , """apply_ocr""" ) ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) _SCREAMING_SNAKE_CASE = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , Image.Image ) # Test not batched input _SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="""pt""" ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) self.assertIsInstance(encoding.words , UpperCAmelCase_ ) self.assertIsInstance(encoding.boxes , UpperCAmelCase_ ) # Test batched _SCREAMING_SNAKE_CASE = image_processing(UpperCAmelCase_ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , numpify=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , np.ndarray ) # Test not batched input _SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched _SCREAMING_SNAKE_CASE = image_processing(UpperCAmelCase_ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase_ , torchify=UpperCAmelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase_ , torch.Tensor ) # Test not batched input _SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched _SCREAMING_SNAKE_CASE = image_processing(UpperCAmelCase_ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = LayoutLMvaImageProcessor() from datasets import load_dataset _SCREAMING_SNAKE_CASE = load_dataset("""hf-internal-testing/fixtures_docvqa""" , split="""test""" ) _SCREAMING_SNAKE_CASE = Image.open(ds[0]["""file"""] ).convert("""RGB""" ) _SCREAMING_SNAKE_CASE = image_processing(UpperCAmelCase_ , return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 _SCREAMING_SNAKE_CASE = [["""11:14""", """to""", """11:39""", """a.m""", """11:39""", """to""", """11:44""", """a.m.""", """11:44""", """a.m.""", """to""", """12:25""", """p.m.""", """12:25""", """to""", """12:58""", """p.m.""", """12:58""", """to""", """4:00""", """p.m.""", """2:00""", """to""", """5:00""", """p.m.""", """Coffee""", """Break""", """Coffee""", """will""", """be""", """served""", """for""", """men""", """and""", """women""", """in""", """the""", """lobby""", """adjacent""", """to""", """exhibit""", """area.""", """Please""", """move""", """into""", """exhibit""", """area.""", """(Exhibits""", """Open)""", """TRRF""", """GENERAL""", """SESSION""", """(PART""", """|)""", """Presiding:""", """Lee""", """A.""", """Waller""", """TRRF""", """Vice""", """President""", """“Introductory""", """Remarks”""", """Lee""", """A.""", """Waller,""", """TRRF""", """Vice""", """Presi-""", """dent""", """Individual""", """Interviews""", """with""", """TRRF""", """Public""", """Board""", """Members""", """and""", """Sci-""", """entific""", """Advisory""", """Council""", """Mem-""", """bers""", """Conducted""", """by""", """TRRF""", """Treasurer""", """Philip""", """G.""", """Kuehn""", """to""", """get""", """answers""", """which""", """the""", """public""", """refrigerated""", """warehousing""", """industry""", """is""", """looking""", """for.""", """Plus""", """questions""", """from""", """the""", """floor.""", """Dr.""", """Emil""", """M.""", """Mrak,""", """University""", """of""", """Cal-""", """ifornia,""", """Chairman,""", """TRRF""", """Board;""", """Sam""", """R.""", """Cecil,""", """University""", """of""", """Georgia""", """College""", """of""", """Agriculture;""", """Dr.""", """Stanley""", """Charm,""", """Tufts""", """University""", """School""", """of""", """Medicine;""", """Dr.""", """Robert""", """H.""", """Cotton,""", """ITT""", """Continental""", """Baking""", """Company;""", """Dr.""", """Owen""", """Fennema,""", """University""", """of""", """Wis-""", """consin;""", """Dr.""", """Robert""", """E.""", """Hardenburg,""", """USDA.""", """Questions""", """and""", """Answers""", """Exhibits""", """Open""", """Capt.""", """Jack""", """Stoney""", """Room""", """TRRF""", """Scientific""", """Advisory""", """Council""", """Meeting""", """Ballroom""", """Foyer"""]] # noqa: E231 _SCREAMING_SNAKE_CASE = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCAmelCase_ ) self.assertListEqual(encoding.boxes , UpperCAmelCase_ ) # with apply_OCR = False _SCREAMING_SNAKE_CASE = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = image_processing(UpperCAmelCase_ , return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
306
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase ) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) self.check_model_type(UpperCAmelCase_ ) def UpperCamelCase ( self: str , **UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} # preprocess args if "points_per_batch" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self: Optional[Any] , UpperCAmelCase_: Tuple , *UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=None , UpperCAmelCase_: Tuple=None , **UpperCAmelCase_: Any ): '''simple docstring''' return super().__call__(UpperCAmelCase_ , *UpperCAmelCase_ , num_workers=UpperCAmelCase_ , batch_size=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict=64 , UpperCAmelCase_: int = 0 , UpperCAmelCase_: float = 512 / 1_500 , UpperCAmelCase_: Optional[int] = 32 , UpperCAmelCase_: Optional[int] = 1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor.size["""longest_edge"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.generate_crop_boxes( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": _SCREAMING_SNAKE_CASE = self.get_inference_context() with inference_context(): _SCREAMING_SNAKE_CASE = self._ensure_tensor_on_device(UpperCAmelCase_ , device=self.device ) _SCREAMING_SNAKE_CASE = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) _SCREAMING_SNAKE_CASE = image_embeddings _SCREAMING_SNAKE_CASE = grid_points.shape[1] _SCREAMING_SNAKE_CASE = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = grid_points[:, i : i + points_per_batch, :, :] _SCREAMING_SNAKE_CASE = input_labels[:, i : i + points_per_batch] _SCREAMING_SNAKE_CASE = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=0.88 , UpperCAmelCase_: Dict=0.95 , UpperCAmelCase_: Tuple=0 , UpperCAmelCase_: str=1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_inputs.pop("""input_boxes""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""is_last""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""original_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = model_inputs.pop("""reshaped_input_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks _SCREAMING_SNAKE_CASE = model_outputs["""pred_masks"""] _SCREAMING_SNAKE_CASE = self.image_processor.post_process_masks( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , binarize=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model_outputs["""iou_scores"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCamelCase ( self: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=False , UpperCAmelCase_: Any=0.7 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.post_process_for_mask_generation( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = defaultdict(UpperCAmelCase_ ) for output in model_outputs: for k, v in output.items(): extra[k].append(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {} if output_rle_mask: _SCREAMING_SNAKE_CASE = rle_mask if output_bboxes_mask: _SCREAMING_SNAKE_CASE = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
306
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
306
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging UpperCamelCase = logging.get_logger(__name__) logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" if "xprophetnet" in prophetnet_checkpoint_path: _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) else: _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) _SCREAMING_SNAKE_CASE = ["""key_proj""", """value_proj""", """query_proj"""] _SCREAMING_SNAKE_CASE = { """self_attn""": """ngram_self_attn""", """cross_attn""": """encoder_attn""", """cross_attn_layer_norm""": """encoder_attn_layer_norm""", """feed_forward_layer_norm""": """final_layer_norm""", """feed_forward""": """""", """intermediate""": """fc1""", """output""": """fc2""", """key_proj""": """k_proj""", """query_proj""": """q_proj""", """value_proj""": """v_proj""", """word_embeddings""": """embed_tokens""", """embeddings_layer_norm""": """emb_layer_norm""", """relative_pos_embeddings""": """relative_linear""", """ngram_embeddings""": """ngram_input_embed""", """position_embeddings""": """embed_positions""", } for key in loading_info["missing_keys"]: _SCREAMING_SNAKE_CASE = key.split(""".""" ) if attributes[0] == "lm_head": _SCREAMING_SNAKE_CASE = prophet _SCREAMING_SNAKE_CASE = prophet_old else: _SCREAMING_SNAKE_CASE = prophet.prophetnet _SCREAMING_SNAKE_CASE = prophet_old.model _SCREAMING_SNAKE_CASE = False for attribute in attributes: if attribute in mapping: _SCREAMING_SNAKE_CASE = mapping[attribute] if not hasattr(snake_case__ ,snake_case__ ) and len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = attribute elif hasattr(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.weight logger.info(F'{attribute} is initialized.' ) _SCREAMING_SNAKE_CASE = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.bias logger.info(F'{attribute} is initialized' ) _SCREAMING_SNAKE_CASE = True break elif attribute in special_keys and hasattr(snake_case__ ,"""in_proj_weight""" ): _SCREAMING_SNAKE_CASE = old_model.in_proj_weight.shape[0] // 3 _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _SCREAMING_SNAKE_CASE = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_12, "We want 512 position_embeddings." _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.embed_positions.weight[:5_12, :] ) _SCREAMING_SNAKE_CASE = True break if attribute.isdigit(): _SCREAMING_SNAKE_CASE = model[int(snake_case__ )] _SCREAMING_SNAKE_CASE = old_model[int(snake_case__ )] else: _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if old_attribute == "": _SCREAMING_SNAKE_CASE = old_model else: if not hasattr(snake_case__ ,snake_case__ ): raise ValueError(F'{old_model} does not have {old_attribute}' ) _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if not is_key_init: raise ValueError(F'{key} was not correctly initialized!' ) print(F'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--prophetnet_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
306
1
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging UpperCamelCase = logging.get_logger(__name__) class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : int = CLIPConfig __snake_case : str = ["CLIPEncoderLayer"] def __init__( self: Dict , UpperCAmelCase_: CLIPConfig ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPVisionModelWithProjection(config.vision_config ) _SCREAMING_SNAKE_CASE = nn.Linear(config.vision_config.projection_dim , 1 ) _SCREAMING_SNAKE_CASE = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def UpperCamelCase ( self: int , UpperCAmelCase_: Any , UpperCAmelCase_: int , UpperCAmelCase_: Dict=0.5 , UpperCAmelCase_: Optional[Any]=0.5 ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.vision_model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = self.p_head(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nsfw_detected.flatten() _SCREAMING_SNAKE_CASE = nsfw_detected > p_threshold _SCREAMING_SNAKE_CASE = nsfw_detected.tolist() if any(UpperCAmelCase_ ): logger.warning( """Potential NSFW content was detected in one or more images. A black image will be returned instead.""" """ Try again with a different prompt and/or seed.""" ) for idx, nsfw_detected_ in enumerate(UpperCAmelCase_ ): if nsfw_detected_: _SCREAMING_SNAKE_CASE = np.zeros(images[idx].shape ) _SCREAMING_SNAKE_CASE = self.w_head(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = watermark_detected.flatten() _SCREAMING_SNAKE_CASE = watermark_detected > w_threshold _SCREAMING_SNAKE_CASE = watermark_detected.tolist() if any(UpperCAmelCase_ ): logger.warning( """Potential watermarked content was detected in one or more images. A black image will be returned instead.""" """ Try again with a different prompt and/or seed.""" ) for idx, watermark_detected_ in enumerate(UpperCAmelCase_ ): if watermark_detected_: _SCREAMING_SNAKE_CASE = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
306
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: _SCREAMING_SNAKE_CASE = i + 1 else: _SCREAMING_SNAKE_CASE = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"{two_pointer([2, 7, 11, 15], 9) = }")
306
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" ,"False" ) ) is not True ,reason="Skipping test because should only be run when releasing minor transformers version" ,) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.g4dn.xlarge", "results": {"train_runtime": 650, "eval_accuracy": 0.6, "eval_loss": 0.9}, }, { "framework": "tensorflow", "script": "run_tf.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.g4dn.xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.3, "eval_loss": 0.9}, }, ] ) class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: Any ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( F'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split() , encoding="""utf-8""" , check=UpperCAmelCase_ , ) assert hasattr(self , """env""" ) def UpperCamelCase ( self: str , UpperCAmelCase_: int=1 ): '''simple docstring''' return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'{self.env.base_job_name}-single' , instance_count=UpperCAmelCase_ , instance_type=self.instance_type , debugger_hook_config=UpperCAmelCase_ , hyperparameters={**self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="""py36""" , ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' TrainingJobAnalytics(UpperCAmelCase_ ).export_csv(F'{self.env.test_path}/{job_name}_metrics.csv' ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.create_estimator() # run training estimator.fit() # result dataframe _SCREAMING_SNAKE_CASE = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _SCREAMING_SNAKE_CASE = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) _SCREAMING_SNAKE_CASE = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _SCREAMING_SNAKE_CASE = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'{estimator.latest_training_job.name}.json' , """w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} , UpperCAmelCase_ )
306
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig UpperCamelCase = logging.get_logger(__name__) # General docstring UpperCamelCase = '''MobileNetV1Config''' # Base docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = [1, 1_024, 7, 7] # Image classification docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = '''tabby, tabby cat''' UpperCamelCase = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = {} if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = model.mobilenet_va else: _SCREAMING_SNAKE_CASE = model _SCREAMING_SNAKE_CASE = """MobilenetV1/Conv2d_0/""" _SCREAMING_SNAKE_CASE = backbone.conv_stem.convolution.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.bias _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_mean _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_var for i in range(13 ): _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = i * 2 _SCREAMING_SNAKE_CASE = backbone.layer[pt_index] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_depthwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var _SCREAMING_SNAKE_CASE = backbone.layer[pt_index + 1] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_pointwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = """MobilenetV1/Logits/Conv2d_1c_1x1/""" _SCREAMING_SNAKE_CASE = model.classifier.weight _SCREAMING_SNAKE_CASE = model.classifier.bias return tf_to_pt_map def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> List[str]: """simple docstring""" try: import numpy as np import tensorflow as tf except ImportError: logger.error( """Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see """ """https://www.tensorflow.org/install/ for installation instructions.""" ) raise # Load weights from TF model _SCREAMING_SNAKE_CASE = tf.train.list_variables(snake_case__ ) _SCREAMING_SNAKE_CASE = {} for name, shape in init_vars: logger.info(F'Loading TF weight {name} with shape {shape}' ) _SCREAMING_SNAKE_CASE = tf.train.load_variable(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = array # Build TF to PyTorch weights loading map _SCREAMING_SNAKE_CASE = _build_tf_to_pytorch_map(snake_case__ ,snake_case__ ,snake_case__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F'Importing {name}' ) if name not in tf_weights: logger.info(F'{name} not in tf pre-trained weights, skipping' ) continue _SCREAMING_SNAKE_CASE = tf_weights[name] if "depthwise_weights" in name: logger.info("""Transposing depthwise""" ) _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(2, 3, 0, 1) ) elif "weights" in name: logger.info("""Transposing""" ) if len(pointer.shape ) == 2: # copying into linear layer _SCREAMING_SNAKE_CASE = array.squeeze().transpose() else: _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'Pointer shape {pointer.shape} and array shape {array.shape} mismatched' ) logger.info(F'Initialize PyTorch weight {name} {array.shape}' ) _SCREAMING_SNAKE_CASE = torch.from_numpy(snake_case__ ) tf_weights.pop(snake_case__ ,snake_case__ ) tf_weights.pop(name + """/RMSProp""" ,snake_case__ ) tf_weights.pop(name + """/RMSProp_1""" ,snake_case__ ) tf_weights.pop(name + """/ExponentialMovingAverage""" ,snake_case__ ) logger.info(F'Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}' ) return model def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> torch.Tensor: """simple docstring""" _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = features.shape[-2:] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.stride _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.kernel_size if in_height % stride_height == 0: _SCREAMING_SNAKE_CASE = max(kernel_height - stride_height ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_height - (in_height % stride_height) ,0 ) if in_width % stride_width == 0: _SCREAMING_SNAKE_CASE = max(kernel_width - stride_width ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_width - (in_width % stride_width) ,0 ) _SCREAMING_SNAKE_CASE = pad_along_width // 2 _SCREAMING_SNAKE_CASE = pad_along_width - pad_left _SCREAMING_SNAKE_CASE = pad_along_height // 2 _SCREAMING_SNAKE_CASE = pad_along_height - pad_top _SCREAMING_SNAKE_CASE = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(snake_case__ ,snake_case__ ,"""constant""" ,0.0 ) class __UpperCAmelCase (nn.Module ): def __init__( self: Optional[Any] , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: bool = False , UpperCAmelCase_: Optional[bool] = True , UpperCAmelCase_: Optional[bool or str] = True , ): '''simple docstring''' super().__init__() _SCREAMING_SNAKE_CASE = config if in_channels % groups != 0: raise ValueError(F'Input channels ({in_channels}) are not divisible by {groups} groups.' ) if out_channels % groups != 0: raise ValueError(F'Output channels ({out_channels}) are not divisible by {groups} groups.' ) _SCREAMING_SNAKE_CASE = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) _SCREAMING_SNAKE_CASE = nn.Convad( in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=UpperCAmelCase_ , groups=UpperCAmelCase_ , bias=UpperCAmelCase_ , padding_mode="""zeros""" , ) if use_normalization: _SCREAMING_SNAKE_CASE = nn.BatchNormad( num_features=UpperCAmelCase_ , eps=config.layer_norm_eps , momentum=0.99_97 , affine=UpperCAmelCase_ , track_running_stats=UpperCAmelCase_ , ) else: _SCREAMING_SNAKE_CASE = None if use_activation: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[use_activation] elif isinstance(config.hidden_act , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: _SCREAMING_SNAKE_CASE = config.hidden_act else: _SCREAMING_SNAKE_CASE = None def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' if self.config.tf_padding: _SCREAMING_SNAKE_CASE = apply_tf_padding(UpperCAmelCase_ , self.convolution ) _SCREAMING_SNAKE_CASE = self.convolution(UpperCAmelCase_ ) if self.normalization is not None: _SCREAMING_SNAKE_CASE = self.normalization(UpperCAmelCase_ ) if self.activation is not None: _SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase_ ) return features class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Dict = MobileNetVaConfig __snake_case : Any = load_tf_weights_in_mobilenet_va __snake_case : Any = "mobilenet_v1" __snake_case : List[Any] = "pixel_values" __snake_case : Any = False def UpperCamelCase ( self: str , UpperCAmelCase_: Union[nn.Linear, nn.Convad] ): '''simple docstring''' if isinstance(UpperCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCAmelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) UpperCamelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' UpperCamelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: bool = True ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config _SCREAMING_SNAKE_CASE = 32 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) _SCREAMING_SNAKE_CASE = MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=config.num_channels , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=2 , ) _SCREAMING_SNAKE_CASE = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] _SCREAMING_SNAKE_CASE = nn.ModuleList() for i in range(13 ): _SCREAMING_SNAKE_CASE = out_channels if strides[i] == 2 or i == 0: depth *= 2 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=strides[i] , groups=UpperCAmelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=1 , ) ) _SCREAMING_SNAKE_CASE = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCamelCase ( self: Dict , UpperCAmelCase_: Tuple ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase ( self: int , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) _SCREAMING_SNAKE_CASE = self.conv_stem(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): _SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase_ ) if output_hidden_states: _SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) _SCREAMING_SNAKE_CASE = hidden_states if self.pooler is not None: _SCREAMING_SNAKE_CASE = torch.flatten(self.pooler(UpperCAmelCase_ ) , start_dim=1 ) else: _SCREAMING_SNAKE_CASE = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase_ , pooler_output=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Dict , UpperCAmelCase_: MobileNetVaConfig ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config.num_labels _SCREAMING_SNAKE_CASE = MobileNetVaModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head _SCREAMING_SNAKE_CASE = nn.Dropout(config.classifier_dropout_prob , inplace=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.Linear(UpperCAmelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict _SCREAMING_SNAKE_CASE = self.mobilenet_va(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.pooler_output if return_dict else outputs[1] _SCREAMING_SNAKE_CASE = self.classifier(self.dropout(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _SCREAMING_SNAKE_CASE = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _SCREAMING_SNAKE_CASE = """single_label_classification""" else: _SCREAMING_SNAKE_CASE = """multi_label_classification""" if self.config.problem_type == "regression": _SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: _SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) elif self.config.problem_type == "single_label_classification": _SCREAMING_SNAKE_CASE = CrossEntropyLoss() _SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: _SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states , )
306
1
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def __lowerCamelCase ( snake_case__ = "" ) -> dict[str, float]: """simple docstring""" _SCREAMING_SNAKE_CASE = url or """https://www.imdb.com/chart/top/?ref_=nv_mv_250""" _SCREAMING_SNAKE_CASE = BeautifulSoup(requests.get(snake_case__ ).text ,"""html.parser""" ) _SCREAMING_SNAKE_CASE = soup.find_all("""td""" ,attrs="""titleColumn""" ) _SCREAMING_SNAKE_CASE = soup.find_all("""td""" ,class_="""ratingColumn imdbRating""" ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(snake_case__ ,snake_case__ ) } def __lowerCamelCase ( snake_case__ = "IMDb_Top_250_Movies.csv" ) -> None: """simple docstring""" _SCREAMING_SNAKE_CASE = get_imdb_top_aaa_movies() with open(snake_case__ ,"""w""" ,newline="""""" ) as out_file: _SCREAMING_SNAKE_CASE = csv.writer(snake_case__ ) writer.writerow(["""Movie title""", """IMDb rating"""] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
306
def __lowerCamelCase ( snake_case__ ) -> list: """simple docstring""" def merge(snake_case__ ,snake_case__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(snake_case__ ) <= 1: return collection _SCREAMING_SNAKE_CASE = len(snake_case__ ) // 2 return merge(merge_sort(collection[:mid] ) ,merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
306
1
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def __lowerCamelCase ( snake_case__ = 3 ) -> qiskit.result.counts.Counts: """simple docstring""" if isinstance(snake_case__ ,snake_case__ ): raise TypeError("""number of qubits must be a integer.""" ) if number_of_qubits <= 0: raise ValueError("""number of qubits must be > 0.""" ) if math.floor(snake_case__ ) != number_of_qubits: raise ValueError("""number of qubits must be exact integer.""" ) if number_of_qubits > 10: raise ValueError("""number of qubits too large to simulate(>10).""" ) _SCREAMING_SNAKE_CASE = QuantumRegister(snake_case__ ,"""qr""" ) _SCREAMING_SNAKE_CASE = ClassicalRegister(snake_case__ ,"""cr""" ) _SCREAMING_SNAKE_CASE = QuantumCircuit(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = number_of_qubits for i in range(snake_case__ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(snake_case__ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) ,snake_case__ ,snake_case__ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(snake_case__ ,number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(snake_case__ ,snake_case__ ) # simulate with 10000 shots _SCREAMING_SNAKE_CASE = Aer.get_backend("""qasm_simulator""" ) _SCREAMING_SNAKE_CASE = execute(snake_case__ ,snake_case__ ,shots=1_00_00 ) return job.result().get_counts(snake_case__ ) if __name__ == "__main__": print( f"Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}" )
306
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length, 2) ,snake_case__ ) else: _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length) ,snake_case__ ) for i, tensor in enumerate(snake_case__ ): if padding_side == "right": if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] return out_tensor.tolist() def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = ord(snake_case__ ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True _SCREAMING_SNAKE_CASE = unicodedata.category(snake_case__ ) if cat.startswith("""P""" ): return True return False @dataclass class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : PreTrainedTokenizerBase __snake_case : Union[bool, str, PaddingStrategy] = True __snake_case : Optional[int] = None __snake_case : Optional[int] = None __snake_case : int = -100 __snake_case : str = "pt" def UpperCamelCase ( self: str , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' import torch _SCREAMING_SNAKE_CASE = """label""" if """label""" in features[0].keys() else """labels""" _SCREAMING_SNAKE_CASE = [feature[label_name] for feature in features] if label_name in features[0].keys() else None _SCREAMING_SNAKE_CASE = self.tokenizer.pad( UpperCAmelCase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" if labels is None else None , ) if labels is None: return batch _SCREAMING_SNAKE_CASE = torch.tensor(batch["""entity_ids"""] ).shape[1] _SCREAMING_SNAKE_CASE = self.tokenizer.padding_side if padding_side == "right": _SCREAMING_SNAKE_CASE = [ list(UpperCAmelCase_ ) + [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) for label in labels ] else: _SCREAMING_SNAKE_CASE = [ [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) + list(UpperCAmelCase_ ) for label in labels ] _SCREAMING_SNAKE_CASE = [feature["""ner_tags"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , -1 , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [feature["""original_entity_spans"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , (-1, -1) , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {k: torch.tensor(UpperCAmelCase_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
306
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def __lowerCamelCase ( snake_case__ ,snake_case__=False ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'blocks.{i}.norm1.weight', F'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'blocks.{i}.norm1.bias', F'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((F'blocks.{i}.attn.proj.weight', F'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((F'blocks.{i}.attn.proj.bias', F'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'blocks.{i}.norm2.weight', F'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'blocks.{i}.norm2.bias', F'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc1.weight', F'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc1.bias', F'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'blocks.{i}.mlp.fc2.weight', F'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'blocks.{i}.mlp.fc2.bias', F'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """vit.embeddings.cls_token"""), ("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _SCREAMING_SNAKE_CASE = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=False ) -> List[str]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _SCREAMING_SNAKE_CASE = """""" else: _SCREAMING_SNAKE_CASE = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _SCREAMING_SNAKE_CASE = state_dict.pop(F'blocks.{i}.attn.qkv.weight' ) _SCREAMING_SNAKE_CASE = state_dict.pop(F'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _SCREAMING_SNAKE_CASE = in_proj_weight[ : config.hidden_size, : ] _SCREAMING_SNAKE_CASE = in_proj_bias[: config.hidden_size] _SCREAMING_SNAKE_CASE = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _SCREAMING_SNAKE_CASE = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _SCREAMING_SNAKE_CASE = in_proj_weight[ -config.hidden_size :, : ] _SCREAMING_SNAKE_CASE = in_proj_bias[-config.hidden_size :] def __lowerCamelCase ( snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(snake_case__ ,snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = dct.pop(snake_case__ ) _SCREAMING_SNAKE_CASE = val def __lowerCamelCase ( ) -> List[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = """http://images.cocodataset.org/val2017/000000039769.jpg""" _SCREAMING_SNAKE_CASE = Image.open(requests.get(snake_case__ ,stream=snake_case__ ).raw ) return im @torch.no_grad() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = ViTConfig() _SCREAMING_SNAKE_CASE = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = int(vit_name[-12:-10] ) _SCREAMING_SNAKE_CASE = int(vit_name[-9:-6] ) else: _SCREAMING_SNAKE_CASE = 10_00 _SCREAMING_SNAKE_CASE = """huggingface/label-files""" _SCREAMING_SNAKE_CASE = """imagenet-1k-id2label.json""" _SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(snake_case__ ,snake_case__ ,repo_type="""dataset""" ) ,"""r""" ) ) _SCREAMING_SNAKE_CASE = {int(snake_case__ ): v for k, v in idalabel.items()} _SCREAMING_SNAKE_CASE = idalabel _SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} _SCREAMING_SNAKE_CASE = int(vit_name[-6:-4] ) _SCREAMING_SNAKE_CASE = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith("""tiny""" ): _SCREAMING_SNAKE_CASE = 1_92 _SCREAMING_SNAKE_CASE = 7_68 _SCREAMING_SNAKE_CASE = 12 _SCREAMING_SNAKE_CASE = 3 elif vit_name[9:].startswith("""small""" ): _SCREAMING_SNAKE_CASE = 3_84 _SCREAMING_SNAKE_CASE = 15_36 _SCREAMING_SNAKE_CASE = 12 _SCREAMING_SNAKE_CASE = 6 else: pass else: if vit_name[4:].startswith("""small""" ): _SCREAMING_SNAKE_CASE = 7_68 _SCREAMING_SNAKE_CASE = 23_04 _SCREAMING_SNAKE_CASE = 8 _SCREAMING_SNAKE_CASE = 8 elif vit_name[4:].startswith("""base""" ): pass elif vit_name[4:].startswith("""large""" ): _SCREAMING_SNAKE_CASE = 10_24 _SCREAMING_SNAKE_CASE = 40_96 _SCREAMING_SNAKE_CASE = 24 _SCREAMING_SNAKE_CASE = 16 elif vit_name[4:].startswith("""huge""" ): _SCREAMING_SNAKE_CASE = 12_80 _SCREAMING_SNAKE_CASE = 51_20 _SCREAMING_SNAKE_CASE = 32 _SCREAMING_SNAKE_CASE = 16 # load original model from timm _SCREAMING_SNAKE_CASE = timm.create_model(snake_case__ ,pretrained=snake_case__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys _SCREAMING_SNAKE_CASE = timm_model.state_dict() if base_model: remove_classification_head_(snake_case__ ) _SCREAMING_SNAKE_CASE = create_rename_keys(snake_case__ ,snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ ,snake_case__ ,snake_case__ ) read_in_q_k_v(snake_case__ ,snake_case__ ,snake_case__ ) # load HuggingFace model if vit_name[-5:] == "in21k": _SCREAMING_SNAKE_CASE = ViTModel(snake_case__ ).eval() else: _SCREAMING_SNAKE_CASE = ViTForImageClassification(snake_case__ ).eval() model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: _SCREAMING_SNAKE_CASE = DeiTImageProcessor(size=config.image_size ) else: _SCREAMING_SNAKE_CASE = ViTImageProcessor(size=config.image_size ) _SCREAMING_SNAKE_CASE = image_processor(images=prepare_img() ,return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = encoding["""pixel_values"""] _SCREAMING_SNAKE_CASE = model(snake_case__ ) if base_model: _SCREAMING_SNAKE_CASE = timm_model.forward_features(snake_case__ ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(snake_case__ ,outputs.pooler_output ,atol=1e-3 ) else: _SCREAMING_SNAKE_CASE = timm_model(snake_case__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(snake_case__ ,outputs.logits ,atol=1e-3 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F'Saving model {vit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case__ ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--vit_name''', default='''vit_base_patch16_224''', type=str, help='''Name of the ViT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) UpperCamelCase = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
306
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = weights[0][0][0] _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[0] ) _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # lsh weights + output _SCREAMING_SNAKE_CASE = weights[0][1] if len(snake_case__ ) < 4: set_layer_weights_in_torch_lsh(snake_case__ ,torch_block.attention ,snake_case__ ) else: set_layer_weights_in_torch_local(snake_case__ ,torch_block.attention ,snake_case__ ) # intermediate weighs _SCREAMING_SNAKE_CASE = weights[2][0][1][2] # Chunked Feed Forward if len(snake_case__ ) == 4: _SCREAMING_SNAKE_CASE = intermediate_weights[2] # layernorm 2 _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # intermediate dense _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) # intermediate out _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = torch_model.reformer # word embeds _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings ,torch.tensor(snake_case__ ) ,) if isinstance(weights[3] ,snake_case__ ): _SCREAMING_SNAKE_CASE = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _SCREAMING_SNAKE_CASE = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(torch.tensor(snake_case__ ) ) _SCREAMING_SNAKE_CASE = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( snake_case__ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _SCREAMING_SNAKE_CASE = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(snake_case__ ,snake_case__ ,snake_case__ ) # output layer norm _SCREAMING_SNAKE_CASE = np.asarray(weights[7][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # output embeddings _SCREAMING_SNAKE_CASE = np.asarray(weights[9][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ReformerConfig.from_json_file(snake_case__ ) print(F'Building PyTorch model from configuration: {config}' ) _SCREAMING_SNAKE_CASE = ReformerModelWithLMHead(snake_case__ ) with open(snake_case__ ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = pickle.load(snake_case__ )["""weights"""] set_model_weights_in_torch(snake_case__ ,snake_case__ ,config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() ,snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
306
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def __lowerCamelCase ( snake_case__ ,snake_case__=False ,snake_case__=False ) -> List[str]: """simple docstring""" _SCREAMING_SNAKE_CASE = """backbone.""" if is_semantic else """""" _SCREAMING_SNAKE_CASE = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'{prefix}blocks.{i}.norm1.weight', F'beit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'{prefix}blocks.{i}.norm1.bias', F'beit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append( (F'{prefix}blocks.{i}.attn.proj.weight', F'beit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append( (F'{prefix}blocks.{i}.attn.proj.bias', F'beit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((F'{prefix}blocks.{i}.norm2.weight', F'beit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'{prefix}blocks.{i}.norm2.bias', F'beit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc1.weight', F'beit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc1.bias', F'beit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc2.weight', F'beit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc2.bias', F'beit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ (F'{prefix}cls_token', """beit.embeddings.cls_token"""), (F'{prefix}patch_embed.proj.weight', """beit.embeddings.patch_embeddings.projection.weight"""), (F'{prefix}patch_embed.proj.bias', """beit.embeddings.patch_embeddings.projection.bias"""), (F'{prefix}pos_embed', """beit.embeddings.position_embeddings"""), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("""mask_token""", """beit.embeddings.mask_token"""), ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ] ) else: # layernorm + classification head rename_keys.extend( [ ("""fc_norm.weight""", """beit.pooler.layernorm.weight"""), ("""fc_norm.bias""", """beit.pooler.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=False ,snake_case__=False ) -> Optional[Any]: """simple docstring""" for i in range(config.num_hidden_layers ): _SCREAMING_SNAKE_CASE = """backbone.""" if is_semantic else """""" # queries, keys and values _SCREAMING_SNAKE_CASE = state_dict.pop(F'{prefix}blocks.{i}.attn.qkv.weight' ) _SCREAMING_SNAKE_CASE = state_dict.pop(F'{prefix}blocks.{i}.attn.q_bias' ) _SCREAMING_SNAKE_CASE = state_dict.pop(F'{prefix}blocks.{i}.attn.v_bias' ) _SCREAMING_SNAKE_CASE = in_proj_weight[ : config.hidden_size, : ] _SCREAMING_SNAKE_CASE = q_bias _SCREAMING_SNAKE_CASE = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _SCREAMING_SNAKE_CASE = in_proj_weight[ -config.hidden_size :, : ] _SCREAMING_SNAKE_CASE = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained _SCREAMING_SNAKE_CASE = state_dict.pop(F'{prefix}blocks.{i}.gamma_1' ) _SCREAMING_SNAKE_CASE = state_dict.pop(F'{prefix}blocks.{i}.gamma_2' ) _SCREAMING_SNAKE_CASE = gamma_a _SCREAMING_SNAKE_CASE = gamma_a def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = dct.pop(snake_case__ ) _SCREAMING_SNAKE_CASE = val def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = """http://images.cocodataset.org/val2017/000000039769.jpg""" _SCREAMING_SNAKE_CASE = Image.open(requests.get(snake_case__ ,stream=snake_case__ ).raw ) return im @torch.no_grad() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=False ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = False if """rvlcdip""" in checkpoint_url else True _SCREAMING_SNAKE_CASE = BeitConfig(use_absolute_position_embeddings=snake_case__ ,use_mask_token=snake_case__ ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: _SCREAMING_SNAKE_CASE = 10_24 _SCREAMING_SNAKE_CASE = 40_96 _SCREAMING_SNAKE_CASE = 24 _SCREAMING_SNAKE_CASE = 16 # labels if "rvlcdip" in checkpoint_url: _SCREAMING_SNAKE_CASE = 16 _SCREAMING_SNAKE_CASE = """huggingface/label-files""" _SCREAMING_SNAKE_CASE = """rvlcdip-id2label.json""" _SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(snake_case__ ,snake_case__ ,repo_type="""dataset""" ) ,"""r""" ) ) _SCREAMING_SNAKE_CASE = {int(snake_case__ ): v for k, v in idalabel.items()} _SCREAMING_SNAKE_CASE = idalabel _SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys _SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(snake_case__ ,map_location="""cpu""" )["""model"""] _SCREAMING_SNAKE_CASE = create_rename_keys(snake_case__ ,has_lm_head=snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ ,snake_case__ ,snake_case__ ) read_in_q_k_v(snake_case__ ,snake_case__ ,has_lm_head=snake_case__ ) # load HuggingFace model _SCREAMING_SNAKE_CASE = BeitForMaskedImageModeling(snake_case__ ) if has_lm_head else BeitForImageClassification(snake_case__ ) model.eval() model.load_state_dict(snake_case__ ) # Check outputs on an image _SCREAMING_SNAKE_CASE = BeitImageProcessor( size=config.image_size ,resample=PILImageResampling.BILINEAR ,do_center_crop=snake_case__ ) _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=snake_case__ ,return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = encoding["""pixel_values"""] _SCREAMING_SNAKE_CASE = model(snake_case__ ) _SCREAMING_SNAKE_CASE = outputs.logits # verify logits _SCREAMING_SNAKE_CASE = [1, 16] if """rvlcdip""" in checkpoint_url else [1, 1_96, 81_92] assert logits.shape == torch.Size(snake_case__ ), "Shape of logits not as expected" Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case__ ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: if has_lm_head: _SCREAMING_SNAKE_CASE = """dit-base""" if """base""" in checkpoint_url else """dit-large""" else: _SCREAMING_SNAKE_CASE = """dit-base-finetuned-rvlcdip""" if """dit-b""" in checkpoint_url else """dit-large-finetuned-rvlcdip""" image_processor.push_to_hub( repo_path_or_name=Path(snake_case__ ,snake_case__ ) ,organization="""nielsr""" ,commit_message="""Add image processor""" ,use_temp_dir=snake_case__ ,) model.push_to_hub( repo_path_or_name=Path(snake_case__ ,snake_case__ ) ,organization="""nielsr""" ,commit_message="""Add model""" ,use_temp_dir=snake_case__ ,) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_url''', default='''https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth''', type=str, help='''URL to the original PyTorch checkpoint (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) UpperCamelCase = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
306
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = TextToVideoSDPipeline __snake_case : Optional[int] = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __snake_case : Optional[int] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase ( self: int ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) _SCREAMING_SNAKE_CASE = CLIPTextModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict=0 ): '''simple docstring''' if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.get_dummy_inputs(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """np""" _SCREAMING_SNAKE_CASE = sd_pipe(**UpperCAmelCase_ ).frames _SCREAMING_SNAKE_CASE = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _SCREAMING_SNAKE_CASE = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=1E-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=25 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
306
1
import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __UpperCAmelCase : @staticmethod def UpperCamelCase ( *UpperCAmelCase_: Optional[Any] , **UpperCAmelCase_: str ): '''simple docstring''' pass @is_pipeline_test @require_torch @require_vision class __UpperCAmelCase (unittest.TestCase ): __snake_case : int = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: Dict , UpperCAmelCase_: Dict , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = pipeline("""visual-question-answering""" , model="""hf-internal-testing/tiny-vilt-random-vqa""" ) _SCREAMING_SNAKE_CASE = [ { """image""": Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ), """question""": """How many cats are there?""", }, { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """question""": """How many cats are there?""", }, ] return vqa_pipeline, examples def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = vqa_pipeline(UpperCAmelCase_ , top_k=1 ) self.assertEqual( UpperCAmelCase_ , [ [{"""score""": ANY(UpperCAmelCase_ ), """answer""": ANY(UpperCAmelCase_ )}], [{"""score""": ANY(UpperCAmelCase_ ), """answer""": ANY(UpperCAmelCase_ )}], ] , ) @require_torch def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = pipeline("""visual-question-answering""" , model="""hf-internal-testing/tiny-vilt-random-vqa""" ) _SCREAMING_SNAKE_CASE = """./tests/fixtures/tests_samples/COCO/000000039769.png""" _SCREAMING_SNAKE_CASE = """How many cats are there?""" _SCREAMING_SNAKE_CASE = vqa_pipeline(image=UpperCAmelCase_ , question="""How many cats are there?""" , top_k=2 ) self.assertEqual( UpperCAmelCase_ , [{"""score""": ANY(UpperCAmelCase_ ), """answer""": ANY(UpperCAmelCase_ )}, {"""score""": ANY(UpperCAmelCase_ ), """answer""": ANY(UpperCAmelCase_ )}] ) _SCREAMING_SNAKE_CASE = vqa_pipeline({"""image""": image, """question""": question} , top_k=2 ) self.assertEqual( UpperCAmelCase_ , [{"""score""": ANY(UpperCAmelCase_ ), """answer""": ANY(UpperCAmelCase_ )}, {"""score""": ANY(UpperCAmelCase_ ), """answer""": ANY(UpperCAmelCase_ )}] ) @slow @require_torch def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = pipeline("""visual-question-answering""" , model="""dandelin/vilt-b32-finetuned-vqa""" ) _SCREAMING_SNAKE_CASE = """./tests/fixtures/tests_samples/COCO/000000039769.png""" _SCREAMING_SNAKE_CASE = """How many cats are there?""" _SCREAMING_SNAKE_CASE = vqa_pipeline(image=UpperCAmelCase_ , question=UpperCAmelCase_ , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{"""score""": 0.87_99, """answer""": """2"""}, {"""score""": 0.2_96, """answer""": """1"""}] ) _SCREAMING_SNAKE_CASE = vqa_pipeline({"""image""": image, """question""": question} , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [{"""score""": 0.87_99, """answer""": """2"""}, {"""score""": 0.2_96, """answer""": """1"""}] ) _SCREAMING_SNAKE_CASE = vqa_pipeline( [{"""image""": image, """question""": question}, {"""image""": image, """question""": question}] , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase_ , decimals=4 ) , [[{"""score""": 0.87_99, """answer""": """2"""}, {"""score""": 0.2_96, """answer""": """1"""}]] * 2 , ) @require_tf @unittest.skip("""Visual question answering not implemented in TF""" ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' pass
306
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class __UpperCAmelCase : def __init__( self: Union[str, Any] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int=13 , UpperCAmelCase_: Optional[int]=7 , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=False , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Optional[int]=33 , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: List[Any]=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Any=37 , UpperCAmelCase_: Optional[Any]="gelu" , UpperCAmelCase_: Dict=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: int=16 , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: Tuple=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: str=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = EsmForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = False __snake_case : Dict = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __snake_case : List[Any] = () __snake_case : Dict = ( { "feature-extraction": EsmModel, "fill-mask": EsmForMaskedLM, "text-classification": EsmForSequenceClassification, "token-classification": EsmForTokenClassification, "zero-shot": EsmForSequenceClassification, } if is_torch_available() else {} ) __snake_case : int = True def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: int ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _SCREAMING_SNAKE_CASE = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _SCREAMING_SNAKE_CASE = create_position_ids_from_input_ids(UpperCAmelCase_ , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.empty(2 , 4 , 30 ) _SCREAMING_SNAKE_CASE = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _SCREAMING_SNAKE_CASE = torch.as_tensor([expected_single_positions, expected_single_positions] ) _SCREAMING_SNAKE_CASE = embeddings.create_position_ids_from_inputs_embeds(UpperCAmelCase_ ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Dict ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Any ): '''simple docstring''' pass @require_torch class __UpperCAmelCase (_UpperCAmelCase ): @slow def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[[8.92_15, -10.58_98, -6.46_71], [-6.39_67, -13.91_14, -1.12_12], [-7.78_12, -13.95_16, -3.74_06]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] # compare the actual values for a slice. _SCREAMING_SNAKE_CASE = torch.tensor( [[[0.14_44, 0.54_13, 0.32_48], [0.30_34, 0.00_53, 0.31_08], [0.32_28, -0.24_99, 0.34_15]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
import random def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = num - 1 _SCREAMING_SNAKE_CASE = 0 while s % 2 == 0: _SCREAMING_SNAKE_CASE = s // 2 t += 1 for _ in range(5 ): _SCREAMING_SNAKE_CASE = random.randrange(2 ,num - 1 ) _SCREAMING_SNAKE_CASE = pow(snake_case__ ,snake_case__ ,snake_case__ ) if v != 1: _SCREAMING_SNAKE_CASE = 0 while v != (num - 1): if i == t - 1: return False else: _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = (v**2) % num return True def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" if num < 2: return False _SCREAMING_SNAKE_CASE = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 1_01, 1_03, 1_07, 1_09, 1_13, 1_27, 1_31, 1_37, 1_39, 1_49, 1_51, 1_57, 1_63, 1_67, 1_73, 1_79, 1_81, 1_91, 1_93, 1_97, 1_99, 2_11, 2_23, 2_27, 2_29, 2_33, 2_39, 2_41, 2_51, 2_57, 2_63, 2_69, 2_71, 2_77, 2_81, 2_83, 2_93, 3_07, 3_11, 3_13, 3_17, 3_31, 3_37, 3_47, 3_49, 3_53, 3_59, 3_67, 3_73, 3_79, 3_83, 3_89, 3_97, 4_01, 4_09, 4_19, 4_21, 4_31, 4_33, 4_39, 4_43, 4_49, 4_57, 4_61, 4_63, 4_67, 4_79, 4_87, 4_91, 4_99, 5_03, 5_09, 5_21, 5_23, 5_41, 5_47, 5_57, 5_63, 5_69, 5_71, 5_77, 5_87, 5_93, 5_99, 6_01, 6_07, 6_13, 6_17, 6_19, 6_31, 6_41, 6_43, 6_47, 6_53, 6_59, 6_61, 6_73, 6_77, 6_83, 6_91, 7_01, 7_09, 7_19, 7_27, 7_33, 7_39, 7_43, 7_51, 7_57, 7_61, 7_69, 7_73, 7_87, 7_97, 8_09, 8_11, 8_21, 8_23, 8_27, 8_29, 8_39, 8_53, 8_57, 8_59, 8_63, 8_77, 8_81, 8_83, 8_87, 9_07, 9_11, 9_19, 9_29, 9_37, 9_41, 9_47, 9_53, 9_67, 9_71, 9_77, 9_83, 9_91, 9_97, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(snake_case__ ) def __lowerCamelCase ( snake_case__ = 10_24 ) -> int: """simple docstring""" while True: _SCREAMING_SNAKE_CASE = random.randrange(2 ** (keysize - 1) ,2 ** (keysize) ) if is_prime_low_num(snake_case__ ): return num if __name__ == "__main__": UpperCamelCase = generate_large_prime() print(('''Prime number:''', num)) print(('''is_prime_low_num:''', is_prime_low_num(num)))
306
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(_UpperCAmelCase ) ,"Tatoeba directory does not exist." ) class __UpperCAmelCase (unittest.TestCase ): @cached_property def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = tempfile.mkdtemp() return TatoebaConverter(save_dir=UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' self.resolver.convert_models(["""heb-eng"""] ) @slow def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.resolver.write_model_card("""opus-mt-he-en""" , dry_run=UpperCAmelCase_ ) assert mmeta["long_pair"] == "heb-eng"
306
def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _SCREAMING_SNAKE_CASE = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
306
1
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class __UpperCAmelCase : def __init__( self: Union[str, Any] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int=13 , UpperCAmelCase_: Optional[int]=7 , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=False , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Optional[int]=33 , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: List[Any]=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Any=37 , UpperCAmelCase_: Optional[Any]="gelu" , UpperCAmelCase_: Dict=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: int=16 , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: Tuple=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: str=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = EsmForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = False __snake_case : Dict = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __snake_case : List[Any] = () __snake_case : Dict = ( { "feature-extraction": EsmModel, "fill-mask": EsmForMaskedLM, "text-classification": EsmForSequenceClassification, "token-classification": EsmForTokenClassification, "zero-shot": EsmForSequenceClassification, } if is_torch_available() else {} ) __snake_case : int = True def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: int ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _SCREAMING_SNAKE_CASE = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _SCREAMING_SNAKE_CASE = create_position_ids_from_input_ids(UpperCAmelCase_ , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.empty(2 , 4 , 30 ) _SCREAMING_SNAKE_CASE = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _SCREAMING_SNAKE_CASE = torch.as_tensor([expected_single_positions, expected_single_positions] ) _SCREAMING_SNAKE_CASE = embeddings.create_position_ids_from_inputs_embeds(UpperCAmelCase_ ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Dict ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Any ): '''simple docstring''' pass @require_torch class __UpperCAmelCase (_UpperCAmelCase ): @slow def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[[8.92_15, -10.58_98, -6.46_71], [-6.39_67, -13.91_14, -1.12_12], [-7.78_12, -13.95_16, -3.74_06]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] # compare the actual values for a slice. _SCREAMING_SNAKE_CASE = torch.tensor( [[[0.14_44, 0.54_13, 0.32_48], [0.30_34, 0.00_53, 0.31_08], [0.32_28, -0.24_99, 0.34_15]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
1
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __UpperCAmelCase : def __init__( self: Optional[int] , UpperCAmelCase_: str , UpperCAmelCase_: str=13 , UpperCAmelCase_: Optional[int]=30 , UpperCAmelCase_: Tuple=2 , UpperCAmelCase_: Optional[Any]=3 , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: Any=5 , UpperCAmelCase_: List[Any]=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: Tuple="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: str=0.1 , UpperCAmelCase_: Union[str, Any]=10 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: List[str]=None , UpperCAmelCase_: List[str]=2 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = image_size _SCREAMING_SNAKE_CASE = patch_size _SCREAMING_SNAKE_CASE = num_channels _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = scope _SCREAMING_SNAKE_CASE = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _SCREAMING_SNAKE_CASE = (image_size // patch_size) ** 2 _SCREAMING_SNAKE_CASE = num_patches + 1 def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self: List[str] ): '''simple docstring''' return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCAmelCase_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: str , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTForMaskedImageModeling(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = ViTForMaskedImageModeling(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Dict , UpperCAmelCase_: int , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.type_sequence_label_size _SCREAMING_SNAKE_CASE = ViTForImageClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = ViTForImageClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : Any = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) __snake_case : Any = ( {"feature-extraction": ViTModel, "image-classification": ViTForImageClassification} if is_torch_available() else {} ) __snake_case : Optional[Any] = True __snake_case : Optional[int] = False __snake_case : List[Any] = False __snake_case : str = False def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , has_text_modality=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def UpperCamelCase ( self: str ): '''simple docstring''' pass def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _SCREAMING_SNAKE_CASE = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase_ , nn.Linear ) ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] _SCREAMING_SNAKE_CASE = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = ViTModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def __lowerCamelCase ( ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __UpperCAmelCase (unittest.TestCase ): @cached_property def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ).to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.default_image_processor _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(**UpperCAmelCase_ ) # verify the logits _SCREAMING_SNAKE_CASE = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor([-0.27_44, 0.82_15, -0.08_36] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTModel.from_pretrained("""facebook/dino-vits8""" ).to(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = ViTImageProcessor.from_pretrained("""facebook/dino-vits8""" , size=480 ) _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = inputs.pixel_values.to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , interpolate_pos_encoding=UpperCAmelCase_ ) # verify the logits _SCREAMING_SNAKE_CASE = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[4.23_40, 4.39_06, -6.66_92], [4.54_63, 1.89_28, -6.72_57], [4.44_29, 0.84_96, -5.85_85]] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ViTModel.from_pretrained("""facebook/dino-vits8""" , torch_dtype=torch.floataa , device_map="""auto""" ) _SCREAMING_SNAKE_CASE = self.default_image_processor _SCREAMING_SNAKE_CASE = prepare_img() _SCREAMING_SNAKE_CASE = image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = inputs.pixel_values.to(UpperCAmelCase_ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )
306
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: UpperCamelCase = None UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1_024, '''facebook/nllb-200-distilled-600M''': 1_024, } # fmt: off UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : List[str] = VOCAB_FILES_NAMES __snake_case : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Tuple = ["input_ids", "attention_mask"] __snake_case : Dict = NllbTokenizer __snake_case : List[int] = [] __snake_case : List[int] = [] def __init__( self: Tuple , UpperCAmelCase_: str=None , UpperCAmelCase_: List[str]=None , UpperCAmelCase_: Tuple="<s>" , UpperCAmelCase_: str="</s>" , UpperCAmelCase_: Union[str, Any]="</s>" , UpperCAmelCase_: int="<s>" , UpperCAmelCase_: Union[str, Any]="<unk>" , UpperCAmelCase_: Union[str, Any]="<pad>" , UpperCAmelCase_: str="<mask>" , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int=None , UpperCAmelCase_: str=False , **UpperCAmelCase_: int , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token _SCREAMING_SNAKE_CASE = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , src_lang=UpperCAmelCase_ , tgt_lang=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , legacy_behaviour=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = vocab_file _SCREAMING_SNAKE_CASE = False if not self.vocab_file else True _SCREAMING_SNAKE_CASE = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"""additional_special_tokens""": _additional_special_tokens} ) _SCREAMING_SNAKE_CASE = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase_ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } _SCREAMING_SNAKE_CASE = src_lang if src_lang is not None else """eng_Latn""" _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(self._src_lang ) _SCREAMING_SNAKE_CASE = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCamelCase ( self: int ): '''simple docstring''' return self._src_lang @src_lang.setter def UpperCamelCase ( self: int , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase ( self: List[str] , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase ( self: Tuple , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] , UpperCAmelCase_: Optional[str] , **UpperCAmelCase_: Any ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = self(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tgt_lang_id return inputs def UpperCamelCase ( self: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str = "eng_Latn" , UpperCAmelCase_: Optional[List[str]] = None , UpperCAmelCase_: str = "fra_Latn" , **UpperCAmelCase_: List[str] , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCAmelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory.' ) return _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) return (out_vocab_file,)
306
1
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class __UpperCAmelCase : __snake_case : Dict = None def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_dict ) _SCREAMING_SNAKE_CASE = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _SCREAMING_SNAKE_CASE = os.path.join(UpperCAmelCase_ , """feat_extract.json""" ) feat_extract_first.to_json_file(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.feature_extraction_class.from_json_file(UpperCAmelCase_ ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _SCREAMING_SNAKE_CASE = feat_extract_first.save_pretrained(UpperCAmelCase_ )[0] check_json_file_has_correct_format(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.feature_extraction_class.from_pretrained(UpperCAmelCase_ ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.feature_extraction_class() self.assertIsNotNone(UpperCAmelCase_ )
306
def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> list: """simple docstring""" _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = [[0] * n for i in range(snake_case__ )] for i in range(snake_case__ ): _SCREAMING_SNAKE_CASE = y_points[i] for i in range(2 ,snake_case__ ): for j in range(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
306
1
import string def __lowerCamelCase ( snake_case__ ) -> None: """simple docstring""" for key in range(len(string.ascii_uppercase ) ): _SCREAMING_SNAKE_CASE = """""" for symbol in message: if symbol in string.ascii_uppercase: _SCREAMING_SNAKE_CASE = string.ascii_uppercase.find(snake_case__ ) _SCREAMING_SNAKE_CASE = num - key if num < 0: _SCREAMING_SNAKE_CASE = num + len(string.ascii_uppercase ) _SCREAMING_SNAKE_CASE = translated + string.ascii_uppercase[num] else: _SCREAMING_SNAKE_CASE = translated + symbol print(F'Decryption using Key #{key}: {translated}' ) def __lowerCamelCase ( ) -> None: """simple docstring""" _SCREAMING_SNAKE_CASE = input("""Encrypted message: """ ) _SCREAMING_SNAKE_CASE = message.upper() decrypt(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
306
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCamelCase = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
1
def _a ( a :int ) -> int: if divisor % 5 == 0 or divisor % 2 == 0: return 0 a = 1 a = 1 while repunit: a = (10 * repunit + 1) % divisor repunit_index += 1 return repunit_index def _a ( a :int = 1_000_000 ) -> int: a = limit - 1 if divisor % 2 == 0: divisor += 1 while least_divisible_repunit(a ) <= limit: divisor += 2 return divisor if __name__ == "__main__": print(f"""{solution() = }""")
0
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class __UpperCAmelCase : __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : int __snake_case : int __snake_case : float __snake_case : float __snake_case : Tuple[int] def UpperCamelCase ( self: str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = torch.arange(self.height * self.width ) _SCREAMING_SNAKE_CASE = torch.stack( [ pixel_indices % self.width, torch.div(UpperCAmelCase_ , self.width , rounding_mode="""trunc""" ), ] , axis=1 , ) return coords @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE = self.shape _SCREAMING_SNAKE_CASE = int(np.prod(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = self.get_image_coords() _SCREAMING_SNAKE_CASE = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) _SCREAMING_SNAKE_CASE = self.get_camera_rays(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = rays.view(UpperCAmelCase_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCamelCase ( self: Any , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] _SCREAMING_SNAKE_CASE = coords.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = self.resolution() _SCREAMING_SNAKE_CASE = self.fov() _SCREAMING_SNAKE_CASE = (flat.float() / (res - 1)) * 2 - 1 _SCREAMING_SNAKE_CASE = fracs * torch.tan(fov / 2 ) _SCREAMING_SNAKE_CASE = fracs.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = ( self.z.view(UpperCAmelCase_ , 1 , 3 ) + self.x.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, 1:] ) _SCREAMING_SNAKE_CASE = directions / directions.norm(dim=-1 , keepdim=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.stack( [ torch.broadcast_to(self.origin.view(UpperCAmelCase_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(UpperCAmelCase_ , *UpperCAmelCase_ , 2 , 3 ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=UpperCAmelCase_ , height=UpperCAmelCase_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __lowerCamelCase ( snake_case__ ) -> DifferentiableProjectiveCamera: """simple docstring""" _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for theta in np.linspace(0 ,2 * np.pi ,num=20 ): _SCREAMING_SNAKE_CASE = np.array([np.sin(snake_case__ ), np.cos(snake_case__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) _SCREAMING_SNAKE_CASE = -z * 4 _SCREAMING_SNAKE_CASE = np.array([np.cos(snake_case__ ), -np.sin(snake_case__ ), 0.0] ) _SCREAMING_SNAKE_CASE = np.cross(snake_case__ ,snake_case__ ) origins.append(snake_case__ ) xs.append(snake_case__ ) ys.append(snake_case__ ) zs.append(snake_case__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,x=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,y=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,z=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,width=snake_case__ ,height=snake_case__ ,x_fov=0.7 ,y_fov=0.7 ,shape=(1, len(snake_case__ )) ,)
306
0
'''simple docstring''' import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class __A : def __init__(self : str , __a : Optional[Any] , __a : Union[str, Any]=14 , __a : List[str]=7 , __a : Union[str, Any]=True , __a : int=True , __a : List[str]=True , __a : str=True , __a : str=True , __a : List[str]=99 , __a : Union[str, Any]=32 , __a : Optional[Any]=5 , __a : int=4 , __a : List[Any]=37 , __a : Optional[Any]="gelu" , __a : Tuple=0.1 , __a : Any=0.1 , __a : int=512 , __a : int=16 , __a : Optional[int]=2 , __a : Any=0.02 , __a : str=3 , __a : Optional[Any]=4 , __a : Optional[Any]=None , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = seq_length UpperCAmelCase_ = is_training UpperCAmelCase_ = use_token_type_ids UpperCAmelCase_ = use_input_mask UpperCAmelCase_ = use_labels UpperCAmelCase_ = use_mc_token_ids UpperCAmelCase_ = vocab_size UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = type_vocab_size UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = num_labels UpperCAmelCase_ = num_choices UpperCAmelCase_ = scope UpperCAmelCase_ = self.vocab_size - 1 def _lowercase (self : Optional[int] ): UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ = None if self.use_input_mask: UpperCAmelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ = None if self.use_token_type_ids: UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase_ = None if self.use_mc_token_ids: UpperCAmelCase_ = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) UpperCAmelCase_ = None UpperCAmelCase_ = None UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase_ = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase_ = self.get_config() UpperCAmelCase_ = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def _lowercase (self : Optional[int] ): return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def _lowercase (self : Dict , __a : List[str] , __a : int , __a : Tuple , __a : int , __a : Union[str, Any] , *__a : List[Any] ): UpperCAmelCase_ = CTRLModel(config=__a ) model.to(__a ) model.eval() model(__a , token_type_ids=__a , head_mask=__a ) model(__a , token_type_ids=__a ) UpperCAmelCase_ = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def _lowercase (self : List[Any] , __a : Tuple , __a : Any , __a : Tuple , __a : str , __a : Tuple , *__a : str ): UpperCAmelCase_ = CTRLLMHeadModel(__a ) model.to(__a ) model.eval() UpperCAmelCase_ = model(__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase (self : List[str] ): UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "head_mask": head_mask} return config, inputs_dict def _lowercase (self : Tuple , __a : str , __a : Union[str, Any] , __a : List[str] , __a : Optional[Any] , *__a : Optional[int] ): UpperCAmelCase_ = self.num_labels UpperCAmelCase_ = CTRLForSequenceClassification(__a ) model.to(__a ) model.eval() UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = model(__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class __A ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): a__ : Dict = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () a__ : Union[str, Any] = (CTRLLMHeadModel,) if is_torch_available() else () a__ : List[Any] = ( { """feature-extraction""": CTRLModel, """text-classification""": CTRLForSequenceClassification, """text-generation""": CTRLLMHeadModel, """zero-shot""": CTRLForSequenceClassification, } if is_torch_available() else {} ) a__ : Any = True a__ : Union[str, Any] = False a__ : Dict = False def _lowercase (self : List[Any] , __a : int , __a : Optional[int] , __a : Dict , __a : int , __a : List[Any] ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def _lowercase (self : Any ): UpperCAmelCase_ = CTRLModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=__a , n_embd=37 ) def _lowercase (self : Union[str, Any] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def _lowercase (self : Optional[int] ): self.config_tester.run_common_tests() def _lowercase (self : Union[str, Any] ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*__a ) def _lowercase (self : str ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*__a ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def _lowercase (self : Optional[int] ): pass @slow def _lowercase (self : Union[str, Any] ): for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = CTRLModel.from_pretrained(__a ) self.assertIsNotNone(__a ) @unittest.skip("The model doesn't support left padding" ) # and it's not used enough to be worth fixing :) def _lowercase (self : List[str] ): pass @require_torch class __A ( unittest.TestCase ): def _lowercase (self : Optional[Any] ): super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def _lowercase (self : Optional[int] ): UpperCAmelCase_ = CTRLLMHeadModel.from_pretrained("ctrl" ) model.to(__a ) UpperCAmelCase_ = torch.tensor( [[11859, 0, 1611, 8]] , dtype=torch.long , device=__a ) # Legal the president is UpperCAmelCase_ = [ 11859, 0, 1611, 8, 5, 150, 26449, 2, 19, 348, 469, 3, 2595, 48, 20740, 246533, 246533, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a UpperCAmelCase_ = model.generate(__a , do_sample=__a ) self.assertListEqual(output_ids[0].tolist() , __a )
1
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __UpperCAmelCase (unittest.TestCase ): def __init__( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any]=13 , UpperCAmelCase_: List[str]=7 , UpperCAmelCase_: Tuple=True , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: str=99 , UpperCAmelCase_: List[Any]=32 , UpperCAmelCase_: Dict=5 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: Optional[int]="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: List[Any]=512 , UpperCAmelCase_: Any=16 , UpperCAmelCase_: Dict=2 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Union[str, Any]=4 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_attention_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_choices def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_attention_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=UpperCAmelCase_ , ) return config, input_ids, attention_mask def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Optional[int] = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase ( self: List[str] ): '''simple docstring''' for model_class_name in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCAmelCase_ ) @require_flax class __UpperCAmelCase (unittest.TestCase ): @slow def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 ) )
306
0
'''simple docstring''' import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate lowerCamelCase : str = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataRow('', '|', '|'), datarow=DataRow('', '|', '|'), padding=1, with_header_hide=None, ) lowerCamelCase : int = [] lowerCamelCase : Optional[int] = [] lowerCamelCase : Tuple = {'type': 'section', 'text': {'type': 'plain_text', 'text': 'No failed tests! 🤗', 'emoji': True}} lowerCamelCase : Union[str, Any] = [ { 'type': 'header', 'text': { 'type': 'plain_text', 'text': f"""🤗 Accelerate nightly {os.environ.get('TEST_TYPE', '')} test results""", 'emoji': True, }, } ] lowerCamelCase : Any = 0 for log in Path().glob('*.log'): lowerCamelCase : int = 0 with open(log, 'r') as f: for line in f: lowerCamelCase : Optional[Any] = json.loads(line) if line.get('nodeid', '') != "": lowerCamelCase : Optional[int] = line['nodeid'] if line.get('duration', None) is not None: lowerCamelCase : Any = f"""{line['duration']:.4f}""" if line.get('outcome', '') == "failed": section_num_failed += 1 failed.append([test, duration, log.name.split('_')[0]]) total_num_failed += 1 group_info.append([str(log), section_num_failed, failed]) lowerCamelCase : Optional[int] = [] log.unlink() lowerCamelCase : Optional[int] = '' lowerCamelCase : int = [] if total_num_failed > 0: for name, num_failed, failed_tests in group_info: if num_failed > 0: if num_failed == 1: message += f"*{name[1:]}: {num_failed} failed test*\n" else: message += f"*{name[1:]}: {num_failed} failed tests*\n" lowerCamelCase : str = [] lowerCamelCase : List[str] = {} for test in failed_tests: lowerCamelCase : Dict = test[0].split('::') lowerCamelCase : Optional[int] = data[0].split('/')[-1] if data[0] not in filesafailed: lowerCamelCase : Optional[Any] = [data[1:]] else: filesafailed[data[0]] += [data[1:]] failed_table.append(data) lowerCamelCase : Any = [test[0] for test in failed_table] lowerCamelCase : List[Any] = list(set(files)) # Count number of instances in failed_tests lowerCamelCase : List[str] = [] for file in individual_files: table.append([file, len(filesafailed[file])]) lowerCamelCase : str = tabulate( table, headers=['Test Location', 'Num Failed'], tablefmt=hf_table_format, stralign='right', ) message += f"\n```\n{failed_table}\n```" all_filesafailed.append(filesafailed) if len(message) > 3_000: lowerCamelCase : Dict = 'Too many failed tests, please see the full report in the Action results.' lowerCamelCase : List[Any] = len(err) + 10 lowerCamelCase : List[Any] = message[: 3_000 - offset] + f"""\n...\n```\n{err}""" print(f"""### {message}""") else: lowerCamelCase : Optional[int] = 'No failed tests! 🤗' print(f"""## {message}""") payload.append(no_error_payload) if os.environ.get('TEST_TYPE', '') != "": from slack_sdk import WebClient lowerCamelCase : Optional[int] = WebClient(token=os.environ['SLACK_API_TOKEN']) if message != "No failed tests! 🤗": lowerCamelCase : str = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': message, }, } payload.append(md_report) lowerCamelCase : Dict = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': '*For more details:*', }, 'accessory': { 'type': 'button', 'text': { 'type': 'plain_text', 'text': 'Check Action results', 'emoji': True, }, 'url': f"""https://github.com/{os.environ['GITHUB_REPOSITORY']}/actions/runs/{os.environ['GITHUB_RUN_ID']}""", }, } payload.append(action_button) lowerCamelCase : Tuple = { 'type': 'context', 'elements': [ { 'type': 'plain_text', 'text': f"""Nightly {os.environ.get('TEST_TYPE')} test results for {date.today()}""", } ], } payload.append(date_report) lowerCamelCase : Any = client.chat_postMessage(channel='#accelerate-ci-daily', text=message, blocks=payload) lowerCamelCase : Tuple = response.data['ts'] for failed_file in all_filesafailed: for test_location, test_failures in failed_file.items(): # Keep only the first instance of the test name lowerCamelCase : Any = '' for i, row in enumerate(test_failures): if row[0] != test_class: lowerCamelCase : Optional[Any] = row[0] else: lowerCamelCase : int = '' lowerCamelCase : Optional[int] = { 'type': 'section', 'text': { 'type': 'mrkdwn', 'text': f"""Test location: {test_location}\n```\n{tabulate(test_failures, headers=['Class', 'Test'], tablefmt=hf_table_format, stralign='right')}\n```""", }, } client.chat_postMessage( channel='#accelerate-ci-daily', thread_ts=ts, blocks=[payload], )
2
import argparse import hashlib # hashlib is only used inside the Test class import struct class __UpperCAmelCase : def __init__( self: List[str] , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = data _SCREAMING_SNAKE_CASE = [0x67_452_301, 0xef_cda_b89, 0x98_bad_cfe, 0x10_325_476, 0xc3_d2e_1f0] @staticmethod def UpperCamelCase ( UpperCAmelCase_: int , UpperCAmelCase_: List[str] ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0xff_fff_fff def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _SCREAMING_SNAKE_CASE = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = list(struct.unpack(""">16L""" , UpperCAmelCase_ ) ) + [0] * 64 for i in range(16 , 80 ): _SCREAMING_SNAKE_CASE = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.padding() _SCREAMING_SNAKE_CASE = self.split_blocks() for block in self.blocks: _SCREAMING_SNAKE_CASE = self.expand_block(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.h for i in range(0 , 80 ): if 0 <= i < 20: _SCREAMING_SNAKE_CASE = (b & c) | ((~b) & d) _SCREAMING_SNAKE_CASE = 0x5a_827_999 elif 20 <= i < 40: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0x6e_d9e_ba1 elif 40 <= i < 60: _SCREAMING_SNAKE_CASE = (b & c) | (b & d) | (c & d) _SCREAMING_SNAKE_CASE = 0x8f_1bb_cdc elif 60 <= i < 80: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0xca_62c_1d6 _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ( self.rotate(UpperCAmelCase_ , 5 ) + f + e + k + expanded_block[i] & 0xff_fff_fff, a, self.rotate(UpperCAmelCase_ , 30 ), c, d, ) _SCREAMING_SNAKE_CASE = ( self.h[0] + a & 0xff_fff_fff, self.h[1] + b & 0xff_fff_fff, self.h[2] + c & 0xff_fff_fff, self.h[3] + d & 0xff_fff_fff, self.h[4] + e & 0xff_fff_fff, ) return ("{:08x}" * 5).format(*self.h ) def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = b"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def __lowerCamelCase ( ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" ,dest="""input_string""" ,default="""Hello World!! Welcome to Cryptography""" ,help="""Hash the string""" ,) parser.add_argument("""--file""" ,dest="""input_file""" ,help="""Hash contents of a file""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = f.read() else: _SCREAMING_SNAKE_CASE = bytes(snake_case__ ,"""utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
306
0
'''simple docstring''' import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=100 , SCREAMING_SNAKE_CASE=13 , SCREAMING_SNAKE_CASE=30 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE=3 , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=32 , SCREAMING_SNAKE_CASE=4 , SCREAMING_SNAKE_CASE=4 , SCREAMING_SNAKE_CASE=37 , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=10 , SCREAMING_SNAKE_CASE=0.02 , SCREAMING_SNAKE_CASE=3 , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=[0, 1, 2, 3] , ) -> Union[str, Any]: """simple docstring""" A : Union[str, Any] = parent A : Optional[int] = 100 A : Optional[int] = batch_size A : Any = image_size A : Tuple = patch_size A : Union[str, Any] = num_channels A : str = is_training A : Dict = use_labels A : Any = hidden_size A : Dict = num_hidden_layers A : List[Any] = num_attention_heads A : Dict = intermediate_size A : Tuple = hidden_act A : Union[str, Any] = hidden_dropout_prob A : Any = attention_probs_dropout_prob A : List[Any] = type_sequence_label_size A : Dict = initializer_range A : int = scope A : Any = out_indices A : Any = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) A : List[Any] = (image_size // patch_size) ** 2 A : Union[str, Any] = num_patches + 1 def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A : Tuple = None A : Any = None if self.use_labels: A : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A : Optional[int] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) A : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self ) -> str: """simple docstring""" return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" A : Union[str, Any] = BeitModel(config=SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : str = model(SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" A : List[Any] = BeitForMaskedImageModeling(config=SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : Dict = model(SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" A : int = self.type_sequence_label_size A : Dict = BeitForImageClassification(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : List[Any] = model(SCREAMING_SNAKE_CASE , labels=SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A : Dict = 1 A : Optional[Any] = BeitForImageClassification(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A : Dict = model(SCREAMING_SNAKE_CASE , labels=SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" A : List[str] = self.num_labels A : int = BeitForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() A : Optional[int] = model(SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) A : List[Any] = model(SCREAMING_SNAKE_CASE , labels=SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Optional[int] = self.prepare_config_and_inputs() A, A, A, A : int = config_and_inputs A : Union[str, Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class A ( __snake_case , __snake_case , unittest.TestCase ): __magic_name__ = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) __magic_name__ = ( { '''feature-extraction''': BeitModel, '''image-classification''': BeitForImageClassification, '''image-segmentation''': BeitForSemanticSegmentation, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : str = BeitModelTester(self ) A : List[Any] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE , has_text_modality=SCREAMING_SNAKE_CASE , hidden_size=37 ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''BEiT does not use inputs_embeds''' ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" pass @require_torch_multi_gpu @unittest.skip(reason='''BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" pass def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A, A : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A : Dict = model_class(SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(SCREAMING_SNAKE_CASE , nn.Linear ) ) def __lowerCAmelCase ( self ) -> int: """simple docstring""" A, A : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A : Dict = model_class(SCREAMING_SNAKE_CASE ) A : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A : Any = [*signature.parameters.keys()] A : Tuple = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" A : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> int: """simple docstring""" A : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" A : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" if not self.model_tester.is_training: return A, A : Dict = self.model_tester.prepare_config_and_inputs_for_common() A : Optional[Any] = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(SCREAMING_SNAKE_CASE ), BeitForMaskedImageModeling]: continue A : Optional[int] = model_class(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.train() A : Union[str, Any] = self._prepare_for_class(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , return_labels=SCREAMING_SNAKE_CASE ) A : List[str] = model(**SCREAMING_SNAKE_CASE ).loss loss.backward() def __lowerCAmelCase ( self ) -> str: """simple docstring""" A, A : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A : Union[str, Any] = False A : Tuple = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(SCREAMING_SNAKE_CASE ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue A : int = model_class(SCREAMING_SNAKE_CASE ) model.gradient_checkpointing_enable() model.to(SCREAMING_SNAKE_CASE ) model.train() A : Dict = self._prepare_for_class(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , return_labels=SCREAMING_SNAKE_CASE ) A : Dict = model(**SCREAMING_SNAKE_CASE ).loss loss.backward() def __lowerCAmelCase ( self ) -> int: """simple docstring""" A, A : int = self.model_tester.prepare_config_and_inputs_for_common() A : List[Any] = _config_zero_init(SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: A : List[str] = model_class(config=SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @slow def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A : int = BeitModel.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsNotNone(SCREAMING_SNAKE_CASE ) def lowerCAmelCase_ ( ): '''simple docstring''' A : int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class A ( unittest.TestCase ): @cached_property def __lowerCAmelCase ( self ) -> int: """simple docstring""" return BeitImageProcessor.from_pretrained('''microsoft/beit-base-patch16-224''' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : Tuple = BeitForMaskedImageModeling.from_pretrained('''microsoft/beit-base-patch16-224-pt22k''' ).to(SCREAMING_SNAKE_CASE ) A : int = self.default_image_processor A : Optional[int] = prepare_img() A : List[Any] = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).pixel_values.to(SCREAMING_SNAKE_CASE ) # prepare bool_masked_pos A : Optional[int] = torch.ones((1, 196) , dtype=torch.bool ).to(SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A : int = model(pixel_values=SCREAMING_SNAKE_CASE , bool_masked_pos=SCREAMING_SNAKE_CASE ) A : Optional[int] = outputs.logits # verify the logits A : str = torch.Size((1, 196, 8192) ) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE ) A : Union[str, Any] = torch.tensor( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ).to(SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , SCREAMING_SNAKE_CASE , atol=1e-2 ) ) @slow def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : List[Any] = BeitForImageClassification.from_pretrained('''microsoft/beit-base-patch16-224''' ).to(SCREAMING_SNAKE_CASE ) A : Tuple = self.default_image_processor A : Tuple = prepare_img() A : Any = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A : Tuple = model(**SCREAMING_SNAKE_CASE ) A : Optional[int] = outputs.logits # verify the logits A : List[str] = torch.Size((1, 1000) ) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE ) A : Optional[Any] = torch.tensor([-1.2_385, -1.0_987, -1.0_108] ).to(SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) ) A : List[Any] = 281 self.assertEqual(logits.argmax(-1 ).item() , SCREAMING_SNAKE_CASE ) @slow def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : List[str] = BeitForImageClassification.from_pretrained('''microsoft/beit-large-patch16-224-pt22k-ft22k''' ).to( SCREAMING_SNAKE_CASE ) A : Union[str, Any] = self.default_image_processor A : List[Any] = prepare_img() A : Optional[Any] = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A : str = model(**SCREAMING_SNAKE_CASE ) A : Dict = outputs.logits # verify the logits A : Union[str, Any] = torch.Size((1, 21841) ) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE ) A : List[Any] = torch.tensor([1.6_881, -0.2_787, 0.5_901] ).to(SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) ) A : str = 2396 self.assertEqual(logits.argmax(-1 ).item() , SCREAMING_SNAKE_CASE ) @slow def __lowerCAmelCase ( self ) -> List[Any]: """simple docstring""" A : Tuple = BeitForSemanticSegmentation.from_pretrained('''microsoft/beit-base-finetuned-ade-640-640''' ) A : Tuple = model.to(SCREAMING_SNAKE_CASE ) A : int = BeitImageProcessor(do_resize=SCREAMING_SNAKE_CASE , size=640 , do_center_crop=SCREAMING_SNAKE_CASE ) A : Union[str, Any] = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) A : List[Any] = Image.open(ds[0]['''file'''] ) A : Optional[Any] = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A : List[Any] = model(**SCREAMING_SNAKE_CASE ) A : Union[str, Any] = outputs.logits # verify the logits A : Union[str, Any] = torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE ) A : int = version.parse(PIL.__version__ ) < version.parse('''9.0.0''' ) if is_pillow_less_than_a: A : Any = torch.tensor( [ [[-4.9_225, -2.3_954, -3.0_522], [-2.8_822, -1.0_046, -1.7_561], [-2.9_549, -1.3_228, -2.1_347]], [[-5.8_168, -3.4_129, -4.0_778], [-3.8_651, -2.2_214, -3.0_277], [-3.8_356, -2.4_643, -3.3_535]], [[-0.0_078, 3.9_952, 4.0_754], [2.9_856, 4.6_944, 5.0_035], [3.2_413, 4.7_813, 4.9_969]], ] , device=SCREAMING_SNAKE_CASE , ) else: A : int = torch.tensor( [ [[-4.8_960, -2.3_688, -3.0_355], [-2.8_478, -0.9_836, -1.7_418], [-2.9_449, -1.3_332, -2.1_456]], [[-5.8_081, -3.4_124, -4.1_006], [-3.8_561, -2.2_081, -3.0_323], [-3.8_365, -2.4_601, -3.3_669]], [[-0.0_309, 3.9_868, 4.0_540], [2.9_640, 4.6_877, 4.9_976], [3.2_081, 4.7_690, 4.9_942]], ] , device=SCREAMING_SNAKE_CASE , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , SCREAMING_SNAKE_CASE , atol=1e-4 ) ) @slow def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : int = BeitForSemanticSegmentation.from_pretrained('''microsoft/beit-base-finetuned-ade-640-640''' ) A : Dict = model.to(SCREAMING_SNAKE_CASE ) A : Optional[int] = BeitImageProcessor(do_resize=SCREAMING_SNAKE_CASE , size=640 , do_center_crop=SCREAMING_SNAKE_CASE ) A : Optional[int] = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) A : Optional[int] = Image.open(ds[0]['''file'''] ) A : List[str] = image_processor(images=SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A : Optional[Any] = model(**SCREAMING_SNAKE_CASE ) A : Any = outputs.logits.detach().cpu() A : List[str] = image_processor.post_process_semantic_segmentation(outputs=SCREAMING_SNAKE_CASE , target_sizes=[(500, 300)] ) A : int = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , SCREAMING_SNAKE_CASE ) A : Dict = image_processor.post_process_semantic_segmentation(outputs=SCREAMING_SNAKE_CASE ) A : str = torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape , SCREAMING_SNAKE_CASE )
3
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Tuple = VOCAB_FILES_NAMES __snake_case : Optional[int] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Optional[Any] = ["input_ids", "attention_mask"] __snake_case : Optional[int] = None def __init__( self: Dict , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int="<unk>" , UpperCAmelCase_: List[str]="<s>" , UpperCAmelCase_: Tuple="</s>" , UpperCAmelCase_: List[Any]="<pad>" , UpperCAmelCase_: Dict=False , UpperCAmelCase_: Dict=False , **UpperCAmelCase_: Dict , ): '''simple docstring''' super().__init__( UpperCAmelCase_ , UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , add_prefix_space=UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCAmelCase_ ) != add_prefix_space: _SCREAMING_SNAKE_CASE = getattr(UpperCAmelCase_ , pre_tok_state.pop("""type""" ) ) _SCREAMING_SNAKE_CASE = add_prefix_space _SCREAMING_SNAKE_CASE = pre_tok_class(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = add_prefix_space def UpperCamelCase ( self: List[str] , *UpperCAmelCase_: Any , **UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Union[str, Any] , *UpperCAmelCase_: Dict , **UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_ ) return tuple(UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: "Conversation" ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) + [self.eos_token_id] ) if len(UpperCAmelCase_ ) > self.model_max_length: _SCREAMING_SNAKE_CASE = input_ids[-self.model_max_length :] return input_ids
306
0
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __snake_case ={ """configuration_efficientnet""": [ """EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientNetConfig""", """EfficientNetOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =["""EfficientNetImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =[ """EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientNetForImageClassification""", """EfficientNetModel""", """EfficientNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys __snake_case =_LazyModule(__name__, globals()["""__file__"""], _import_structure)
4
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __UpperCAmelCase : def __init__( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int]=13 , UpperCAmelCase_: str=7 , UpperCAmelCase_: int=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Dict=True , UpperCAmelCase_: Any=True , UpperCAmelCase_: Tuple=99 , UpperCAmelCase_: Optional[Any]=32 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Tuple=37 , UpperCAmelCase_: Union[str, Any]="gelu" , UpperCAmelCase_: List[str]=0.1 , UpperCAmelCase_: int=0.1 , UpperCAmelCase_: str=512 , UpperCAmelCase_: Union[str, Any]=16 , UpperCAmelCase_: List[Any]=2 , UpperCAmelCase_: str=0.02 , UpperCAmelCase_: int=False , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: Optional[Any]="None" , UpperCAmelCase_: Optional[int]=3 , UpperCAmelCase_: Any=4 , UpperCAmelCase_: Optional[int]=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = relative_attention _SCREAMING_SNAKE_CASE = position_biased_input _SCREAMING_SNAKE_CASE = pos_att_type _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=UpperCAmelCase_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _SCREAMING_SNAKE_CASE = [input_ids, input_mask] _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForMaskedLM(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Any , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForSequenceClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForTokenClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: Any , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForQuestionAnswering(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) __snake_case : Union[str, Any] = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) __snake_case : Dict = False __snake_case : Optional[Any] = False def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(UpperCAmelCase_ ) @require_tf class __UpperCAmelCase (unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = tf.constant( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 )
306
0
import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet import VOCAB_FILES_NAMES, ProphetNetTokenizer from transformers.testing_utils import require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase__ ( lowerCAmelCase , unittest.TestCase): SCREAMING_SNAKE_CASE__ = ProphetNetTokenizer SCREAMING_SNAKE_CASE__ = False def __A (self ) -> Optional[int]: super().setUp() _lowercase =[ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] _lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __A (self , UpperCAmelCase ) -> int: _lowercase ='''UNwant\u00E9d,running''' _lowercase ='''unwanted, running''' return input_text, output_text def __A (self ) -> Optional[int]: _lowercase =self.tokenizer_class(self.vocab_file ) _lowercase =tokenizer.tokenize('''UNwant\u00E9d,running''' ) self.assertListEqual(UpperCAmelCase , ['''un''', '''##want''', '''##ed''', ''',''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [9, 6, 7, 1_2, 1_0, 1_1] ) def __A (self ) -> Union[str, Any]: _lowercase =BasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) , ['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def __A (self ) -> List[str]: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def __A (self ) -> Dict: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase , strip_accents=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''h\u00E9llo'''] ) def __A (self ) -> List[str]: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase , strip_accents=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def __A (self ) -> Optional[int]: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def __A (self ) -> str: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def __A (self ) -> Union[str, Any]: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase , strip_accents=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def __A (self ) -> int: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase , strip_accents=UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def __A (self ) -> List[Any]: _lowercase =BasicTokenizer(do_lower_case=UpperCAmelCase , never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def __A (self ) -> List[str]: _lowercase =['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] _lowercase ={} for i, token in enumerate(UpperCAmelCase ): _lowercase =i _lowercase =WordpieceTokenizer(vocab=UpperCAmelCase , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) , ['''[UNK]''', '''runn''', '''##ing'''] ) @require_torch def __A (self ) -> Union[str, Any]: _lowercase =self.tokenizer_class.from_pretrained('''microsoft/prophetnet-large-uncased''' ) _lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] _lowercase =[1_0_3_7, 2_1_4_6, 2_0_4_2_3, 2_0_0_5, 7_6_8_0, 7_8_4_9, 3_9_8_9, 1_0_1_2, 1_0_2] _lowercase =tokenizer(UpperCAmelCase , padding=UpperCAmelCase , return_tensors='''pt''' ) self.assertIsInstance(UpperCAmelCase , UpperCAmelCase ) _lowercase =list(batch.input_ids.numpy()[0] ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) def __A (self ) -> int: self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def __A (self ) -> Optional[int]: self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def __A (self ) -> Any: self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) @slow def __A (self ) -> int: _lowercase =self.tokenizer_class.from_pretrained('''microsoft/prophetnet-large-uncased''' ) _lowercase =tokenizer.encode('''sequence builders''' , add_special_tokens=UpperCAmelCase ) _lowercase =tokenizer.encode('''multi-sequence build''' , add_special_tokens=UpperCAmelCase ) _lowercase =tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) _lowercase =tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == text + [1_0_2] assert encoded_pair == text + [1_0_2] + text_a + [1_0_2]
5
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __lowerCamelCase ( snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = gather(snake_case__ ) assert gathered_tensor.tolist() == list(range(1 ,state.num_processes**2 + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = [state.process_index] _SCREAMING_SNAKE_CASE = gather_object(snake_case__ ) assert len(snake_case__ ) == state.num_processes, F'{gathered_obj}, {len(snake_case__ )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = broadcast(snake_case__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 ,state.num_processes + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Tuple: """simple docstring""" if state.is_main_process: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes + 1 ).to(state.device ) else: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes ).to(state.device ) _SCREAMING_SNAKE_CASE = pad_across_processes(snake_case__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 ,state.num_processes ) ) + [0] def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""sum""" ) _SCREAMING_SNAKE_CASE = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""mean""" ) _SCREAMING_SNAKE_CASE = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> str: """simple docstring""" main() def __lowerCamelCase ( ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = PartialState() state.print(F'State: {state}' ) state.print("""testing gather""" ) test_gather(snake_case__ ) state.print("""testing gather_object""" ) test_gather_object(snake_case__ ) state.print("""testing broadcast""" ) test_broadcast(snake_case__ ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(snake_case__ ) state.print("""testing reduce_sum""" ) test_reduce_sum(snake_case__ ) state.print("""testing reduce_mean""" ) test_reduce_mean(snake_case__ ) if __name__ == "__main__": main()
306
0
import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class __A: def __init__( self , _snake_case=2 , _snake_case=3 , _snake_case=64 , _snake_case=None ) -> Union[str, Any]: '''simple docstring''' __a = np.random.default_rng(_snake_case ) __a = length __a = rng.normal(size=(length,) ).astype(np.floataa ) __a = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self ) -> Tuple: '''simple docstring''' return self.length def __getitem__( self , _snake_case ) -> Union[str, Any]: '''simple docstring''' return {"x": self.x[i], "y": self.y[i]} class __A( torch.nn.Module ): def __init__( self , _snake_case=0 , _snake_case=0 , _snake_case=False ) -> Any: '''simple docstring''' super().__init__() __a = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __a = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __a = True def SCREAMING_SNAKE_CASE_ ( self , _snake_case=None ) -> Tuple: '''simple docstring''' if self.first_batch: print(F"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) __a = False return x * self.a[0] + self.b[0] class __A( torch.nn.Module ): def __init__( self , _snake_case=0 , _snake_case=0 , _snake_case=False ) -> Dict: '''simple docstring''' super().__init__() __a = torch.nn.Parameter(torch.tensor(_snake_case ).float() ) __a = torch.nn.Parameter(torch.tensor(_snake_case ).float() ) __a = True def SCREAMING_SNAKE_CASE_ ( self , _snake_case=None ) -> List[Any]: '''simple docstring''' if self.first_batch: print(F"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) __a = False return x * self.a + self.b def __lowerCAmelCase ( a__ , a__ = 16 ) -> List[str]: from datasets import load_dataset from transformers import AutoTokenizer __a = AutoTokenizer.from_pretrained('''bert-base-cased''' ) __a = {'''train''': '''tests/test_samples/MRPC/train.csv''', '''validation''': '''tests/test_samples/MRPC/dev.csv'''} __a = load_dataset('''csv''' , data_files=a__ ) __a = datasets['''train'''].unique('''label''' ) __a = {v: i for i, v in enumerate(a__ )} def tokenize_function(a__ ): # max_length=None => use the model max length (it's actually the default) __a = tokenizer( examples['''sentence1'''] , examples['''sentence2'''] , truncation=a__ , max_length=a__ , padding='''max_length''' ) if "label" in examples: __a = [label_to_id[l] for l in examples['''label''']] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __a = datasets.map( a__ , batched=a__ , remove_columns=['''sentence1''', '''sentence2''', '''label'''] , ) def collate_fn(a__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(a__ , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(a__ , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __a = DataLoader(tokenized_datasets['''train'''] , shuffle=a__ , collate_fn=a__ , batch_size=2 ) __a = DataLoader(tokenized_datasets['''validation'''] , shuffle=a__ , collate_fn=a__ , batch_size=1 ) return train_dataloader, eval_dataloader
6
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __lowerCamelCase ( ) -> tuple[list[int], int]: """simple docstring""" _SCREAMING_SNAKE_CASE = [randint(-10_00 ,10_00 ) for i in range(10 )] _SCREAMING_SNAKE_CASE = randint(-50_00 ,50_00 ) return (arr, r) UpperCamelCase = make_dataset() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(snake_case__ ,3 ): if sum(snake_case__ ) == target: return tuple(sorted(snake_case__ ) ) return (0, 0, 0) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, int, int]: """simple docstring""" arr.sort() _SCREAMING_SNAKE_CASE = len(snake_case__ ) for i in range(n - 1 ): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __lowerCamelCase ( ) -> tuple[float, float]: """simple docstring""" _SCREAMING_SNAKE_CASE = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _SCREAMING_SNAKE_CASE = """ triplet_sum1(*dataset) """ _SCREAMING_SNAKE_CASE = """ triplet_sum2(*dataset) """ _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) return (min(snake_case__ ), min(snake_case__ )) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
306
0
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed lowercase_ = "true" def _snake_case( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[Any]=82 , SCREAMING_SNAKE_CASE__ : Optional[int]=16 ) -> Optional[Any]: '''simple docstring''' set_seed(42 ) A__ = RegressionModel() A__ = deepcopy(SCREAMING_SNAKE_CASE__ ) A__ = RegressionDataset(length=SCREAMING_SNAKE_CASE__ ) A__ = DataLoader(SCREAMING_SNAKE_CASE__ , batch_size=SCREAMING_SNAKE_CASE__ ) model.to(accelerator.device ) A__ , A__ = accelerator.prepare(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return model, ddp_model, dataloader def _snake_case( SCREAMING_SNAKE_CASE__ : Accelerator , SCREAMING_SNAKE_CASE__ : Tuple=False ) -> int: '''simple docstring''' A__ = AutoTokenizer.from_pretrained('hf-internal-testing/mrpc-bert-base-cased' ) A__ = load_dataset('glue' , 'mrpc' , split='validation' ) def tokenize_function(SCREAMING_SNAKE_CASE__ : List[Any] ): A__ = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=SCREAMING_SNAKE_CASE__ , max_length=SCREAMING_SNAKE_CASE__ ) return outputs with accelerator.main_process_first(): A__ = dataset.map( SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ , remove_columns=['idx', 'sentence1', 'sentence2'] , ) A__ = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(SCREAMING_SNAKE_CASE__ : Dict ): if use_longest: return tokenizer.pad(SCREAMING_SNAKE_CASE__ , padding='longest' , return_tensors='pt' ) return tokenizer.pad(SCREAMING_SNAKE_CASE__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return DataLoader(SCREAMING_SNAKE_CASE__ , shuffle=SCREAMING_SNAKE_CASE__ , collate_fn=SCREAMING_SNAKE_CASE__ , batch_size=16 ) def _snake_case( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Any ) -> str: '''simple docstring''' A__ = Accelerator(dispatch_batches=SCREAMING_SNAKE_CASE__ , split_batches=SCREAMING_SNAKE_CASE__ ) A__ = get_dataloader(SCREAMING_SNAKE_CASE__ , not dispatch_batches ) A__ = AutoModelForSequenceClassification.from_pretrained( 'hf-internal-testing/mrpc-bert-base-cased' , return_dict=SCREAMING_SNAKE_CASE__ ) A__ , A__ = accelerator.prepare(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _snake_case( SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[str]: '''simple docstring''' A__ = [] for batch in dataloader: A__ , A__ = batch.values() with torch.no_grad(): A__ = model(SCREAMING_SNAKE_CASE__ ) A__ , A__ = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) A__ , A__ = [], [] for logit, targ in logits_and_targets: logits.append(SCREAMING_SNAKE_CASE__ ) targs.append(SCREAMING_SNAKE_CASE__ ) A__ , A__ = torch.cat(SCREAMING_SNAKE_CASE__ ), torch.cat(SCREAMING_SNAKE_CASE__ ) return logits, targs def _snake_case( SCREAMING_SNAKE_CASE__ : Accelerator , SCREAMING_SNAKE_CASE__ : int=82 , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Any=False , SCREAMING_SNAKE_CASE__ : Tuple=16 ) -> List[Any]: '''simple docstring''' A__ , A__ , A__ = get_basic_setup(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) A__ , A__ = generate_predictions(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) assert ( len(SCREAMING_SNAKE_CASE__ ) == num_samples ), f'Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(SCREAMING_SNAKE_CASE__ )}' def _snake_case( SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False ) -> str: '''simple docstring''' A__ = evaluate.load('glue' , 'mrpc' ) A__ , A__ = get_mrpc_setup(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # First do baseline A__ , A__ , A__ = setup['no'] model.to(SCREAMING_SNAKE_CASE__ ) model.eval() for batch in dataloader: batch.to(SCREAMING_SNAKE_CASE__ ) with torch.inference_mode(): A__ = model(**SCREAMING_SNAKE_CASE__ ) A__ = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=SCREAMING_SNAKE_CASE__ , references=batch['labels'] ) A__ = metric.compute() # Then do distributed A__ , A__ , A__ = setup['ddp'] model.eval() for batch in dataloader: with torch.inference_mode(): A__ = model(**SCREAMING_SNAKE_CASE__ ) A__ = outputs.logits.argmax(dim=-1 ) A__ = batch['labels'] A__ , A__ = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=SCREAMING_SNAKE_CASE__ , references=SCREAMING_SNAKE_CASE__ ) A__ = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f'Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n' def _snake_case( ) -> Optional[Any]: '''simple docstring''' A__ = Accelerator(split_batches=SCREAMING_SNAKE_CASE__ , dispatch_batches=SCREAMING_SNAKE_CASE__ ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('**Testing gather_for_metrics**' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f'With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`' ) test_mrpc(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('**Test torch metrics**' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: A__ = Accelerator(split_batches=SCREAMING_SNAKE_CASE__ , dispatch_batches=SCREAMING_SNAKE_CASE__ ) if accelerator.is_local_main_process: print(f'With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99' ) test_torch_metrics(SCREAMING_SNAKE_CASE__ , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('**Test last batch is not dropped when perfectly divisible**' ) A__ = Accelerator() test_torch_metrics(SCREAMING_SNAKE_CASE__ , 512 ) accelerator.state._reset_state() def _snake_case( SCREAMING_SNAKE_CASE__ : List[Any] ) -> Union[str, Any]: '''simple docstring''' main() if __name__ == "__main__": main()
7
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase ) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) self.check_model_type(UpperCAmelCase_ ) def UpperCamelCase ( self: str , **UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} # preprocess args if "points_per_batch" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self: Optional[Any] , UpperCAmelCase_: Tuple , *UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=None , UpperCAmelCase_: Tuple=None , **UpperCAmelCase_: Any ): '''simple docstring''' return super().__call__(UpperCAmelCase_ , *UpperCAmelCase_ , num_workers=UpperCAmelCase_ , batch_size=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict=64 , UpperCAmelCase_: int = 0 , UpperCAmelCase_: float = 512 / 1_500 , UpperCAmelCase_: Optional[int] = 32 , UpperCAmelCase_: Optional[int] = 1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor.size["""longest_edge"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.generate_crop_boxes( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": _SCREAMING_SNAKE_CASE = self.get_inference_context() with inference_context(): _SCREAMING_SNAKE_CASE = self._ensure_tensor_on_device(UpperCAmelCase_ , device=self.device ) _SCREAMING_SNAKE_CASE = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) _SCREAMING_SNAKE_CASE = image_embeddings _SCREAMING_SNAKE_CASE = grid_points.shape[1] _SCREAMING_SNAKE_CASE = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = grid_points[:, i : i + points_per_batch, :, :] _SCREAMING_SNAKE_CASE = input_labels[:, i : i + points_per_batch] _SCREAMING_SNAKE_CASE = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=0.88 , UpperCAmelCase_: Dict=0.95 , UpperCAmelCase_: Tuple=0 , UpperCAmelCase_: str=1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_inputs.pop("""input_boxes""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""is_last""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""original_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = model_inputs.pop("""reshaped_input_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks _SCREAMING_SNAKE_CASE = model_outputs["""pred_masks"""] _SCREAMING_SNAKE_CASE = self.image_processor.post_process_masks( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , binarize=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model_outputs["""iou_scores"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCamelCase ( self: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=False , UpperCAmelCase_: Any=0.7 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.post_process_for_mask_generation( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = defaultdict(UpperCAmelCase_ ) for output in model_outputs: for k, v in output.items(): extra[k].append(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {} if output_rle_mask: _SCREAMING_SNAKE_CASE = rle_mask if output_bboxes_mask: _SCREAMING_SNAKE_CASE = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
306
0
from __future__ import annotations from collections.abc import Callable def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 100 , ): snake_case_ = x_start snake_case_ = fnc(SCREAMING_SNAKE_CASE__ ) snake_case_ = 0.0 for _ in range(SCREAMING_SNAKE_CASE__ ): # Approximates small segments of curve as linear and solve # for trapezoidal area snake_case_ = (x_end - x_start) / steps + xa snake_case_ = fnc(SCREAMING_SNAKE_CASE__ ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step snake_case_ = xa snake_case_ = fxa return area if __name__ == "__main__": def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') lowerCAmelCase_ = 10 while i <= 10_00_00: print(f"""with {i} steps: {trapezoidal_area(f, -5, 5, i)}""") i *= 10
8
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging UpperCamelCase = logging.get_logger(__name__) logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" if "xprophetnet" in prophetnet_checkpoint_path: _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) else: _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) _SCREAMING_SNAKE_CASE = ["""key_proj""", """value_proj""", """query_proj"""] _SCREAMING_SNAKE_CASE = { """self_attn""": """ngram_self_attn""", """cross_attn""": """encoder_attn""", """cross_attn_layer_norm""": """encoder_attn_layer_norm""", """feed_forward_layer_norm""": """final_layer_norm""", """feed_forward""": """""", """intermediate""": """fc1""", """output""": """fc2""", """key_proj""": """k_proj""", """query_proj""": """q_proj""", """value_proj""": """v_proj""", """word_embeddings""": """embed_tokens""", """embeddings_layer_norm""": """emb_layer_norm""", """relative_pos_embeddings""": """relative_linear""", """ngram_embeddings""": """ngram_input_embed""", """position_embeddings""": """embed_positions""", } for key in loading_info["missing_keys"]: _SCREAMING_SNAKE_CASE = key.split(""".""" ) if attributes[0] == "lm_head": _SCREAMING_SNAKE_CASE = prophet _SCREAMING_SNAKE_CASE = prophet_old else: _SCREAMING_SNAKE_CASE = prophet.prophetnet _SCREAMING_SNAKE_CASE = prophet_old.model _SCREAMING_SNAKE_CASE = False for attribute in attributes: if attribute in mapping: _SCREAMING_SNAKE_CASE = mapping[attribute] if not hasattr(snake_case__ ,snake_case__ ) and len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = attribute elif hasattr(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.weight logger.info(F'{attribute} is initialized.' ) _SCREAMING_SNAKE_CASE = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.bias logger.info(F'{attribute} is initialized' ) _SCREAMING_SNAKE_CASE = True break elif attribute in special_keys and hasattr(snake_case__ ,"""in_proj_weight""" ): _SCREAMING_SNAKE_CASE = old_model.in_proj_weight.shape[0] // 3 _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _SCREAMING_SNAKE_CASE = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_12, "We want 512 position_embeddings." _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.embed_positions.weight[:5_12, :] ) _SCREAMING_SNAKE_CASE = True break if attribute.isdigit(): _SCREAMING_SNAKE_CASE = model[int(snake_case__ )] _SCREAMING_SNAKE_CASE = old_model[int(snake_case__ )] else: _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if old_attribute == "": _SCREAMING_SNAKE_CASE = old_model else: if not hasattr(snake_case__ ,snake_case__ ): raise ValueError(F'{old_model} does not have {old_attribute}' ) _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if not is_key_init: raise ValueError(F'{key} was not correctly initialized!' ) print(F'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--prophetnet_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
306
0
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def _UpperCamelCase ( lowercase__ , lowercase__=False ): try: __SCREAMING_SNAKE_CASE : str = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __SCREAMING_SNAKE_CASE : Any = default else: # KEY is set, convert it to True or False. try: __SCREAMING_SNAKE_CASE : List[Any] = strtobool(lowercase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value __lowerCAmelCase : Dict =parse_flag_from_env('RUN_SLOW', default=False) def _UpperCamelCase ( lowercase__ ): return unittest.skip('''Test was skipped''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowercase__ ) def _UpperCamelCase ( lowercase__=None , lowercase__=None ): if test_case is None: return partial(lowercase__ , version=lowercase__ ) return unittest.skipUnless(is_torch_version('''>=''' , lowercase__ ) , F'''test requires torch version >= {version}''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowercase__ ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowercase__ ) __lowerCAmelCase : Optional[Any] =( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def _UpperCamelCase ( lowercase__ ): return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowercase__ ) class _lowercase ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = True @classmethod def __magic_name__( cls :Optional[Any] ) -> List[Any]: __SCREAMING_SNAKE_CASE : Optional[Any] = tempfile.mkdtemp() @classmethod def __magic_name__( cls :List[Any] ) -> List[str]: if os.path.exists(cls.tmpdir ): shutil.rmtree(cls.tmpdir ) def __magic_name__( self :List[Any] ) -> List[str]: if self.clear_on_setup: for path in Path(self.tmpdir ).glob('''**/*''' ): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(lowerCAmelCase__ ) class _lowercase ( unittest.TestCase ): '''simple docstring''' def __magic_name__( self :List[str] ) -> Any: super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class _lowercase ( unittest.TestCase ): '''simple docstring''' def __magic_name__( self :str , lowerCAmelCase__ :Union[mock.Mock, List[mock.Mock]] ) -> Tuple: __SCREAMING_SNAKE_CASE : List[str] = mocks if isinstance(lowerCAmelCase__ , (tuple, list) ) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop ) def _UpperCamelCase ( lowercase__ ): __SCREAMING_SNAKE_CASE : int = AcceleratorState() __SCREAMING_SNAKE_CASE : Optional[int] = tensor[None].clone().to(state.device ) __SCREAMING_SNAKE_CASE : List[str] = gather(lowercase__ ).cpu() __SCREAMING_SNAKE_CASE : Union[str, Any] = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowercase__ ): return False return True class _lowercase : '''simple docstring''' def __init__( self :Union[str, Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :int , lowerCAmelCase__ :str ) -> List[str]: __SCREAMING_SNAKE_CASE : List[str] = returncode __SCREAMING_SNAKE_CASE : Optional[int] = stdout __SCREAMING_SNAKE_CASE : Dict = stderr async def _UpperCamelCase ( lowercase__ , lowercase__ ): while True: __SCREAMING_SNAKE_CASE : Tuple = await stream.readline() if line: callback(lowercase__ ) else: break async def _UpperCamelCase ( lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=False , lowercase__=False ): if echo: print('''\nRunning: ''' , ''' '''.join(lowercase__ ) ) __SCREAMING_SNAKE_CASE : Union[str, Any] = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowercase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowercase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __SCREAMING_SNAKE_CASE : Tuple = [] __SCREAMING_SNAKE_CASE : Union[str, Any] = [] def tee(lowercase__ , lowercase__ , lowercase__ , lowercase__="" ): __SCREAMING_SNAKE_CASE : Tuple = line.decode('''utf-8''' ).rstrip() sink.append(lowercase__ ) if not quiet: print(lowercase__ , lowercase__ , file=lowercase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowercase__ : tee(lowercase__ , lowercase__ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowercase__ : tee(lowercase__ , lowercase__ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowercase__ , ) return _RunOutput(await p.wait() , lowercase__ , lowercase__ ) def _UpperCamelCase ( lowercase__ , lowercase__=None , lowercase__=None , lowercase__=180 , lowercase__=False , lowercase__=True ): __SCREAMING_SNAKE_CASE : Union[str, Any] = asyncio.get_event_loop() __SCREAMING_SNAKE_CASE : Dict = loop.run_until_complete( _stream_subprocess(lowercase__ , env=lowercase__ , stdin=lowercase__ , timeout=lowercase__ , quiet=lowercase__ , echo=lowercase__ ) ) __SCREAMING_SNAKE_CASE : Optional[int] = ''' '''.join(lowercase__ ) if result.returncode > 0: __SCREAMING_SNAKE_CASE : int = '''\n'''.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) return result class _lowercase ( A__ ): '''simple docstring''' pass def _UpperCamelCase ( lowercase__ , lowercase__=False ): try: __SCREAMING_SNAKE_CASE : Union[str, Any] = subprocess.check_output(lowercase__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowercase__ , '''decode''' ): __SCREAMING_SNAKE_CASE : List[str] = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F'''Command `{' '.join(lowercase__ )}` failed with the following error:\n\n{e.output.decode()}''' ) from e
9
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: _SCREAMING_SNAKE_CASE = i + 1 else: _SCREAMING_SNAKE_CASE = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"{two_pointer([2, 7, 11, 15], 9) = }")
306
0
class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : Any , UpperCAmelCase_ : str = "" , UpperCAmelCase_ : bool = False) ->None: '''simple docstring''' lowerCamelCase__: dict[str, RadixNode] ={} # A node will be a leaf if the tree contains its word lowerCamelCase__: Any =is_leaf lowerCamelCase__: List[str] =prefix def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : str) ->tuple[str, str, str]: '''simple docstring''' lowerCamelCase__: Any =0 for q, w in zip(self.prefix , UpperCAmelCase_): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def SCREAMING_SNAKE_CASE_ (self : Optional[Any] , UpperCAmelCase_ : list[str]) ->None: '''simple docstring''' for word in words: self.insert(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : str) ->None: '''simple docstring''' if self.prefix == word: lowerCamelCase__: Optional[int] =True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCamelCase__: Any =RadixNode(prefix=UpperCAmelCase_ , is_leaf=UpperCAmelCase_) else: lowerCamelCase__: Union[str, Any] =self.nodes[word[0]] lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: Tuple =incoming_node.match( UpperCAmelCase_) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase_) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCamelCase__: Union[str, Any] =remaining_prefix lowerCamelCase__: Optional[int] =self.nodes[matching_string[0]] lowerCamelCase__: Dict =RadixNode(UpperCAmelCase_ , UpperCAmelCase_) lowerCamelCase__: Tuple =aux_node if remaining_word == "": lowerCamelCase__: Dict =True else: self.nodes[matching_string[0]].insert(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : str) ->bool: '''simple docstring''' lowerCamelCase__: Dict =self.nodes.get(word[0] , UpperCAmelCase_) if not incoming_node: return False else: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: str =incoming_node.match( UpperCAmelCase_) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : str) ->bool: '''simple docstring''' lowerCamelCase__: Union[str, Any] =self.nodes.get(word[0] , UpperCAmelCase_) if not incoming_node: return False else: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: List[Any] =incoming_node.match( UpperCAmelCase_) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase_) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes) == 1 and not self.is_leaf: lowerCamelCase__: int =list(self.nodes.values())[0] lowerCamelCase__: Any =merging_node.is_leaf self.prefix += merging_node.prefix lowerCamelCase__: Tuple =merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes) > 1: lowerCamelCase__: Dict =False # If there is 1 edge, we merge it with its child else: lowerCamelCase__: str =list(incoming_node.nodes.values())[0] lowerCamelCase__: Any =merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCamelCase__: Optional[Any] =merging_node.nodes return True def SCREAMING_SNAKE_CASE_ (self : List[Any] , UpperCAmelCase_ : int = 0) ->None: '''simple docstring''' if self.prefix != "": print("-" * height , self.prefix , " (leaf)" if self.is_leaf else "") for value in self.nodes.values(): value.print_tree(height + 1) def lowerCAmelCase_ ( ) -> bool: """simple docstring""" lowerCamelCase__: str ="banana bananas bandana band apple all beast".split() lowerCamelCase__: List[Any] =RadixNode() root.insert_many(__a ) assert all(root.find(__a ) for word in words ) assert not root.find("bandanas" ) assert not root.find("apps" ) root.delete("all" ) assert not root.find("all" ) root.delete("banana" ) assert not root.find("banana" ) assert root.find("bananas" ) return True def lowerCAmelCase_ ( ) -> None: """simple docstring""" assert test_trie() def lowerCAmelCase_ ( ) -> None: """simple docstring""" lowerCamelCase__: Optional[int] =RadixNode() lowerCamelCase__: Optional[int] ="banana bananas bandanas bandana band apple all beast".split() root.insert_many(__a ) print("Words:" , __a ) print("Tree:" ) root.print_tree() if __name__ == "__main__": main()
10
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig UpperCamelCase = logging.get_logger(__name__) # General docstring UpperCamelCase = '''MobileNetV1Config''' # Base docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = [1, 1_024, 7, 7] # Image classification docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = '''tabby, tabby cat''' UpperCamelCase = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = {} if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = model.mobilenet_va else: _SCREAMING_SNAKE_CASE = model _SCREAMING_SNAKE_CASE = """MobilenetV1/Conv2d_0/""" _SCREAMING_SNAKE_CASE = backbone.conv_stem.convolution.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.bias _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_mean _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_var for i in range(13 ): _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = i * 2 _SCREAMING_SNAKE_CASE = backbone.layer[pt_index] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_depthwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var _SCREAMING_SNAKE_CASE = backbone.layer[pt_index + 1] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_pointwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = """MobilenetV1/Logits/Conv2d_1c_1x1/""" _SCREAMING_SNAKE_CASE = model.classifier.weight _SCREAMING_SNAKE_CASE = model.classifier.bias return tf_to_pt_map def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> List[str]: """simple docstring""" try: import numpy as np import tensorflow as tf except ImportError: logger.error( """Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see """ """https://www.tensorflow.org/install/ for installation instructions.""" ) raise # Load weights from TF model _SCREAMING_SNAKE_CASE = tf.train.list_variables(snake_case__ ) _SCREAMING_SNAKE_CASE = {} for name, shape in init_vars: logger.info(F'Loading TF weight {name} with shape {shape}' ) _SCREAMING_SNAKE_CASE = tf.train.load_variable(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = array # Build TF to PyTorch weights loading map _SCREAMING_SNAKE_CASE = _build_tf_to_pytorch_map(snake_case__ ,snake_case__ ,snake_case__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F'Importing {name}' ) if name not in tf_weights: logger.info(F'{name} not in tf pre-trained weights, skipping' ) continue _SCREAMING_SNAKE_CASE = tf_weights[name] if "depthwise_weights" in name: logger.info("""Transposing depthwise""" ) _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(2, 3, 0, 1) ) elif "weights" in name: logger.info("""Transposing""" ) if len(pointer.shape ) == 2: # copying into linear layer _SCREAMING_SNAKE_CASE = array.squeeze().transpose() else: _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'Pointer shape {pointer.shape} and array shape {array.shape} mismatched' ) logger.info(F'Initialize PyTorch weight {name} {array.shape}' ) _SCREAMING_SNAKE_CASE = torch.from_numpy(snake_case__ ) tf_weights.pop(snake_case__ ,snake_case__ ) tf_weights.pop(name + """/RMSProp""" ,snake_case__ ) tf_weights.pop(name + """/RMSProp_1""" ,snake_case__ ) tf_weights.pop(name + """/ExponentialMovingAverage""" ,snake_case__ ) logger.info(F'Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}' ) return model def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> torch.Tensor: """simple docstring""" _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = features.shape[-2:] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.stride _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.kernel_size if in_height % stride_height == 0: _SCREAMING_SNAKE_CASE = max(kernel_height - stride_height ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_height - (in_height % stride_height) ,0 ) if in_width % stride_width == 0: _SCREAMING_SNAKE_CASE = max(kernel_width - stride_width ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_width - (in_width % stride_width) ,0 ) _SCREAMING_SNAKE_CASE = pad_along_width // 2 _SCREAMING_SNAKE_CASE = pad_along_width - pad_left _SCREAMING_SNAKE_CASE = pad_along_height // 2 _SCREAMING_SNAKE_CASE = pad_along_height - pad_top _SCREAMING_SNAKE_CASE = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(snake_case__ ,snake_case__ ,"""constant""" ,0.0 ) class __UpperCAmelCase (nn.Module ): def __init__( self: Optional[Any] , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: bool = False , UpperCAmelCase_: Optional[bool] = True , UpperCAmelCase_: Optional[bool or str] = True , ): '''simple docstring''' super().__init__() _SCREAMING_SNAKE_CASE = config if in_channels % groups != 0: raise ValueError(F'Input channels ({in_channels}) are not divisible by {groups} groups.' ) if out_channels % groups != 0: raise ValueError(F'Output channels ({out_channels}) are not divisible by {groups} groups.' ) _SCREAMING_SNAKE_CASE = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) _SCREAMING_SNAKE_CASE = nn.Convad( in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=UpperCAmelCase_ , groups=UpperCAmelCase_ , bias=UpperCAmelCase_ , padding_mode="""zeros""" , ) if use_normalization: _SCREAMING_SNAKE_CASE = nn.BatchNormad( num_features=UpperCAmelCase_ , eps=config.layer_norm_eps , momentum=0.99_97 , affine=UpperCAmelCase_ , track_running_stats=UpperCAmelCase_ , ) else: _SCREAMING_SNAKE_CASE = None if use_activation: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[use_activation] elif isinstance(config.hidden_act , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: _SCREAMING_SNAKE_CASE = config.hidden_act else: _SCREAMING_SNAKE_CASE = None def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' if self.config.tf_padding: _SCREAMING_SNAKE_CASE = apply_tf_padding(UpperCAmelCase_ , self.convolution ) _SCREAMING_SNAKE_CASE = self.convolution(UpperCAmelCase_ ) if self.normalization is not None: _SCREAMING_SNAKE_CASE = self.normalization(UpperCAmelCase_ ) if self.activation is not None: _SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase_ ) return features class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Dict = MobileNetVaConfig __snake_case : Any = load_tf_weights_in_mobilenet_va __snake_case : Any = "mobilenet_v1" __snake_case : List[Any] = "pixel_values" __snake_case : Any = False def UpperCamelCase ( self: str , UpperCAmelCase_: Union[nn.Linear, nn.Convad] ): '''simple docstring''' if isinstance(UpperCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCAmelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) UpperCamelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' UpperCamelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: bool = True ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config _SCREAMING_SNAKE_CASE = 32 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) _SCREAMING_SNAKE_CASE = MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=config.num_channels , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=2 , ) _SCREAMING_SNAKE_CASE = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] _SCREAMING_SNAKE_CASE = nn.ModuleList() for i in range(13 ): _SCREAMING_SNAKE_CASE = out_channels if strides[i] == 2 or i == 0: depth *= 2 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=strides[i] , groups=UpperCAmelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=1 , ) ) _SCREAMING_SNAKE_CASE = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCamelCase ( self: Dict , UpperCAmelCase_: Tuple ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase ( self: int , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) _SCREAMING_SNAKE_CASE = self.conv_stem(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): _SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase_ ) if output_hidden_states: _SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) _SCREAMING_SNAKE_CASE = hidden_states if self.pooler is not None: _SCREAMING_SNAKE_CASE = torch.flatten(self.pooler(UpperCAmelCase_ ) , start_dim=1 ) else: _SCREAMING_SNAKE_CASE = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase_ , pooler_output=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Dict , UpperCAmelCase_: MobileNetVaConfig ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config.num_labels _SCREAMING_SNAKE_CASE = MobileNetVaModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head _SCREAMING_SNAKE_CASE = nn.Dropout(config.classifier_dropout_prob , inplace=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.Linear(UpperCAmelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict _SCREAMING_SNAKE_CASE = self.mobilenet_va(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.pooler_output if return_dict else outputs[1] _SCREAMING_SNAKE_CASE = self.classifier(self.dropout(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _SCREAMING_SNAKE_CASE = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _SCREAMING_SNAKE_CASE = """single_label_classification""" else: _SCREAMING_SNAKE_CASE = """multi_label_classification""" if self.config.problem_type == "regression": _SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: _SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) elif self.config.problem_type == "single_label_classification": _SCREAMING_SNAKE_CASE = CrossEntropyLoss() _SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: _SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states , )
306
0
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False")) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env") @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 650, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ]) class lowerCAmelCase__ ( unittest.TestCase): '''simple docstring''' def _lowerCamelCase ( self) -> str: if self.framework == "pytorch": subprocess.run( F"cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py".split() , encoding="utf-8" , check=__lowerCamelCase , ) assert hasattr(self , "env") def _lowerCamelCase ( self , __lowerCamelCase) -> Tuple: _A : Dict = F"{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}" # distributed data settings _A : Optional[Any] = {"smdistributed": {"dataparallel": {"enabled": True}}} if self.script != "run_ddp.py" else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=__lowerCamelCase , instance_count=__lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=__lowerCamelCase , hyperparameters={**self.env.distributed_hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=__lowerCamelCase , py_version="py36" , ) def _lowerCamelCase ( self , __lowerCamelCase) -> Optional[Any]: TrainingJobAnalytics(__lowerCamelCase).export_csv(F"{self.env.test_path}/{job_name}_metrics.csv") @parameterized.expand([(2,)]) def _lowerCamelCase ( self , __lowerCamelCase) -> Any: # create estimator _A : Union[str, Any] = self.create_estimator(__lowerCamelCase) # run training estimator.fit() # result dataframe _A : Optional[Any] = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis _A : List[Any] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) _A : Dict = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping _A : Optional[Any] = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 9_9_9_9_9_9) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(F"{estimator.latest_training_job.name}.json" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __lowerCamelCase)
11
def __lowerCamelCase ( snake_case__ ) -> list: """simple docstring""" def merge(snake_case__ ,snake_case__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(snake_case__ ) <= 1: return collection _SCREAMING_SNAKE_CASE = len(snake_case__ ) // 2 return merge(merge_sort(collection[:mid] ) ,merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
306
0
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = 'https://openaipublic.azureedge.net/jukebox/models/' UpperCAmelCase_ = { 'jukebox-1b-lyrics': [ '5b/vqvae.pth.tar', '5b/prior_level_0.pth.tar', '5b/prior_level_1.pth.tar', '1b_lyrics/prior_level_2.pth.tar', ], 'jukebox-5b-lyrics': [ '5b/vqvae.pth.tar', '5b/prior_level_0.pth.tar', '5b/prior_level_1.pth.tar', '5b_lyrics/prior_level_2.pth.tar', ], } def lowerCamelCase__ ( A__ : Tuple ): '''simple docstring''' if key.endswith(""".model.1.bias""" ) and len(key.split(""".""" ) ) > 10: __lowerCamelCase = key.replace(""".model.1.bias""" , """.conv1d_1.bias""" ) elif key.endswith(""".model.1.weight""" ) and len(key.split(""".""" ) ) > 10: __lowerCamelCase = key.replace(""".model.1.weight""" , """.conv1d_1.weight""" ) elif key.endswith(""".model.3.bias""" ) and len(key.split(""".""" ) ) > 10: __lowerCamelCase = key.replace(""".model.3.bias""" , """.conv1d_2.bias""" ) elif key.endswith(""".model.3.weight""" ) and len(key.split(""".""" ) ) > 10: __lowerCamelCase = key.replace(""".model.3.weight""" , """.conv1d_2.weight""" ) if "conditioner_blocks.0." in key: __lowerCamelCase = key.replace("""conditioner_blocks.0""" , """conditioner_blocks""" ) if "prime_prior" in key: __lowerCamelCase = key.replace("""prime_prior""" , """encoder""" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: __lowerCamelCase = key.replace(""".emb.""" , """.""" ) if key.endswith("""k""" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(""".k""" , """.codebook""" ) if "y_emb." in key: return key.replace("""y_emb.""" , """metadata_embedding.""" ) if "x_emb.emb." in key: __lowerCamelCase = key.replace("""0.x_emb.emb""" , """embed_tokens""" ) if "prime_state_ln" in key: return key.replace("""prime_state_ln""" , """encoder.final_layer_norm""" ) if ".ln" in key: return key.replace(""".ln""" , """.layer_norm""" ) if "_ln" in key: return key.replace("""_ln""" , """_layer_norm""" ) if "prime_state_proj" in key: return key.replace("""prime_state_proj""" , """encoder.proj_in""" ) if "prime_x_out" in key: return key.replace("""prime_x_out""" , """encoder.lm_head""" ) if "prior.x_out" in key: return key.replace("""x_out""" , """fc_proj_out""" ) if "x_emb" in key: return key.replace("""x_emb""" , """embed_tokens""" ) return key def lowerCamelCase__ ( A__ : List[str] , A__ : Dict , A__ : Optional[Any] , A__ : List[Any] ): '''simple docstring''' __lowerCamelCase = {} import re __lowerCamelCase = re.compile(R"""encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)""" ) __lowerCamelCase = re.compile( R"""encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)""" ) __lowerCamelCase = re.compile(R"""encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)""" ) __lowerCamelCase = re.compile(R"""decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)""" ) __lowerCamelCase = re.compile( R"""decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)""" ) __lowerCamelCase = re.compile(R"""decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)""" ) __lowerCamelCase = re.compile(R"""conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)""" ) __lowerCamelCase = re.compile( R"""conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)""" ) __lowerCamelCase = re.compile(R"""conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)""" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(A__ ): __lowerCamelCase = re_encoder_block_conv_in.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) __lowerCamelCase = f'encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}' __lowerCamelCase = re_encoder_block_conv_in.sub(A__ , A__ ) elif re_encoder_block_resnet.fullmatch(A__ ): __lowerCamelCase = re_encoder_block_resnet.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) __lowerCamelCase = {"""1""": 1, """3""": 2}[groups[-2]] __lowerCamelCase = f'encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.' __lowerCamelCase = f'resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}' __lowerCamelCase = prefix + resnet_block __lowerCamelCase = re_encoder_block_resnet.sub(A__ , A__ ) elif re_encoder_block_proj_out.fullmatch(A__ ): __lowerCamelCase = re_encoder_block_proj_out.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = f'encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}' __lowerCamelCase = re_encoder_block_proj_out.sub(A__ , A__ ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(A__ ): __lowerCamelCase = re_decoder_block_conv_out.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowerCamelCase = f'decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}' __lowerCamelCase = re_decoder_block_conv_out.sub(A__ , A__ ) elif re_decoder_block_resnet.fullmatch(A__ ): __lowerCamelCase = re_decoder_block_resnet.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowerCamelCase = {"""1""": 1, """3""": 2}[groups[-2]] __lowerCamelCase = f'decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.' __lowerCamelCase = f'resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}' __lowerCamelCase = prefix + resnet_block __lowerCamelCase = re_decoder_block_resnet.sub(A__ , A__ ) elif re_decoder_block_proj_in.fullmatch(A__ ): __lowerCamelCase = re_decoder_block_proj_in.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = f'decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}' __lowerCamelCase = re_decoder_block_proj_in.sub(A__ , A__ ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(A__ ): __lowerCamelCase = re_prior_cond_conv_out.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowerCamelCase = f'conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}' __lowerCamelCase = re_prior_cond_conv_out.sub(A__ , A__ ) elif re_prior_cond_resnet.fullmatch(A__ ): __lowerCamelCase = re_prior_cond_resnet.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowerCamelCase = {"""1""": 1, """3""": 2}[groups[-2]] __lowerCamelCase = f'conditioner_blocks.upsampler.upsample_block.{block_index}.' __lowerCamelCase = f'resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}' __lowerCamelCase = prefix + resnet_block __lowerCamelCase = re_prior_cond_resnet.sub(A__ , A__ ) elif re_prior_cond_proj_in.fullmatch(A__ ): __lowerCamelCase = re_prior_cond_proj_in.match(A__ ) __lowerCamelCase = regex_match.groups() __lowerCamelCase = f'conditioner_blocks.upsampler.proj_in.{groups[-1]}' __lowerCamelCase = re_prior_cond_proj_in.sub(A__ , A__ ) # keep original key else: __lowerCamelCase = original_key __lowerCamelCase = replace_key(A__ ) if f'{key_prefix}.{key}' not in model_state_dict or key is None: print(f'failed converting {original_key} to {key}, does not match' ) # handle missmatched shape elif value.shape != model_state_dict[f'{key_prefix}.{key}'].shape: __lowerCamelCase = model_state_dict[f'{key_prefix}.{key}'] print(f'{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match' ) __lowerCamelCase = original_key __lowerCamelCase = original_key __lowerCamelCase = value return new_dict @torch.no_grad() def lowerCamelCase__ ( A__ : str=None , A__ : List[Any]=None ): '''simple docstring''' for file in MODEL_MAPPING[model_name]: if not os.path.isfile(f'{pytorch_dump_folder_path}/{file.split("/" )[-1]}' ): __lowerCamelCase = requests.get(f'{PREFIX}{file}' , allow_redirects=A__ ) os.makedirs(f'{pytorch_dump_folder_path}/' , exist_ok=A__ ) open(f'{pytorch_dump_folder_path}/{file.split("/" )[-1]}' , """wb""" ).write(r.content ) __lowerCamelCase = MODEL_MAPPING[model_name.split("""/""" )[-1]] __lowerCamelCase = JukeboxConfig.from_pretrained(A__ ) __lowerCamelCase = JukeboxModel(A__ ) __lowerCamelCase = [] __lowerCamelCase = {} for i, dict_name in enumerate(A__ ): __lowerCamelCase = torch.load(f'{pytorch_dump_folder_path}/{dict_name.split("/" )[-1]}' )["""model"""] __lowerCamelCase = {} for k in old_dic.keys(): if k.endswith(""".b""" ): __lowerCamelCase = old_dic[k] elif k.endswith(""".w""" ): __lowerCamelCase = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: __lowerCamelCase = old_dic[k] else: __lowerCamelCase = old_dic[k] __lowerCamelCase = """vqvae""" if i == 0 else f'priors.{3 - i}' __lowerCamelCase = fix_jukebox_keys(A__ , model.state_dict() , A__ , A__ ) weight_dict.append(A__ ) __lowerCamelCase = weight_dict.pop(0 ) model.vqvae.load_state_dict(A__ ) for i in range(len(A__ ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(A__ ).mkdir(exist_ok=A__ ) with open(f'{pytorch_dump_folder_path}/mapping.json' , """w""" ) as txtfile: json.dump(A__ , A__ ) print(f'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(A__ ) return weight_dict if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='jukebox-5b-lyrics', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='jukebox-5b-lyrics-converted', type=str, help='Path to the output PyTorch model directory.', ) UpperCAmelCase_ = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
12
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length, 2) ,snake_case__ ) else: _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length) ,snake_case__ ) for i, tensor in enumerate(snake_case__ ): if padding_side == "right": if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] return out_tensor.tolist() def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = ord(snake_case__ ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True _SCREAMING_SNAKE_CASE = unicodedata.category(snake_case__ ) if cat.startswith("""P""" ): return True return False @dataclass class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : PreTrainedTokenizerBase __snake_case : Union[bool, str, PaddingStrategy] = True __snake_case : Optional[int] = None __snake_case : Optional[int] = None __snake_case : int = -100 __snake_case : str = "pt" def UpperCamelCase ( self: str , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' import torch _SCREAMING_SNAKE_CASE = """label""" if """label""" in features[0].keys() else """labels""" _SCREAMING_SNAKE_CASE = [feature[label_name] for feature in features] if label_name in features[0].keys() else None _SCREAMING_SNAKE_CASE = self.tokenizer.pad( UpperCAmelCase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" if labels is None else None , ) if labels is None: return batch _SCREAMING_SNAKE_CASE = torch.tensor(batch["""entity_ids"""] ).shape[1] _SCREAMING_SNAKE_CASE = self.tokenizer.padding_side if padding_side == "right": _SCREAMING_SNAKE_CASE = [ list(UpperCAmelCase_ ) + [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) for label in labels ] else: _SCREAMING_SNAKE_CASE = [ [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) + list(UpperCAmelCase_ ) for label in labels ] _SCREAMING_SNAKE_CASE = [feature["""ner_tags"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , -1 , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [feature["""original_entity_spans"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , (-1, -1) , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {k: torch.tensor(UpperCAmelCase_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
306
0
import logging from transformers import PretrainedConfig lowerCAmelCase : Optional[Any] = logging.getLogger(__name__) lowerCAmelCase : Optional[int] = { """bertabs-finetuned-cnndm""": """https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json""", } class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Dict = '''bertabs''' def __init__( self : Any , lowerCAmelCase__ : Any=3_0522 , lowerCAmelCase__ : List[Any]=512 , lowerCAmelCase__ : Union[str, Any]=6 , lowerCAmelCase__ : Union[str, Any]=512 , lowerCAmelCase__ : Optional[int]=8 , lowerCAmelCase__ : int=512 , lowerCAmelCase__ : Dict=0.2 , lowerCAmelCase__ : str=6 , lowerCAmelCase__ : str=768 , lowerCAmelCase__ : Any=8 , lowerCAmelCase__ : Union[str, Any]=2048 , lowerCAmelCase__ : Union[str, Any]=0.2 , **lowerCAmelCase__ : List[Any] , ): super().__init__(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = vocab_size SCREAMING_SNAKE_CASE_: List[str] = max_pos SCREAMING_SNAKE_CASE_: List[str] = enc_layers SCREAMING_SNAKE_CASE_: Union[str, Any] = enc_hidden_size SCREAMING_SNAKE_CASE_: Union[str, Any] = enc_heads SCREAMING_SNAKE_CASE_: List[Any] = enc_ff_size SCREAMING_SNAKE_CASE_: Dict = enc_dropout SCREAMING_SNAKE_CASE_: List[str] = dec_layers SCREAMING_SNAKE_CASE_: Any = dec_hidden_size SCREAMING_SNAKE_CASE_: Union[str, Any] = dec_heads SCREAMING_SNAKE_CASE_: Union[str, Any] = dec_ff_size SCREAMING_SNAKE_CASE_: Union[str, Any] = dec_dropout
13
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = weights[0][0][0] _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[0] ) _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # lsh weights + output _SCREAMING_SNAKE_CASE = weights[0][1] if len(snake_case__ ) < 4: set_layer_weights_in_torch_lsh(snake_case__ ,torch_block.attention ,snake_case__ ) else: set_layer_weights_in_torch_local(snake_case__ ,torch_block.attention ,snake_case__ ) # intermediate weighs _SCREAMING_SNAKE_CASE = weights[2][0][1][2] # Chunked Feed Forward if len(snake_case__ ) == 4: _SCREAMING_SNAKE_CASE = intermediate_weights[2] # layernorm 2 _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # intermediate dense _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) # intermediate out _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = torch_model.reformer # word embeds _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings ,torch.tensor(snake_case__ ) ,) if isinstance(weights[3] ,snake_case__ ): _SCREAMING_SNAKE_CASE = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _SCREAMING_SNAKE_CASE = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(torch.tensor(snake_case__ ) ) _SCREAMING_SNAKE_CASE = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( snake_case__ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _SCREAMING_SNAKE_CASE = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(snake_case__ ,snake_case__ ,snake_case__ ) # output layer norm _SCREAMING_SNAKE_CASE = np.asarray(weights[7][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # output embeddings _SCREAMING_SNAKE_CASE = np.asarray(weights[9][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ReformerConfig.from_json_file(snake_case__ ) print(F'Building PyTorch model from configuration: {config}' ) _SCREAMING_SNAKE_CASE = ReformerModelWithLMHead(snake_case__ ) with open(snake_case__ ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = pickle.load(snake_case__ )["""weights"""] set_model_weights_in_torch(snake_case__ ,snake_case__ ,config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() ,snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
306
0
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' UpperCAmelCase__ = 42 class UpperCamelCase_ ( UpperCAmelCase__ , UpperCAmelCase__ ): '''simple docstring''' @register_to_config def __init__( self : Dict , UpperCAmelCase__ : int = 32 , UpperCAmelCase__ : int = 64 , UpperCAmelCase__ : int = 20 , UpperCAmelCase__ : int = 768 , UpperCAmelCase__ : str=77 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : str = "silu" , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = None , UpperCAmelCase__ : Optional[str] = "linear" , UpperCAmelCase__ : Optional[str] = "prd" , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[int] = None , ) ->List[str]: '''simple docstring''' super().__init__() A__ = num_attention_heads A__ = attention_head_dim A__ = num_attention_heads * attention_head_dim A__ = additional_embeddings A__ = time_embed_dim or inner_dim A__ = embedding_proj_dim or embedding_dim A__ = clip_embed_dim or embedding_dim A__ = Timesteps(UpperCAmelCase__ , UpperCAmelCase__ , 0) A__ = TimestepEmbedding(UpperCAmelCase__ , UpperCAmelCase__ , out_dim=UpperCAmelCase__ , act_fn=UpperCAmelCase__) A__ = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__) if embedding_proj_norm_type is None: A__ = None elif embedding_proj_norm_type == "layer": A__ = nn.LayerNorm(UpperCAmelCase__) else: raise ValueError(f"""unsupported embedding_proj_norm_type: {embedding_proj_norm_type}""") A__ = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__) if encoder_hid_proj_type is None: A__ = None elif encoder_hid_proj_type == "linear": A__ = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__) else: raise ValueError(f"""unsupported encoder_hid_proj_type: {encoder_hid_proj_type}""") A__ = nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , UpperCAmelCase__)) if added_emb_type == "prd": A__ = nn.Parameter(torch.zeros(1 , 1 , UpperCAmelCase__)) elif added_emb_type is None: A__ = None else: raise ValueError( f"""`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `'prd'` or `None`.""") A__ = nn.ModuleList( [ BasicTransformerBlock( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , dropout=UpperCAmelCase__ , activation_fn='''gelu''' , attention_bias=UpperCAmelCase__ , ) for d in range(UpperCAmelCase__) ]) if norm_in_type == "layer": A__ = nn.LayerNorm(UpperCAmelCase__) elif norm_in_type is None: A__ = None else: raise ValueError(f"""Unsupported norm_in_type: {norm_in_type}.""") A__ = nn.LayerNorm(UpperCAmelCase__) A__ = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__) A__ = torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -10000.0) causal_attention_mask.triu_(1) A__ = causal_attention_mask[None, ...] self.register_buffer('''causal_attention_mask''' , UpperCAmelCase__ , persistent=UpperCAmelCase__) A__ = nn.Parameter(torch.zeros(1 , UpperCAmelCase__)) A__ = nn.Parameter(torch.zeros(1 , UpperCAmelCase__)) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def SCREAMING_SNAKE_CASE ( self : int) ->Dict[str, AttentionProcessor]: '''simple docstring''' A__ = {} def fn_recursive_add_processors(UpperCAmelCase__ : str , UpperCAmelCase__ : torch.nn.Module , UpperCAmelCase__ : Dict[str, AttentionProcessor]): if hasattr(UpperCAmelCase__ , '''set_processor'''): A__ = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f"""{name}.{sub_name}""" , UpperCAmelCase__ , UpperCAmelCase__) return processors for name, module in self.named_children(): fn_recursive_add_processors(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) return processors def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Union[AttentionProcessor, Dict[str, AttentionProcessor]]) ->Optional[Any]: '''simple docstring''' A__ = len(self.attn_processors.keys()) if isinstance(UpperCAmelCase__ , UpperCAmelCase__) and len(UpperCAmelCase__) != count: raise ValueError( f"""A dict of processors was passed, but the number of processors {len(UpperCAmelCase__)} does not match the""" f""" number of attention layers: {count}. Please make sure to pass {count} processor classes.""") def fn_recursive_attn_processor(UpperCAmelCase__ : str , UpperCAmelCase__ : torch.nn.Module , UpperCAmelCase__ : List[str]): if hasattr(UpperCAmelCase__ , '''set_processor'''): if not isinstance(UpperCAmelCase__ , UpperCAmelCase__): module.set_processor(UpperCAmelCase__) else: module.set_processor(processor.pop(f"""{name}.processor""")) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f"""{name}.{sub_name}""" , UpperCAmelCase__ , UpperCAmelCase__) for name, module in self.named_children(): fn_recursive_attn_processor(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Dict: '''simple docstring''' self.set_attn_processor(AttnProcessor()) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[torch.Tensor, float, int] , UpperCAmelCase__ : torch.FloatTensor , UpperCAmelCase__ : Optional[torch.FloatTensor] = None , UpperCAmelCase__ : Optional[torch.BoolTensor] = None , UpperCAmelCase__ : bool = True , ) ->Tuple: '''simple docstring''' A__ = hidden_states.shape[0] A__ = timestep if not torch.is_tensor(UpperCAmelCase__): A__ = torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device) elif torch.is_tensor(UpperCAmelCase__) and len(timesteps.shape) == 0: A__ = timesteps[None].to(hidden_states.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML A__ = timesteps * torch.ones(UpperCAmelCase__ , dtype=timesteps.dtype , device=timesteps.device) A__ = self.time_proj(UpperCAmelCase__) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. A__ = timesteps_projected.to(dtype=self.dtype) A__ = self.time_embedding(UpperCAmelCase__) if self.embedding_proj_norm is not None: A__ = self.embedding_proj_norm(UpperCAmelCase__) A__ = self.embedding_proj(UpperCAmelCase__) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: A__ = self.encoder_hidden_states_proj(UpperCAmelCase__) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError('''`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set''') A__ = self.proj_in(UpperCAmelCase__) A__ = self.positional_embedding.to(hidden_states.dtype) A__ = [] A__ = 0 if encoder_hidden_states is not None: additional_embeds.append(UpperCAmelCase__) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape) == 2: A__ = proj_embeddings[:, None, :] if len(hidden_states.shape) == 2: A__ = hidden_states[:, None, :] A__ = additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: A__ = self.prd_embedding.to(hidden_states.dtype).expand(UpperCAmelCase__ , -1 , -1) additional_embeds.append(UpperCAmelCase__) A__ = torch.cat( UpperCAmelCase__ , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens A__ = additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: A__ = F.pad( UpperCAmelCase__ , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) A__ = hidden_states + positional_embeddings if attention_mask is not None: A__ = (1 - attention_mask.to(hidden_states.dtype)) * -10000.0 A__ = F.pad(UpperCAmelCase__ , (0, self.additional_embeddings) , value=0.0) A__ = (attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype) A__ = attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0) if self.norm_in is not None: A__ = self.norm_in(UpperCAmelCase__) for block in self.transformer_blocks: A__ = block(UpperCAmelCase__ , attention_mask=UpperCAmelCase__) A__ = self.norm_out(UpperCAmelCase__) if self.prd_embedding is not None: A__ = hidden_states[:, -1] else: A__ = hidden_states[:, additional_embeddings_len:] A__ = self.proj_to_clip_embeddings(UpperCAmelCase__) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Any) ->Dict: '''simple docstring''' A__ = (prior_latents * self.clip_std) + self.clip_mean return prior_latents
14
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = TextToVideoSDPipeline __snake_case : Optional[int] = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __snake_case : Optional[int] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase ( self: int ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) _SCREAMING_SNAKE_CASE = CLIPTextModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict=0 ): '''simple docstring''' if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.get_dummy_inputs(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """np""" _SCREAMING_SNAKE_CASE = sd_pipe(**UpperCAmelCase_ ).frames _SCREAMING_SNAKE_CASE = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _SCREAMING_SNAKE_CASE = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=1E-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=25 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
306
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE :Optional[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :List[Any] = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = "yolos" def __init__( self : Any ,A : Optional[Any]=7_68 ,A : Dict=12 ,A : Any=12 ,A : str=30_72 ,A : Any="gelu" ,A : str=0.0 ,A : List[str]=0.0 ,A : Dict=0.02 ,A : int=1E-12 ,A : Tuple=[5_12, 8_64] ,A : List[Any]=16 ,A : str=3 ,A : str=True ,A : Any=1_00 ,A : Dict=True ,A : Dict=False ,A : Tuple=1 ,A : Union[str, Any]=5 ,A : Optional[Any]=2 ,A : Union[str, Any]=5 ,A : int=2 ,A : int=0.1 ,**A : List[str] ,): super().__init__(**A ) __A = hidden_size __A = num_hidden_layers __A = num_attention_heads __A = intermediate_size __A = hidden_act __A = hidden_dropout_prob __A = attention_probs_dropout_prob __A = initializer_range __A = layer_norm_eps __A = image_size __A = patch_size __A = num_channels __A = qkv_bias __A = num_detection_tokens __A = use_mid_position_embeddings __A = auxiliary_loss # Hungarian matcher __A = class_cost __A = bbox_cost __A = giou_cost # Loss coefficients __A = bbox_loss_coefficient __A = giou_loss_coefficient __A = eos_coefficient class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = version.parse("1.11" ) @property def UpperCamelCase_ ( self : str ): return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def UpperCamelCase_ ( self : List[Any] ): return 1E-4 @property def UpperCamelCase_ ( self : Optional[Any] ): return 12
15
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class __UpperCAmelCase : def __init__( self: Union[str, Any] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int=13 , UpperCAmelCase_: Optional[int]=7 , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=False , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Optional[int]=33 , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: List[Any]=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Any=37 , UpperCAmelCase_: Optional[Any]="gelu" , UpperCAmelCase_: Dict=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: int=16 , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: Tuple=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: str=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = EsmForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = False __snake_case : Dict = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __snake_case : List[Any] = () __snake_case : Dict = ( { "feature-extraction": EsmModel, "fill-mask": EsmForMaskedLM, "text-classification": EsmForSequenceClassification, "token-classification": EsmForTokenClassification, "zero-shot": EsmForSequenceClassification, } if is_torch_available() else {} ) __snake_case : int = True def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: int ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _SCREAMING_SNAKE_CASE = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _SCREAMING_SNAKE_CASE = create_position_ids_from_input_ids(UpperCAmelCase_ , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.empty(2 , 4 , 30 ) _SCREAMING_SNAKE_CASE = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _SCREAMING_SNAKE_CASE = torch.as_tensor([expected_single_positions, expected_single_positions] ) _SCREAMING_SNAKE_CASE = embeddings.create_position_ids_from_inputs_embeds(UpperCAmelCase_ ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Dict ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Any ): '''simple docstring''' pass @require_torch class __UpperCAmelCase (_UpperCAmelCase ): @slow def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[[8.92_15, -10.58_98, -6.46_71], [-6.39_67, -13.91_14, -1.12_12], [-7.78_12, -13.95_16, -3.74_06]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] # compare the actual values for a slice. _SCREAMING_SNAKE_CASE = torch.tensor( [[[0.14_44, 0.54_13, 0.32_48], [0.30_34, 0.00_53, 0.31_08], [0.32_28, -0.24_99, 0.34_15]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase_ = logging.get_logger(__name__) class __A ( A_ ): '''simple docstring''' lowerCAmelCase : List[str] = ["pixel_values"] def __init__( self : Any ,_snake_case : bool = True ,_snake_case : Dict[str, int] = None ,_snake_case : PILImageResampling = PILImageResampling.BICUBIC ,_snake_case : bool = True ,_snake_case : Union[int, float] = 1 / 255 ,_snake_case : bool = True ,_snake_case : Optional[Union[float, List[float]]] = None ,_snake_case : Optional[Union[float, List[float]]] = None ,_snake_case : bool = True ,**_snake_case : Union[str, Any] ,) -> None: """simple docstring""" super().__init__(**_snake_case ) lowercase__ : List[str] = size if size is not None else {'''height''': 384, '''width''': 384} lowercase__ : Optional[Any] = get_size_dict(_snake_case ,default_to_square=_snake_case ) lowercase__ : Optional[int] = do_resize lowercase__ : str = size lowercase__ : Optional[Any] = resample lowercase__ : Optional[Any] = do_rescale lowercase__ : Union[str, Any] = rescale_factor lowercase__ : Dict = do_normalize lowercase__ : List[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowercase__ : List[Any] = image_std if image_std is not None else OPENAI_CLIP_STD lowercase__ : Optional[Any] = do_convert_rgb def UpperCAmelCase ( self : Union[str, Any] ,_snake_case : np.ndarray ,_snake_case : Dict[str, int] ,_snake_case : PILImageResampling = PILImageResampling.BICUBIC ,_snake_case : Optional[Union[str, ChannelDimension]] = None ,**_snake_case : Optional[Any] ,) -> np.ndarray: """simple docstring""" lowercase__ : List[str] = get_size_dict(_snake_case ,default_to_square=_snake_case ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) lowercase__ : Tuple = (size['''height'''], size['''width''']) return resize(_snake_case ,size=_snake_case ,resample=_snake_case ,data_format=_snake_case ,**_snake_case ) def UpperCAmelCase ( self : Optional[int] ,_snake_case : np.ndarray ,_snake_case : Union[int, float] ,_snake_case : Optional[Union[str, ChannelDimension]] = None ,**_snake_case : Optional[Any] ,) -> int: """simple docstring""" return rescale(_snake_case ,scale=_snake_case ,data_format=_snake_case ,**_snake_case ) def UpperCAmelCase ( self : Any ,_snake_case : np.ndarray ,_snake_case : Union[float, List[float]] ,_snake_case : Union[float, List[float]] ,_snake_case : Optional[Union[str, ChannelDimension]] = None ,**_snake_case : Union[str, Any] ,) -> np.ndarray: """simple docstring""" return normalize(_snake_case ,mean=_snake_case ,std=_snake_case ,data_format=_snake_case ,**_snake_case ) def UpperCAmelCase ( self : Any ,_snake_case : ImageInput ,_snake_case : Optional[bool] = None ,_snake_case : Optional[Dict[str, int]] = None ,_snake_case : PILImageResampling = None ,_snake_case : Optional[bool] = None ,_snake_case : Optional[float] = None ,_snake_case : Optional[bool] = None ,_snake_case : Optional[Union[float, List[float]]] = None ,_snake_case : Optional[Union[float, List[float]]] = None ,_snake_case : Optional[Union[str, TensorType]] = None ,_snake_case : bool = None ,_snake_case : ChannelDimension = ChannelDimension.FIRST ,**_snake_case : Dict ,) -> PIL.Image.Image: """simple docstring""" lowercase__ : List[str] = do_resize if do_resize is not None else self.do_resize lowercase__ : Tuple = resample if resample is not None else self.resample lowercase__ : int = do_rescale if do_rescale is not None else self.do_rescale lowercase__ : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ : Dict = do_normalize if do_normalize is not None else self.do_normalize lowercase__ : Tuple = image_mean if image_mean is not None else self.image_mean lowercase__ : str = image_std if image_std is not None else self.image_std lowercase__ : Optional[int] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowercase__ : List[Any] = size if size is not None else self.size lowercase__ : str = get_size_dict(_snake_case ,default_to_square=_snake_case ) lowercase__ : Union[str, Any] = make_list_of_images(_snake_case ) if not valid_images(_snake_case ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: lowercase__ : List[str] = [convert_to_rgb(_snake_case ) for image in images] # All transformations expect numpy arrays. lowercase__ : List[Any] = [to_numpy_array(_snake_case ) for image in images] if do_resize: lowercase__ : Optional[Any] = [self.resize(image=_snake_case ,size=_snake_case ,resample=_snake_case ) for image in images] if do_rescale: lowercase__ : Tuple = [self.rescale(image=_snake_case ,scale=_snake_case ) for image in images] if do_normalize: lowercase__ : Dict = [self.normalize(image=_snake_case ,mean=_snake_case ,std=_snake_case ) for image in images] lowercase__ : List[str] = [to_channel_dimension_format(_snake_case ,_snake_case ) for image in images] lowercase__ : Optional[Any] = BatchFeature(data={'''pixel_values''': images} ,tensor_type=_snake_case ) return encoded_outputs
16
import random def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = num - 1 _SCREAMING_SNAKE_CASE = 0 while s % 2 == 0: _SCREAMING_SNAKE_CASE = s // 2 t += 1 for _ in range(5 ): _SCREAMING_SNAKE_CASE = random.randrange(2 ,num - 1 ) _SCREAMING_SNAKE_CASE = pow(snake_case__ ,snake_case__ ,snake_case__ ) if v != 1: _SCREAMING_SNAKE_CASE = 0 while v != (num - 1): if i == t - 1: return False else: _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = (v**2) % num return True def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" if num < 2: return False _SCREAMING_SNAKE_CASE = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 1_01, 1_03, 1_07, 1_09, 1_13, 1_27, 1_31, 1_37, 1_39, 1_49, 1_51, 1_57, 1_63, 1_67, 1_73, 1_79, 1_81, 1_91, 1_93, 1_97, 1_99, 2_11, 2_23, 2_27, 2_29, 2_33, 2_39, 2_41, 2_51, 2_57, 2_63, 2_69, 2_71, 2_77, 2_81, 2_83, 2_93, 3_07, 3_11, 3_13, 3_17, 3_31, 3_37, 3_47, 3_49, 3_53, 3_59, 3_67, 3_73, 3_79, 3_83, 3_89, 3_97, 4_01, 4_09, 4_19, 4_21, 4_31, 4_33, 4_39, 4_43, 4_49, 4_57, 4_61, 4_63, 4_67, 4_79, 4_87, 4_91, 4_99, 5_03, 5_09, 5_21, 5_23, 5_41, 5_47, 5_57, 5_63, 5_69, 5_71, 5_77, 5_87, 5_93, 5_99, 6_01, 6_07, 6_13, 6_17, 6_19, 6_31, 6_41, 6_43, 6_47, 6_53, 6_59, 6_61, 6_73, 6_77, 6_83, 6_91, 7_01, 7_09, 7_19, 7_27, 7_33, 7_39, 7_43, 7_51, 7_57, 7_61, 7_69, 7_73, 7_87, 7_97, 8_09, 8_11, 8_21, 8_23, 8_27, 8_29, 8_39, 8_53, 8_57, 8_59, 8_63, 8_77, 8_81, 8_83, 8_87, 9_07, 9_11, 9_19, 9_29, 9_37, 9_41, 9_47, 9_53, 9_67, 9_71, 9_77, 9_83, 9_91, 9_97, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(snake_case__ ) def __lowerCamelCase ( snake_case__ = 10_24 ) -> int: """simple docstring""" while True: _SCREAMING_SNAKE_CASE = random.randrange(2 ** (keysize - 1) ,2 ** (keysize) ) if is_prime_low_num(snake_case__ ): return num if __name__ == "__main__": UpperCamelCase = generate_large_prime() print(('''Prime number:''', num)) print(('''is_prime_low_num:''', is_prime_low_num(num)))
306
0
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( lowercase ,lowercase ,unittest.TestCase ): """simple docstring""" __UpperCAmelCase : List[str] = StableDiffusionSAGPipeline __UpperCAmelCase : Dict = TEXT_TO_IMAGE_PARAMS __UpperCAmelCase : int = TEXT_TO_IMAGE_BATCH_PARAMS __UpperCAmelCase : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS __UpperCAmelCase : List[str] = TEXT_TO_IMAGE_IMAGE_PARAMS __UpperCAmelCase : Optional[int] = False def _lowercase ( self : Any ): torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( block_out_channels=(3_2, 6_4), layers_per_block=2, sample_size=3_2, in_channels=4, out_channels=4, down_block_types=("DownBlock2D", "CrossAttnDownBlock2D"), up_block_types=("CrossAttnUpBlock2D", "UpBlock2D"), cross_attention_dim=3_2, ) __lowercase = DDIMScheduler( beta_start=0.00_085, beta_end=0.012, beta_schedule="scaled_linear", clip_sample=UpperCAmelCase__, set_alpha_to_one=UpperCAmelCase__, ) torch.manual_seed(0 ) __lowercase = AutoencoderKL( block_out_channels=[3_2, 6_4], in_channels=3, out_channels=3, down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"], up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"], latent_channels=4, ) torch.manual_seed(0 ) __lowercase = CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=3_2, intermediate_size=3_7, layer_norm_eps=1E-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=1_0_0_0, ) __lowercase = CLIPTextModel(UpperCAmelCase__ ) __lowercase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) __lowercase = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def _lowercase ( self : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : Optional[int]=0 ): if str(UpperCAmelCase__ ).startswith("mps" ): __lowercase = torch.manual_seed(UpperCAmelCase__ ) else: __lowercase = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) __lowercase = { "prompt": ".", "generator": generator, "num_inference_steps": 2, "guidance_scale": 1.0, "sag_scale": 1.0, "output_type": "numpy", } return inputs def _lowercase ( self : Tuple ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : List[Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : int ): __lowercase = StableDiffusionSAGPipeline.from_pretrained("CompVis/stable-diffusion-v1-4" ) __lowercase = sag_pipe.to(UpperCAmelCase__ ) sag_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = "." __lowercase = torch.manual_seed(0 ) __lowercase = sag_pipe( [prompt], generator=UpperCAmelCase__, guidance_scale=7.5, sag_scale=1.0, num_inference_steps=2_0, output_type="np" ) __lowercase = output.images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array([0.1_568, 0.1_738, 0.1_695, 0.1_693, 0.1_507, 0.1_705, 0.1_547, 0.1_751, 0.1_949] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-2 def _lowercase ( self : Union[str, Any] ): __lowercase = StableDiffusionSAGPipeline.from_pretrained("stabilityai/stable-diffusion-2-1-base" ) __lowercase = sag_pipe.to(UpperCAmelCase__ ) sag_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = "." __lowercase = torch.manual_seed(0 ) __lowercase = sag_pipe( [prompt], generator=UpperCAmelCase__, guidance_scale=7.5, sag_scale=1.0, num_inference_steps=2_0, output_type="np" ) __lowercase = output.images __lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array([0.3_459, 0.2_876, 0.2_537, 0.3_002, 0.2_671, 0.2_160, 0.3_026, 0.2_262, 0.2_371] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-2 def _lowercase ( self : int ): __lowercase = StableDiffusionSAGPipeline.from_pretrained("stabilityai/stable-diffusion-2-1-base" ) __lowercase = sag_pipe.to(UpperCAmelCase__ ) sag_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) __lowercase = "." __lowercase = torch.manual_seed(0 ) __lowercase = sag_pipe( [prompt], width=7_6_8, height=5_1_2, generator=UpperCAmelCase__, guidance_scale=7.5, sag_scale=1.0, num_inference_steps=2_0, output_type="np", ) __lowercase = output.images assert image.shape == (1, 5_1_2, 7_6_8, 3)
17
def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _SCREAMING_SNAKE_CASE = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
306
0
import csv import tweepy # Twitter API credentials __lowerCamelCase : Dict = '''''' __lowerCamelCase : Union[str, Any] = '''''' __lowerCamelCase : Dict = '''''' __lowerCamelCase : List[Any] = '''''' def _snake_case ( lowerCAmelCase : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = tweepy.OAuthHandler(lowerCAmelCase , lowerCAmelCase ) auth.set_access_token(lowerCAmelCase , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = tweepy.API(lowerCAmelCase ) # initialize a list to hold all the tweepy Tweets SCREAMING_SNAKE_CASE_ : int = [] # make initial request for most recent tweets (200 is the maximum allowed count) SCREAMING_SNAKE_CASE_ : List[Any] = api.user_timeline(screen_name=lowerCAmelCase , count=2_0_0 ) # save most recent tweets alltweets.extend(lowerCAmelCase ) # save the id of the oldest tweet less one SCREAMING_SNAKE_CASE_ : List[Any] = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(lowerCAmelCase ) > 0: print(f'getting tweets before {oldest}' ) # all subsequent requests use the max_id param to prevent duplicates SCREAMING_SNAKE_CASE_ : int = api.user_timeline( screen_name=lowerCAmelCase , count=2_0_0 , max_id=lowerCAmelCase ) # save most recent tweets alltweets.extend(lowerCAmelCase ) # update the id of the oldest tweet less one SCREAMING_SNAKE_CASE_ : str = alltweets[-1].id - 1 print(f'...{len(lowerCAmelCase )} tweets downloaded so far' ) # transform the tweepy tweets into a 2D array that will populate the csv SCREAMING_SNAKE_CASE_ : Dict = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(f'new_{screen_name}_tweets.csv' , "w" ) as f: SCREAMING_SNAKE_CASE_ : Union[str, Any] = csv.writer(lowerCAmelCase ) writer.writerow(["id", "created_at", "text"] ) writer.writerows(lowerCAmelCase ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('''FirePing32''')
18
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPanoramaPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() @skip_mps class _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ , unittest.TestCase ): lowerCAmelCase__ = StableDiffusionPanoramaPipeline lowerCAmelCase__ = TEXT_TO_IMAGE_PARAMS lowerCAmelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCAmelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCAmelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: torch.manual_seed(0 ) lowerCamelCase_ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) lowerCamelCase_ = DDIMScheduler() torch.manual_seed(0 ) lowerCamelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCamelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) lowerCamelCase_ = CLIPTextModel(lowercase ) lowerCamelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCamelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase=0 ) -> List[str]: lowerCamelCase_ = torch.manual_seed(lowercase ) lowerCamelCase_ = { "prompt": "a photo of the dolomites", "generator": generator, # Setting height and width to None to prevent OOMs on CPU. "height": None, "width": None, "num_inference_steps": 1, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.get_dummy_components() lowerCamelCase_ = StableDiffusionPanoramaPipeline(**lowercase ) lowerCamelCase_ = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = self.get_dummy_inputs(lowercase ) lowerCamelCase_ = sd_pipe(**lowercase ).images lowerCamelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ = np.array([0.6_1_8_6, 0.5_3_7_4, 0.4_9_1_5, 0.4_1_3_5, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_7, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE_( self ) -> int: super().test_inference_batch_consistent(batch_sizes=[1, 2] ) def SCREAMING_SNAKE_CASE_( self ) -> Tuple: super().test_inference_batch_single_identical(batch_size=2 , expected_max_diff=3.25e-3 ) def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: lowerCamelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.get_dummy_components() lowerCamelCase_ = StableDiffusionPanoramaPipeline(**lowercase ) lowerCamelCase_ = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = self.get_dummy_inputs(lowercase ) lowerCamelCase_ = "french fries" lowerCamelCase_ = sd_pipe(**lowercase , negative_prompt=lowercase ) lowerCamelCase_ = output.images lowerCamelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.get_dummy_components() lowerCamelCase_ = StableDiffusionPanoramaPipeline(**lowercase ) lowerCamelCase_ = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = self.get_dummy_inputs(lowercase ) lowerCamelCase_ = sd_pipe(**lowercase , view_batch_size=2 ) lowerCamelCase_ = output.images lowerCamelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: lowerCamelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.get_dummy_components() lowerCamelCase_ = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" ) lowerCamelCase_ = StableDiffusionPanoramaPipeline(**lowercase ) lowerCamelCase_ = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = self.get_dummy_inputs(lowercase ) lowerCamelCase_ = sd_pipe(**lowercase ).images lowerCamelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ = np.array([0.4_0_2_4, 0.6_5_1_0, 0.4_9_0_1, 0.5_3_7_8, 0.5_8_1_3, 0.5_6_2_2, 0.4_7_9_5, 0.4_4_6_7, 0.4_9_5_2] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: lowerCamelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ = self.get_dummy_components() lowerCamelCase_ = PNDMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" , skip_prk_steps=lowercase ) lowerCamelCase_ = StableDiffusionPanoramaPipeline(**lowercase ) lowerCamelCase_ = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) lowerCamelCase_ = self.get_dummy_inputs(lowercase ) lowerCamelCase_ = sd_pipe(**lowercase ).images lowerCamelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCamelCase_ = np.array([0.6_3_9_1, 0.6_2_9_1, 0.4_8_6_1, 0.5_1_3_4, 0.5_5_5_2, 0.4_5_7_8, 0.5_0_3_2, 0.5_0_2_3, 0.4_5_3_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def SCREAMING_SNAKE_CASE_( self ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE_( self , lowercase=0 ) -> Any: lowerCamelCase_ = torch.manual_seed(lowercase ) lowerCamelCase_ = { "prompt": "a photo of the dolomites", "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = "stabilityai/stable-diffusion-2-base" lowerCamelCase_ = DDIMScheduler.from_pretrained(lowercase , subfolder="scheduler" ) lowerCamelCase_ = StableDiffusionPanoramaPipeline.from_pretrained(lowercase , scheduler=lowercase , safety_checker=lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() lowerCamelCase_ = self.get_inputs() lowerCamelCase_ = pipe(**lowercase ).images lowerCamelCase_ = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 2048, 3) lowerCamelCase_ = np.array( [ 0.3_6_9_6_8_3_9_2, 0.2_7_0_2_5_3_7_2, 0.3_2_4_4_6_7_6_6, 0.2_8_3_7_9_3_8_7, 0.3_6_3_6_3_2_7_4, 0.3_0_7_3_3_3_4_7, 0.2_7_1_0_0_0_2_7, 0.2_7_0_5_4_1_2_5, 0.2_5_5_3_6_0_9_6, ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = StableDiffusionPanoramaPipeline.from_pretrained( "stabilityai/stable-diffusion-2-base" , safety_checker=lowercase ) lowerCamelCase_ = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() lowerCamelCase_ = self.get_inputs() lowerCamelCase_ = pipe(**lowercase ).images lowerCamelCase_ = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 2048, 3) lowerCamelCase_ = np.array( [ [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ] ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def SCREAMING_SNAKE_CASE_( self ) -> Any: lowerCamelCase_ = 0 def callback_fn(lowercase , lowercase , lowercase ) -> None: lowerCamelCase_ = True nonlocal number_of_steps number_of_steps += 1 if step == 1: lowerCamelCase_ = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 256) lowerCamelCase_ = latents[0, -3:, -3:, -1] lowerCamelCase_ = np.array( [ 0.1_8_6_8_1_8_6_9, 0.3_3_9_0_7_8_1_6, 0.5_3_6_1_2_7_6, 0.1_4_4_3_2_8_6_5, -0.0_2_8_5_6_6_1_1, -0.7_3_9_4_1_1_2_3, 0.2_3_3_9_7_9_8_7, 0.4_7_3_2_2_6_8_2, -0.3_7_8_2_3_1_6_4, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: lowerCamelCase_ = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 256) lowerCamelCase_ = latents[0, -3:, -3:, -1] lowerCamelCase_ = np.array( [ 0.1_8_5_3_9_6_4_5, 0.3_3_9_8_7_2_4_8, 0.5_3_7_8_5_5_9, 0.1_4_4_3_7_1_4_2, -0.0_2_4_5_5_2_6_1, -0.7_3_3_8_3_1_7, 0.2_3_9_9_0_7_5_5, 0.4_7_3_5_6_2_7_2, -0.3_7_8_6_5_0_5, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 lowerCamelCase_ = False lowerCamelCase_ = "stabilityai/stable-diffusion-2-base" lowerCamelCase_ = DDIMScheduler.from_pretrained(lowercase , subfolder="scheduler" ) lowerCamelCase_ = StableDiffusionPanoramaPipeline.from_pretrained(lowercase , scheduler=lowercase , safety_checker=lowercase ) lowerCamelCase_ = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() lowerCamelCase_ = self.get_inputs() pipe(**lowercase , callback=lowercase , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def SCREAMING_SNAKE_CASE_( self ) -> Tuple: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCamelCase_ = "stabilityai/stable-diffusion-2-base" lowerCamelCase_ = DDIMScheduler.from_pretrained(lowercase , subfolder="scheduler" ) lowerCamelCase_ = StableDiffusionPanoramaPipeline.from_pretrained(lowercase , scheduler=lowercase , safety_checker=lowercase ) lowerCamelCase_ = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowerCamelCase_ = self.get_inputs() lowerCamelCase_ = pipe(**lowercase ) lowerCamelCase_ = torch.cuda.max_memory_allocated() # make sure that less than 5.2 GB is allocated assert mem_bytes < 5.5 * 10**9
19
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: UpperCamelCase = None UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json''' ), }, } UpperCamelCase = { '''facebook/nllb-large-en-ro''': 1_024, '''facebook/nllb-200-distilled-600M''': 1_024, } # fmt: off UpperCamelCase = ['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : List[str] = VOCAB_FILES_NAMES __snake_case : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Tuple = ["input_ids", "attention_mask"] __snake_case : Dict = NllbTokenizer __snake_case : List[int] = [] __snake_case : List[int] = [] def __init__( self: Tuple , UpperCAmelCase_: str=None , UpperCAmelCase_: List[str]=None , UpperCAmelCase_: Tuple="<s>" , UpperCAmelCase_: str="</s>" , UpperCAmelCase_: Union[str, Any]="</s>" , UpperCAmelCase_: int="<s>" , UpperCAmelCase_: Union[str, Any]="<unk>" , UpperCAmelCase_: Union[str, Any]="<pad>" , UpperCAmelCase_: str="<mask>" , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int=None , UpperCAmelCase_: str=False , **UpperCAmelCase_: int , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token _SCREAMING_SNAKE_CASE = legacy_behaviour super().__init__( vocab_file=UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , src_lang=UpperCAmelCase_ , tgt_lang=UpperCAmelCase_ , additional_special_tokens=UpperCAmelCase_ , legacy_behaviour=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = vocab_file _SCREAMING_SNAKE_CASE = False if not self.vocab_file else True _SCREAMING_SNAKE_CASE = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"""additional_special_tokens""": _additional_special_tokens} ) _SCREAMING_SNAKE_CASE = { lang_code: self.convert_tokens_to_ids(UpperCAmelCase_ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } _SCREAMING_SNAKE_CASE = src_lang if src_lang is not None else """eng_Latn""" _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(self._src_lang ) _SCREAMING_SNAKE_CASE = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCamelCase ( self: int ): '''simple docstring''' return self._src_lang @src_lang.setter def UpperCamelCase ( self: int , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase ( self: List[str] , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[int] , UpperCAmelCase_: Optional[List[int]] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase ( self: Tuple , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] , UpperCAmelCase_: Optional[str] , **UpperCAmelCase_: Any ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = self(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = tgt_lang_id return inputs def UpperCamelCase ( self: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str = "eng_Latn" , UpperCAmelCase_: Optional[List[str]] = None , UpperCAmelCase_: str = "fra_Latn" , **UpperCAmelCase_: List[str] , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = src_lang _SCREAMING_SNAKE_CASE = tgt_lang return super().prepare_seqaseq_batch(UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Optional[int] , UpperCAmelCase_: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.convert_tokens_to_ids(UpperCAmelCase_ ) if self.legacy_behaviour: _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [self.eos_token_id, self.cur_lang_code] else: _SCREAMING_SNAKE_CASE = [self.cur_lang_code] _SCREAMING_SNAKE_CASE = [self.eos_token_id] _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.prefix_tokens ) _SCREAMING_SNAKE_CASE = self.convert_ids_to_tokens(self.suffix_tokens ) _SCREAMING_SNAKE_CASE = processors.TemplateProcessing( single=prefix_tokens_str + ["""$A"""] + suffix_tokens_str , pair=prefix_tokens_str + ["""$A""", """$B"""] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCAmelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory.' ) return _SCREAMING_SNAKE_CASE = os.path.join( UpperCAmelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ): copyfile(self.vocab_file , UpperCAmelCase_ ) return (out_vocab_file,)
306
0
class __snake_case : def __init__( self ): '''simple docstring''' lowercase : Optional[Any] = """""" lowercase : Tuple = """""" lowercase : List[Any] = [] def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' if m == -1: return n + 1 elif n == -1: return m + 1 elif self.dp[m][n] > -1: return self.dp[m][n] else: if self.worda[m] == self.worda[n]: lowercase : List[str] = self.__min_dist_top_down_dp(m - 1 ,n - 1 ) else: lowercase : Any = self.__min_dist_top_down_dp(snake_case ,n - 1 ) lowercase : List[Any] = self.__min_dist_top_down_dp(m - 1 ,snake_case ) lowercase : Tuple = self.__min_dist_top_down_dp(m - 1 ,n - 1 ) lowercase : Union[str, Any] = 1 + min(snake_case ,snake_case ,snake_case ) return self.dp[m][n] def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : str = worda lowercase : Dict = worda lowercase : str = [[-1 for _ in range(len(snake_case ) )] for _ in range(len(snake_case ) )] return self.__min_dist_top_down_dp(len(snake_case ) - 1 ,len(snake_case ) - 1 ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : Optional[int] = worda lowercase : int = worda lowercase : List[Any] = len(snake_case ) lowercase : Tuple = len(snake_case ) lowercase : str = [[0 for _ in range(n + 1 )] for _ in range(m + 1 )] for i in range(m + 1 ): for j in range(n + 1 ): if i == 0: # first string is empty lowercase : str = j elif j == 0: # second string is empty lowercase : Dict = i elif worda[i - 1] == worda[j - 1]: # last characters are equal lowercase : str = self.dp[i - 1][j - 1] else: lowercase : Dict = self.dp[i][j - 1] lowercase : Optional[int] = self.dp[i - 1][j] lowercase : Dict = self.dp[i - 1][j - 1] lowercase : int = 1 + min(snake_case ,snake_case ,snake_case ) return self.dp[m][n] if __name__ == "__main__": lowercase : Union[str, Any] = EditDistance() print("""****************** Testing Edit Distance DP Algorithm ******************""") print() lowercase : Optional[Any] = input("""Enter the first string: """).strip() lowercase : str = input("""Enter the second string: """).strip() print() print(F'''The minimum edit distance is: {solver.min_dist_top_down(Sa, Sa)}''') print(F'''The minimum edit distance is: {solver.min_dist_bottom_up(Sa, Sa)}''') print() print("""*************** End of Testing Edit Distance DP Algorithm ***************""")
20
def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> list: """simple docstring""" _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = [[0] * n for i in range(snake_case__ )] for i in range(snake_case__ ): _SCREAMING_SNAKE_CASE = y_points[i] for i in range(2 ,snake_case__ ): for j in range(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
306
0
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Tuple = None lowercase_ : Tuple = BloomTokenizerFast lowercase_ : Optional[int] = BloomTokenizerFast lowercase_ : int = True lowercase_ : str = False lowercase_ : Dict = """tokenizer_file""" lowercase_ : int = {"""bos_token""": """<s>""", """eos_token""": """</s>""", """unk_token""": """<unk>""", """pad_token""": """<pad>"""} def UpperCamelCase ( self) -> Tuple: """simple docstring""" super().setUp() _lowercase : str = BloomTokenizerFast.from_pretrained('bigscience/tokenizer') tokenizer.save_pretrained(self.tmpdirname) def UpperCamelCase ( self, **lowerCamelCase) -> List[Any]: """simple docstring""" kwargs.update(self.special_tokens_map) return BloomTokenizerFast.from_pretrained(self.tmpdirname, **lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : List[Any] = self.get_rust_tokenizer() _lowercase : Any = ['The quick brown fox</s>', 'jumps over the lazy dog</s>'] _lowercase : List[str] = [[21_75, 2_37_14, 7_31_73, 14_42_52, 2], [77, 13_26_19, 34_78, 3_68, 10_95_86, 3_54_33, 2]] _lowercase : List[str] = tokenizer.batch_encode_plus(lowerCamelCase)['input_ids'] self.assertListEqual(lowerCamelCase, lowerCamelCase) _lowercase : Dict = tokenizer.batch_decode(lowerCamelCase) self.assertListEqual(lowerCamelCase, lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase=6) -> List[str]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})'''): _lowercase : Optional[int] = self.rust_tokenizer_class.from_pretrained(lowerCamelCase, **lowerCamelCase) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input _lowercase : Any = 'This is a simple input' _lowercase : Union[str, Any] = ['This is a simple input 1', 'This is a simple input 2'] _lowercase : Optional[int] = ('This is a simple input', 'This is a pair') _lowercase : Tuple = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests try: tokenizer_r.encode(lowerCamelCase, max_length=lowerCamelCase) tokenizer_r.encode_plus(lowerCamelCase, max_length=lowerCamelCase) tokenizer_r.batch_encode_plus(lowerCamelCase, max_length=lowerCamelCase) tokenizer_r.encode(lowerCamelCase, max_length=lowerCamelCase) tokenizer_r.batch_encode_plus(lowerCamelCase, max_length=lowerCamelCase) except ValueError: self.fail('Bloom Tokenizer should be able to deal with padding') _lowercase : Union[str, Any] = None # Hotfixing padding = None self.assertRaises(lowerCamelCase, tokenizer_r.encode, lowerCamelCase, max_length=lowerCamelCase, padding='max_length') # Simple input self.assertRaises(lowerCamelCase, tokenizer_r.encode_plus, lowerCamelCase, max_length=lowerCamelCase, padding='max_length') # Simple input self.assertRaises( lowerCamelCase, tokenizer_r.batch_encode_plus, lowerCamelCase, max_length=lowerCamelCase, padding='max_length', ) # Pair input self.assertRaises(lowerCamelCase, tokenizer_r.encode, lowerCamelCase, max_length=lowerCamelCase, padding='max_length') # Pair input self.assertRaises(lowerCamelCase, tokenizer_r.encode_plus, lowerCamelCase, max_length=lowerCamelCase, padding='max_length') # Pair input self.assertRaises( lowerCamelCase, tokenizer_r.batch_encode_plus, lowerCamelCase, max_length=lowerCamelCase, padding='max_length', ) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : str = self.get_rust_tokenizer() _lowercase : Optional[int] = load_dataset('xnli', 'all_languages', split='test', streaming=lowerCamelCase) _lowercase : Optional[Any] = next(iter(lowerCamelCase))['premise'] # pick up one data _lowercase : Optional[int] = list(sample_data.values()) _lowercase : List[str] = list(map(tokenizer.encode, lowerCamelCase)) _lowercase : Dict = [tokenizer.decode(lowerCamelCase, clean_up_tokenization_spaces=lowerCamelCase) for x in output_tokens] self.assertListEqual(lowerCamelCase, lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map), 1) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values())[0]), 1)
21
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCamelCase = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE :Optional[Any] = { '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE :str = [ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE :Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
22
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class __UpperCAmelCase : __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : int __snake_case : int __snake_case : float __snake_case : float __snake_case : Tuple[int] def UpperCamelCase ( self: str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = torch.arange(self.height * self.width ) _SCREAMING_SNAKE_CASE = torch.stack( [ pixel_indices % self.width, torch.div(UpperCAmelCase_ , self.width , rounding_mode="""trunc""" ), ] , axis=1 , ) return coords @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE = self.shape _SCREAMING_SNAKE_CASE = int(np.prod(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = self.get_image_coords() _SCREAMING_SNAKE_CASE = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) _SCREAMING_SNAKE_CASE = self.get_camera_rays(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = rays.view(UpperCAmelCase_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCamelCase ( self: Any , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] _SCREAMING_SNAKE_CASE = coords.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = self.resolution() _SCREAMING_SNAKE_CASE = self.fov() _SCREAMING_SNAKE_CASE = (flat.float() / (res - 1)) * 2 - 1 _SCREAMING_SNAKE_CASE = fracs * torch.tan(fov / 2 ) _SCREAMING_SNAKE_CASE = fracs.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = ( self.z.view(UpperCAmelCase_ , 1 , 3 ) + self.x.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, 1:] ) _SCREAMING_SNAKE_CASE = directions / directions.norm(dim=-1 , keepdim=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.stack( [ torch.broadcast_to(self.origin.view(UpperCAmelCase_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(UpperCAmelCase_ , *UpperCAmelCase_ , 2 , 3 ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=UpperCAmelCase_ , height=UpperCAmelCase_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __lowerCamelCase ( snake_case__ ) -> DifferentiableProjectiveCamera: """simple docstring""" _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for theta in np.linspace(0 ,2 * np.pi ,num=20 ): _SCREAMING_SNAKE_CASE = np.array([np.sin(snake_case__ ), np.cos(snake_case__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) _SCREAMING_SNAKE_CASE = -z * 4 _SCREAMING_SNAKE_CASE = np.array([np.cos(snake_case__ ), -np.sin(snake_case__ ), 0.0] ) _SCREAMING_SNAKE_CASE = np.cross(snake_case__ ,snake_case__ ) origins.append(snake_case__ ) xs.append(snake_case__ ) ys.append(snake_case__ ) zs.append(snake_case__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,x=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,y=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,z=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,width=snake_case__ ,height=snake_case__ ,x_fov=0.7 ,y_fov=0.7 ,shape=(1, len(snake_case__ )) ,)
306
0
'''simple docstring''' from math import ceil def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] ) -> Tuple: UpperCAmelCase : List[Any] = list(range(0 , _lowerCAmelCase ) ) UpperCAmelCase : Union[str, Any] = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check UpperCAmelCase : Optional[Any] = [] for i in device_map_blocks: if device_map_blocks.count(_lowerCAmelCase ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(_lowerCAmelCase ) # Missing blocks UpperCAmelCase : Optional[int] = [i for i in blocks if i not in device_map_blocks] UpperCAmelCase : int = [i for i in device_map_blocks if i not in blocks] if len(_lowerCAmelCase ) != 0: raise ValueError( '''Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device.''' ''' These attention blocks were specified more than once: ''' + str(_lowerCAmelCase ) ) if len(_lowerCAmelCase ) != 0: raise ValueError( '''There are attention blocks for this model that are not specified in the device_map. Add these attention ''' '''blocks to a device on the device_map: ''' + str(_lowerCAmelCase ) ) if len(_lowerCAmelCase ) != 0: raise ValueError( '''The device_map contains more attention blocks than this model has. Remove these from the device_map:''' + str(_lowerCAmelCase ) ) def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> int: UpperCAmelCase : Dict = list(range(_lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = int(ceil(n_layers / len(_lowerCAmelCase ) ) ) UpperCAmelCase : List[str] = [layers[i : i + n_blocks] for i in range(0 , _lowerCAmelCase , _lowerCAmelCase )] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) )
23
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class __UpperCAmelCase (unittest.TestCase ): def __init__( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any]=13 , UpperCAmelCase_: List[str]=7 , UpperCAmelCase_: Tuple=True , UpperCAmelCase_: List[Any]=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: str=99 , UpperCAmelCase_: List[Any]=32 , UpperCAmelCase_: Dict=5 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Optional[Any]=37 , UpperCAmelCase_: Optional[int]="gelu" , UpperCAmelCase_: Optional[Any]=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: List[Any]=512 , UpperCAmelCase_: Any=16 , UpperCAmelCase_: Dict=2 , UpperCAmelCase_: Union[str, Any]=0.02 , UpperCAmelCase_: Union[str, Any]=4 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_attention_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_choices def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_attention_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=UpperCAmelCase_ , ) return config, input_ids, attention_mask def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : Optional[int] = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase ( self: List[str] ): '''simple docstring''' for model_class_name in self.all_model_classes: _SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(UpperCAmelCase_ ) @require_flax class __UpperCAmelCase (unittest.TestCase ): @slow def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) _SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = np.array([[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 ) )
306
0
import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__(self : Any , a__ : List[Any] , a__ : Dict=13 , a__ : str=32 , a__ : Tuple=3 , a__ : Optional[Any]=4 , a__ : Optional[int]=[10, 20, 30, 40] , a__ : List[Any]=[2, 2, 3, 2] , a__ : List[Any]=True , a__ : int=True , a__ : List[Any]=37 , a__ : Any="gelu" , a__ : int=10 , a__ : Dict=0.0_2 , a__ : Dict=["stage2", "stage3", "stage4"] , a__ : Tuple=[2, 3, 4] , a__ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = image_size __snake_case = num_channels __snake_case = num_stages __snake_case = hidden_sizes __snake_case = depths __snake_case = is_training __snake_case = use_labels __snake_case = intermediate_size __snake_case = hidden_act __snake_case = num_labels __snake_case = initializer_range __snake_case = out_features __snake_case = out_indices __snake_case = scope def a (self : Dict ): """simple docstring""" __snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.num_labels ) __snake_case = self.get_config() return config, pixel_values, labels def a (self : List[str] ): """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=a__ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def a (self : str , a__ : Union[str, Any] , a__ : List[str] , a__ : List[Any] ): """simple docstring""" __snake_case = ConvNextModel(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a (self : Optional[Any] , a__ : List[Any] , a__ : str , a__ : List[Any] ): """simple docstring""" __snake_case = ConvNextForImageClassification(a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a (self : Tuple , a__ : List[Any] , a__ : List[str] , a__ : List[str] ): """simple docstring""" __snake_case = ConvNextBackbone(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __snake_case = None __snake_case = ConvNextBackbone(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def a (self : Tuple ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case = config_and_inputs __snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A_ : Dict = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) A_ : Optional[Any] = ( {'feature-extraction': ConvNextModel, 'image-classification': ConvNextForImageClassification} if is_torch_available() else {} ) A_ : Dict = True A_ : Optional[Any] = False A_ : int = False A_ : int = False A_ : List[str] = False def a (self : List[str] ): """simple docstring""" __snake_case = ConvNextModelTester(self ) __snake_case = ConfigTester(self , config_class=a__ , has_text_modality=a__ , hidden_size=37 ) def a (self : Tuple ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a (self : str ): """simple docstring""" return @unittest.skip(reason='''ConvNext does not use inputs_embeds''' ) def a (self : int ): """simple docstring""" pass @unittest.skip(reason='''ConvNext does not support input and output embeddings''' ) def a (self : Dict ): """simple docstring""" pass @unittest.skip(reason='''ConvNext does not use feedforward chunking''' ) def a (self : List[Any] ): """simple docstring""" pass def a (self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a__ ) __snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case = [*signature.parameters.keys()] __snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , a__ ) def a (self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def a (self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*a__ ) def a (self : Dict ): """simple docstring""" def check_hidden_states_output(a__ : List[str] , a__ : str , a__ : Tuple ): __snake_case = model_class(a__ ) model.to(a__ ) model.eval() with torch.no_grad(): __snake_case = model(**self._prepare_for_class(a__ , a__ ) ) __snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __snake_case = self.model_tester.num_stages self.assertEqual(len(a__ ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = True check_hidden_states_output(a__ , a__ , a__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case = True check_hidden_states_output(a__ , a__ , a__ ) def a (self : Optional[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a__ ) @slow def a (self : Any ): """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = ConvNextModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) def lowerCamelCase__ ( ) -> List[str]: __snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def a (self : Tuple ): """simple docstring""" return AutoImageProcessor.from_pretrained('''facebook/convnext-tiny-224''' ) if is_vision_available() else None @slow def a (self : Optional[Any] ): """simple docstring""" __snake_case = ConvNextForImageClassification.from_pretrained('''facebook/convnext-tiny-224''' ).to(a__ ) __snake_case = self.default_image_processor __snake_case = prepare_img() __snake_case = image_processor(images=a__ , return_tensors='''pt''' ).to(a__ ) # forward pass with torch.no_grad(): __snake_case = model(**a__ ) # verify the logits __snake_case = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , a__ ) __snake_case = torch.tensor([-0.0_2_6_0, -0.4_7_3_9, 0.1_9_1_1] ).to(a__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a__ , atol=1E-4 ) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase , _UpperCAmelCase ): A_ : Union[str, Any] = (ConvNextBackbone,) if is_torch_available() else () A_ : List[Any] = ConvNextConfig A_ : Optional[Any] = False def a (self : Optional[int] ): """simple docstring""" __snake_case = ConvNextModelTester(self )
24
import argparse import hashlib # hashlib is only used inside the Test class import struct class __UpperCAmelCase : def __init__( self: List[str] , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = data _SCREAMING_SNAKE_CASE = [0x67_452_301, 0xef_cda_b89, 0x98_bad_cfe, 0x10_325_476, 0xc3_d2e_1f0] @staticmethod def UpperCamelCase ( UpperCAmelCase_: int , UpperCAmelCase_: List[str] ): '''simple docstring''' return ((n << b) | (n >> (32 - b))) & 0xff_fff_fff def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = B"""\x80""" + B"""\x00""" * (63 - (len(self.data ) + 8) % 64) _SCREAMING_SNAKE_CASE = self.data + padding + struct.pack(""">Q""" , 8 * len(self.data ) ) return padded_data def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = list(struct.unpack(""">16L""" , UpperCAmelCase_ ) ) + [0] * 64 for i in range(16 , 80 ): _SCREAMING_SNAKE_CASE = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.padding() _SCREAMING_SNAKE_CASE = self.split_blocks() for block in self.blocks: _SCREAMING_SNAKE_CASE = self.expand_block(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.h for i in range(0 , 80 ): if 0 <= i < 20: _SCREAMING_SNAKE_CASE = (b & c) | ((~b) & d) _SCREAMING_SNAKE_CASE = 0x5a_827_999 elif 20 <= i < 40: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0x6e_d9e_ba1 elif 40 <= i < 60: _SCREAMING_SNAKE_CASE = (b & c) | (b & d) | (c & d) _SCREAMING_SNAKE_CASE = 0x8f_1bb_cdc elif 60 <= i < 80: _SCREAMING_SNAKE_CASE = b ^ c ^ d _SCREAMING_SNAKE_CASE = 0xca_62c_1d6 _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ( self.rotate(UpperCAmelCase_ , 5 ) + f + e + k + expanded_block[i] & 0xff_fff_fff, a, self.rotate(UpperCAmelCase_ , 30 ), c, d, ) _SCREAMING_SNAKE_CASE = ( self.h[0] + a & 0xff_fff_fff, self.h[1] + b & 0xff_fff_fff, self.h[2] + c & 0xff_fff_fff, self.h[3] + d & 0xff_fff_fff, self.h[4] + e & 0xff_fff_fff, ) return ("{:08x}" * 5).format(*self.h ) def __lowerCamelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = b"""Test String""" assert SHAaHash(snake_case__ ).final_hash() == hashlib.shaa(snake_case__ ).hexdigest() # noqa: S324 def __lowerCamelCase ( ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Process some strings or files""" ) parser.add_argument( """--string""" ,dest="""input_string""" ,default="""Hello World!! Welcome to Cryptography""" ,help="""Hash the string""" ,) parser.add_argument("""--file""" ,dest="""input_file""" ,help="""Hash contents of a file""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = f.read() else: _SCREAMING_SNAKE_CASE = bytes(snake_case__ ,"""utf-8""" ) print(SHAaHash(snake_case__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
306
0
"""simple docstring""" from math import pi, sqrt def lowercase_ ( _snake_case ): if num <= 0: raise ValueError("""math domain error""" ) if num > 171.5: raise OverflowError("""math range error""" ) elif num - int(_snake_case ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(_snake_case ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowercase_ ( ): assert gamma(0.5 ) == sqrt(_snake_case ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() UpperCAmelCase__ : Dict = 1.0 while num: UpperCAmelCase__ : List[Any] = float(input('Gamma of: ')) print(f"""gamma({num}) = {gamma(num)}""") print('\nEnter 0 to exit...')
25
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''tokenizer_file''': '''tokenizer.json'''} UpperCamelCase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Tuple = VOCAB_FILES_NAMES __snake_case : Optional[int] = PRETRAINED_VOCAB_FILES_MAP __snake_case : Optional[Any] = ["input_ids", "attention_mask"] __snake_case : Optional[int] = None def __init__( self: Dict , UpperCAmelCase_: Union[str, Any]=None , UpperCAmelCase_: str=None , UpperCAmelCase_: Optional[int]=None , UpperCAmelCase_: int="<unk>" , UpperCAmelCase_: List[str]="<s>" , UpperCAmelCase_: Tuple="</s>" , UpperCAmelCase_: List[Any]="<pad>" , UpperCAmelCase_: Dict=False , UpperCAmelCase_: Dict=False , **UpperCAmelCase_: Dict , ): '''simple docstring''' super().__init__( UpperCAmelCase_ , UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , add_prefix_space=UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ , **UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCAmelCase_ ) != add_prefix_space: _SCREAMING_SNAKE_CASE = getattr(UpperCAmelCase_ , pre_tok_state.pop("""type""" ) ) _SCREAMING_SNAKE_CASE = add_prefix_space _SCREAMING_SNAKE_CASE = pre_tok_class(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = add_prefix_space def UpperCamelCase ( self: List[str] , *UpperCAmelCase_: Any , **UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Union[str, Any] , *UpperCAmelCase_: Dict , **UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.get("""is_split_into_words""" , UpperCAmelCase_ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Optional[str] = None ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_ ) return tuple(UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: "Conversation" ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) + [self.eos_token_id] ) if len(UpperCAmelCase_ ) > self.model_max_length: _SCREAMING_SNAKE_CASE = input_ids[-self.model_max_length :] return input_ids
306
0
import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def lowerCAmelCase_ ( ): _A : Optional[int] = argparse.ArgumentParser() parser.add_argument("""--model_ckpt""",type=snake_case_,default="""microsoft/unixcoder-base-nine""" ) parser.add_argument("""--num_epochs""",type=snake_case_,default=5 ) parser.add_argument("""--batch_size""",type=snake_case_,default=6 ) parser.add_argument("""--gradient_accumulation_steps""",type=snake_case_,default=1 ) parser.add_argument("""--freeze""",type=snake_case_,default=snake_case_ ) parser.add_argument("""--learning_rate""",type=snake_case_,default=5e-4 ) parser.add_argument("""--seed""",type=snake_case_,default=0 ) parser.add_argument("""--lr_scheduler_type""",type=snake_case_,default="""cosine""" ) parser.add_argument("""--num_warmup_steps""",type=snake_case_,default=10 ) parser.add_argument("""--weight_decay""",type=snake_case_,default=0.01 ) parser.add_argument("""--output_dir""",type=snake_case_,default="""./results""" ) return parser.parse_args() _snake_case = load("accuracy") def lowerCAmelCase_ ( snake_case_ ): _A , _A : str = eval_pred _A : Optional[Any] = np.argmax(snake_case_,axis=1 ) return metric.compute(predictions=snake_case_,references=snake_case_ ) class lowercase ( UpperCamelCase__ ): def __init__( self , _a ) -> None: super().__init__() _A : Dict = trainer def a__ ( self , _a , _a , _a , **_a ) -> str: if control.should_evaluate: _A : Tuple = deepcopy(_a ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix="""train""" ) return control_copy def lowerCAmelCase_ ( ): _A : List[Any] = get_args() set_seed(args.seed ) _A : Optional[int] = load_dataset("""codeparrot/codecomplex""",split="""train""" ) _A : Optional[int] = dataset.train_test_split(test_size=0.2 ) _A : str = train_test["""test"""].train_test_split(test_size=0.5 ) _A : Optional[int] = DatasetDict( { """train""": train_test["""train"""], """test""": test_validation["""train"""], """valid""": test_validation["""test"""], } ) print("""Loading tokenizer and model""" ) _A : List[Any] = AutoTokenizer.from_pretrained(args.model_ckpt ) _A : List[str] = tokenizer.eos_token _A : int = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt,num_labels=7 ) _A : Union[str, Any] = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): _A : int = False _A : int = ClassLabel(num_classes=7,names=list(set(train_test_validation["""train"""]["""complexity"""] ) ) ) def tokenize(snake_case_ ): _A : List[str] = tokenizer(example["""src"""],truncation=snake_case_,max_length=1024 ) _A : List[Any] = labels.straint(example["""complexity"""] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } _A : Optional[Any] = train_test_validation.map( snake_case_,batched=snake_case_,remove_columns=train_test_validation["""train"""].column_names,) _A : List[str] = DataCollatorWithPadding(tokenizer=snake_case_ ) _A : int = TrainingArguments( output_dir=args.output_dir,learning_rate=args.learning_rate,lr_scheduler_type=args.lr_scheduler_type,evaluation_strategy="""epoch""",save_strategy="""epoch""",logging_strategy="""epoch""",per_device_train_batch_size=args.batch_size,per_device_eval_batch_size=args.batch_size,num_train_epochs=args.num_epochs,gradient_accumulation_steps=args.gradient_accumulation_steps,weight_decay=0.01,metric_for_best_model="""accuracy""",run_name="""complexity-java""",report_to="""wandb""",) _A : Dict = Trainer( model=snake_case_,args=snake_case_,train_dataset=tokenized_datasets["""train"""],eval_dataset=tokenized_datasets["""valid"""],tokenizer=snake_case_,data_collator=snake_case_,compute_metrics=snake_case_,) print("""Training...""" ) trainer.add_callback(CustomCallback(snake_case_ ) ) trainer.train() if __name__ == "__main__": main()
26
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __UpperCAmelCase : def __init__( self: Any , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int]=13 , UpperCAmelCase_: str=7 , UpperCAmelCase_: int=True , UpperCAmelCase_: List[str]=True , UpperCAmelCase_: Dict=True , UpperCAmelCase_: Any=True , UpperCAmelCase_: Tuple=99 , UpperCAmelCase_: Optional[Any]=32 , UpperCAmelCase_: Optional[int]=2 , UpperCAmelCase_: Tuple=4 , UpperCAmelCase_: Tuple=37 , UpperCAmelCase_: Union[str, Any]="gelu" , UpperCAmelCase_: List[str]=0.1 , UpperCAmelCase_: int=0.1 , UpperCAmelCase_: str=512 , UpperCAmelCase_: Union[str, Any]=16 , UpperCAmelCase_: List[Any]=2 , UpperCAmelCase_: str=0.02 , UpperCAmelCase_: int=False , UpperCAmelCase_: Union[str, Any]=True , UpperCAmelCase_: Optional[Any]="None" , UpperCAmelCase_: Optional[int]=3 , UpperCAmelCase_: Any=4 , UpperCAmelCase_: Optional[int]=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = relative_attention _SCREAMING_SNAKE_CASE = position_biased_input _SCREAMING_SNAKE_CASE = pos_att_type _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=UpperCAmelCase_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} _SCREAMING_SNAKE_CASE = [input_ids, input_mask] _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: Tuple , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: int , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForMaskedLM(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Any , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForSequenceClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: Optional[int] , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = TFDebertaVaForTokenClassification(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: str , UpperCAmelCase_: str , UpperCAmelCase_: Any , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaForQuestionAnswering(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : int = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) __snake_case : Union[str, Any] = ( { "feature-extraction": TFDebertaVaModel, "fill-mask": TFDebertaVaForMaskedLM, "question-answering": TFDebertaVaForQuestionAnswering, "text-classification": TFDebertaVaForSequenceClassification, "token-classification": TFDebertaVaForTokenClassification, "zero-shot": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) __snake_case : Dict = False __snake_case : Optional[Any] = False def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(UpperCAmelCase_ ) @require_tf class __UpperCAmelCase (unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def UpperCamelCase ( self: Tuple ): '''simple docstring''' pass @slow def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) _SCREAMING_SNAKE_CASE = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = tf.constant( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , UpperCAmelCase_ , atol=1E-4 )
306
0
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class __UpperCamelCase ( unittest.TestCase ): def __UpperCAmelCase ( self , __a ): '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] , model_result['ss'] ): __a : List[Any] = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[str] = 'sshleifer/tiny-gpt2' __a : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=__a , multi_process=__a , ) __a : List[Any] = TensorFlowBenchmark(__a ) __a : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = 'sgugger/tiny-distilbert-classification' __a : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , only_pretrain_model=__a , ) __a : Dict = TensorFlowBenchmark(__a ) __a : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = 'sshleifer/tiny-gpt2' __a : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) __a : str = TensorFlowBenchmark(__a ) __a : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : str = 'sshleifer/tiny-gpt2' __a : Optional[Any] = AutoConfig.from_pretrained(__a ) __a : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=__a , multi_process=__a , ) __a : Tuple = TensorFlowBenchmark(__a , [config] ) __a : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = 'sshleifer/tiny-gpt2' __a : Optional[int] = AutoConfig.from_pretrained(__a ) __a : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) __a : Tuple = TensorFlowBenchmark(__a , [config] ) __a : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[Any] = 'sshleifer/tiny-gpt2' __a : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) __a : Optional[Any] = TensorFlowBenchmark(__a ) __a : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = 'sshleifer/tiny-gpt2' __a : Tuple = AutoConfig.from_pretrained(__a ) __a : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) __a : Optional[Any] = TensorFlowBenchmark(__a , [config] ) __a : Any = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = 'patrickvonplaten/t5-tiny-random' __a : Optional[int] = AutoConfig.from_pretrained(__a ) __a : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) __a : str = TensorFlowBenchmark(__a , configs=[config] ) __a : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('GPU' ) ) == 0 , 'Cannot do xla on CPU.' ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = 'sshleifer/tiny-gpt2' __a : str = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , use_xla=__a , multi_process=__a , ) __a : int = TensorFlowBenchmark(__a ) __a : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: __a : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=__a , save_to_csv=__a , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(__a , 'inf_time.csv' ) , inference_memory_csv_file=os.path.join(__a , 'inf_mem.csv' ) , env_info_csv_file=os.path.join(__a , 'env.csv' ) , multi_process=__a , ) __a : int = TensorFlowBenchmark(__a ) benchmark.run() self.assertTrue(Path(os.path.join(__a , 'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__a , 'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__a , 'env.csv' ) ).exists() ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(__a ): self.assertTrue(hasattr(__a , 'sequential' ) ) self.assertTrue(hasattr(__a , 'cumulative' ) ) self.assertTrue(hasattr(__a , 'current' ) ) self.assertTrue(hasattr(__a , 'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: __a : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(__a , 'log.txt' ) , log_print=__a , trace_memory_line_by_line=__a , eager_mode=__a , multi_process=__a , ) __a : Any = TensorFlowBenchmark(__a ) __a : str = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(__a , 'log.txt' ) ).exists() )
27
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __lowerCamelCase ( snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = gather(snake_case__ ) assert gathered_tensor.tolist() == list(range(1 ,state.num_processes**2 + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = [state.process_index] _SCREAMING_SNAKE_CASE = gather_object(snake_case__ ) assert len(snake_case__ ) == state.num_processes, F'{gathered_obj}, {len(snake_case__ )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = broadcast(snake_case__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 ,state.num_processes + 1 ) ) def __lowerCamelCase ( snake_case__ ) -> Tuple: """simple docstring""" if state.is_main_process: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes + 1 ).to(state.device ) else: _SCREAMING_SNAKE_CASE = torch.arange(state.num_processes ).to(state.device ) _SCREAMING_SNAKE_CASE = pad_across_processes(snake_case__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 ,state.num_processes ) ) + [0] def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""sum""" ) _SCREAMING_SNAKE_CASE = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" if state.num_processes != 2: return _SCREAMING_SNAKE_CASE = create_tensor(snake_case__ ) _SCREAMING_SNAKE_CASE = reduce(snake_case__ ,"""mean""" ) _SCREAMING_SNAKE_CASE = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(snake_case__ ,snake_case__ ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( snake_case__ ) -> str: """simple docstring""" main() def __lowerCamelCase ( ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = PartialState() state.print(F'State: {state}' ) state.print("""testing gather""" ) test_gather(snake_case__ ) state.print("""testing gather_object""" ) test_gather_object(snake_case__ ) state.print("""testing broadcast""" ) test_broadcast(snake_case__ ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(snake_case__ ) state.print("""testing reduce_sum""" ) test_reduce_sum(snake_case__ ) state.print("""testing reduce_mean""" ) test_reduce_mean(snake_case__ ) if __name__ == "__main__": main()
306
0
'''simple docstring''' from __future__ import annotations class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Any , UpperCamelCase__ : list[list[int]] ): """simple docstring""" UpperCamelCase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(UpperCamelCase__ ) != 0: UpperCamelCase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(UpperCamelCase__ ) != cols: raise error for value in row: if not isinstance(UpperCamelCase__ , (int, float) ): raise error UpperCamelCase = rows else: UpperCamelCase = [] def A ( self : Dict ): """simple docstring""" return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def A ( self : int ): """simple docstring""" return len(self.rows ) @property def A ( self : Dict ): """simple docstring""" return len(self.rows[0] ) @property def A ( self : int ): """simple docstring""" return (self.num_rows, self.num_columns) @property def A ( self : List[Any] ): """simple docstring""" return self.order[0] == self.order[1] def A ( self : List[Any] ): """simple docstring""" UpperCamelCase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(UpperCamelCase__ ) def A ( self : str ): """simple docstring""" if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def A ( self : Optional[Any] ): """simple docstring""" return bool(self.determinant() ) def A ( self : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" UpperCamelCase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(UpperCamelCase__ ).determinant() def A ( self : Optional[Any] , UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" if (row + column) % 2 == 0: return self.get_minor(UpperCamelCase__ , UpperCamelCase__ ) return -1 * self.get_minor(UpperCamelCase__ , UpperCamelCase__ ) def A ( self : Any ): """simple docstring""" return Matrix( [ [self.get_minor(UpperCamelCase__ , UpperCamelCase__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def A ( self : Optional[Any] ): """simple docstring""" return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def A ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(UpperCamelCase__ ) def A ( self : List[Any] ): """simple docstring""" UpperCamelCase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self : Union[str, Any] ): """simple docstring""" return str(self.rows ) def __str__( self : Any ): """simple docstring""" if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(UpperCamelCase__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def A ( self : Optional[Any] , UpperCamelCase__ : list[int] , UpperCamelCase__ : int | None = None ): """simple docstring""" UpperCamelCase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise type_error for value in row: if not isinstance(UpperCamelCase__ , (int, float) ): raise type_error if len(UpperCamelCase__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(UpperCamelCase__ ) else: UpperCamelCase = self.rows[0:position] + [row] + self.rows[position:] def A ( self : str , UpperCamelCase__ : list[int] , UpperCamelCase__ : int | None = None ): """simple docstring""" UpperCamelCase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise type_error for value in column: if not isinstance(UpperCamelCase__ , (int, float) ): raise type_error if len(UpperCamelCase__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: UpperCamelCase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCamelCase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self : Optional[Any] , UpperCamelCase__ : object ): """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): return NotImplemented return self.rows == other.rows def __ne__( self : int , UpperCamelCase__ : object ): """simple docstring""" return not self == other def __neg__( self : Optional[Any] ): """simple docstring""" return self * -1 def __add__( self : Optional[int] , UpperCamelCase__ : Matrix ): """simple docstring""" if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self : int , UpperCamelCase__ : Matrix ): """simple docstring""" if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self : List[str] , UpperCamelCase__ : Matrix | int | float ): """simple docstring""" if isinstance(UpperCamelCase__ , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(UpperCamelCase__ , UpperCamelCase__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self : Tuple , UpperCamelCase__ : int ): """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) UpperCamelCase = self for _ in range(other - 1 ): result *= self return result @classmethod def A ( cls : Optional[int] , UpperCamelCase__ : list[int] , UpperCamelCase__ : list[int] ): """simple docstring""" return sum(row[i] * column[i] for i in range(len(UpperCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
28
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __lowerCamelCase ( ) -> tuple[list[int], int]: """simple docstring""" _SCREAMING_SNAKE_CASE = [randint(-10_00 ,10_00 ) for i in range(10 )] _SCREAMING_SNAKE_CASE = randint(-50_00 ,50_00 ) return (arr, r) UpperCamelCase = make_dataset() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(snake_case__ ,3 ): if sum(snake_case__ ) == target: return tuple(sorted(snake_case__ ) ) return (0, 0, 0) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> tuple[int, int, int]: """simple docstring""" arr.sort() _SCREAMING_SNAKE_CASE = len(snake_case__ ) for i in range(n - 1 ): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __lowerCamelCase ( ) -> tuple[float, float]: """simple docstring""" _SCREAMING_SNAKE_CASE = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ _SCREAMING_SNAKE_CASE = """ triplet_sum1(*dataset) """ _SCREAMING_SNAKE_CASE = """ triplet_sum2(*dataset) """ _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) _SCREAMING_SNAKE_CASE = repeat(setup=snake_case__ ,stmt=snake_case__ ,repeat=5 ,number=1_00_00 ) return (min(snake_case__ ), min(snake_case__ )) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
306
0
def lowercase__ ( __snake_case : int ): '''simple docstring''' if n == 1 or not isinstance(__snake_case , __snake_case ): return 0 elif n == 2: return 1 else: UpperCAmelCase_ : Tuple = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowercase__ ( __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : int = 0 UpperCAmelCase_ : Any = 2 while digits < n: index += 1 UpperCAmelCase_ : Tuple = len(str(fibonacci(__snake_case ) ) ) return index def lowercase__ ( __snake_case : int = 1_000 ): '''simple docstring''' return fibonacci_digits_index(__snake_case ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
29
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase ) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' super().__init__(**UpperCAmelCase_ ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) self.check_model_type(UpperCAmelCase_ ) def UpperCamelCase ( self: str , **UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = {} # preprocess args if "points_per_batch" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self: Optional[Any] , UpperCAmelCase_: Tuple , *UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=None , UpperCAmelCase_: Tuple=None , **UpperCAmelCase_: Any ): '''simple docstring''' return super().__call__(UpperCAmelCase_ , *UpperCAmelCase_ , num_workers=UpperCAmelCase_ , batch_size=UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[str] , UpperCAmelCase_: Dict=64 , UpperCAmelCase_: int = 0 , UpperCAmelCase_: float = 512 / 1_500 , UpperCAmelCase_: Optional[int] = 32 , UpperCAmelCase_: Optional[int] = 1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor.size["""longest_edge"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.generate_crop_boxes( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.image_processor(images=UpperCAmelCase_ , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": _SCREAMING_SNAKE_CASE = self.get_inference_context() with inference_context(): _SCREAMING_SNAKE_CASE = self._ensure_tensor_on_device(UpperCAmelCase_ , device=self.device ) _SCREAMING_SNAKE_CASE = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) _SCREAMING_SNAKE_CASE = image_embeddings _SCREAMING_SNAKE_CASE = grid_points.shape[1] _SCREAMING_SNAKE_CASE = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = grid_points[:, i : i + points_per_batch, :, :] _SCREAMING_SNAKE_CASE = input_labels[:, i : i + points_per_batch] _SCREAMING_SNAKE_CASE = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCamelCase ( self: Any , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Optional[Any]=0.88 , UpperCAmelCase_: Dict=0.95 , UpperCAmelCase_: Tuple=0 , UpperCAmelCase_: str=1 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_inputs.pop("""input_boxes""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""is_last""" ) _SCREAMING_SNAKE_CASE = model_inputs.pop("""original_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = model_inputs.pop("""reshaped_input_sizes""" ).tolist() _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks _SCREAMING_SNAKE_CASE = model_outputs["""pred_masks"""] _SCREAMING_SNAKE_CASE = self.image_processor.post_process_masks( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , binarize=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model_outputs["""iou_scores"""] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCamelCase ( self: Any , UpperCAmelCase_: List[Any] , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=False , UpperCAmelCase_: Any=0.7 , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.cat(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = self.image_processor.post_process_for_mask_generation( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = defaultdict(UpperCAmelCase_ ) for output in model_outputs: for k, v in output.items(): extra[k].append(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {} if output_rle_mask: _SCREAMING_SNAKE_CASE = rle_mask if output_bboxes_mask: _SCREAMING_SNAKE_CASE = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
306
0
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def a ( ): '''simple docstring''' lowercase_ = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowercase_ = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(snake_case__ ) DownloadCommand.register_subcommand(snake_case__ ) EnvironmentCommand.register_subcommand(snake_case__ ) RunCommand.register_subcommand(snake_case__ ) ServeCommand.register_subcommand(snake_case__ ) UserCommands.register_subcommand(snake_case__ ) AddNewModelCommand.register_subcommand(snake_case__ ) AddNewModelLikeCommand.register_subcommand(snake_case__ ) LfsCommands.register_subcommand(snake_case__ ) PTtoTFCommand.register_subcommand(snake_case__ ) # Let's go lowercase_ = parser.parse_args() if not hasattr(snake_case__ , '''func''' ): parser.print_help() exit(1 ) # Run lowercase_ = args.func(snake_case__ ) service.run() if __name__ == "__main__": main()
30
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging UpperCamelCase = logging.get_logger(__name__) logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" if "xprophetnet" in prophetnet_checkpoint_path: _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = XLMProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) else: _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGenerationOld.from_pretrained(snake_case__ ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ProphetNetForConditionalGeneration.from_pretrained( snake_case__ ,output_loading_info=snake_case__ ) _SCREAMING_SNAKE_CASE = ["""key_proj""", """value_proj""", """query_proj"""] _SCREAMING_SNAKE_CASE = { """self_attn""": """ngram_self_attn""", """cross_attn""": """encoder_attn""", """cross_attn_layer_norm""": """encoder_attn_layer_norm""", """feed_forward_layer_norm""": """final_layer_norm""", """feed_forward""": """""", """intermediate""": """fc1""", """output""": """fc2""", """key_proj""": """k_proj""", """query_proj""": """q_proj""", """value_proj""": """v_proj""", """word_embeddings""": """embed_tokens""", """embeddings_layer_norm""": """emb_layer_norm""", """relative_pos_embeddings""": """relative_linear""", """ngram_embeddings""": """ngram_input_embed""", """position_embeddings""": """embed_positions""", } for key in loading_info["missing_keys"]: _SCREAMING_SNAKE_CASE = key.split(""".""" ) if attributes[0] == "lm_head": _SCREAMING_SNAKE_CASE = prophet _SCREAMING_SNAKE_CASE = prophet_old else: _SCREAMING_SNAKE_CASE = prophet.prophetnet _SCREAMING_SNAKE_CASE = prophet_old.model _SCREAMING_SNAKE_CASE = False for attribute in attributes: if attribute in mapping: _SCREAMING_SNAKE_CASE = mapping[attribute] if not hasattr(snake_case__ ,snake_case__ ) and len(snake_case__ ) > 0: _SCREAMING_SNAKE_CASE = attribute elif hasattr(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.weight logger.info(F'{attribute} is initialized.' ) _SCREAMING_SNAKE_CASE = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" _SCREAMING_SNAKE_CASE = old_model.bias logger.info(F'{attribute} is initialized' ) _SCREAMING_SNAKE_CASE = True break elif attribute in special_keys and hasattr(snake_case__ ,"""in_proj_weight""" ): _SCREAMING_SNAKE_CASE = old_model.in_proj_weight.shape[0] // 3 _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) _SCREAMING_SNAKE_CASE = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_12, "We want 512 position_embeddings." _SCREAMING_SNAKE_CASE = nn.Parameter(old_model.embed_positions.weight[:5_12, :] ) _SCREAMING_SNAKE_CASE = True break if attribute.isdigit(): _SCREAMING_SNAKE_CASE = model[int(snake_case__ )] _SCREAMING_SNAKE_CASE = old_model[int(snake_case__ )] else: _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if old_attribute == "": _SCREAMING_SNAKE_CASE = old_model else: if not hasattr(snake_case__ ,snake_case__ ): raise ValueError(F'{old_model} does not have {old_attribute}' ) _SCREAMING_SNAKE_CASE = getattr(snake_case__ ,snake_case__ ) if not is_key_init: raise ValueError(F'{key} was not correctly initialized!' ) print(F'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--prophetnet_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
306
0
'''simple docstring''' import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class lowerCamelCase_ (snake_case__ , snake_case__ ): '''simple docstring''' __UpperCamelCase: Optional[Any] = 1 @register_to_config def __init__( self : Optional[int] , A : int = 1000 , A : Optional[Union[np.ndarray, List[float]]] = None ): # set `betas`, `alphas`, `timesteps` self.set_timesteps(A ) # standard deviation of the initial noise distribution _UpperCAmelCase : int = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. _UpperCAmelCase : int = 4 # running values _UpperCAmelCase : Dict = [] def _A ( self : Optional[int] , A : int , A : Union[str, torch.device] = None ): _UpperCAmelCase : int = num_inference_steps _UpperCAmelCase : Union[str, Any] = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] _UpperCAmelCase : Any = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: _UpperCAmelCase : str = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: _UpperCAmelCase : Dict = torch.sin(steps * math.pi / 2 ) ** 2 _UpperCAmelCase : List[Any] = (1.0 - self.betas**2) ** 0.5 _UpperCAmelCase : List[str] = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] _UpperCAmelCase : Dict = timesteps.to(A ) _UpperCAmelCase : Dict = [] def _A ( self : Optional[int] , A : torch.FloatTensor , A : int , A : torch.FloatTensor , A : bool = True , ): if self.num_inference_steps is None: raise ValueError( "Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler" ) _UpperCAmelCase : Tuple = (self.timesteps == timestep).nonzero().item() _UpperCAmelCase : Optional[Any] = timestep_index + 1 _UpperCAmelCase : int = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(A ) if len(self.ets ) == 1: _UpperCAmelCase : List[Any] = self.ets[-1] elif len(self.ets ) == 2: _UpperCAmelCase : str = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: _UpperCAmelCase : Tuple = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: _UpperCAmelCase : Union[str, Any] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) _UpperCAmelCase : Union[str, Any] = self._get_prev_sample(A , A , A , A ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=A ) def _A ( self : Union[str, Any] , A : torch.FloatTensor , *A : Union[str, Any] , **A : Dict ): return sample def _A ( self : Optional[Any] , A : Optional[int] , A : int , A : Optional[Any] , A : List[str] ): _UpperCAmelCase : List[str] = self.alphas[timestep_index] _UpperCAmelCase : List[Any] = self.betas[timestep_index] _UpperCAmelCase : Optional[Any] = self.alphas[prev_timestep_index] _UpperCAmelCase : Dict = self.betas[prev_timestep_index] _UpperCAmelCase : Tuple = (sample - sigma * ets) / max(A , 1E-8 ) _UpperCAmelCase : List[str] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Union[str, Any] ): return self.config.num_train_timesteps
31
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> list[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = len(snake_case__ ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: _SCREAMING_SNAKE_CASE = i + 1 else: _SCREAMING_SNAKE_CASE = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"{two_pointer([2, 7, 11, 15], 9) = }")
306
0
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( lowercase__ , unittest.TestCase ): snake_case__ : Union[str, Any] = LayoutLMTokenizer snake_case__ : Union[str, Any] = LayoutLMTokenizerFast snake_case__ : Union[str, Any] = True snake_case__ : str = True def SCREAMING_SNAKE_CASE ( self : Tuple ) -> int: super().setUp() a_ : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] a_ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , **SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Union[str, Any]: return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Any , SCREAMING_SNAKE_CASE__ : List[str] ) -> Dict: a_ : int = 'UNwant\u00E9d,running' a_ : Optional[int] = 'unwanted, running' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Any: a_ : List[str] = self.tokenizer_class(self.vocab_file ) a_ : Tuple = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) , [7, 4, 5, 1_0, 8, 9] ) def SCREAMING_SNAKE_CASE ( self : int ) -> str: pass
32
from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig UpperCamelCase = logging.get_logger(__name__) # General docstring UpperCamelCase = '''MobileNetV1Config''' # Base docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = [1, 1_024, 7, 7] # Image classification docstring UpperCamelCase = '''google/mobilenet_v1_1.0_224''' UpperCamelCase = '''tabby, tabby cat''' UpperCamelCase = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = {} if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = model.mobilenet_va else: _SCREAMING_SNAKE_CASE = model _SCREAMING_SNAKE_CASE = """MobilenetV1/Conv2d_0/""" _SCREAMING_SNAKE_CASE = backbone.conv_stem.convolution.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.bias _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.weight _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_mean _SCREAMING_SNAKE_CASE = backbone.conv_stem.normalization.running_var for i in range(13 ): _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = i * 2 _SCREAMING_SNAKE_CASE = backbone.layer[pt_index] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_depthwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var _SCREAMING_SNAKE_CASE = backbone.layer[pt_index + 1] _SCREAMING_SNAKE_CASE = F'MobilenetV1/Conv2d_{tf_index}_pointwise/' _SCREAMING_SNAKE_CASE = pointer.convolution.weight _SCREAMING_SNAKE_CASE = pointer.normalization.bias _SCREAMING_SNAKE_CASE = pointer.normalization.weight _SCREAMING_SNAKE_CASE = pointer.normalization.running_mean _SCREAMING_SNAKE_CASE = pointer.normalization.running_var if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = """MobilenetV1/Logits/Conv2d_1c_1x1/""" _SCREAMING_SNAKE_CASE = model.classifier.weight _SCREAMING_SNAKE_CASE = model.classifier.bias return tf_to_pt_map def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> List[str]: """simple docstring""" try: import numpy as np import tensorflow as tf except ImportError: logger.error( """Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see """ """https://www.tensorflow.org/install/ for installation instructions.""" ) raise # Load weights from TF model _SCREAMING_SNAKE_CASE = tf.train.list_variables(snake_case__ ) _SCREAMING_SNAKE_CASE = {} for name, shape in init_vars: logger.info(F'Loading TF weight {name} with shape {shape}' ) _SCREAMING_SNAKE_CASE = tf.train.load_variable(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = array # Build TF to PyTorch weights loading map _SCREAMING_SNAKE_CASE = _build_tf_to_pytorch_map(snake_case__ ,snake_case__ ,snake_case__ ) for name, pointer in tf_to_pt_map.items(): logger.info(F'Importing {name}' ) if name not in tf_weights: logger.info(F'{name} not in tf pre-trained weights, skipping' ) continue _SCREAMING_SNAKE_CASE = tf_weights[name] if "depthwise_weights" in name: logger.info("""Transposing depthwise""" ) _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(2, 3, 0, 1) ) elif "weights" in name: logger.info("""Transposing""" ) if len(pointer.shape ) == 2: # copying into linear layer _SCREAMING_SNAKE_CASE = array.squeeze().transpose() else: _SCREAMING_SNAKE_CASE = np.transpose(snake_case__ ,(3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'Pointer shape {pointer.shape} and array shape {array.shape} mismatched' ) logger.info(F'Initialize PyTorch weight {name} {array.shape}' ) _SCREAMING_SNAKE_CASE = torch.from_numpy(snake_case__ ) tf_weights.pop(snake_case__ ,snake_case__ ) tf_weights.pop(name + """/RMSProp""" ,snake_case__ ) tf_weights.pop(name + """/RMSProp_1""" ,snake_case__ ) tf_weights.pop(name + """/ExponentialMovingAverage""" ,snake_case__ ) logger.info(F'Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}' ) return model def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> torch.Tensor: """simple docstring""" _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = features.shape[-2:] _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.stride _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = conv_layer.kernel_size if in_height % stride_height == 0: _SCREAMING_SNAKE_CASE = max(kernel_height - stride_height ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_height - (in_height % stride_height) ,0 ) if in_width % stride_width == 0: _SCREAMING_SNAKE_CASE = max(kernel_width - stride_width ,0 ) else: _SCREAMING_SNAKE_CASE = max(kernel_width - (in_width % stride_width) ,0 ) _SCREAMING_SNAKE_CASE = pad_along_width // 2 _SCREAMING_SNAKE_CASE = pad_along_width - pad_left _SCREAMING_SNAKE_CASE = pad_along_height // 2 _SCREAMING_SNAKE_CASE = pad_along_height - pad_top _SCREAMING_SNAKE_CASE = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(snake_case__ ,snake_case__ ,"""constant""" ,0.0 ) class __UpperCAmelCase (nn.Module ): def __init__( self: Optional[Any] , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: Optional[int] = 1 , UpperCAmelCase_: bool = False , UpperCAmelCase_: Optional[bool] = True , UpperCAmelCase_: Optional[bool or str] = True , ): '''simple docstring''' super().__init__() _SCREAMING_SNAKE_CASE = config if in_channels % groups != 0: raise ValueError(F'Input channels ({in_channels}) are not divisible by {groups} groups.' ) if out_channels % groups != 0: raise ValueError(F'Output channels ({out_channels}) are not divisible by {groups} groups.' ) _SCREAMING_SNAKE_CASE = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) _SCREAMING_SNAKE_CASE = nn.Convad( in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=UpperCAmelCase_ , stride=UpperCAmelCase_ , padding=UpperCAmelCase_ , groups=UpperCAmelCase_ , bias=UpperCAmelCase_ , padding_mode="""zeros""" , ) if use_normalization: _SCREAMING_SNAKE_CASE = nn.BatchNormad( num_features=UpperCAmelCase_ , eps=config.layer_norm_eps , momentum=0.99_97 , affine=UpperCAmelCase_ , track_running_stats=UpperCAmelCase_ , ) else: _SCREAMING_SNAKE_CASE = None if use_activation: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[use_activation] elif isinstance(config.hidden_act , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = ACTaFN[config.hidden_act] else: _SCREAMING_SNAKE_CASE = config.hidden_act else: _SCREAMING_SNAKE_CASE = None def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' if self.config.tf_padding: _SCREAMING_SNAKE_CASE = apply_tf_padding(UpperCAmelCase_ , self.convolution ) _SCREAMING_SNAKE_CASE = self.convolution(UpperCAmelCase_ ) if self.normalization is not None: _SCREAMING_SNAKE_CASE = self.normalization(UpperCAmelCase_ ) if self.activation is not None: _SCREAMING_SNAKE_CASE = self.activation(UpperCAmelCase_ ) return features class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : Dict = MobileNetVaConfig __snake_case : Any = load_tf_weights_in_mobilenet_va __snake_case : Any = "mobilenet_v1" __snake_case : List[Any] = "pixel_values" __snake_case : Any = False def UpperCamelCase ( self: str , UpperCAmelCase_: Union[nn.Linear, nn.Convad] ): '''simple docstring''' if isinstance(UpperCAmelCase_ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(UpperCAmelCase_ , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) UpperCamelCase = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' UpperCamelCase = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare MobileNetV1 model outputting raw hidden-states without any specific head on top." ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Any , UpperCAmelCase_: MobileNetVaConfig , UpperCAmelCase_: bool = True ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config _SCREAMING_SNAKE_CASE = 32 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) _SCREAMING_SNAKE_CASE = MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=config.num_channels , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=2 , ) _SCREAMING_SNAKE_CASE = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] _SCREAMING_SNAKE_CASE = nn.ModuleList() for i in range(13 ): _SCREAMING_SNAKE_CASE = out_channels if strides[i] == 2 or i == 0: depth *= 2 _SCREAMING_SNAKE_CASE = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=3 , stride=strides[i] , groups=UpperCAmelCase_ , ) ) self.layer.append( MobileNetVaConvLayer( UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , kernel_size=1 , ) ) _SCREAMING_SNAKE_CASE = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def UpperCamelCase ( self: Dict , UpperCAmelCase_: Tuple ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase ( self: int , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("""You have to specify pixel_values""" ) _SCREAMING_SNAKE_CASE = self.conv_stem(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): _SCREAMING_SNAKE_CASE = layer_module(UpperCAmelCase_ ) if output_hidden_states: _SCREAMING_SNAKE_CASE = all_hidden_states + (hidden_states,) _SCREAMING_SNAKE_CASE = hidden_states if self.pooler is not None: _SCREAMING_SNAKE_CASE = torch.flatten(self.pooler(UpperCAmelCase_ ) , start_dim=1 ) else: _SCREAMING_SNAKE_CASE = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCAmelCase_ , pooler_output=UpperCAmelCase_ , hidden_states=UpperCAmelCase_ , ) @add_start_docstrings( "\n MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " ,_UpperCAmelCase ,) class __UpperCAmelCase (_UpperCAmelCase ): def __init__( self: Dict , UpperCAmelCase_: MobileNetVaConfig ): '''simple docstring''' super().__init__(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config.num_labels _SCREAMING_SNAKE_CASE = MobileNetVaModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head _SCREAMING_SNAKE_CASE = nn.Dropout(config.classifier_dropout_prob , inplace=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = nn.Linear(UpperCAmelCase_ , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCAmelCase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCAmelCase_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , UpperCAmelCase_: Optional[torch.Tensor] = None , UpperCAmelCase_: Optional[bool] = None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = return_dict if return_dict is not None else self.config.use_return_dict _SCREAMING_SNAKE_CASE = self.mobilenet_va(UpperCAmelCase_ , output_hidden_states=UpperCAmelCase_ , return_dict=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.pooler_output if return_dict else outputs[1] _SCREAMING_SNAKE_CASE = self.classifier(self.dropout(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _SCREAMING_SNAKE_CASE = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _SCREAMING_SNAKE_CASE = """single_label_classification""" else: _SCREAMING_SNAKE_CASE = """multi_label_classification""" if self.config.problem_type == "regression": _SCREAMING_SNAKE_CASE = MSELoss() if self.num_labels == 1: _SCREAMING_SNAKE_CASE = loss_fct(logits.squeeze() , labels.squeeze() ) else: _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) elif self.config.problem_type == "single_label_classification": _SCREAMING_SNAKE_CASE = CrossEntropyLoss() _SCREAMING_SNAKE_CASE = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _SCREAMING_SNAKE_CASE = BCEWithLogitsLoss() _SCREAMING_SNAKE_CASE = loss_fct(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: _SCREAMING_SNAKE_CASE = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=UpperCAmelCase_ , logits=UpperCAmelCase_ , hidden_states=outputs.hidden_states , )
306
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A : Optional[int] = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Optional[Any] = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys __A : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
33
def __lowerCamelCase ( snake_case__ ) -> list: """simple docstring""" def merge(snake_case__ ,snake_case__ ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(snake_case__ ) <= 1: return collection _SCREAMING_SNAKE_CASE = len(snake_case__ ) // 2 return merge(merge_sort(collection[:mid] ) ,merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
306
0
'''simple docstring''' # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys A =subprocess.check_output('git merge-base main HEAD'.split()).decode('utf-8') A =subprocess.check_output(f"""git diff --name-only {fork_point_sha}""".split()).decode('utf-8').split() A ='|'.join(sys.argv[1:]) A =re.compile(rf"""^({joined_dirs}).*?\.py$""") A =[x for x in modified_files if regex.match(x)] print(' '.join(relevant_modified_files), end='')
34
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length, 2) ,snake_case__ ) else: _SCREAMING_SNAKE_CASE = np.full((len(snake_case__ ), sequence_length) ,snake_case__ ) for i, tensor in enumerate(snake_case__ ): if padding_side == "right": if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: if isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = tensor[:sequence_length] else: _SCREAMING_SNAKE_CASE = tensor[:sequence_length] return out_tensor.tolist() def __lowerCamelCase ( snake_case__ ) -> Dict: """simple docstring""" _SCREAMING_SNAKE_CASE = ord(snake_case__ ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 1_23 and cp <= 1_26): return True _SCREAMING_SNAKE_CASE = unicodedata.category(snake_case__ ) if cat.startswith("""P""" ): return True return False @dataclass class __UpperCAmelCase (_UpperCAmelCase ): __snake_case : PreTrainedTokenizerBase __snake_case : Union[bool, str, PaddingStrategy] = True __snake_case : Optional[int] = None __snake_case : Optional[int] = None __snake_case : int = -100 __snake_case : str = "pt" def UpperCamelCase ( self: str , UpperCAmelCase_: Optional[Any] ): '''simple docstring''' import torch _SCREAMING_SNAKE_CASE = """label""" if """label""" in features[0].keys() else """labels""" _SCREAMING_SNAKE_CASE = [feature[label_name] for feature in features] if label_name in features[0].keys() else None _SCREAMING_SNAKE_CASE = self.tokenizer.pad( UpperCAmelCase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" if labels is None else None , ) if labels is None: return batch _SCREAMING_SNAKE_CASE = torch.tensor(batch["""entity_ids"""] ).shape[1] _SCREAMING_SNAKE_CASE = self.tokenizer.padding_side if padding_side == "right": _SCREAMING_SNAKE_CASE = [ list(UpperCAmelCase_ ) + [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) for label in labels ] else: _SCREAMING_SNAKE_CASE = [ [self.label_pad_token_id] * (sequence_length - len(UpperCAmelCase_ )) + list(UpperCAmelCase_ ) for label in labels ] _SCREAMING_SNAKE_CASE = [feature["""ner_tags"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , -1 , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = [feature["""original_entity_spans"""] for feature in features] _SCREAMING_SNAKE_CASE = padding_tensor(UpperCAmelCase_ , (-1, -1) , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {k: torch.tensor(UpperCAmelCase_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
306
0
'''simple docstring''' import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class UpperCAmelCase_ ( _a ): """simple docstring""" def __get__( self : str , snake_case_ : Union[str, Any] , snake_case_ : Tuple=None ): # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError("""unreadable attribute""" ) snake_case__ : Optional[int] = """__cached_""" + self.fget.__name__ snake_case__ : int = getattr(snake_case_ , snake_case_ , snake_case_ ) if cached is None: snake_case__ : Union[str, Any] = self.fget(snake_case_ ) setattr(snake_case_ , snake_case_ , snake_case_ ) return cached def __snake_case( _lowerCAmelCase ) -> Tuple: snake_case__ : Optional[int] = val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f"invalid truth value {val!r}" ) def __snake_case( _lowerCAmelCase ) -> Any: if is_torch_fx_proxy(_lowerCAmelCase ): return True if is_torch_available(): import torch if isinstance(_lowerCAmelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(_lowerCAmelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(_lowerCAmelCase , (jnp.ndarray, Tracer) ): return True return isinstance(_lowerCAmelCase , np.ndarray ) def __snake_case( _lowerCAmelCase ) -> Optional[Any]: return isinstance(_lowerCAmelCase , np.ndarray ) def __snake_case( _lowerCAmelCase ) -> int: return _is_numpy(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> List[str]: import torch return isinstance(_lowerCAmelCase , torch.Tensor ) def __snake_case( _lowerCAmelCase ) -> str: return False if not is_torch_available() else _is_torch(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> List[str]: import torch return isinstance(_lowerCAmelCase , torch.device ) def __snake_case( _lowerCAmelCase ) -> List[Any]: return False if not is_torch_available() else _is_torch_device(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> List[Any]: import torch if isinstance(_lowerCAmelCase , _lowerCAmelCase ): if hasattr(_lowerCAmelCase , _lowerCAmelCase ): snake_case__ : Union[str, Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) else: return False return isinstance(_lowerCAmelCase , torch.dtype ) def __snake_case( _lowerCAmelCase ) -> List[Any]: return False if not is_torch_available() else _is_torch_dtype(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> List[str]: import tensorflow as tf return isinstance(_lowerCAmelCase , tf.Tensor ) def __snake_case( _lowerCAmelCase ) -> Any: return False if not is_tf_available() else _is_tensorflow(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> Optional[int]: import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(_lowerCAmelCase , """is_symbolic_tensor""" ): return tf.is_symbolic_tensor(_lowerCAmelCase ) return type(_lowerCAmelCase ) == tf.Tensor def __snake_case( _lowerCAmelCase ) -> Optional[Any]: return False if not is_tf_available() else _is_tf_symbolic_tensor(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> Dict: import jax.numpy as jnp # noqa: F811 return isinstance(_lowerCAmelCase , jnp.ndarray ) def __snake_case( _lowerCAmelCase ) -> List[str]: return False if not is_flax_available() else _is_jax(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> Any: if isinstance(_lowerCAmelCase , (dict, UserDict) ): return {k: to_py_obj(_lowerCAmelCase ) for k, v in obj.items()} elif isinstance(_lowerCAmelCase , (list, tuple) ): return [to_py_obj(_lowerCAmelCase ) for o in obj] elif is_tf_tensor(_lowerCAmelCase ): return obj.numpy().tolist() elif is_torch_tensor(_lowerCAmelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(_lowerCAmelCase ): return np.asarray(_lowerCAmelCase ).tolist() elif isinstance(_lowerCAmelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def __snake_case( _lowerCAmelCase ) -> List[Any]: if isinstance(_lowerCAmelCase , (dict, UserDict) ): return {k: to_numpy(_lowerCAmelCase ) for k, v in obj.items()} elif isinstance(_lowerCAmelCase , (list, tuple) ): return np.array(_lowerCAmelCase ) elif is_tf_tensor(_lowerCAmelCase ): return obj.numpy() elif is_torch_tensor(_lowerCAmelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(_lowerCAmelCase ): return np.asarray(_lowerCAmelCase ) else: return obj class UpperCAmelCase_ ( _a ): """simple docstring""" def lowerCamelCase ( self : Union[str, Any] ): snake_case__ : List[str] = fields(self ) # Safety and consistency checks if not len(snake_case_ ): raise ValueError(f"{self.__class__.__name__} has no fields." ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f"{self.__class__.__name__} should not have more than one required field." ) snake_case__ : List[str] = getattr(self , class_fields[0].name ) snake_case__ : int = all(getattr(self , field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(snake_case_ ): if isinstance(snake_case_ , snake_case_ ): snake_case__ : List[str] = first_field.items() snake_case__ : Optional[int] = True else: try: snake_case__ : Optional[Any] = iter(snake_case_ ) snake_case__ : Optional[int] = True except TypeError: snake_case__ : Dict = False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(snake_case_ ): if ( not isinstance(snake_case_ , (list, tuple) ) or not len(snake_case_ ) == 2 or not isinstance(element[0] , snake_case_ ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute snake_case__ : int = first_field else: # If we have a mixed iterator, raise an error raise ValueError( f"Cannot set key/value for {element}. It needs to be a tuple (key, value)." ) break setattr(self , element[0] , element[1] ) if element[1] is not None: snake_case__ : Union[str, Any] = element[1] elif first_field is not None: snake_case__ : List[str] = first_field else: for field in class_fields: snake_case__ : List[Any] = getattr(self , field.name ) if v is not None: snake_case__ : Dict = v def __delitem__( self : Union[str, Any] , *snake_case_ : Tuple , **snake_case_ : int ): raise Exception(f"You cannot use ``__delitem__`` on a {self.__class__.__name__} instance." ) def lowerCamelCase ( self : List[str] , *snake_case_ : List[Any] , **snake_case_ : Optional[Any] ): raise Exception(f"You cannot use ``setdefault`` on a {self.__class__.__name__} instance." ) def lowerCamelCase ( self : Dict , *snake_case_ : List[Any] , **snake_case_ : str ): raise Exception(f"You cannot use ``pop`` on a {self.__class__.__name__} instance." ) def lowerCamelCase ( self : List[str] , *snake_case_ : Tuple , **snake_case_ : Tuple ): raise Exception(f"You cannot use ``update`` on a {self.__class__.__name__} instance." ) def __getitem__( self : Tuple , snake_case_ : Optional[Any] ): if isinstance(snake_case_ , snake_case_ ): snake_case__ : int = dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : List[str] , snake_case_ : Tuple , snake_case_ : Union[str, Any] ): if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(snake_case_ , snake_case_ ) super().__setattr__(snake_case_ , snake_case_ ) def __setitem__( self : str , snake_case_ : Union[str, Any] , snake_case_ : int ): # Will raise a KeyException if needed super().__setitem__(snake_case_ , snake_case_ ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(snake_case_ , snake_case_ ) def lowerCamelCase ( self : int ): return tuple(self[k] for k in self.keys() ) class UpperCAmelCase_ ( _a , _a ): """simple docstring""" @classmethod def lowerCamelCase ( cls : int , snake_case_ : List[Any] ): raise ValueError( f"{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}" ) class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = "longest" lowercase = "max_length" lowercase = "do_not_pad" class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = "pt" lowercase = "tf" lowercase = "np" lowercase = "jax" class UpperCAmelCase_ : """simple docstring""" def __init__( self : int , snake_case_ : List[ContextManager] ): snake_case__ : Any = context_managers snake_case__ : int = ExitStack() def __enter__( self : Union[str, Any] ): for context_manager in self.context_managers: self.stack.enter_context(snake_case_ ) def __exit__( self : Optional[int] , *snake_case_ : List[Any] , **snake_case_ : Any ): self.stack.__exit__(*snake_case_ , **snake_case_ ) def __snake_case( _lowerCAmelCase ) -> List[str]: snake_case__ : Optional[Any] = infer_framework(_lowerCAmelCase ) if framework == "tf": snake_case__ : Optional[Any] = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": snake_case__ : str = inspect.signature(model_class.forward ) # PyTorch models else: snake_case__ : str = inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def __snake_case( _lowerCAmelCase ) -> Dict: snake_case__ : Optional[int] = model_class.__name__ snake_case__ : int = infer_framework(_lowerCAmelCase ) if framework == "tf": snake_case__ : List[str] = inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": snake_case__ : int = inspect.signature(model_class.forward ) # PyTorch models else: snake_case__ : Any = inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def __snake_case( _lowerCAmelCase , _lowerCAmelCase = "" , _lowerCAmelCase = "." ) -> List[str]: def _flatten_dict(_lowerCAmelCase , _lowerCAmelCase="" , _lowerCAmelCase="." ): for k, v in d.items(): snake_case__ : Optional[int] = str(_lowerCAmelCase ) + delimiter + str(_lowerCAmelCase ) if parent_key else k if v and isinstance(_lowerCAmelCase , _lowerCAmelCase ): yield from flatten_dict(_lowerCAmelCase , _lowerCAmelCase , delimiter=_lowerCAmelCase ).items() else: yield key, v return dict(_flatten_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) ) @contextmanager def __snake_case( _lowerCAmelCase , _lowerCAmelCase = False ) -> Tuple: if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def __snake_case( _lowerCAmelCase , _lowerCAmelCase=None ) -> Optional[Any]: if is_numpy_array(_lowerCAmelCase ): return np.transpose(_lowerCAmelCase , axes=_lowerCAmelCase ) elif is_torch_tensor(_lowerCAmelCase ): return array.T if axes is None else array.permute(*_lowerCAmelCase ) elif is_tf_tensor(_lowerCAmelCase ): import tensorflow as tf return tf.transpose(_lowerCAmelCase , perm=_lowerCAmelCase ) elif is_jax_tensor(_lowerCAmelCase ): return jnp.transpose(_lowerCAmelCase , axes=_lowerCAmelCase ) else: raise ValueError(f"Type not supported for transpose: {type(_lowerCAmelCase )}." ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: if is_numpy_array(_lowerCAmelCase ): return np.reshape(_lowerCAmelCase , _lowerCAmelCase ) elif is_torch_tensor(_lowerCAmelCase ): return array.reshape(*_lowerCAmelCase ) elif is_tf_tensor(_lowerCAmelCase ): import tensorflow as tf return tf.reshape(_lowerCAmelCase , _lowerCAmelCase ) elif is_jax_tensor(_lowerCAmelCase ): return jnp.reshape(_lowerCAmelCase , _lowerCAmelCase ) else: raise ValueError(f"Type not supported for reshape: {type(_lowerCAmelCase )}." ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase=None ) -> Dict: if is_numpy_array(_lowerCAmelCase ): return np.squeeze(_lowerCAmelCase , axis=_lowerCAmelCase ) elif is_torch_tensor(_lowerCAmelCase ): return array.squeeze() if axis is None else array.squeeze(dim=_lowerCAmelCase ) elif is_tf_tensor(_lowerCAmelCase ): import tensorflow as tf return tf.squeeze(_lowerCAmelCase , axis=_lowerCAmelCase ) elif is_jax_tensor(_lowerCAmelCase ): return jnp.squeeze(_lowerCAmelCase , axis=_lowerCAmelCase ) else: raise ValueError(f"Type not supported for squeeze: {type(_lowerCAmelCase )}." ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: if is_numpy_array(_lowerCAmelCase ): return np.expand_dims(_lowerCAmelCase , _lowerCAmelCase ) elif is_torch_tensor(_lowerCAmelCase ): return array.unsqueeze(dim=_lowerCAmelCase ) elif is_tf_tensor(_lowerCAmelCase ): import tensorflow as tf return tf.expand_dims(_lowerCAmelCase , axis=_lowerCAmelCase ) elif is_jax_tensor(_lowerCAmelCase ): return jnp.expand_dims(_lowerCAmelCase , axis=_lowerCAmelCase ) else: raise ValueError(f"Type not supported for expand_dims: {type(_lowerCAmelCase )}." ) def __snake_case( _lowerCAmelCase ) -> Optional[Any]: if is_numpy_array(_lowerCAmelCase ): return np.size(_lowerCAmelCase ) elif is_torch_tensor(_lowerCAmelCase ): return array.numel() elif is_tf_tensor(_lowerCAmelCase ): import tensorflow as tf return tf.size(_lowerCAmelCase ) elif is_jax_tensor(_lowerCAmelCase ): return array.size else: raise ValueError(f"Type not supported for expand_dims: {type(_lowerCAmelCase )}." ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> Dict: for key, value in auto_map.items(): if isinstance(_lowerCAmelCase , (tuple, list) ): snake_case__ : Tuple = [f"{repo_id}--{v}" if (v is not None and """--""" not in v) else v for v in value] elif value is not None and "--" not in value: snake_case__ : str = f"{repo_id}--{value}" return auto_map def __snake_case( _lowerCAmelCase ) -> int: for base_class in inspect.getmro(_lowerCAmelCase ): snake_case__ : int = base_class.__module__ snake_case__ : Any = base_class.__name__ if module.startswith("""tensorflow""" ) or module.startswith("""keras""" ) or name == "TFPreTrainedModel": return "tf" elif module.startswith("""torch""" ) or name == "PreTrainedModel": return "pt" elif module.startswith("""flax""" ) or module.startswith("""jax""" ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f"Could not infer framework from class {model_class}." )
35
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__=None ) -> Optional[int]: """simple docstring""" assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(snake_case__ ) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> str: """simple docstring""" _SCREAMING_SNAKE_CASE = np.asarray(weights[0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[2] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.key ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.self_attention.value ,torch.tensor(snake_case__ ).transpose(1 ,2 ).contiguous().view(-1 ,snake_case__ ) ,) set_param( torch_layer.output.dense ,torch.tensor(snake_case__ ).view(-1 ,snake_case__ ).contiguous().transpose(0 ,1 ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = weights[0][0][0] _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[0] ) _SCREAMING_SNAKE_CASE = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # lsh weights + output _SCREAMING_SNAKE_CASE = weights[0][1] if len(snake_case__ ) < 4: set_layer_weights_in_torch_lsh(snake_case__ ,torch_block.attention ,snake_case__ ) else: set_layer_weights_in_torch_local(snake_case__ ,torch_block.attention ,snake_case__ ) # intermediate weighs _SCREAMING_SNAKE_CASE = weights[2][0][1][2] # Chunked Feed Forward if len(snake_case__ ) == 4: _SCREAMING_SNAKE_CASE = intermediate_weights[2] # layernorm 2 _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # intermediate dense _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) # intermediate out _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][0] ) _SCREAMING_SNAKE_CASE = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = torch_model.reformer # word embeds _SCREAMING_SNAKE_CASE = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings ,torch.tensor(snake_case__ ) ,) if isinstance(weights[3] ,snake_case__ ): _SCREAMING_SNAKE_CASE = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _SCREAMING_SNAKE_CASE = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' _SCREAMING_SNAKE_CASE = nn.Parameter(torch.tensor(snake_case__ ) ) _SCREAMING_SNAKE_CASE = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( snake_case__ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _SCREAMING_SNAKE_CASE = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(snake_case__ ,snake_case__ ,snake_case__ ) # output layer norm _SCREAMING_SNAKE_CASE = np.asarray(weights[7][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm ,torch.tensor(snake_case__ ) ,torch.tensor(snake_case__ ) ,) # output embeddings _SCREAMING_SNAKE_CASE = np.asarray(weights[9][0] ) _SCREAMING_SNAKE_CASE = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder ,torch.tensor(snake_case__ ).transpose(0 ,1 ).contiguous() ,torch.tensor(snake_case__ ) ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ReformerConfig.from_json_file(snake_case__ ) print(F'Building PyTorch model from configuration: {config}' ) _SCREAMING_SNAKE_CASE = ReformerModelWithLMHead(snake_case__ ) with open(snake_case__ ,"""rb""" ) as f: _SCREAMING_SNAKE_CASE = pickle.load(snake_case__ )["""weights"""] set_model_weights_in_torch(snake_case__ ,snake_case__ ,config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() ,snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
306
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { "configuration_table_transformer": [ "TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TableTransformerConfig", "TableTransformerOnnxConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TableTransformerForObjectDetection", "TableTransformerModel", "TableTransformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
36
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __UpperCAmelCase (_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = TextToVideoSDPipeline __snake_case : Optional[int] = TEXT_TO_IMAGE_PARAMS __snake_case : Dict = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __snake_case : Optional[int] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def UpperCamelCase ( self: int ): '''simple docstring''' torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) _SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase_ , set_alpha_to_one=UpperCAmelCase_ , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) _SCREAMING_SNAKE_CASE = CLIPTextModel(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict=0 ): '''simple docstring''' if str(UpperCAmelCase_ ).startswith("""mps""" ): _SCREAMING_SNAKE_CASE = torch.manual_seed(UpperCAmelCase_ ) else: _SCREAMING_SNAKE_CASE = torch.Generator(device=UpperCAmelCase_ ).manual_seed(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE = self.get_dummy_components() _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = sd_pipe.to(UpperCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.get_dummy_inputs(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = """np""" _SCREAMING_SNAKE_CASE = sd_pipe(**UpperCAmelCase_ ).frames _SCREAMING_SNAKE_CASE = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _SCREAMING_SNAKE_CASE = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=UpperCAmelCase_ , expected_max_diff=1E-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def UpperCamelCase ( self: int ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' pass def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __UpperCAmelCase (unittest.TestCase ): def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=25 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) _SCREAMING_SNAKE_CASE = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) _SCREAMING_SNAKE_CASE = pipe.to("""cuda""" ) _SCREAMING_SNAKE_CASE = """Spiderman is surfing""" _SCREAMING_SNAKE_CASE = torch.Generator(device="""cpu""" ).manual_seed(0 ) _SCREAMING_SNAKE_CASE = pipe(UpperCAmelCase_ , generator=UpperCAmelCase_ , num_inference_steps=2 , output_type="""pt""" ).frames _SCREAMING_SNAKE_CASE = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
306
0
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase_: '''simple docstring''' def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=2 ,__UpperCAmelCase=8 ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=99 ,__UpperCAmelCase=16 ,__UpperCAmelCase=5 ,__UpperCAmelCase=2 ,__UpperCAmelCase=36 ,__UpperCAmelCase="gelu" ,__UpperCAmelCase=0.0 ,__UpperCAmelCase=0.0 ,__UpperCAmelCase=512 ,__UpperCAmelCase=16 ,__UpperCAmelCase=2 ,__UpperCAmelCase=0.0_2 ,__UpperCAmelCase=3 ,__UpperCAmelCase=4 ,__UpperCAmelCase=None ,) -> Dict: lowerCAmelCase__ : Dict = parent lowerCAmelCase__ : Optional[int] = batch_size lowerCAmelCase__ : Optional[int] = seq_length lowerCAmelCase__ : Any = is_training lowerCAmelCase__ : str = use_input_mask lowerCAmelCase__ : Any = use_token_type_ids lowerCAmelCase__ : Union[str, Any] = use_labels lowerCAmelCase__ : List[Any] = vocab_size lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : Tuple = num_hidden_layers lowerCAmelCase__ : Any = num_attention_heads lowerCAmelCase__ : Union[str, Any] = intermediate_size lowerCAmelCase__ : List[Any] = hidden_act lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : int = attention_probs_dropout_prob lowerCAmelCase__ : Dict = max_position_embeddings lowerCAmelCase__ : Optional[int] = type_vocab_size lowerCAmelCase__ : Optional[int] = type_sequence_label_size lowerCAmelCase__ : int = initializer_range lowerCAmelCase__ : Dict = num_labels lowerCAmelCase__ : List[str] = num_choices lowerCAmelCase__ : str = scope def UpperCAmelCase_ ( self ) -> int: lowerCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase__ : Union[str, Any] = None if self.use_input_mask: lowerCAmelCase__ : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : Tuple = None if self.use_token_type_ids: lowerCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) lowerCAmelCase__ : int = None lowerCAmelCase__ : str = None lowerCAmelCase__ : int = None if self.use_labels: lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowerCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) lowerCAmelCase__ : Any = ids_tensor([self.batch_size] ,self.num_choices ) lowerCAmelCase__ : Any = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self ) -> int: return MraConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=__UpperCAmelCase ,initializer_range=self.initializer_range ,) def UpperCAmelCase_ ( self ) -> Optional[int]: lowerCAmelCase__ : Dict = self.get_config() lowerCAmelCase__ : Union[str, Any] = 300 return config def UpperCAmelCase_ ( self ) -> Optional[int]: ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : int = self.prepare_config_and_inputs() lowerCAmelCase__ : int = True lowerCAmelCase__ : Any = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> Optional[int]: lowerCAmelCase__ : Optional[int] = MraModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : Optional[int] = model(__UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ) lowerCAmelCase__ : List[Any] = model(__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ) lowerCAmelCase__ : List[Any] = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,) -> int: lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Union[str, Any] = MraModel(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : Optional[Any] = model( __UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,encoder_hidden_states=__UpperCAmelCase ,encoder_attention_mask=__UpperCAmelCase ,) lowerCAmelCase__ : str = model( __UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,encoder_hidden_states=__UpperCAmelCase ,) lowerCAmelCase__ : Any = model(__UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> int: lowerCAmelCase__ : Tuple = MraForMaskedLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : Any = model(__UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> List[str]: lowerCAmelCase__ : List[Any] = MraForQuestionAnswering(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : int = model( __UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,start_positions=__UpperCAmelCase ,end_positions=__UpperCAmelCase ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> str: lowerCAmelCase__ : Optional[Any] = self.num_labels lowerCAmelCase__ : Union[str, Any] = MraForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : int = model(__UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> Optional[int]: lowerCAmelCase__ : List[Any] = self.num_labels lowerCAmelCase__ : Optional[Any] = MraForTokenClassification(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : Any = model(__UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> Optional[Any]: lowerCAmelCase__ : Any = self.num_choices lowerCAmelCase__ : Optional[Any] = MraForMultipleChoice(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() lowerCAmelCase__ : int = input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() lowerCAmelCase__ : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() lowerCAmelCase__ : List[Any] = input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() lowerCAmelCase__ : Tuple = model( __UpperCAmelCase ,attention_mask=__UpperCAmelCase ,token_type_ids=__UpperCAmelCase ,labels=__UpperCAmelCase ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def UpperCAmelCase_ ( self ) -> List[str]: lowerCAmelCase__ : Optional[int] = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Optional[Any] = config_and_inputs lowerCAmelCase__ : Tuple = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' __lowercase : Dict = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __lowercase : str = False __lowercase : Union[str, Any] = False __lowercase : Optional[Any] = False __lowercase : int = False __lowercase : int = () def UpperCAmelCase_ ( self ) -> Tuple: lowerCAmelCase__ : List[str] = MraModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self ,config_class=__UpperCAmelCase ,hidden_size=37 ) def UpperCAmelCase_ ( self ) -> Tuple: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self ) -> List[Any]: lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCAmelCase__ : str = type self.model_tester.create_and_check_model(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> str: lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> str: lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Any = MraModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) @unittest.skip(reason="""MRA does not output attentions""" ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: return @require_torch class lowerCAmelCase_( unittest.TestCase ): '''simple docstring''' @slow def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : str = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) lowerCAmelCase__ : List[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): lowerCAmelCase__ : int = model(__UpperCAmelCase )[0] lowerCAmelCase__ : Optional[int] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape ,__UpperCAmelCase ) lowerCAmelCase__ : str = torch.tensor( [[[-0.0_1_4_0, 0.0_8_3_0, -0.0_3_8_1], [0.1_5_4_6, 0.1_4_0_2, 0.0_2_2_0], [0.1_1_6_2, 0.0_8_5_1, 0.0_1_6_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,__UpperCAmelCase ,atol=1E-4 ) ) @slow def UpperCAmelCase_ ( self ) -> Tuple: lowerCAmelCase__ : Any = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) lowerCAmelCase__ : Tuple = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(__UpperCAmelCase )[0] lowerCAmelCase__ : List[str] = 5_0265 lowerCAmelCase__ : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape ,__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = torch.tensor( [[[9.2_5_9_5, -3.6_0_3_8, 1_1.8_8_1_9], [9.3_8_6_9, -3.2_6_9_3, 1_1.0_9_5_6], [1_1.8_5_2_4, -3.4_9_3_8, 1_3.1_2_1_0]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,__UpperCAmelCase ,atol=1E-4 ) ) @slow def UpperCAmelCase_ ( self ) -> Tuple: lowerCAmelCase__ : Union[str, Any] = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) lowerCAmelCase__ : Optional[int] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): lowerCAmelCase__ : Optional[int] = model(__UpperCAmelCase )[0] lowerCAmelCase__ : Optional[Any] = 5_0265 lowerCAmelCase__ : Tuple = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape ,__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = torch.tensor( [[[5.4_7_8_9, -2.3_5_6_4, 7.5_0_6_4], [7.9_0_6_7, -1.3_3_6_9, 9.9_6_6_8], [9.0_7_1_2, -1.8_1_0_6, 7.0_3_8_0]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] ,__UpperCAmelCase ,atol=1E-4 ) )
37
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class __UpperCAmelCase : def __init__( self: Union[str, Any] , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: int=13 , UpperCAmelCase_: Optional[int]=7 , UpperCAmelCase_: List[str]=False , UpperCAmelCase_: str=True , UpperCAmelCase_: Union[str, Any]=False , UpperCAmelCase_: Optional[Any]=True , UpperCAmelCase_: Optional[int]=33 , UpperCAmelCase_: Tuple=32 , UpperCAmelCase_: List[Any]=5 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: Any=37 , UpperCAmelCase_: Optional[Any]="gelu" , UpperCAmelCase_: Dict=0.1 , UpperCAmelCase_: List[Any]=0.1 , UpperCAmelCase_: Dict=512 , UpperCAmelCase_: int=16 , UpperCAmelCase_: Optional[Any]=2 , UpperCAmelCase_: Optional[Any]=0.02 , UpperCAmelCase_: Tuple=3 , UpperCAmelCase_: Union[str, Any]=4 , UpperCAmelCase_: str=None , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = parent _SCREAMING_SNAKE_CASE = batch_size _SCREAMING_SNAKE_CASE = seq_length _SCREAMING_SNAKE_CASE = is_training _SCREAMING_SNAKE_CASE = use_input_mask _SCREAMING_SNAKE_CASE = use_token_type_ids _SCREAMING_SNAKE_CASE = use_labels _SCREAMING_SNAKE_CASE = vocab_size _SCREAMING_SNAKE_CASE = hidden_size _SCREAMING_SNAKE_CASE = num_hidden_layers _SCREAMING_SNAKE_CASE = num_attention_heads _SCREAMING_SNAKE_CASE = intermediate_size _SCREAMING_SNAKE_CASE = hidden_act _SCREAMING_SNAKE_CASE = hidden_dropout_prob _SCREAMING_SNAKE_CASE = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE = max_position_embeddings _SCREAMING_SNAKE_CASE = type_vocab_size _SCREAMING_SNAKE_CASE = type_sequence_label_size _SCREAMING_SNAKE_CASE = initializer_range _SCREAMING_SNAKE_CASE = num_labels _SCREAMING_SNAKE_CASE = num_choices _SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _SCREAMING_SNAKE_CASE = None if self.use_input_mask: _SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = None if self.use_labels: _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) _SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: str , UpperCAmelCase_: List[str] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: List[str] , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: int , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: List[Any] , UpperCAmelCase_: int , UpperCAmelCase_: List[str] , UpperCAmelCase_: str , UpperCAmelCase_: Union[str, Any] , UpperCAmelCase_: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.num_labels _SCREAMING_SNAKE_CASE = EsmForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) = config_and_inputs _SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): __snake_case : List[Any] = False __snake_case : Dict = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __snake_case : List[Any] = () __snake_case : Dict = ( { "feature-extraction": EsmModel, "fill-mask": EsmForMaskedLM, "text-classification": EsmForSequenceClassification, "token-classification": EsmForTokenClassification, "zero-shot": EsmForSequenceClassification, } if is_torch_available() else {} ) __snake_case : int = True def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = EsmModelTester(self ) _SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def UpperCamelCase ( self: int ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def UpperCamelCase ( self: Any ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def UpperCamelCase ( self: int ): '''simple docstring''' for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def UpperCamelCase ( self: str ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) _SCREAMING_SNAKE_CASE = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) _SCREAMING_SNAKE_CASE = create_position_ids_from_input_ids(UpperCAmelCase_ , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) def UpperCamelCase ( self: List[str] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs()[0] _SCREAMING_SNAKE_CASE = EsmEmbeddings(config=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.empty(2 , 4 , 30 ) _SCREAMING_SNAKE_CASE = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] _SCREAMING_SNAKE_CASE = torch.as_tensor([expected_single_positions, expected_single_positions] ) _SCREAMING_SNAKE_CASE = embeddings.create_position_ids_from_inputs_embeds(UpperCAmelCase_ ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(UpperCAmelCase_ , UpperCAmelCase_ ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' pass @unittest.skip("""Esm does not support embedding resizing""" ) def UpperCamelCase ( self: Dict ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def UpperCamelCase ( self: Any ): '''simple docstring''' pass @require_torch class __UpperCAmelCase (_UpperCAmelCase ): @slow def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 2, 3, 4, 5]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] _SCREAMING_SNAKE_CASE = 33 _SCREAMING_SNAKE_CASE = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.tensor( [[[8.92_15, -10.58_98, -6.46_71], [-6.39_67, -13.91_14, -1.12_12], [-7.78_12, -13.95_16, -3.74_06]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def UpperCamelCase ( self: Dict ): '''simple docstring''' with torch.no_grad(): _SCREAMING_SNAKE_CASE = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() _SCREAMING_SNAKE_CASE = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ )[0] # compare the actual values for a slice. _SCREAMING_SNAKE_CASE = torch.tensor( [[[0.14_44, 0.54_13, 0.32_48], [0.30_34, 0.00_53, 0.31_08], [0.32_28, -0.24_99, 0.34_15]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) )
306
0
def SCREAMING_SNAKE_CASE_ ( __magic_name__ : str , __magic_name__ : int ) -> list[str]: """simple docstring""" return [sentence[i : i + ngram_size] for i in range(len(__magic_name__ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
38
import random def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = num - 1 _SCREAMING_SNAKE_CASE = 0 while s % 2 == 0: _SCREAMING_SNAKE_CASE = s // 2 t += 1 for _ in range(5 ): _SCREAMING_SNAKE_CASE = random.randrange(2 ,num - 1 ) _SCREAMING_SNAKE_CASE = pow(snake_case__ ,snake_case__ ,snake_case__ ) if v != 1: _SCREAMING_SNAKE_CASE = 0 while v != (num - 1): if i == t - 1: return False else: _SCREAMING_SNAKE_CASE = i + 1 _SCREAMING_SNAKE_CASE = (v**2) % num return True def __lowerCamelCase ( snake_case__ ) -> bool: """simple docstring""" if num < 2: return False _SCREAMING_SNAKE_CASE = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 1_01, 1_03, 1_07, 1_09, 1_13, 1_27, 1_31, 1_37, 1_39, 1_49, 1_51, 1_57, 1_63, 1_67, 1_73, 1_79, 1_81, 1_91, 1_93, 1_97, 1_99, 2_11, 2_23, 2_27, 2_29, 2_33, 2_39, 2_41, 2_51, 2_57, 2_63, 2_69, 2_71, 2_77, 2_81, 2_83, 2_93, 3_07, 3_11, 3_13, 3_17, 3_31, 3_37, 3_47, 3_49, 3_53, 3_59, 3_67, 3_73, 3_79, 3_83, 3_89, 3_97, 4_01, 4_09, 4_19, 4_21, 4_31, 4_33, 4_39, 4_43, 4_49, 4_57, 4_61, 4_63, 4_67, 4_79, 4_87, 4_91, 4_99, 5_03, 5_09, 5_21, 5_23, 5_41, 5_47, 5_57, 5_63, 5_69, 5_71, 5_77, 5_87, 5_93, 5_99, 6_01, 6_07, 6_13, 6_17, 6_19, 6_31, 6_41, 6_43, 6_47, 6_53, 6_59, 6_61, 6_73, 6_77, 6_83, 6_91, 7_01, 7_09, 7_19, 7_27, 7_33, 7_39, 7_43, 7_51, 7_57, 7_61, 7_69, 7_73, 7_87, 7_97, 8_09, 8_11, 8_21, 8_23, 8_27, 8_29, 8_39, 8_53, 8_57, 8_59, 8_63, 8_77, 8_81, 8_83, 8_87, 9_07, 9_11, 9_19, 9_29, 9_37, 9_41, 9_47, 9_53, 9_67, 9_71, 9_77, 9_83, 9_91, 9_97, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(snake_case__ ) def __lowerCamelCase ( snake_case__ = 10_24 ) -> int: """simple docstring""" while True: _SCREAMING_SNAKE_CASE = random.randrange(2 ** (keysize - 1) ,2 ** (keysize) ) if is_prime_low_num(snake_case__ ): return num if __name__ == "__main__": UpperCamelCase = generate_large_prime() print(('''Prime number:''', num)) print(('''is_prime_low_num:''', is_prime_low_num(num)))
306
0
import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _a = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. _a = direct_transformers_import(PATH_TO_TRANSFORMERS) _a = transformers.models.auto.configuration_auto.CONFIG_MAPPING _a = { # used to compute the property `self.chunk_length` '''EncodecConfig''': ['''overlap'''], # used as `self.bert_model = BertModel(config, ...)` '''DPRConfig''': True, # not used in modeling files, but it's an important information '''FSMTConfig''': ['''langs'''], # used internally in the configuration class file '''GPTNeoConfig''': ['''attention_types'''], # used internally in the configuration class file '''EsmConfig''': ['''is_folding_model'''], # used during training (despite we don't have training script for these models yet) '''Mask2FormerConfig''': ['''ignore_value'''], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) '''OneFormerConfig''': ['''ignore_value''', '''norm'''], # used during preprocessing and collation, see `collating_graphormer.py` '''GraphormerConfig''': ['''spatial_pos_max'''], # used internally in the configuration class file '''T5Config''': ['''feed_forward_proj'''], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally '''MT5Config''': ['''feed_forward_proj''', '''tokenizer_class'''], '''UMT5Config''': ['''feed_forward_proj''', '''tokenizer_class'''], # used internally in the configuration class file '''LongT5Config''': ['''feed_forward_proj'''], # used internally in the configuration class file '''SwitchTransformersConfig''': ['''feed_forward_proj'''], # having default values other than `1e-5` - we can't fix them without breaking '''BioGptConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''GLPNConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''SegformerConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''CvtConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''PerceiverConfig''': ['''layer_norm_eps'''], # used internally to calculate the feature size '''InformerConfig''': ['''num_static_real_features''', '''num_time_features'''], # used internally to calculate the feature size '''TimeSeriesTransformerConfig''': ['''num_static_real_features''', '''num_time_features'''], # used internally to calculate the feature size '''AutoformerConfig''': ['''num_static_real_features''', '''num_time_features'''], # used internally to calculate `mlp_dim` '''SamVisionConfig''': ['''mlp_ratio'''], # For (head) training, but so far not implemented '''ClapAudioConfig''': ['''num_classes'''], # Not used, but providing useful information to users '''SpeechT5HifiGanConfig''': ['''sampling_rate'''], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { '''CLIPSegConfig''': True, '''DeformableDetrConfig''': True, '''DetaConfig''': True, '''DinatConfig''': True, '''DonutSwinConfig''': True, '''EfficientFormerConfig''': True, '''FSMTConfig''': True, '''JukeboxConfig''': True, '''LayoutLMv2Config''': True, '''MaskFormerSwinConfig''': True, '''MT5Config''': True, '''NatConfig''': True, '''OneFormerConfig''': True, '''PerceiverConfig''': True, '''RagConfig''': True, '''SpeechT5Config''': True, '''SwinConfig''': True, '''Swin2SRConfig''': True, '''Swinv2Config''': True, '''SwitchTransformersConfig''': True, '''TableTransformerConfig''': True, '''TapasConfig''': True, '''TransfoXLConfig''': True, '''UniSpeechConfig''': True, '''UniSpeechSatConfig''': True, '''WavLMConfig''': True, '''WhisperConfig''': True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) '''JukeboxPriorConfig''': True, # TODO: @Younes (for `is_decoder`) '''Pix2StructTextConfig''': True, } ) def __A ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase )-> Dict: """simple docstring""" _UpperCAmelCase = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( F"""config.{attribute}""" in modeling_source or F"""getattr(config, \"{attribute}\"""" in modeling_source or F"""getattr(self.config, \"{attribute}\"""" in modeling_source ): _UpperCAmelCase = True # Deal with multi-line cases elif ( re.search( RF"""getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*\"{attribute}\"""" , __lowerCAmelCase , ) is not None ): _UpperCAmelCase = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: _UpperCAmelCase = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files _UpperCAmelCase = [ 'bos_index', 'eos_index', 'pad_index', 'unk_index', 'mask_index', 'image_size', 'use_cache', 'out_features', 'out_indices', ] _UpperCAmelCase = ['encoder_no_repeat_ngram_size'] # Special cases to be allowed _UpperCAmelCase = True if not attribute_used: _UpperCAmelCase = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: _UpperCAmelCase = True elif attribute in ["tie_word_embeddings"] and default_value is False: _UpperCAmelCase = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: _UpperCAmelCase = True elif attribute.endswith('_token_id' ): _UpperCAmelCase = True # configuration class specific cases if not case_allowed: _UpperCAmelCase = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__ , [] ) _UpperCAmelCase = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def __A ( __lowerCAmelCase )-> Optional[int]: """simple docstring""" _UpperCAmelCase = dict(inspect.signature(config_class.__init__ ).parameters ) _UpperCAmelCase = [x for x in list(signature.keys() ) if x not in ['self', 'kwargs']] _UpperCAmelCase = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass _UpperCAmelCase = {} if len(config_class.attribute_map ) > 0: _UpperCAmelCase = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files _UpperCAmelCase = inspect.getsourcefile(__lowerCAmelCase ) _UpperCAmelCase = os.path.dirname(__lowerCAmelCase ) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. _UpperCAmelCase = [os.path.join(__lowerCAmelCase , __lowerCAmelCase ) for fn in os.listdir(__lowerCAmelCase ) if fn.startswith('modeling_' )] # Get the source code strings _UpperCAmelCase = [] for path in modeling_paths: if os.path.isfile(__lowerCAmelCase ): with open(__lowerCAmelCase ) as fp: modeling_sources.append(fp.read() ) _UpperCAmelCase = [] for config_param, default_value in zip(__lowerCAmelCase , __lowerCAmelCase ): # `attributes` here is all the variant names for `config_param` _UpperCAmelCase = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param] ) if not check_attribute_being_used(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): unused_attributes.append(attributes[0] ) return sorted(__lowerCAmelCase ) def __A ( )-> int: """simple docstring""" _UpperCAmelCase = {} for _config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) _UpperCAmelCase = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class ) , lambda __lowerCAmelCase : inspect.isclass(__lowerCAmelCase ) and issubclass(__lowerCAmelCase , __lowerCAmelCase ) and inspect.getmodule(__lowerCAmelCase ) == inspect.getmodule(_config_class ) , ) ] for config_class in config_classes_in_module: _UpperCAmelCase = check_config_attributes_being_used(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: _UpperCAmelCase = unused_attributes if len(__lowerCAmelCase ) > 0: _UpperCAmelCase = 'The following configuration classes contain unused attributes in the corresponding modeling files:\n' for name, attributes in configs_with_unused_attributes.items(): error += F"""{name}: {attributes}\n""" raise ValueError(__lowerCAmelCase ) if __name__ == "__main__": check_config_attributes()
39
def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _SCREAMING_SNAKE_CASE = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
306
0
"""simple docstring""" def lowercase ( A_ )-> int: '''simple docstring''' a : int = hex_num.strip() if not hex_num: raise ValueError("No value was passed to the function" ) a : str = hex_num[0] == "-" if is_negative: a : str = hex_num[1:] try: a : int = int(A_ , 16 ) except ValueError: raise ValueError("Invalid value was passed to the function" ) a : str = "" while int_num > 0: a : List[Any] = str(int_num % 2 ) + bin_str int_num >>= 1 return int(("-" + bin_str) if is_negative else bin_str ) if __name__ == "__main__": import doctest doctest.testmod()
40
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
306
0