code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : List[str] = { 'configuration_informer': [ 'INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'InformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : List[Any] = [ 'INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'InformerForPrediction', 'InformerModel', 'InformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys A__ : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
103
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=3 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=10 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[1, 1, 2, 1] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="relu" , _lowerCamelCase=3 , _lowerCamelCase=None , ) -> List[str]: A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> Optional[Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : Any = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase_ ( self ) -> int: super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: pass def UpperCAmelCase_ ( self ) -> int: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase , _lowerCamelCase ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) def UpperCAmelCase_ ( self ) -> str: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> int: return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""tf""" ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 )
344
0
from queue import PriorityQueue from typing import Any import numpy as np def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, ): """simple docstring""" for nxt, d in graph[v]: if nxt in visited_forward: continue SCREAMING_SNAKE_CASE =cst_fwd.get(a_, np.inf ) SCREAMING_SNAKE_CASE =cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) SCREAMING_SNAKE_CASE =new_cost_f SCREAMING_SNAKE_CASE =v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: SCREAMING_SNAKE_CASE =cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE =-1 SCREAMING_SNAKE_CASE =set() SCREAMING_SNAKE_CASE =set() SCREAMING_SNAKE_CASE ={source: 0} SCREAMING_SNAKE_CASE ={destination: 0} SCREAMING_SNAKE_CASE ={source: None} SCREAMING_SNAKE_CASE ={destination: None} SCREAMING_SNAKE_CASE =PriorityQueue() SCREAMING_SNAKE_CASE =PriorityQueue() SCREAMING_SNAKE_CASE =np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): SCREAMING_SNAKE_CASE =queue_forward.get() visited_forward.add(a_ ) SCREAMING_SNAKE_CASE =queue_backward.get() visited_backward.add(a_ ) SCREAMING_SNAKE_CASE =pass_and_relaxation( a_, a_, a_, a_, a_, a_, a_, a_, a_, ) SCREAMING_SNAKE_CASE =pass_and_relaxation( a_, a_, a_, a_, a_, a_, a_, a_, a_, ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: SCREAMING_SNAKE_CASE =shortest_distance return shortest_path_distance _lowerCamelCase ={ 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } _lowerCamelCase ={ 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
334
'''simple docstring''' def UpperCAmelCase ( a_ = 1_0_0 ) -> int: """simple docstring""" A_ : Dict = n * (n + 1) * (2 * n + 1) / 6 A_ : Optional[int] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'{solution() = }')
344
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCAmelCase = logging.get_logger(__name__) class A ( __A , __A ): UpperCamelCase_ : Optional[Any] ='''maskformer-swin''' UpperCamelCase_ : List[str] ={ '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__(self , lowerCAmelCase=2_2_4 , lowerCAmelCase=4 , lowerCAmelCase=3 , lowerCAmelCase=9_6 , lowerCAmelCase=[2, 2, 6, 2] , lowerCAmelCase=[3, 6, 1_2, 2_4] , lowerCAmelCase=7 , lowerCAmelCase=4.0 , lowerCAmelCase=True , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=0.1 , lowerCAmelCase="gelu" , lowerCAmelCase=False , lowerCAmelCase=0.02 , lowerCAmelCase=1E-5 , lowerCAmelCase=None , lowerCAmelCase=None , **lowerCAmelCase , ): super().__init__(**_lowerCamelCase ) __lowercase= image_size __lowercase= patch_size __lowercase= num_channels __lowercase= embed_dim __lowercase= depths __lowercase= len(_lowerCamelCase ) __lowercase= num_heads __lowercase= window_size __lowercase= mlp_ratio __lowercase= qkv_bias __lowercase= hidden_dropout_prob __lowercase= attention_probs_dropout_prob __lowercase= drop_path_rate __lowercase= hidden_act __lowercase= use_absolute_embeddings __lowercase= layer_norm_eps __lowercase= initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __lowercase= int(embed_dim * 2 ** (len(_lowerCamelCase ) - 1) ) __lowercase= ["""stem"""] + [f'stage{idx}' for idx in range(1 , len(_lowerCamelCase ) + 1 )] __lowercase= get_aligned_output_features_output_indices( out_features=_lowerCamelCase , out_indices=_lowerCamelCase , stage_names=self.stage_names )
295
'''simple docstring''' from typing import TYPE_CHECKING from ..utils import _LazyModule UpperCamelCase__ : int = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys UpperCamelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class A : __magic_name__ = 42 # setable values __magic_name__ = 42 __magic_name__ = 42 __magic_name__ = None @classmethod def __lowerCAmelCase ( cls , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" return cls(common=_lowerCamelCase , init_noise_sigma=_lowerCamelCase , timesteps=_lowerCamelCase ) @dataclass class A ( __A ): __magic_name__ = 42 class A ( __A , __A ): __magic_name__ = [e.name for e in FlaxKarrasDiffusionSchedulers] __magic_name__ = 42 @property def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" return True @register_to_config def __init__( self , SCREAMING_SNAKE_CASE = 1000 , SCREAMING_SNAKE_CASE = 0.0_001 , SCREAMING_SNAKE_CASE = 0.02 , SCREAMING_SNAKE_CASE = "linear" , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = "fixed_small" , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = "epsilon" , SCREAMING_SNAKE_CASE = jnp.floataa , ) -> Tuple: """simple docstring""" A : Tuple = dtype def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE = None ) -> DDPMSchedulerState: """simple docstring""" if common is None: A : str = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution A : Any = jnp.array(1.0 , dtype=self.dtype ) A : Optional[Any] = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_lowerCamelCase , init_noise_sigma=_lowerCamelCase , timesteps=_lowerCamelCase , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> jnp.ndarray: """simple docstring""" return sample def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = () ) -> DDPMSchedulerState: """simple docstring""" A : Tuple = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 A : Tuple = (jnp.arange(0 , _lowerCamelCase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_lowerCamelCase , timesteps=_lowerCamelCase , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None ) -> Any: """simple docstring""" A : Union[str, Any] = state.common.alphas_cumprod[t] A : str = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample A : List[str] = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: A : Optional[Any] = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": A : str = jnp.clip(_lowerCamelCase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": A : Tuple = jnp.log(jnp.clip(_lowerCamelCase , a_min=1e-20 ) ) elif variance_type == "fixed_large": A : Union[str, Any] = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log A : int = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": A : Any = variance A : Union[str, Any] = state.common.betas[t] A : Optional[Any] = (predicted_variance + 1) / 2 A : int = frac * max_log + (1 - frac) * min_log return variance def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: """simple docstring""" A : Optional[int] = timestep if key is None: A : List[str] = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: A : List[Any] = jnp.split(_lowerCamelCase , sample.shape[1] , axis=1 ) else: A : Union[str, Any] = None # 1. compute alphas, betas A : Optional[Any] = state.common.alphas_cumprod[t] A : Tuple = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) A : str = 1 - alpha_prod_t A : Dict = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": A : Any = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": A : str = model_output elif self.config.prediction_type == "v_prediction": A : Any = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ' ''' for the FlaxDDPMScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: A : Optional[Any] = jnp.clip(_lowerCamelCase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf A : Union[str, Any] = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t A : Tuple = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf A : Tuple = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): A : Dict = jax.random.split(_lowerCamelCase , num=1 ) A : int = jax.random.normal(_lowerCamelCase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_lowerCamelCase , _lowerCamelCase , predicted_variance=_lowerCamelCase ) ** 0.5) * noise A : str = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) A : Dict = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_lowerCamelCase , state=_lowerCamelCase ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> jnp.ndarray: """simple docstring""" return add_noise_common(state.common , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> jnp.ndarray: """simple docstring""" return get_velocity_common(state.common , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __len__( self ) -> str: """simple docstring""" return self.config.num_train_timesteps
3
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCAmelCase ( a_ ) -> Dict[str, torch.Tensor]: """simple docstring""" A_ : List[str] = [] A_ : Dict = [] A_ : List[Any] = [] for rt in rc.restypes: A_ : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) A_ : Union[str, Any] = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) A_ : Tuple = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : Optional[int] = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : List[Any] = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) A_ : Optional[int] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein A_ : Dict = restype_atomaa_to_atomaa[protein_aatype] A_ : Optional[Any] = restype_atomaa_mask[protein_aatype] A_ : Any = residx_atomaa_mask A_ : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back A_ : Tuple = restype_atomaa_to_atomaa[protein_aatype] A_ : Tuple = residx_atomaa_to_atomaa.long() # create the corresponding mask A_ : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): A_ : Optional[Any] = rc.restype_atoa[restype_letter] A_ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: A_ : Any = rc.atom_order[atom_name] A_ : Optional[int] = 1 A_ : Optional[int] = restype_atomaa_mask[protein_aatype] A_ : Dict = residx_atomaa_mask return protein def UpperCAmelCase ( a_ ) -> Dict[str, np.ndarray]: """simple docstring""" A_ : Union[str, Any] = tree_map(lambda a_ : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) A_ : Optional[int] = tensor_tree_map(lambda a_ : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
344
0
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class __lowerCAmelCase ( unittest.TestCase ): def __init__( self :str , __magic_name__ :Any , __magic_name__ :str=7 , __magic_name__ :List[Any]=3 , __magic_name__ :str=30 , __magic_name__ :Optional[int]=400 , __magic_name__ :Union[str, Any]=True , __magic_name__ :str=None , __magic_name__ :List[str]=True , __magic_name__ :int=[0.5, 0.5, 0.5] , __magic_name__ :List[Any]=[0.5, 0.5, 0.5] , __magic_name__ :Tuple=True , __magic_name__ :Optional[int]=1 / 255 , __magic_name__ :str=True , ): '''simple docstring''' a = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 1333} a = parent a = batch_size a = num_channels a = min_resolution a = max_resolution a = do_resize a = size a = do_normalize a = image_mean a = image_std a = do_rescale a = rescale_factor a = do_pad def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def lowerCamelCase__ ( self :Any , __magic_name__ :Dict , __magic_name__ :Optional[Any]=False ): '''simple docstring''' if not batched: a = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image ): a = image.size else: a = image.shape[1], image.shape[2] if w < h: a = int(self.size["""shortest_edge"""] * h / w ) a = self.size["""shortest_edge"""] elif w > h: a = self.size["""shortest_edge"""] a = int(self.size["""shortest_edge"""] * w / h ) else: a = self.size["""shortest_edge"""] a = self.size["""shortest_edge"""] else: a = [] for image in image_inputs: a = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) a = max(_lowerCamelCase , key=lambda __magic_name__ : item[0] )[0] a = max(_lowerCamelCase , key=lambda __magic_name__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class __lowerCAmelCase ( __A , unittest.TestCase ): UpperCamelCase__ = DetaImageProcessor if is_vision_available() else None def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = DetaImageProcessingTester(self ) @property def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_rescale""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_pad""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) def lowerCamelCase__ ( self :Dict ): '''simple docstring''' a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 1333} ) self.assertEqual(image_processor.do_pad , _lowerCamelCase ) def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' pass def lowerCamelCase__ ( self :int ): '''simple docstring''' a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images a = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input a = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values a = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched a = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) a = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors a = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input a = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values a = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched a = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values a = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors a = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input a = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values a = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched a = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values a = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' a = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: a = json.loads(f.read() ) a = {"""image_id""": 3_9769, """annotations""": target} # encode them a = DetaImageProcessor() a = image_processing(images=_lowerCamelCase , annotations=_lowerCamelCase , return_tensors="""pt""" ) # verify pixel values a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["""pixel_values"""].shape , _lowerCamelCase ) a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , _lowerCamelCase , atol=1E-4 ) ) # verify area a = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , _lowerCamelCase ) ) # verify boxes a = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , _lowerCamelCase ) a = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , _lowerCamelCase , atol=1E-3 ) ) # verify image_id a = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , _lowerCamelCase ) ) # verify is_crowd a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , _lowerCamelCase ) ) # verify class_labels a = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , _lowerCamelCase ) ) # verify orig_size a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , _lowerCamelCase ) ) # verify size a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , _lowerCamelCase ) ) @slow def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' a = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: a = json.loads(f.read() ) a = {"""file_name""": """000000039769.png""", """image_id""": 3_9769, """segments_info""": target} a = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them a = DetaImageProcessor(format="""coco_panoptic""" ) a = image_processing(images=_lowerCamelCase , annotations=_lowerCamelCase , masks_path=_lowerCamelCase , return_tensors="""pt""" ) # verify pixel values a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["""pixel_values"""].shape , _lowerCamelCase ) a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , _lowerCamelCase , atol=1E-4 ) ) # verify area a = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , _lowerCamelCase ) ) # verify boxes a = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , _lowerCamelCase ) a = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , _lowerCamelCase , atol=1E-3 ) ) # verify image_id a = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , _lowerCamelCase ) ) # verify is_crowd a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , _lowerCamelCase ) ) # verify class_labels a = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , _lowerCamelCase ) ) # verify masks a = 82_2873 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , _lowerCamelCase ) # verify orig_size a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , _lowerCamelCase ) ) # verify size a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , _lowerCamelCase ) )
228
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ) -> str: A_ : Optional[int] = parent A_ : Dict = batch_size A_ : List[Any] = image_size A_ : Optional[int] = patch_size A_ : List[str] = num_channels A_ : List[Any] = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_size A_ : str = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Union[str, Any] = intermediate_size A_ : Any = hidden_act A_ : Optional[Any] = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Dict = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : str = scope A_ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) A_ : Tuple = (image_size // patch_size) ** 2 A_ : Union[str, Any] = num_patches + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> int: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[str] = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : int = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A_ : Dict = 1 A_ : Optional[int] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Tuple = self.type_sequence_label_size A_ : Tuple = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : Dict = 1 A_ : Any = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : List[str] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = DeiTModelTester(self ) A_ : str = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: pass def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ , A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[str] = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Union[str, Any] = [*signature.parameters.keys()] A_ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: A_ : int = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase_ ( self ) -> Optional[Any]: if not self.model_tester.is_training: return A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[str] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : List[str] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> int: A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A_ : Any = False A_ : Union[str, Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue A_ : List[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Union[str, Any] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Tuple: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): A_ : Dict = problem_type["""title"""] A_ : List[Any] = problem_type["""num_labels"""] A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: A_ : Tuple = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) A_ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: A_ : List[str] = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Optional[Any]: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) A_ : Optional[int] = self.default_image_processor A_ : str = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Any = model(**_lowerCamelCase ) # verify the logits A_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : List[Any] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : int = self.default_image_processor A_ : List[str] = prepare_img() A_ : List[Any] = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) A_ : Union[str, Any] = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A_ : List[Any] = model(_lowerCamelCase )
344
0
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = [] _snake_case = [] _snake_case = [] for rt in rc.restypes: _snake_case = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) _snake_case = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) _snake_case = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) _snake_case = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) _snake_case = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) _snake_case = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein _snake_case = restype_atomaa_to_atomaa[protein_aatype] _snake_case = restype_atomaa_mask[protein_aatype] _snake_case = residx_atomaa_mask _snake_case = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back _snake_case = restype_atomaa_to_atomaa[protein_aatype] _snake_case = residx_atomaa_to_atomaa.long() # create the corresponding mask _snake_case = torch.zeros([21, 37] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): _snake_case = rc.restype_atoa[restype_letter] _snake_case = rc.residue_atoms[restype_name] for atom_name in atom_names: _snake_case = rc.atom_order[atom_name] _snake_case = 1 _snake_case = restype_atomaa_mask[protein_aatype] _snake_case = residx_atomaa_mask return protein def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = tree_map(lambda _SCREAMING_SNAKE_CASE : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) _snake_case = tensor_tree_map(lambda _SCREAMING_SNAKE_CASE : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
341
'''simple docstring''' import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 32 , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = True , _lowerCamelCase = [0.4814_5466, 0.457_8275, 0.4082_1073] , _lowerCamelCase = [0.2686_2954, 0.2613_0258, 0.2757_7711] , _lowerCamelCase = True , _lowerCamelCase=7 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=3 , ) -> Union[str, Any]: A_ : Optional[int] = parent A_ : Union[str, Any] = do_resize A_ : Optional[Any] = size if size is not None else {"""shortest_edge""": 288} A_ : Tuple = size_divisor A_ : List[Any] = do_rescale A_ : Dict = rescale_factor A_ : List[Any] = do_normalize A_ : Dict = do_center_crop A_ : Optional[Any] = image_mean A_ : List[str] = image_std A_ : str = do_pad A_ : Any = batch_size A_ : List[str] = num_channels A_ : List[str] = min_resolution A_ : Union[str, Any] = max_resolution def UpperCAmelCase_ ( self ) -> Any: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase=False ) -> Optional[int]: if not batched: A_ : Union[str, Any] = self.size["""shortest_edge"""] A_ : Dict = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image ): A_ , A_ : Optional[Any] = image.size else: A_ , A_ : int = image.shape[1], image.shape[2] A_ : Optional[int] = size / min(_lowerCamelCase , _lowerCamelCase ) if h < w: A_ , A_ : Optional[Any] = size, scale * w else: A_ , A_ : Dict = scale * h, size A_ : Union[str, Any] = int((1333 / 800) * size ) if max(_lowerCamelCase , _lowerCamelCase ) > max_size: A_ : str = max_size / max(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = newh * scale A_ : Dict = neww * scale A_ , A_ : str = int(newh + 0.5 ), int(neww + 0.5 ) A_ , A_ : Dict = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: A_ : Tuple = [] for image in image_inputs: A_ , A_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) A_ : List[Any] = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[0] )[0] A_ : Tuple = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : int = BridgeTowerImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size_divisor""" ) ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image processor A_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : int = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: # Initialize image processor A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image processor A_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
344
0
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json', # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class lowerCAmelCase_ ( __A ): __lowerCamelCase : int = "blenderbot-small" __lowerCamelCase : List[str] = ["past_key_values"] __lowerCamelCase : Tuple = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self , _lowerCAmelCase=50265 , _lowerCAmelCase=512 , _lowerCAmelCase=8 , _lowerCAmelCase=2048 , _lowerCAmelCase=16 , _lowerCAmelCase=8 , _lowerCAmelCase=2048 , _lowerCAmelCase=16 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase="gelu" , _lowerCAmelCase=512 , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.02 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=0 , _lowerCAmelCase=1 , _lowerCAmelCase=2 , _lowerCAmelCase=2 , **_lowerCAmelCase , ) -> List[Any]: _lowerCAmelCase = vocab_size _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = d_model _lowerCAmelCase = encoder_ffn_dim _lowerCAmelCase = encoder_layers _lowerCAmelCase = encoder_attention_heads _lowerCAmelCase = decoder_ffn_dim _lowerCAmelCase = decoder_layers _lowerCAmelCase = decoder_attention_heads _lowerCAmelCase = dropout _lowerCAmelCase = attention_dropout _lowerCAmelCase = activation_dropout _lowerCAmelCase = activation_function _lowerCAmelCase = init_std _lowerCAmelCase = encoder_layerdrop _lowerCAmelCase = decoder_layerdrop _lowerCAmelCase = use_cache _lowerCAmelCase = encoder_layers _lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , is_encoder_decoder=_lowerCamelCase , decoder_start_token_id=_lowerCamelCase , forced_eos_token_id=_lowerCamelCase , **_lowerCamelCase , ) class lowerCAmelCase_ ( __A ): @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: _lowerCAmelCase = {0: """batch"""} _lowerCAmelCase = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: _lowerCAmelCase = {0: """batch""", 1: """decoder_sequence"""} _lowerCAmelCase = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(_lowerCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. _lowerCAmelCase = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: _lowerCAmelCase = self.num_layers for i in range(_lowerCamelCase ): _lowerCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} _lowerCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} else: _lowerCAmelCase = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase = super().outputs else: _lowerCAmelCase = super(_lowerCamelCase , self ).outputs if self.use_past: _lowerCAmelCase = self.num_layers for i in range(_lowerCamelCase ): _lowerCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} _lowerCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = -1 , _lowerCAmelCase = -1 , _lowerCAmelCase = False , _lowerCAmelCase = None , ) -> Mapping[str, Any]: _lowerCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # Generate decoder inputs _lowerCAmelCase = seq_length if not self.use_past else 1 _lowerCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase = {f'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} _lowerCAmelCase = dict(**_lowerCamelCase , **_lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch _lowerCAmelCase = common_inputs["""input_ids"""].shape _lowerCAmelCase = common_inputs["""decoder_input_ids"""].shape[1] _lowerCAmelCase = self.num_attention_heads _lowerCAmelCase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) _lowerCAmelCase = decoder_seq_length + 3 _lowerCAmelCase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) _lowerCAmelCase = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(_lowerCamelCase , _lowerCamelCase )] , dim=1 ) _lowerCAmelCase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered _lowerCAmelCase = self.num_layers _lowerCAmelCase = min(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase = max(_lowerCamelCase , _lowerCamelCase ) - min_num_layers _lowerCAmelCase = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(_lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase ), ) ) # TODO: test this. _lowerCAmelCase = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(_lowerCamelCase , _lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase )) ) return common_inputs def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = -1 , _lowerCAmelCase = -1 , _lowerCAmelCase = False , _lowerCAmelCase = None , ) -> Mapping[str, Any]: _lowerCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch _lowerCAmelCase = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values _lowerCAmelCase = seqlen + 2 _lowerCAmelCase = self.num_layers _lowerCAmelCase = self.num_attention_heads _lowerCAmelCase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) _lowerCAmelCase = common_inputs["""attention_mask"""].dtype _lowerCAmelCase = torch.cat( [common_inputs["attention_mask"], torch.ones(_lowerCamelCase , _lowerCamelCase , dtype=_lowerCamelCase )] , dim=1 ) _lowerCAmelCase = [ (torch.zeros(_lowerCamelCase ), torch.zeros(_lowerCamelCase )) for _ in range(_lowerCamelCase ) ] return common_inputs def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = -1 , _lowerCAmelCase = -1 , _lowerCAmelCase = False , _lowerCAmelCase = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _lowerCAmelCase = compute_effective_axis_dimension( _lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _lowerCAmelCase = tokenizer.num_special_tokens_to_add(_lowerCamelCase ) _lowerCAmelCase = compute_effective_axis_dimension( _lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence _lowerCAmelCase = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size _lowerCAmelCase = dict(tokenizer(_lowerCamelCase , return_tensors=_lowerCamelCase ) ) return common_inputs def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = -1 , _lowerCAmelCase = -1 , _lowerCAmelCase = False , _lowerCAmelCase = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) elif self.task == "causal-lm": _lowerCAmelCase = self._generate_dummy_inputs_for_causal_lm( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) else: _lowerCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) return common_inputs def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> int: if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase = super()._flatten_past_key_values_(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) else: _lowerCAmelCase = super(_lowerCamelCase , self )._flatten_past_key_values_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase )
158
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(a_ ): for j in range(a_ ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) , end="""\t""" ) else: print("""INF""" , end="""\t""" ) print() def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" A_ : List[str] = [[float("""inf""" ) for _ in range(a_ )] for _ in range(a_ )] for i in range(a_ ): for j in range(a_ ): A_ : List[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(a_ ): # looping through rows of graph array for i in range(a_ ): # looping through columns of graph array for j in range(a_ ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): A_ : List[str] = dist[i][k] + dist[k][j] _print_dist(a_ , a_ ) return dist, v if __name__ == "__main__": UpperCamelCase__ : Tuple = int(input('Enter number of vertices: ')) UpperCamelCase__ : int = int(input('Enter number of edges: ')) UpperCamelCase__ : Dict = [[float('inf') for i in range(v)] for j in range(v)] for i in range(v): UpperCamelCase__ : Union[str, Any] = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print('\nEdge ', i + 1) UpperCamelCase__ : Union[str, Any] = int(input('Enter source:')) UpperCamelCase__ : int = int(input('Enter destination:')) UpperCamelCase__ : Optional[Any] = float(input('Enter weight:')) UpperCamelCase__ : Any = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
344
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __A : List[str] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'GroupViTTextConfig', 'GroupViTVisionConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ 'GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'GroupViTModel', 'GroupViTPreTrainedModel', 'GroupViTTextModel', 'GroupViTVisionModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ 'TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFGroupViTModel', 'TFGroupViTPreTrainedModel', 'TFGroupViTTextModel', 'TFGroupViTVisionModel', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
138
'''simple docstring''' import datasets from .evaluate import evaluate UpperCamelCase__ : int = '\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n' UpperCamelCase__ : Any = '\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n' UpperCamelCase__ : Optional[Any] = '\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the SQuAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> squad_metric = datasets.load_metric("squad")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: A_ : Optional[Any] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} A_ : List[Any] = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] A_ : int = evaluate(dataset=_lowerCamelCase , predictions=_lowerCamelCase ) return score
344
0
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class A_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case = None , snake_case = None ): super().__init__() lowercase = pad_token_id lowercase = max_length lowercase = vocab lowercase = merges lowercase = BytePairTokenizer(_lowerCamelCase , _lowerCamelCase , sequence_length=_lowerCamelCase ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case , *snake_case , **snake_case ): lowercase = [""" """.join(_lowerCamelCase ) for m in tokenizer.bpe_ranks.keys()] lowercase = tokenizer.get_vocab() return cls(_lowerCamelCase , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case , *snake_case , **snake_case ): lowercase = GPTaTokenizer.from_pretrained(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) return cls.from_tokenizer(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case ): return cls(**_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = self.tf_tokenizer(_lowerCamelCase ) lowercase = tf.ones_like(_lowerCamelCase ) if self.pad_token_id is not None: # pad the tokens up to max length lowercase = max_length if max_length is not None else self.max_length if max_length is not None: lowercase = pad_model_inputs( _lowerCamelCase , max_seq_length=_lowerCamelCase , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
195
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ : Any = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Optional[Any] = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] UpperCamelCase__ : List[str] = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] UpperCamelCase__ : str = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): UpperCamelCase__ : List[str] = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCamelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class lowerCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self ) -> int: UpperCamelCase : str = tempfile.mkdtemp() UpperCamelCase : Union[str, Any] = BlipImageProcessor() UpperCamelCase : Dict = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) UpperCamelCase : Dict = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) UpperCamelCase : Any = InstructBlipProcessor(_lowerCamelCase, _lowerCamelCase, _lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def snake_case_ ( self, **SCREAMING_SNAKE_CASE_ ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname, **_lowerCamelCase ).tokenizer def snake_case_ ( self, **SCREAMING_SNAKE_CASE_ ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname, **_lowerCamelCase ).image_processor def snake_case_ ( self, **SCREAMING_SNAKE_CASE_ ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname, **_lowerCamelCase ).qformer_tokenizer def snake_case_ ( self ) -> int: shutil.rmtree(self.tmpdirname ) def snake_case_ ( self ) -> List[str]: UpperCamelCase : Optional[Any] = [np.random.randint(255, size=(3, 30, 400), dtype=np.uinta )] UpperCamelCase : str = [Image.fromarray(np.moveaxis(_lowerCamelCase, 0, -1 ) ) for x in image_inputs] return image_inputs def snake_case_ ( self ) -> Any: UpperCamelCase : Tuple = InstructBlipProcessor( tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor(), qformer_tokenizer=self.get_qformer_tokenizer(), ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase : Optional[Any] = self.get_tokenizer(bos_token='(BOS)', eos_token='(EOS)' ) UpperCamelCase : Union[str, Any] = self.get_image_processor(do_normalize=_lowerCamelCase, padding_value=1.0 ) UpperCamelCase : str = InstructBlipProcessor.from_pretrained( self.tmpdirname, bos_token='(BOS)', eos_token='(EOS)', do_normalize=_lowerCamelCase, padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, _lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, _lowerCamelCase ) self.assertIsInstance(processor.qformer_tokenizer, _lowerCamelCase ) def snake_case_ ( self ) -> Any: UpperCamelCase : Optional[Any] = self.get_image_processor() UpperCamelCase : Any = self.get_tokenizer() UpperCamelCase : List[str] = self.get_qformer_tokenizer() UpperCamelCase : Any = InstructBlipProcessor( tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase, qformer_tokenizer=_lowerCamelCase ) UpperCamelCase : Tuple = self.prepare_image_inputs() UpperCamelCase : Any = image_processor(_lowerCamelCase, return_tensors='np' ) UpperCamelCase : List[Any] = processor(images=_lowerCamelCase, return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1e-2 ) def snake_case_ ( self ) -> Dict: UpperCamelCase : Any = self.get_image_processor() UpperCamelCase : Dict = self.get_tokenizer() UpperCamelCase : List[str] = self.get_qformer_tokenizer() UpperCamelCase : List[str] = InstructBlipProcessor( tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase, qformer_tokenizer=_lowerCamelCase ) UpperCamelCase : int = """lower newer""" UpperCamelCase : Optional[int] = processor(text=_lowerCamelCase ) UpperCamelCase : Dict = tokenizer(_lowerCamelCase, return_token_type_ids=_lowerCamelCase ) UpperCamelCase : Any = qformer_tokenizer(_lowerCamelCase, return_token_type_ids=_lowerCamelCase ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key], encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key], encoded_processor['qformer_' + key] ) def snake_case_ ( self ) -> Any: UpperCamelCase : List[Any] = self.get_image_processor() UpperCamelCase : str = self.get_tokenizer() UpperCamelCase : Optional[Any] = self.get_qformer_tokenizer() UpperCamelCase : Union[str, Any] = InstructBlipProcessor( tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase, qformer_tokenizer=_lowerCamelCase ) UpperCamelCase : Union[str, Any] = """lower newer""" UpperCamelCase : Optional[Any] = self.prepare_image_inputs() UpperCamelCase : Optional[int] = processor(text=_lowerCamelCase, images=_lowerCamelCase ) self.assertListEqual( list(inputs.keys() ), ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'], ) # test if it raises when no input is passed with pytest.raises(_lowerCamelCase ): processor() def snake_case_ ( self ) -> Tuple: UpperCamelCase : List[str] = self.get_image_processor() UpperCamelCase : Any = self.get_tokenizer() UpperCamelCase : Optional[Any] = self.get_qformer_tokenizer() UpperCamelCase : Dict = InstructBlipProcessor( tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase, qformer_tokenizer=_lowerCamelCase ) UpperCamelCase : int = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase : Optional[Any] = processor.batch_decode(_lowerCamelCase ) UpperCamelCase : Any = tokenizer.batch_decode(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase, _lowerCamelCase ) def snake_case_ ( self ) -> str: UpperCamelCase : List[Any] = self.get_image_processor() UpperCamelCase : List[str] = self.get_tokenizer() UpperCamelCase : Optional[int] = self.get_qformer_tokenizer() UpperCamelCase : Union[str, Any] = InstructBlipProcessor( tokenizer=_lowerCamelCase, image_processor=_lowerCamelCase, qformer_tokenizer=_lowerCamelCase ) UpperCamelCase : Union[str, Any] = """lower newer""" UpperCamelCase : Optional[int] = self.prepare_image_inputs() UpperCamelCase : int = processor(text=_lowerCamelCase, images=_lowerCamelCase ) self.assertListEqual( list(inputs.keys() ), ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'], )
119
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''vision-encoder-decoder''' lowerCamelCase = True def __init__( self , **_lowerCamelCase ) -> str: super().__init__(**_lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F"A configuraton of type {self.model_type} cannot be instantiated because " F"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) A_ : Optional[int] = kwargs.pop("""encoder""" ) A_ : List[str] = encoder_config.pop("""model_type""" ) A_ : str = kwargs.pop("""decoder""" ) A_ : Optional[Any] = decoder_config.pop("""model_type""" ) A_ : List[str] = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : str = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : Any = True @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A_ : int = True A_ : List[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = copy.deepcopy(self.__dict__ ) A_ : List[str] = self.encoder.to_dict() A_ : Union[str, Any] = self.decoder.to_dict() A_ : str = self.__class__.model_type return output class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> float: return 1e-4 @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: A_ : Optional[Any] = OrderedDict() A_ : Any = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : Optional[int] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = -1 , _lowerCamelCase = -1 , _lowerCamelCase = False , _lowerCamelCase = None , ) -> Mapping[str, Any]: import torch A_ : Optional[int] = OrderedDict() A_ : List[Any] = super().generate_dummy_inputs( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) A_ , A_ : str = dummy_input["""input_ids"""].shape A_ : Optional[int] = (batch, encoder_sequence, self._config.encoder_hidden_size) A_ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A_ : List[str] = dummy_input.pop("""attention_mask""" ) A_ : Optional[int] = torch.zeros(_lowerCamelCase ) return common_inputs class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> None: pass def UpperCAmelCase_ ( self , _lowerCamelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = "default" ) -> OnnxConfig: A_ : List[Any] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_lowerCamelCase , _lowerCamelCase )
344
0
"""simple docstring""" from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase : def __init__( self : Optional[int] , _lowerCamelCase : List[str] , _lowerCamelCase : Union[str, Any]=3 , _lowerCamelCase : str=32 , _lowerCamelCase : Union[str, Any]=3 , _lowerCamelCase : str=10 , _lowerCamelCase : str=[10, 20, 30, 40] , _lowerCamelCase : Optional[int]=[1, 1, 2, 1] , _lowerCamelCase : Dict=True , _lowerCamelCase : int=True , _lowerCamelCase : Optional[Any]="relu" , _lowerCamelCase : Any=3 , _lowerCamelCase : Optional[Any]=None , ): """simple docstring""" A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def a_ ( self : str ): """simple docstring""" A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def a_ ( self : List[Any] ): """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def a_ ( self : List[Any] , _lowerCamelCase : Any , _lowerCamelCase : Any , _lowerCamelCase : Union[str, Any] ): """simple docstring""" A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a_ ( self : List[Any] , _lowerCamelCase : List[Any] , _lowerCamelCase : Tuple , _lowerCamelCase : int ): """simple docstring""" A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a_ ( self : Dict ): """simple docstring""" A_ : Any = self.prepare_config_and_inputs() A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class lowercase ( __A , __A , unittest.TestCase): __lowerCAmelCase : Any = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () __lowerCAmelCase : str = ( {"""feature-extraction""": TFRegNetModel, """image-classification""": TFRegNetForImageClassification} if is_tf_available() else {} ) __lowerCAmelCase : List[Any] = False __lowerCAmelCase : Union[str, Any] = False __lowerCAmelCase : Dict = False __lowerCAmelCase : Any = False __lowerCAmelCase : Tuple = False def a_ ( self : str ): """simple docstring""" A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def a_ ( self : List[str] ): """simple docstring""" return @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def a_ ( self : Union[str, Any] ): """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , reason='''TF does not support backprop for grouped convolutions on CPU.''' , ) @slow def a_ ( self : Optional[Any] ): """simple docstring""" super().test_keras_fit() @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def a_ ( self : Union[str, Any] ): """simple docstring""" pass def a_ ( self : str ): """simple docstring""" A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def a_ ( self : int ): """simple docstring""" A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def a_ ( self : Dict ): """simple docstring""" def check_hidden_states_output(_lowerCamelCase : Dict , _lowerCamelCase : int , _lowerCamelCase : Optional[Any] ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def a_ ( self : str ): """simple docstring""" A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase : Tuple , _lowerCamelCase : int , _lowerCamelCase : List[Any] , _lowerCamelCase : Optional[Any]={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase : Optional[Any] , _lowerCamelCase : str ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( '''Tuple and dict output are not equal. Difference:''' F""" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}""" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {'''output_hidden_states''': True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {'''output_hidden_states''': True} ) def a_ ( self : Optional[int] ): """simple docstring""" A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def a_ ( self : Union[str, Any] ): """simple docstring""" for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def lowercase_ ( ): """simple docstring""" A_ : Optional[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowercase ( unittest.TestCase): @cached_property def a_ ( self : Any ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def a_ ( self : Union[str, Any] ): """simple docstring""" A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors='''tf''' ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1E-4 )
167
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCamelCase__ : Any = '\\n Text data.\n Second line of data.' UpperCamelCase__ : List[Any] = 'file' @pytest.fixture(scope="""session""" ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : int = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") A_ : int = bytes(a_ , """utf-8""" ) with zstd.open(a_ , """wb""" ) as f: f.write(a_ ) return path @pytest.fixture def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , a_ ) , """w""" ) as f: f.write(a_ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> Optional[int]: """simple docstring""" A_ : List[str] = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} A_ : Any = input_paths[compression_format] A_ : Tuple = tmp_path / """cache""" A_ : Tuple = DownloadConfig(cache_dir=a_ , extract_compressed_file=a_ ) A_ : Dict = cached_path(a_ , download_config=a_ ) with open(a_ ) as f: A_ : Optional[Any] = f.read() with open(a_ ) as f: A_ : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Union[str, Any] = """custom_cache""" A_ : List[str] = """custom_extracted_dir""" A_ : Optional[Any] = tmp_path / """custom_extracted_path""" if default_extracted: A_ : Any = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , a_ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(a_ ) ) A_ : Union[str, Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) A_ : List[Any] = xz_file A_ : Optional[int] = ( DownloadConfig(extract_compressed_file=a_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=a_ ) ) A_ : Union[str, Any] = cached_path(a_ , download_config=a_ ) assert Path(a_ ).parent.parts[-2:] == expected def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : str = str(Path(a_ ).resolve() ) assert cached_path(a_ ) == text_file # relative path A_ : List[str] = str(Path(a_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(a_ ) == text_file def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : Optional[Any] = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(a_ ): cached_path(a_ ) # relative path A_ : Tuple = """./__missing_file__.txt""" with pytest.raises(a_ ): cached_path(a_ ) def UpperCAmelCase ( a_ ) -> Tuple: """simple docstring""" A_ : Any = get_from_cache(F"tmp://{tmpfs_file}" ) with open(a_ ) as f: A_ : List[str] = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( ) -> List[str]: """simple docstring""" with pytest.raises(a_ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Union[str, Any]: """simple docstring""" A_ : List[str] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): http_get("""https://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): ftp_get("""ftp://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): fsspec_get("""s3://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): fsspec_head("""s3://huggingface.co""" )
344
0
from ...configuration_utils import PretrainedConfig from ...utils import logging A__ : Tuple = logging.get_logger(__name__) A__ : Any = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class __snake_case ( __A ): _a = '''open-llama''' def __init__( self : List[Any] , A_ : str=1_0_0_0_0_0 , A_ : List[str]=4_0_9_6 , A_ : Any=1_1_0_0_8 , A_ : List[str]=3_2 , A_ : List[str]=3_2 , A_ : List[str]="silu" , A_ : Union[str, Any]=2_0_4_8 , A_ : Dict=0.02 , A_ : Union[str, Any]=1e-6 , A_ : Union[str, Any]=True , A_ : str=0 , A_ : List[Any]=1 , A_ : List[str]=2 , A_ : str=False , A_ : Dict=True , A_ : Optional[int]=0.1 , A_ : Tuple=0.1 , A_ : Union[str, Any]=True , A_ : str=True , A_ : List[Any]=None , **A_ : List[str] , ): lowerCAmelCase_ : List[str] = vocab_size lowerCAmelCase_ : Dict = max_position_embeddings lowerCAmelCase_ : Dict = hidden_size lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : int = num_hidden_layers lowerCAmelCase_ : Dict = num_attention_heads lowerCAmelCase_ : Any = hidden_act lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Dict = rms_norm_eps lowerCAmelCase_ : Optional[Any] = use_cache lowerCAmelCase_ : int = kwargs.pop( '''use_memorry_efficient_attention''' , _lowerCamelCase) lowerCAmelCase_ : Dict = hidden_dropout_prob lowerCAmelCase_ : Optional[Any] = attention_dropout_prob lowerCAmelCase_ : Optional[int] = use_stable_embedding lowerCAmelCase_ : List[Any] = shared_input_output_embedding lowerCAmelCase_ : int = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , tie_word_embeddings=_lowerCamelCase , **_lowerCamelCase , ) def UpperCAmelCase__ ( self : Optional[int]): if self.rope_scaling is None: return if not isinstance(self.rope_scaling , _lowerCamelCase) or len(self.rope_scaling) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' F"""got {self.rope_scaling}""") lowerCAmelCase_ : List[Any] = self.rope_scaling.get('''type''' , _lowerCamelCase) lowerCAmelCase_ : Union[str, Any] = self.rope_scaling.get('''factor''' , _lowerCamelCase) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""") if rope_scaling_factor is None or not isinstance(_lowerCamelCase , _lowerCamelCase) or rope_scaling_factor <= 1.0: raise ValueError(F"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""")
103
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) UpperCamelCase__ : int = {'processing_layoutxlm': ['LayoutXLMProcessor']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Tuple = ['LayoutXLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : List[Any] = ['LayoutXLMTokenizerFast'] if TYPE_CHECKING: from .processing_layoutxlm import LayoutXLMProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm import LayoutXLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm_fast import LayoutXLMTokenizerFast else: import sys UpperCamelCase__ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor _lowerCamelCase =logging.get_logger(__name__) class a_ ( __A ): """simple docstring""" def __init__( self : Dict ,*snake_case : str ,**snake_case : Optional[Any] ): warnings.warn( 'The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use PoolFormerImageProcessor instead.' ,_lowerCamelCase ,) super().__init__(*_lowerCamelCase ,**_lowerCamelCase )
334
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ : Any = logging.get_logger(__name__) UpperCamelCase__ : Optional[int] = { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/config.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/config.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json' ), 'distilbert-base-uncased-finetuned-sst-2-english': ( 'https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json' ), } class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''distilbert''' lowerCamelCase = { '''hidden_size''': '''dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', } def __init__( self , _lowerCamelCase=3_0522 , _lowerCamelCase=512 , _lowerCamelCase=False , _lowerCamelCase=6 , _lowerCamelCase=12 , _lowerCamelCase=768 , _lowerCamelCase=4 * 768 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase="gelu" , _lowerCamelCase=0.02 , _lowerCamelCase=0.1 , _lowerCamelCase=0.2 , _lowerCamelCase=0 , **_lowerCamelCase , ) -> Optional[Any]: A_ : Tuple = vocab_size A_ : List[Any] = max_position_embeddings A_ : int = sinusoidal_pos_embds A_ : int = n_layers A_ : str = n_heads A_ : Optional[int] = dim A_ : int = hidden_dim A_ : Tuple = dropout A_ : List[Any] = attention_dropout A_ : int = activation A_ : Dict = initializer_range A_ : List[Any] = qa_dropout A_ : int = seq_classif_dropout super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": A_ : Union[str, Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A_ : int = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
344
0
import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers lowerCAmelCase = 'python tqdm regex requests packaging filelock numpy tokenizers'.split() if sys.version_info < (3, 7): pkgs_to_check_at_runtime.append('''dataclasses''') if sys.version_info < (3, 8): pkgs_to_check_at_runtime.append('''importlib_metadata''') for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(F'can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py') def _lowerCamelCase( lowercase__ , lowercase__=None ) -> Dict: '''simple docstring''' require_version(deps[pkg] , a_ )
295
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ : int = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) UpperCamelCase__ : Any = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', f'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', f'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qpos_proj.weight', f'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kpos_proj.weight', f'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.weight', f'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', f'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', f'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kpos_proj.weight', f'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.weight', f'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', f'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', f'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', f'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_qpos_proj.bias', f'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_kpos_proj.bias', f'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.bias', f'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', f'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', f'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_kpos_proj.bias', f'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.bias', f'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', f'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ('transformer.decoder.ref_point_head.layers.0.weight', 'decoder.ref_point_head.layers.0.weight'), ('transformer.decoder.ref_point_head.layers.0.bias', 'decoder.ref_point_head.layers.0.bias'), ('transformer.decoder.ref_point_head.layers.1.weight', 'decoder.ref_point_head.layers.1.weight'), ('transformer.decoder.ref_point_head.layers.1.bias', 'decoder.ref_point_head.layers.1.bias'), ('transformer.decoder.query_scale.layers.0.weight', 'decoder.query_scale.layers.0.weight'), ('transformer.decoder.query_scale.layers.0.bias', 'decoder.query_scale.layers.0.bias'), ('transformer.decoder.query_scale.layers.1.weight', 'decoder.query_scale.layers.1.weight'), ('transformer.decoder.query_scale.layers.1.bias', 'decoder.query_scale.layers.1.bias'), ('transformer.decoder.layers.0.ca_qpos_proj.weight', 'decoder.layers.0.ca_qpos_proj.weight'), ('transformer.decoder.layers.0.ca_qpos_proj.bias', 'decoder.layers.0.ca_qpos_proj.bias'), ] ) def UpperCAmelCase ( a_ , a_ , a_ ) -> Optional[Any]: """simple docstring""" A_ : int = state_dict.pop(a_ ) A_ : Tuple = val def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : Union[str, Any] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: A_ : Optional[int] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) A_ : str = value else: A_ : int = value return new_state_dict def UpperCAmelCase ( a_ , a_=False ) -> Optional[int]: """simple docstring""" A_ : List[Any] = """""" if is_panoptic: A_ : Any = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) A_ : Optional[int] = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) A_ : str = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict A_ : Optional[Any] = in_proj_weight[:2_5_6, :] A_ : Tuple = in_proj_bias[:2_5_6] A_ : Dict = in_proj_weight[2_5_6:5_1_2, :] A_ : int = in_proj_bias[2_5_6:5_1_2] A_ : int = in_proj_weight[-2_5_6:, :] A_ : Optional[int] = in_proj_bias[-2_5_6:] def UpperCAmelCase ( ) -> Dict: """simple docstring""" A_ : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A_ : List[Any] = Image.open(requests.get(a_ , stream=a_ ).raw ) return im @torch.no_grad() def UpperCAmelCase ( a_ , a_ ) -> Dict: """simple docstring""" A_ : int = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: A_ : str = """resnet101""" if "dc5" in model_name: A_ : List[Any] = True A_ : str = """panoptic""" in model_name if is_panoptic: A_ : Dict = 2_5_0 else: A_ : Union[str, Any] = 9_1 A_ : str = """huggingface/label-files""" A_ : Union[str, Any] = """coco-detection-id2label.json""" A_ : Optional[Any] = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) A_ : str = {int(a_ ): v for k, v in idalabel.items()} A_ : Optional[int] = idalabel A_ : Tuple = {v: k for k, v in idalabel.items()} # load image processor A_ : List[Any] = """coco_panoptic""" if is_panoptic else """coco_detection""" A_ : Any = ConditionalDetrImageProcessor(format=a_ ) # prepare image A_ : Tuple = prepare_img() A_ : Any = image_processor(images=a_ , return_tensors="""pt""" ) A_ : Optional[int] = encoding["""pixel_values"""] logger.info(F"Converting model {model_name}..." ) # load original model from torch hub A_ : int = torch.hub.load("""DeppMeng/ConditionalDETR""" , a_ , pretrained=a_ ).eval() A_ : List[Any] = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: A_ : Union[str, Any] = """conditional_detr.""" + src rename_key(a_ , a_ , a_ ) A_ : Any = rename_backbone_keys(a_ ) # query, key and value matrices need special treatment read_in_q_k_v(a_ , is_panoptic=a_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them A_ : List[str] = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): A_ : Dict = state_dict.pop(a_ ) A_ : List[Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: A_ : str = state_dict.pop(a_ ) A_ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: A_ : Optional[int] = state_dict.pop(a_ ) A_ : str = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): A_ : Tuple = state_dict.pop(a_ ) A_ : Dict = val # finally, create HuggingFace model and load state dict A_ : Union[str, Any] = ConditionalDetrForSegmentation(a_ ) if is_panoptic else ConditionalDetrForObjectDetection(a_ ) model.load_state_dict(a_ ) model.eval() model.push_to_hub(repo_id=a_ , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion A_ : str = conditional_detr(a_ ) A_ : str = model(a_ ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(a_ ).mkdir(exist_ok=a_ ) model.save_pretrained(a_ ) image_processor.save_pretrained(a_ ) if __name__ == "__main__": UpperCamelCase__ : int = argparse.ArgumentParser() parser.add_argument( '--model_name', default='conditional_detr_resnet50', type=str, help='Name of the CONDITIONAL_DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) UpperCamelCase__ : Optional[Any] = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
344
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class A ( unittest.TestCase ): @slow def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A : Dict = XLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) A : Tuple = torch.tensor([[0, 581, 10269, 83, 99942, 136, 60742, 23, 70, 80583, 18276, 2]] ) # The dog is cute and lives in the garden house A : Union[str, Any] = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim A : List[str] = torch.tensor( [[-0.0_101, 0.1_218, -0.0_803, 0.0_801, 0.1_327, 0.0_776, -0.1_215, 0.2_383, 0.3_338, 0.3_106, 0.0_300, 0.0_252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): A : Optional[Any] = model(_lowerCamelCase )["""last_hidden_state"""].detach() self.assertEqual(output.shape , _lowerCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowerCamelCase , atol=1e-3 ) ) @slow def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" A : List[Any] = XLMRobertaModel.from_pretrained('''xlm-roberta-large''' ) A : List[Any] = torch.tensor([[0, 581, 10269, 83, 99942, 136, 60742, 23, 70, 80583, 18276, 2]] ) # The dog is cute and lives in the garden house A : List[Any] = torch.Size((1, 12, 1024) ) # batch_size, sequence_length, embedding_vector_dim A : Optional[int] = torch.tensor( [[-0.0_699, -0.0_318, 0.0_705, -0.1_241, 0.0_999, -0.0_520, 0.1_004, -0.1_838, -0.4_704, 0.1_437, 0.0_821, 0.0_126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): A : int = model(_lowerCamelCase )["""last_hidden_state"""].detach() self.assertEqual(output.shape , _lowerCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowerCamelCase , atol=1e-3 ) )
3
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = (UnCLIPScheduler,) def UpperCAmelCase_ ( self , **_lowerCamelCase ) -> List[Any]: A_ : Union[str, Any] = { """num_train_timesteps""": 1000, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**_lowerCamelCase ) return config def UpperCAmelCase_ ( self ) -> List[Any]: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample in [True, False]: self.check_over_configs(clip_sample=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[int]: for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=_lowerCamelCase , prev_timestep=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: A_ : Optional[int] = self.scheduler_classes[0] A_ : Any = self.get_scheduler_config(variance_type="""fixed_small_log""" ) A_ : List[Any] = scheduler_class(**_lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.054_9625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.999_4987 ) ) < 1e-5 def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : List[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config(variance_type="""learned_range""" ) A_ : Dict = scheduler_class(**_lowerCamelCase ) A_ : Dict = 0.5 assert scheduler._get_variance(1 , predicted_variance=_lowerCamelCase ) - -10.171_2790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=_lowerCamelCase ) - -5.799_8052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=_lowerCamelCase ) - -0.001_0011 < 1e-5 def UpperCAmelCase_ ( self ) -> Any: A_ : Optional[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config() A_ : Optional[Any] = scheduler_class(**_lowerCamelCase ) A_ : int = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : str = self.dummy_sample_deter A_ : Optional[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Any = model(_lowerCamelCase , _lowerCamelCase ) # 2. predict previous mean of sample x_t-1 A_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : List[Any] = pred_prev_sample A_ : Any = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Optional[Any] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 252.268_2495 ) < 1e-2 assert abs(result_mean.item() - 0.328_4743 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Dict: A_ : Union[str, Any] = self.scheduler_classes[0] A_ : Dict = self.get_scheduler_config() A_ : Tuple = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(25 ) A_ : List[str] = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : List[Any] = self.dummy_sample_deter A_ : List[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Optional[Any] = model(_lowerCamelCase , _lowerCamelCase ) if i + 1 == timesteps.shape[0]: A_ : List[str] = None else: A_ : Dict = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 A_ : str = scheduler.step( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , prev_timestep=_lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : Optional[Any] = pred_prev_sample A_ : Dict = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : List[str] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 258.204_4983 ) < 1e-2 assert abs(result_mean.item() - 0.336_2038 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> int: pass
344
0
import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Any: a = FunnelConfig.from_json_file(a_ ) print(f'Building PyTorch model from configuration: {config}' ) a = FunnelBaseModel(a_ ) if base_model else FunnelModel(a_ ) # Load weights from tf checkpoint load_tf_weights_in_funnel(a_ , a_ , a_ ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , a_ ) if __name__ == "__main__": __UpperCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) __UpperCamelCase : Tuple = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
228
'''simple docstring''' import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=7 , _lowerCamelCase=3 , _lowerCamelCase=18 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=False , ) -> Optional[int]: A_ : Union[str, Any] = size if size is not None else {"""height""": 20, """width""": 20} A_ : Tuple = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} A_ : Optional[Any] = parent A_ : Optional[int] = batch_size A_ : Union[str, Any] = num_channels A_ : str = image_size A_ : Tuple = min_resolution A_ : Dict = max_resolution A_ : str = do_resize A_ : Tuple = size A_ : int = do_center_crop A_ : Dict = crop_size A_ : Tuple = do_normalize A_ : List[str] = image_mean A_ : Optional[Any] = image_std A_ : Any = do_reduce_labels def UpperCAmelCase_ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def UpperCAmelCase ( ) -> List[str]: """simple docstring""" A_ : Any = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(dataset[0]["""file"""] ) A_ : Dict = Image.open(dataset[1]["""file"""] ) return image, map def UpperCAmelCase ( ) -> Optional[int]: """simple docstring""" A_ : Tuple = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(ds[0]["""file"""] ) A_ : List[Any] = Image.open(ds[1]["""file"""] ) A_ : Any = Image.open(ds[2]["""file"""] ) A_ : str = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BeitImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : List[Any] = BeitImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) A_ : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=_lowerCamelCase ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> Dict: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> str: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : Union[str, Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> Optional[int]: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) A_ : Optional[int] = [] for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) A_ , A_ : List[Any] = prepare_semantic_single_inputs() A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) A_ , A_ : str = prepare_semantic_batch_inputs() A_ : Any = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image_processing A_ : Any = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 A_ , A_ : Tuple = prepare_semantic_single_inputs() A_ : str = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) A_ : str = True A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
344
0
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class _lowerCAmelCase ( datasets.BuilderConfig ): '''simple docstring''' lowerCAmelCase_ = None class _lowerCAmelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' lowerCAmelCase_ = PandasConfig def lowercase (self ) -> Tuple: return datasets.DatasetInfo(features=self.config.features ) def lowercase (self , UpperCAmelCase ) -> List[Any]: if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) _snake_case = dl_manager.download_and_extract(self.config.data_files ) if isinstance(_lowerCamelCase , (str, list, tuple) ): _snake_case = data_files if isinstance(_lowerCamelCase , _lowerCamelCase ): _snake_case = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _snake_case = [dl_manager.iter_files(_lowerCamelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] _snake_case = [] for split_name, files in data_files.items(): if isinstance(_lowerCamelCase , _lowerCamelCase ): _snake_case = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _snake_case = [dl_manager.iter_files(_lowerCamelCase ) for file in files] splits.append(datasets.SplitGenerator(name=_lowerCamelCase , gen_kwargs={"""files""": files} ) ) return splits def lowercase (self , UpperCAmelCase ) -> pa.Table: if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example _snake_case = table_cast(_lowerCamelCase , self.config.features.arrow_schema ) return pa_table def lowercase (self , UpperCAmelCase ) -> Optional[int]: for i, file in enumerate(itertools.chain.from_iterable(_lowerCamelCase ) ): with open(_lowerCamelCase , """rb""" ) as f: _snake_case = pa.Table.from_pandas(pd.read_pickle(_lowerCamelCase ) ) yield i, self._cast_table(_lowerCamelCase )
341
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class _lowerCAmelCase ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None ) -> str: super().__init__() A_ : Optional[Any] = pad_token_id A_ : List[Any] = max_length A_ : str = vocab A_ : Union[str, Any] = merges A_ : List[Any] = BytePairTokenizer(_lowerCamelCase , _lowerCamelCase , sequence_length=_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> int: A_ : Tuple = [""" """.join(_lowerCamelCase ) for m in tokenizer.bpe_ranks.keys()] A_ : Dict = tokenizer.get_vocab() return cls(_lowerCamelCase , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> str: A_ : Tuple = GPTaTokenizer.from_pretrained(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) return cls.from_tokenizer(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase ) -> List[Any]: return cls(**_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ) -> Any: A_ : List[Any] = self.tf_tokenizer(_lowerCamelCase ) A_ : Any = tf.ones_like(_lowerCamelCase ) if self.pad_token_id is not None: # pad the tokens up to max length A_ : List[Any] = max_length if max_length is not None else self.max_length if max_length is not None: A_ , A_ : Tuple = pad_model_inputs( _lowerCamelCase , max_seq_length=_lowerCamelCase , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
344
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'RWKV/rwkv-4-169m-pile': 'https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json', 'RWKV/rwkv-4-430m-pile': 'https://huggingface.co/RWKV/rwkv-4-430m-pile/resolve/main/config.json', 'RWKV/rwkv-4-1b5-pile': 'https://huggingface.co/RWKV/rwkv-4-1b5-pile/resolve/main/config.json', 'RWKV/rwkv-4-3b-pile': 'https://huggingface.co/RWKV/rwkv-4-3b-pile/resolve/main/config.json', 'RWKV/rwkv-4-7b-pile': 'https://huggingface.co/RWKV/rwkv-4-7b-pile/resolve/main/config.json', 'RWKV/rwkv-4-14b-pile': 'https://huggingface.co/RWKV/rwkv-4-14b-pile/resolve/main/config.json', 'RWKV/rwkv-raven-1b5': 'https://huggingface.co/RWKV/rwkv-raven-1b5/resolve/main/config.json', 'RWKV/rwkv-raven-3b': 'https://huggingface.co/RWKV/rwkv-raven-3b/resolve/main/config.json', 'RWKV/rwkv-raven-7b': 'https://huggingface.co/RWKV/rwkv-raven-7b/resolve/main/config.json', 'RWKV/rwkv-raven-14b': 'https://huggingface.co/RWKV/rwkv-raven-14b/resolve/main/config.json', } class lowerCAmelCase_ ( __A ): __lowerCamelCase : int = "rwkv" __lowerCamelCase : str = {"max_position_embeddings": "context_length"} def __init__( self , _lowerCAmelCase=50277 , _lowerCAmelCase=1024 , _lowerCAmelCase=4096 , _lowerCAmelCase=32 , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=1E-5 , _lowerCAmelCase=0 , _lowerCAmelCase=0 , _lowerCAmelCase=6 , _lowerCAmelCase=False , _lowerCAmelCase=True , **_lowerCAmelCase , ) -> Optional[Any]: _lowerCAmelCase = vocab_size _lowerCAmelCase = context_length _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = attention_hidden_size if attention_hidden_size is not None else hidden_size _lowerCAmelCase = intermediate_size if intermediate_size is not None else 4 * hidden_size _lowerCAmelCase = layer_norm_epsilon _lowerCAmelCase = rescale_every _lowerCAmelCase = use_cache _lowerCAmelCase = bos_token_id _lowerCAmelCase = eos_token_id super().__init__( tie_word_embeddings=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , **_lowerCamelCase )
158
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase__ : Optional[int] = {'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : int = ['YolosFeatureExtractor'] UpperCamelCase__ : int = ['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Dict = [ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys UpperCamelCase__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
from __future__ import annotations from fractions import Fraction def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> bool: '''simple docstring''' return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> list[str]: '''simple docstring''' lowerCAmelCase : Tuple = [] lowerCAmelCase : Optional[Any] = 11 lowerCAmelCase : str = int('1' + '0' * digit_len ) for num in range(a_, a_ ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(a_, a_ ): solutions.append(f"{num}/{den}" ) den += 1 num += 1 lowerCAmelCase : Any = 10 return solutions def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase = 2 ) -> int: '''simple docstring''' lowerCAmelCase : int = 1.0 for fraction in fraction_list(a_ ): lowerCAmelCase : List[str] = Fraction(a_ ) result *= frac.denominator / frac.numerator return int(a_ ) if __name__ == "__main__": print(solution())
138
'''simple docstring''' class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Optional[Any] = name A_ : Dict = value A_ : Union[str, Any] = weight def __repr__( self ) -> List[str]: return F"{self.__class__.__name__}({self.name}, {self.value}, {self.weight})" def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.value def UpperCAmelCase_ ( self ) -> List[str]: return self.name def UpperCAmelCase_ ( self ) -> Tuple: return self.weight def UpperCAmelCase_ ( self ) -> Optional[int]: return self.value / self.weight def UpperCAmelCase ( a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Optional[int] = [] for i in range(len(a_ ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def UpperCAmelCase ( a_ , a_ , a_ ) -> List[Any]: """simple docstring""" A_ : Optional[Any] = sorted(a_ , key=a_ , reverse=a_ ) A_ : str = [] A_ , A_ : Dict = 0.0, 0.0 for i in range(len(a_ ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
344
0
import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A_ ( __A , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Optional[Any] = XLMTokenizer _UpperCamelCase : List[Any] = False def SCREAMING_SNAKE_CASE__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] lowercase = dict(zip(_lowerCamelCase , range(len(_lowerCamelCase ) ) ) ) lowercase = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(_lowerCamelCase ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(_lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = """lower newer""" lowercase = """lower newer""" return input_text, output_text def SCREAMING_SNAKE_CASE__ ( self ): lowercase = XLMTokenizer(self.vocab_file , self.merges_file ) lowercase = """lower""" lowercase = ["""low""", """er</w>"""] lowercase = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) lowercase = tokens + ["""<unk>"""] lowercase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , _lowerCamelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = XLMTokenizer.from_pretrained('xlm-mlm-en-2048' ) lowercase = tokenizer.encode('sequence builders' , add_special_tokens=_lowerCamelCase ) lowercase = tokenizer.encode('multi-sequence build' , add_special_tokens=_lowerCamelCase ) lowercase = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase ) lowercase = tokenizer.build_inputs_with_special_tokens(_lowerCamelCase , _lowerCamelCase ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
195
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCAmelCase ( a_ , a_ ) -> tuple: """simple docstring""" if inductance <= 0: raise ValueError("""Inductance cannot be 0 or negative""" ) elif capacitance <= 0: raise ValueError("""Capacitance cannot be 0 or negative""" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
344
0
__UpperCAmelCase = {} def UpperCamelCase ( snake_case__ : int , snake_case__ : Dict , snake_case__ : List[Any] ) -> int: if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on UpperCamelCase : Optional[Any] = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one UpperCamelCase : Optional[Any] = _calculate(days - 1 , a_ , late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 UpperCamelCase : Tuple = _calculate(days - 1 , absent + 1 , 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter UpperCamelCase : str = _calculate(days - 1 , a_ , 0 ) UpperCamelCase : Union[str, Any] = state_late + state_absent + state_ontime UpperCamelCase : List[str] = prizestrings return prizestrings def UpperCamelCase ( snake_case__ : Union[str, Any] = 30 ) -> int: return _calculate(a_ , absent=0 , late=0 ) if __name__ == "__main__": print(solution())
119
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class _lowerCAmelCase ( __A ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase ) -> Any: A_ : List[Any] = parent A_ : int = config_class A_ : int = has_text_modality A_ : str = kwargs A_ : int = common_properties def UpperCAmelCase_ ( self ) -> str: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : Optional[int] = ( ["""hidden_size""", """num_attention_heads""", """num_hidden_layers"""] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["""vocab_size"""] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(_lowerCamelCase , _lowerCamelCase ) , msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(_lowerCamelCase ): try: setattr(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(_lowerCamelCase ): try: A_ : List[str] = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = self.config_class(**self.inputs_dict ) A_ : Optional[int] = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : str = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A_ : List[Any] = os.path.join(_lowerCamelCase , """config.json""" ) config_first.to_json_file(_lowerCamelCase ) A_ : Dict = self.config_class.from_json_file(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Any = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(_lowerCamelCase ) A_ : Union[str, Any] = self.config_class.from_pretrained(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : List[Any] = """test""" with tempfile.TemporaryDirectory() as tmpdirname: A_ : Any = os.path.join(_lowerCamelCase , _lowerCamelCase ) config_first.save_pretrained(_lowerCamelCase ) A_ : Any = self.config_class.from_pretrained(_lowerCamelCase , subfolder=_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Tuple = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) A_ : str = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def UpperCAmelCase_ ( self ) -> Optional[Any]: if self.config_class.is_composition: return A_ : Dict = self.config_class() self.parent.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : Any = copy.deepcopy(_lowerCamelCase ) A_ : Tuple = self.config_class(**_lowerCamelCase ) A_ : Optional[Any] = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("""torch_dtype""", config.torch_dtype, torch.floataa) ) elif getattr(_lowerCamelCase , _lowerCamelCase ) != value: wrong_values.append((key, getattr(_lowerCamelCase , _lowerCamelCase ), value) ) if len(_lowerCamelCase ) > 0: A_ : List[Any] = """\n""".join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def UpperCAmelCase_ ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
344
0
"""simple docstring""" import os import sys _lowerCamelCase : Optional[Any] = os.path.join(os.path.dirname(__file__), 'src') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) _lowerCamelCase : int = [ 'torch', 'numpy', 'tokenizers', 'filelock', 'requests', 'tqdm', 'regex', 'sentencepiece', 'sacremoses', 'importlib_metadata', 'huggingface_hub', ] @add_start_docstrings(AutoConfig.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoConfig.from_pretrained(*a_ , **a_ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoTokenizer.from_pretrained(*a_ , **a_ ) @add_start_docstrings(AutoModel.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoModel.from_pretrained(*a_ , **a_ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoModelForCausalLM.from_pretrained(*a_ , **a_ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoModelForMaskedLM.from_pretrained(*a_ , **a_ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoModelForSequenceClassification.from_pretrained(*a_ , **a_ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def lowercase_ ( *_UpperCAmelCase , **_UpperCAmelCase ): """simple docstring""" return AutoModelForQuestionAnswering.from_pretrained(*a_ , **a_ )
167
'''simple docstring''' from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" try: with open(a_ , """rb""" ) as flax_state_f: A_ : Tuple = from_bytes(a_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(a_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"Unable to convert {model_file} to Flax deserializable object. " ) return load_flax_weights_in_pytorch_model(a_ , a_ ) def UpperCAmelCase ( a_ , a_ ) -> Any: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights A_ : List[Any] = flatten_dict(jax.tree_util.tree_map(lambda a_ : x.dtype == jnp.bfloataa , a_ ) ).values() if any(a_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) A_ : str = jax.tree_util.tree_map( lambda a_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , a_ ) A_ : Any = """""" A_ : Optional[int] = flatten_dict(a_ , sep=""".""" ) A_ : List[str] = pt_model.state_dict() # keep track of unexpected & missing keys A_ : Union[str, Any] = [] A_ : Dict = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): A_ : List[Any] = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: A_ : Optional[Any] = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[Any] = jnp.transpose(a_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": A_ : int = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[int] = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": A_ : Any = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(a_ ): A_ : Tuple = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) A_ : Dict = """.""".join(a_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected " F"to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}." ) else: # add weight to pytorch dict A_ : Optional[Any] = np.asarray(a_ ) if not isinstance(a_ , np.ndarray ) else flax_tensor A_ : Tuple = torch.from_numpy(a_ ) # remove from missing keys missing_keys.remove(a_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(a_ ) pt_model.load_state_dict(a_ ) # re-transform missing_keys to list A_ : Dict = list(a_ ) if len(a_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing" F" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(a_ ) > 0: logger.warning( F"Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly" F" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to" """ use it for predictions and inference.""" ) return pt_model
344
0
import math def UpperCamelCase( __UpperCamelCase : Optional[Any] ): return math.sqrt(a_ ) * math.sqrt(a_ ) == num def UpperCamelCase( __UpperCamelCase : Any ): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : List[Any] = n while left <= right: lowerCAmelCase_ : Union[str, Any] = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: lowerCAmelCase_ : Any = mid - 1 else: lowerCAmelCase_ : Tuple = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
103
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=3 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=10 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[1, 1, 2, 1] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="relu" , _lowerCamelCase=3 , _lowerCamelCase=None , ) -> List[str]: A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> Optional[Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : Any = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase_ ( self ) -> int: super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: pass def UpperCAmelCase_ ( self ) -> int: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase , _lowerCamelCase ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) def UpperCAmelCase_ ( self ) -> str: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> int: return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""tf""" ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 )
344
0
import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging _lowerCamelCase =( 'https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py' ) _lowerCamelCase =logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case__ ( ): """simple docstring""" SCREAMING_SNAKE_CASE ="""https://pypi.org/pypi/diffusers/json""" SCREAMING_SNAKE_CASE =json.loads(request.urlopen(a_ ).read() )["""releases"""].keys() return sorted(a_, key=lambda lowerCAmelCase_ : version.Version(a_ ) ) def snake_case__ ( ): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(a_ ) os.makedirs(a_, exist_ok=a_ ) SCREAMING_SNAKE_CASE =Path(a_ ) / """__init__.py""" if not init_path.exists(): init_path.touch() def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" init_hf_modules() SCREAMING_SNAKE_CASE =Path(a_ ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(a_, exist_ok=a_ ) SCREAMING_SNAKE_CASE =dynamic_module_path / """__init__.py""" if not init_path.exists(): init_path.touch() def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" with open(a_, 'r', encoding='utf-8' ) as f: SCREAMING_SNAKE_CASE =f.read() # Imports of the form `import .xxx` SCREAMING_SNAKE_CASE =re.findall('^\s*import\s+\.(\S+)\s*$', a_, flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import', a_, flags=re.MULTILINE ) # Unique-ify return list(set(a_ ) ) def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE =False SCREAMING_SNAKE_CASE =[module_file] SCREAMING_SNAKE_CASE =[] # Let's recurse through all relative imports while not no_change: SCREAMING_SNAKE_CASE =[] for f in files_to_check: new_imports.extend(get_relative_imports(a_ ) ) SCREAMING_SNAKE_CASE =Path(a_ ).parent SCREAMING_SNAKE_CASE =[str(module_path / m ) for m in new_imports] SCREAMING_SNAKE_CASE =[f for f in new_import_files if f not in all_relative_imports] SCREAMING_SNAKE_CASE =[F'{f}.py' for f in new_import_files] SCREAMING_SNAKE_CASE =len(a_ ) == 0 all_relative_imports.extend(a_ ) return all_relative_imports def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" with open(a_, 'r', encoding='utf-8' ) as f: SCREAMING_SNAKE_CASE =f.read() # Imports of the form `import xxx` SCREAMING_SNAKE_CASE =re.findall('^\s*import\s+(\S+)\s*$', a_, flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import', a_, flags=re.MULTILINE ) # Only keep the top-level module SCREAMING_SNAKE_CASE =[imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all SCREAMING_SNAKE_CASE =list(set(a_ ) ) SCREAMING_SNAKE_CASE =[] for imp in imports: try: importlib.import_module(a_ ) except ImportError: missing_packages.append(a_ ) if len(a_ ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' F'{", ".join(a_ )}. Run `pip install {" ".join(a_ )}`' ) return get_relative_imports(a_ ) def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE =module_path.replace(os.path.sep, '.' ) SCREAMING_SNAKE_CASE =importlib.import_module(a_ ) if class_name is None: return find_pipeline_class(a_ ) return getattr(a_, a_ ) def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" from ..pipelines import DiffusionPipeline SCREAMING_SNAKE_CASE =dict(inspect.getmembers(a_, inspect.isclass ) ) SCREAMING_SNAKE_CASE =None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls, a_ ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( F'Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:' F' {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in' F' {loaded_module}.' ) SCREAMING_SNAKE_CASE =cls return pipeline_class def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ = None, lowerCAmelCase_ = False, lowerCAmelCase_ = False, lowerCAmelCase_ = None, lowerCAmelCase_ = None, lowerCAmelCase_ = None, lowerCAmelCase_ = False, ): """simple docstring""" SCREAMING_SNAKE_CASE =str(a_ ) SCREAMING_SNAKE_CASE =os.path.join(a_, a_ ) if os.path.isfile(a_ ): SCREAMING_SNAKE_CASE =module_file_or_url SCREAMING_SNAKE_CASE ="""local""" elif pretrained_model_name_or_path.count('/' ) == 0: SCREAMING_SNAKE_CASE =get_diffusers_versions() # cut ".dev0" SCREAMING_SNAKE_CASE ="""v""" + """.""".join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: SCREAMING_SNAKE_CASE =latest_version if latest_version[1:] in available_versions else """main""" logger.info(F'Defaulting to latest_version: {revision}.' ) elif revision in available_versions: SCREAMING_SNAKE_CASE =F'v{revision}' elif revision == "main": SCREAMING_SNAKE_CASE =revision else: raise ValueError( F'`custom_revision`: {revision} does not exist. Please make sure to choose one of' F' {", ".join(available_versions + ["main"] )}.' ) # community pipeline on GitHub SCREAMING_SNAKE_CASE =COMMUNITY_PIPELINES_URL.format(revision=a_, pipeline=a_ ) try: SCREAMING_SNAKE_CASE =cached_download( a_, cache_dir=a_, force_download=a_, proxies=a_, resume_download=a_, local_files_only=a_, use_auth_token=a_, ) SCREAMING_SNAKE_CASE ="""git""" SCREAMING_SNAKE_CASE =pretrained_model_name_or_path + """.py""" except EnvironmentError: logger.error(F'Could not locate the {module_file} inside {pretrained_model_name_or_path}.' ) raise else: try: # Load from URL or cache if already cached SCREAMING_SNAKE_CASE =hf_hub_download( a_, a_, cache_dir=a_, force_download=a_, proxies=a_, resume_download=a_, local_files_only=a_, use_auth_token=a_, ) SCREAMING_SNAKE_CASE =os.path.join('local', '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(F'Could not locate the {module_file} inside {pretrained_model_name_or_path}.' ) raise # Check we have all the requirements in our environment SCREAMING_SNAKE_CASE =check_imports(a_ ) # Now we move the module inside our cached dynamic modules. SCREAMING_SNAKE_CASE =DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(a_ ) SCREAMING_SNAKE_CASE =Path(a_ ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(a_, submodule_path / module_file ) for module_needed in modules_needed: SCREAMING_SNAKE_CASE =F'{module_needed}.py' shutil.copy(os.path.join(a_, a_ ), submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(a_, a_ ): SCREAMING_SNAKE_CASE =use_auth_token elif use_auth_token is True: SCREAMING_SNAKE_CASE =HfFolder.get_token() else: SCREAMING_SNAKE_CASE =None SCREAMING_SNAKE_CASE =model_info(a_, revision=a_, token=a_ ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. SCREAMING_SNAKE_CASE =submodule_path / commit_hash SCREAMING_SNAKE_CASE =full_submodule + os.path.sep + commit_hash create_dynamic_module(a_ ) if not (submodule_path / module_file).exists(): shutil.copy(a_, submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( a_, F'{module_needed}.py', cache_dir=a_, force_download=a_, resume_download=a_, proxies=a_, use_auth_token=a_, revision=a_, local_files_only=a_, ) return os.path.join(a_, a_ ) def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ = None, lowerCAmelCase_ = None, lowerCAmelCase_ = False, lowerCAmelCase_ = False, lowerCAmelCase_ = None, lowerCAmelCase_ = None, lowerCAmelCase_ = None, lowerCAmelCase_ = False, **lowerCAmelCase_, ): """simple docstring""" SCREAMING_SNAKE_CASE =get_cached_module_file( a_, a_, cache_dir=a_, force_download=a_, resume_download=a_, proxies=a_, use_auth_token=a_, revision=a_, local_files_only=a_, ) return get_class_in_module(a_, final_module.replace('.py', '' ) )
334
'''simple docstring''' def UpperCAmelCase ( a_ = 1_0_0 ) -> int: """simple docstring""" A_ : Dict = n * (n + 1) * (2 * n + 1) / 6 A_ : Optional[int] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'{solution() = }')
344
0
import string from math import logaa def _lowerCamelCase( lowercase__ , lowercase__ ) -> int: '''simple docstring''' __lowercase= document.translate( str.maketrans('' , '' , string.punctuation ) ).replace('\n' , '' ) __lowercase= document_without_punctuation.split(' ' ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def _lowerCamelCase( lowercase__ , lowercase__ ) -> tuple[int, int]: '''simple docstring''' __lowercase= corpus.lower().translate( str.maketrans('' , '' , string.punctuation ) ) # strip all punctuation and replace it with '' __lowercase= corpus_without_punctuation.split('\n' ) __lowercase= term.lower() return (len([doc for doc in docs if term in doc] ), len(a_ )) def _lowerCamelCase( lowercase__ , lowercase__ , lowercase__=False ) -> float: '''simple docstring''' if smoothing: if n == 0: raise ValueError('log10(0) is undefined.' ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError('df must be > 0' ) elif n == 0: raise ValueError('log10(0) is undefined.' ) return round(logaa(n / df ) , 3 ) def _lowerCamelCase( lowercase__ , lowercase__ ) -> float: '''simple docstring''' return round(tf * idf , 3 )
295
'''simple docstring''' from typing import TYPE_CHECKING from ..utils import _LazyModule UpperCamelCase__ : int = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys UpperCamelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
'''simple docstring''' from __future__ import annotations from cmath import sqrt def lowerCAmelCase_ ( snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' if a == 0: raise ValueError('''Coefficient \'a\' must not be zero.''' ) A : List[Any] = b * b - 4 * a * c A : Tuple = (-b + sqrt(a_ )) / (2 * a) A : List[Any] = (-b - sqrt(a_ )) / (2 * a) return ( root_a.real if not root_a.imag else root_a, root_a.real if not root_a.imag else root_a, ) def lowerCAmelCase_ ( ): '''simple docstring''' A : List[Any] = quadratic_roots(a=5 , b=6 , c=1 ) print(F'The solutions are: {solutiona} and {solutiona}' ) if __name__ == "__main__": main()
3
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCAmelCase ( a_ ) -> Dict[str, torch.Tensor]: """simple docstring""" A_ : List[str] = [] A_ : Dict = [] A_ : List[Any] = [] for rt in rc.restypes: A_ : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) A_ : Union[str, Any] = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) A_ : Tuple = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : Optional[int] = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : List[Any] = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) A_ : Optional[int] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein A_ : Dict = restype_atomaa_to_atomaa[protein_aatype] A_ : Optional[Any] = restype_atomaa_mask[protein_aatype] A_ : Any = residx_atomaa_mask A_ : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back A_ : Tuple = restype_atomaa_to_atomaa[protein_aatype] A_ : Tuple = residx_atomaa_to_atomaa.long() # create the corresponding mask A_ : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): A_ : Optional[Any] = rc.restype_atoa[restype_letter] A_ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: A_ : Any = rc.atom_order[atom_name] A_ : Optional[int] = 1 A_ : Optional[int] = restype_atomaa_mask[protein_aatype] A_ : Dict = residx_atomaa_mask return protein def UpperCAmelCase ( a_ ) -> Dict[str, np.ndarray]: """simple docstring""" A_ : Union[str, Any] = tree_map(lambda a_ : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) A_ : Optional[int] = tensor_tree_map(lambda a_ : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
344
0
import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType __UpperCamelCase : Dict = False, False, False @dataclass class __lowerCAmelCase : UpperCamelCase__ = None UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = None # Automatically constructed UpperCamelCase__ = '''dict''' UpperCamelCase__ = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) UpperCamelCase__ = field(default='''Audio''' , init=__A , repr=__A ) def __call__( self :Any ): '''simple docstring''' return self.pa_type def lowerCamelCase__ ( self :Optional[Any] , __magic_name__ :List[str] ): '''simple docstring''' try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("""To support encoding audio data, please install 'soundfile'.""" ) from err if isinstance(_lowerCamelCase , _lowerCamelCase ): return {"bytes": None, "path": value} elif isinstance(_lowerCamelCase , _lowerCamelCase ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes a = BytesIO() sf.write(_lowerCamelCase , value["""array"""] , value["""sampling_rate"""] , format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("""pcm""" ): # "PCM" only has raw audio bytes if value.get("""sampling_rate""" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("""To use PCM files, please specify a 'sampling_rate' in Audio object""" ) if value.get("""bytes""" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) a = np.frombuffer(value["""bytes"""] , dtype=np.intaa ).astype(np.floataa ) / 3_2767 else: a = np.memmap(value["""path"""] , dtype="""h""" , mode="""r""" ).astype(np.floataa ) / 3_2767 a = BytesIO(bytes() ) sf.write(_lowerCamelCase , _lowerCamelCase , value["""sampling_rate"""] , format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( F'An audio sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.' ) def lowerCamelCase__ ( self :str , __magic_name__ :str , __magic_name__ :str = None ): '''simple docstring''' if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Audio(decode=True) instead.""" ) a = (value["""path"""], BytesIO(value["""bytes"""] )) if value["""bytes"""] is not None else (value["""path"""], None) if path is None and file is None: raise ValueError(F'An audio sample should have one of \'path\' or \'bytes\' but both are None in {value}.' ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("""To support decoding audio files, please install 'librosa' and 'soundfile'.""" ) from err a = xsplitext(_lowerCamelCase )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( """Decoding 'opus' files requires system library 'libsndfile'>=1.0.31, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( """Decoding 'mp3' files requires system library 'libsndfile'>=1.1.0, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) if file is None: a = token_per_repo_id or {} a = path.split("""::""" )[-1] try: a = string_to_dict(_lowerCamelCase , config.HUB_DATASETS_URL )["""repo_id"""] a = token_per_repo_id[repo_id] except (ValueError, KeyError): a = None with xopen(_lowerCamelCase , """rb""" , use_auth_token=_lowerCamelCase ) as f: a = sf.read(_lowerCamelCase ) else: a = sf.read(_lowerCamelCase ) a = array.T if self.mono: a = librosa.to_mono(_lowerCamelCase ) if self.sampling_rate and self.sampling_rate != sampling_rate: a = librosa.resample(_lowerCamelCase , orig_sr=_lowerCamelCase , target_sr=self.sampling_rate ) a = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def lowerCamelCase__ ( self :str ): '''simple docstring''' from .features import Value if self.decode: raise ValueError("""Cannot flatten a decoded Audio feature.""" ) return { "bytes": Value("""binary""" ), "path": Value("""string""" ), } def lowerCamelCase__ ( self :str , __magic_name__ :Any ): '''simple docstring''' if pa.types.is_string(storage.type ): a = pa.array([None] * len(_lowerCamelCase ) , type=pa.binary() ) a = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): a = pa.array([None] * len(_lowerCamelCase ) , type=pa.string() ) a = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("""array""" ): a = pa.array([Audio().encode_example(_lowerCamelCase ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: a = storage.field("""bytes""" ) else: a = pa.array([None] * len(_lowerCamelCase ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: a = storage.field("""path""" ) else: a = pa.array([None] * len(_lowerCamelCase ) , type=pa.string() ) a = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) return array_cast(_lowerCamelCase , self.pa_type ) def lowerCamelCase__ ( self :Any , __magic_name__ :str ): '''simple docstring''' @no_op_if_value_is_null def path_to_bytes(__magic_name__ :Any ): with xopen(_lowerCamelCase , """rb""" ) as f: a = f.read() return bytes_ a = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) a = pa.array( [os.path.basename(_lowerCamelCase ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) a = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(_lowerCamelCase , self.pa_type )
228
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ) -> str: A_ : Optional[int] = parent A_ : Dict = batch_size A_ : List[Any] = image_size A_ : Optional[int] = patch_size A_ : List[str] = num_channels A_ : List[Any] = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_size A_ : str = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Union[str, Any] = intermediate_size A_ : Any = hidden_act A_ : Optional[Any] = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Dict = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : str = scope A_ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) A_ : Tuple = (image_size // patch_size) ** 2 A_ : Union[str, Any] = num_patches + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> int: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[str] = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : int = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A_ : Dict = 1 A_ : Optional[int] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Tuple = self.type_sequence_label_size A_ : Tuple = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : Dict = 1 A_ : Any = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : List[str] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = DeiTModelTester(self ) A_ : str = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: pass def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ , A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[str] = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Union[str, Any] = [*signature.parameters.keys()] A_ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: A_ : int = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase_ ( self ) -> Optional[Any]: if not self.model_tester.is_training: return A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[str] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : List[str] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> int: A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A_ : Any = False A_ : Union[str, Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue A_ : List[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Union[str, Any] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Tuple: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): A_ : Dict = problem_type["""title"""] A_ : List[Any] = problem_type["""num_labels"""] A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: A_ : Tuple = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) A_ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: A_ : List[str] = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Optional[Any]: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) A_ : Optional[int] = self.default_image_processor A_ : str = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Any = model(**_lowerCamelCase ) # verify the logits A_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : List[Any] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : int = self.default_image_processor A_ : List[str] = prepare_img() A_ : List[Any] = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) A_ : Union[str, Any] = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A_ : List[Any] = model(_lowerCamelCase )
344
0
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 1 , _SCREAMING_SNAKE_CASE = 1000 ): _snake_case = 1 _snake_case = 0 for divide_by_number in range(a_ , digit + 1 ): _snake_case = [] _snake_case = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(a_ ): _snake_case = len(a_ ) _snake_case = divide_by_number else: has_been_divided.append(a_ ) _snake_case = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
341
'''simple docstring''' import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 32 , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = True , _lowerCamelCase = [0.4814_5466, 0.457_8275, 0.4082_1073] , _lowerCamelCase = [0.2686_2954, 0.2613_0258, 0.2757_7711] , _lowerCamelCase = True , _lowerCamelCase=7 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=3 , ) -> Union[str, Any]: A_ : Optional[int] = parent A_ : Union[str, Any] = do_resize A_ : Optional[Any] = size if size is not None else {"""shortest_edge""": 288} A_ : Tuple = size_divisor A_ : List[Any] = do_rescale A_ : Dict = rescale_factor A_ : List[Any] = do_normalize A_ : Dict = do_center_crop A_ : Optional[Any] = image_mean A_ : List[str] = image_std A_ : str = do_pad A_ : Any = batch_size A_ : List[str] = num_channels A_ : List[str] = min_resolution A_ : Union[str, Any] = max_resolution def UpperCAmelCase_ ( self ) -> Any: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase=False ) -> Optional[int]: if not batched: A_ : Union[str, Any] = self.size["""shortest_edge"""] A_ : Dict = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image ): A_ , A_ : Optional[Any] = image.size else: A_ , A_ : int = image.shape[1], image.shape[2] A_ : Optional[int] = size / min(_lowerCamelCase , _lowerCamelCase ) if h < w: A_ , A_ : Optional[Any] = size, scale * w else: A_ , A_ : Dict = scale * h, size A_ : Union[str, Any] = int((1333 / 800) * size ) if max(_lowerCamelCase , _lowerCamelCase ) > max_size: A_ : str = max_size / max(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = newh * scale A_ : Dict = neww * scale A_ , A_ : str = int(newh + 0.5 ), int(neww + 0.5 ) A_ , A_ : Dict = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: A_ : Tuple = [] for image in image_inputs: A_ , A_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) A_ : List[Any] = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[0] )[0] A_ : Tuple = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : int = BridgeTowerImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size_divisor""" ) ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image processor A_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : int = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: # Initialize image processor A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image processor A_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
344
0
'''simple docstring''' from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import BaseOutput, is_torch_available, is_transformers_available @dataclass class lowerCAmelCase_ ( __A ): __lowerCamelCase : Any = 42 __lowerCamelCase : Dict = 42 if is_transformers_available() and is_torch_available(): from .pipeline_semantic_stable_diffusion import SemanticStableDiffusionPipeline
158
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(a_ ): for j in range(a_ ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) , end="""\t""" ) else: print("""INF""" , end="""\t""" ) print() def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" A_ : List[str] = [[float("""inf""" ) for _ in range(a_ )] for _ in range(a_ )] for i in range(a_ ): for j in range(a_ ): A_ : List[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(a_ ): # looping through rows of graph array for i in range(a_ ): # looping through columns of graph array for j in range(a_ ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): A_ : List[str] = dist[i][k] + dist[k][j] _print_dist(a_ , a_ ) return dist, v if __name__ == "__main__": UpperCamelCase__ : Tuple = int(input('Enter number of vertices: ')) UpperCamelCase__ : int = int(input('Enter number of edges: ')) UpperCamelCase__ : Dict = [[float('inf') for i in range(v)] for j in range(v)] for i in range(v): UpperCamelCase__ : Union[str, Any] = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print('\nEdge ', i + 1) UpperCamelCase__ : Union[str, Any] = int(input('Enter source:')) UpperCamelCase__ : int = int(input('Enter destination:')) UpperCamelCase__ : Optional[Any] = float(input('Enter weight:')) UpperCamelCase__ : Any = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
344
0
from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class __A ( __A ): def __lt__( self : Any , UpperCAmelCase_ : Optional[int] ): return self[-1] < other[-1] def __eq__( self : Any , UpperCAmelCase_ : Dict ): return self[-1] == other[-1] def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> list: '''simple docstring''' lowerCAmelCase : list[Stack] = [] # sort into stacks for element in collection: lowerCAmelCase : Any = Stack([element] ) lowerCAmelCase : Optional[int] = bisect_left(a_, a_ ) if i != len(a_ ): stacks[i].append(a_ ) else: stacks.append(a_ ) # use a heap-based merge to merge stack efficiently lowerCAmelCase : Any = merge(*(reversed(a_ ) for stack in stacks) ) return collection if __name__ == "__main__": __A : Union[str, Any] = input('''Enter numbers separated by a comma:\n''').strip() __A : List[str] = [int(item) for item in user_input.split(''',''')] print(patience_sort(unsorted))
138
'''simple docstring''' import datasets from .evaluate import evaluate UpperCamelCase__ : int = '\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n' UpperCamelCase__ : Any = '\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n' UpperCamelCase__ : Optional[Any] = '\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the SQuAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> squad_metric = datasets.load_metric("squad")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: A_ : Optional[Any] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} A_ : List[Any] = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] A_ : int = evaluate(dataset=_lowerCamelCase , predictions=_lowerCamelCase ) return score
344
0
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): stooge(a_ , 0 , len(a_ ) - 1 ) return arr def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: lowercase = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: lowercase = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(a_ , a_ , (h - t) ) # Recursively sort last 2/3 elements stooge(a_ , i + t , (a_) ) # Recursively sort first 2/3 elements stooge(a_ , a_ , (h - t) ) if __name__ == "__main__": UpperCAmelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
195
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ : Any = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Optional[Any] = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] UpperCamelCase__ : List[str] = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] UpperCamelCase__ : str = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): UpperCamelCase__ : List[str] = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCamelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
def UpperCamelCase ( snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : Optional[Any] ) -> float: UpperCamelCase : Optional[Any] = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def UpperCamelCase ( ) -> Union[str, Any]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
119
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''vision-encoder-decoder''' lowerCamelCase = True def __init__( self , **_lowerCamelCase ) -> str: super().__init__(**_lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F"A configuraton of type {self.model_type} cannot be instantiated because " F"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) A_ : Optional[int] = kwargs.pop("""encoder""" ) A_ : List[str] = encoder_config.pop("""model_type""" ) A_ : str = kwargs.pop("""decoder""" ) A_ : Optional[Any] = decoder_config.pop("""model_type""" ) A_ : List[str] = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : str = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : Any = True @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A_ : int = True A_ : List[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = copy.deepcopy(self.__dict__ ) A_ : List[str] = self.encoder.to_dict() A_ : Union[str, Any] = self.decoder.to_dict() A_ : str = self.__class__.model_type return output class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> float: return 1e-4 @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: A_ : Optional[Any] = OrderedDict() A_ : Any = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : Optional[int] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = -1 , _lowerCamelCase = -1 , _lowerCamelCase = False , _lowerCamelCase = None , ) -> Mapping[str, Any]: import torch A_ : Optional[int] = OrderedDict() A_ : List[Any] = super().generate_dummy_inputs( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) A_ , A_ : str = dummy_input["""input_ids"""].shape A_ : Optional[int] = (batch, encoder_sequence, self._config.encoder_hidden_size) A_ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A_ : List[str] = dummy_input.pop("""attention_mask""" ) A_ : Optional[int] = torch.zeros(_lowerCamelCase ) return common_inputs class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> None: pass def UpperCAmelCase_ ( self , _lowerCamelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = "default" ) -> OnnxConfig: A_ : List[Any] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_lowerCamelCase , _lowerCamelCase )
344
0
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() _lowerCamelCase : Tuple = logging.get_logger(__name__) _lowerCamelCase : Dict = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" for attribute in key.split('''.''' ): A_ : List[str] = getattr(a_ , a_ ) if weight_type is not None: A_ : int = getattr(a_ , a_ ).shape else: A_ : str = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": A_ : Tuple = value elif weight_type == "weight_g": A_ : Dict = value elif weight_type == "weight_v": A_ : Dict = value elif weight_type == "bias": A_ : str = value else: A_ : Optional[Any] = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Union[str, Any] = [] A_ : int = fairseq_model.state_dict() A_ : Optional[int] = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): A_ : Any = False if "conv_layers" in name: load_conv_layer( a_ , a_ , a_ , a_ , hf_model.config.feat_extract_norm == '''group''' , ) A_ : Tuple = True else: for key, mapped_key in MAPPING.items(): A_ : Dict = """hubert.""" + mapped_key if (is_finetuned and mapped_key != """lm_head""") else mapped_key if key in name or (key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0] and not is_finetuned): A_ : Dict = True if "*" in mapped_key: A_ : Optional[int] = name.split(a_ )[0].split('''.''' )[-2] A_ : int = mapped_key.replace('''*''' , a_ ) if "weight_g" in name: A_ : Any = """weight_g""" elif "weight_v" in name: A_ : List[str] = """weight_v""" elif "weight" in name: A_ : int = """weight""" elif "bias" in name: A_ : Union[str, Any] = """bias""" else: A_ : Any = None set_recursively(a_ , a_ , a_ , a_ , a_ ) continue if not is_used: unused_weights.append(a_ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : str = full_name.split('''conv_layers.''' )[-1] A_ : Tuple = name.split('''.''' ) A_ : Tuple = int(items[0] ) A_ : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) A_ : Any = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) A_ : Union[str, Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) A_ : Optional[Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) A_ : str = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(a_ ) @torch.no_grad() def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=True ): """simple docstring""" if config_path is not None: A_ : Tuple = HubertConfig.from_pretrained(a_ ) else: A_ : List[str] = HubertConfig() if is_finetuned: if dict_path: A_ : Dict = Dictionary.load(a_ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq A_ : Optional[Any] = target_dict.pad_index A_ : List[Any] = target_dict.bos_index A_ : Any = target_dict.eos_index A_ : Optional[Any] = len(target_dict.symbols ) A_ : Any = os.path.join(a_ , '''vocab.json''' ) if not os.path.isdir(a_ ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(a_ ) ) return os.makedirs(a_ , exist_ok=a_ ) with open(a_ , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , a_ ) A_ : Optional[Any] = WavaVecaCTCTokenizer( a_ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=a_ , ) A_ : int = True if config.feat_extract_norm == """layer""" else False A_ : List[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=a_ , return_attention_mask=a_ , ) A_ : int = WavaVecaProcessor(feature_extractor=a_ , tokenizer=a_ ) processor.save_pretrained(a_ ) A_ : int = HubertForCTC(a_ ) else: A_ : Tuple = HubertModel(a_ ) if is_finetuned: A_ : Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: A_ : Any = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) A_ : Optional[Any] = model[0].eval() recursively_load_weights(a_ , a_ , a_ ) hf_wavavec.save_pretrained(a_ ) if __name__ == "__main__": _lowerCamelCase : List[str] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) _lowerCamelCase : int = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
167
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCamelCase__ : Any = '\\n Text data.\n Second line of data.' UpperCamelCase__ : List[Any] = 'file' @pytest.fixture(scope="""session""" ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : int = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") A_ : int = bytes(a_ , """utf-8""" ) with zstd.open(a_ , """wb""" ) as f: f.write(a_ ) return path @pytest.fixture def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , a_ ) , """w""" ) as f: f.write(a_ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> Optional[int]: """simple docstring""" A_ : List[str] = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} A_ : Any = input_paths[compression_format] A_ : Tuple = tmp_path / """cache""" A_ : Tuple = DownloadConfig(cache_dir=a_ , extract_compressed_file=a_ ) A_ : Dict = cached_path(a_ , download_config=a_ ) with open(a_ ) as f: A_ : Optional[Any] = f.read() with open(a_ ) as f: A_ : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Union[str, Any] = """custom_cache""" A_ : List[str] = """custom_extracted_dir""" A_ : Optional[Any] = tmp_path / """custom_extracted_path""" if default_extracted: A_ : Any = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , a_ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(a_ ) ) A_ : Union[str, Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) A_ : List[Any] = xz_file A_ : Optional[int] = ( DownloadConfig(extract_compressed_file=a_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=a_ ) ) A_ : Union[str, Any] = cached_path(a_ , download_config=a_ ) assert Path(a_ ).parent.parts[-2:] == expected def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : str = str(Path(a_ ).resolve() ) assert cached_path(a_ ) == text_file # relative path A_ : List[str] = str(Path(a_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(a_ ) == text_file def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : Optional[Any] = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(a_ ): cached_path(a_ ) # relative path A_ : Tuple = """./__missing_file__.txt""" with pytest.raises(a_ ): cached_path(a_ ) def UpperCAmelCase ( a_ ) -> Tuple: """simple docstring""" A_ : Any = get_from_cache(F"tmp://{tmpfs_file}" ) with open(a_ ) as f: A_ : List[str] = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( ) -> List[str]: """simple docstring""" with pytest.raises(a_ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Union[str, Any]: """simple docstring""" A_ : List[str] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): http_get("""https://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): ftp_get("""ftp://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): fsspec_get("""s3://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): fsspec_head("""s3://huggingface.co""" )
344
0
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class __snake_case ( unittest.TestCase ): def __init__( self : Dict , A_ : Dict , A_ : Optional[Any] = True , A_ : Dict = None , A_ : Tuple = 3_2 , A_ : List[str] = True , A_ : Union[str, Any] = 1 / 2_5_5 , A_ : List[str] = True , A_ : Optional[int] = True , A_ : Tuple = [0.4814_5466, 0.457_8275, 0.4082_1073] , A_ : Any = [0.2686_2954, 0.2613_0258, 0.2757_7711] , A_ : List[Any] = True , A_ : str=7 , A_ : Optional[int]=3_0 , A_ : Dict=4_0_0 , A_ : Optional[int]=3 , ): lowerCAmelCase_ : Optional[int] = parent lowerCAmelCase_ : Union[str, Any] = do_resize lowerCAmelCase_ : Optional[Any] = size if size is not None else {"""shortest_edge""": 2_8_8} lowerCAmelCase_ : Tuple = size_divisor lowerCAmelCase_ : List[Any] = do_rescale lowerCAmelCase_ : Dict = rescale_factor lowerCAmelCase_ : List[Any] = do_normalize lowerCAmelCase_ : Dict = do_center_crop lowerCAmelCase_ : Optional[Any] = image_mean lowerCAmelCase_ : List[str] = image_std lowerCAmelCase_ : str = do_pad lowerCAmelCase_ : Any = batch_size lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : List[str] = min_resolution lowerCAmelCase_ : Union[str, Any] = max_resolution def UpperCAmelCase__ ( self : Tuple): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase__ ( self : str , A_ : Union[str, Any] , A_ : List[Any]=False): if not batched: lowerCAmelCase_ : Union[str, Any] = self.size["""shortest_edge"""] lowerCAmelCase_ : Dict = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image): lowerCAmelCase_ : Optional[Any] = image.size else: lowerCAmelCase_ : int = image.shape[1], image.shape[2] lowerCAmelCase_ : Optional[int] = size / min(_lowerCamelCase , _lowerCamelCase) if h < w: lowerCAmelCase_ : Optional[Any] = size, scale * w else: lowerCAmelCase_ : Dict = scale * h, size lowerCAmelCase_ : Union[str, Any] = int((1_3_3_3 / 8_0_0) * size) if max(_lowerCamelCase , _lowerCamelCase) > max_size: lowerCAmelCase_ : str = max_size / max(_lowerCamelCase , _lowerCamelCase) lowerCAmelCase_ : Dict = newh * scale lowerCAmelCase_ : Dict = neww * scale lowerCAmelCase_ : str = int(newh + 0.5), int(neww + 0.5) lowerCAmelCase_ : Dict = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: lowerCAmelCase_ : Tuple = [] for image in image_inputs: lowerCAmelCase_ : Tuple = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) lowerCAmelCase_ : List[Any] = max(_lowerCamelCase , key=lambda A_: item[0])[0] lowerCAmelCase_ : Tuple = max(_lowerCamelCase , key=lambda A_: item[1])[1] return expected_height, expected_width @require_torch @require_vision class __snake_case ( __A ,unittest.TestCase ): _a = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : List[str]): lowerCAmelCase_ : int = BridgeTowerImageProcessingTester(self) @property def UpperCAmelCase__ ( self : str): return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Optional[int]): lowerCAmelCase_ : int = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(_lowerCamelCase , '''image_mean''')) self.assertTrue(hasattr(_lowerCamelCase , '''image_std''')) self.assertTrue(hasattr(_lowerCamelCase , '''do_normalize''')) self.assertTrue(hasattr(_lowerCamelCase , '''do_resize''')) self.assertTrue(hasattr(_lowerCamelCase , '''size''')) self.assertTrue(hasattr(_lowerCamelCase , '''size_divisor''')) def UpperCAmelCase__ ( self : Optional[Any]): pass def UpperCAmelCase__ ( self : Optional[Any]): # Initialize image processor lowerCAmelCase_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images lowerCAmelCase_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image) # Test not batched input lowerCAmelCase_ : int = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values lowerCAmelCase_ : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCamelCase) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase_ : Optional[Any] = image_processing(_lowerCamelCase , return_tensors='''pt''').pixel_values lowerCAmelCase_ : int = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : int): # Initialize image processor lowerCAmelCase_ : Tuple = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values lowerCAmelCase_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase_ : int = image_processing(_lowerCamelCase , return_tensors='''pt''').pixel_values lowerCAmelCase_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase__ ( self : List[str]): # Initialize image processor lowerCAmelCase_ : Dict = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor) # Test not batched input lowerCAmelCase_ : Any = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values lowerCAmelCase_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched lowerCAmelCase_ : List[Any] = image_processing(_lowerCamelCase , return_tensors='''pt''').pixel_values lowerCAmelCase_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
103
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) UpperCamelCase__ : int = {'processing_layoutxlm': ['LayoutXLMProcessor']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Tuple = ['LayoutXLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : List[Any] = ['LayoutXLMTokenizerFast'] if TYPE_CHECKING: from .processing_layoutxlm import LayoutXLMProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm import LayoutXLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm_fast import LayoutXLMTokenizerFast else: import sys UpperCamelCase__ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase ={ 'facebook/mask2former-swin-small-coco-instance': ( 'https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json' ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } _lowerCamelCase =logging.get_logger(__name__) class a_ ( __A ): """simple docstring""" __UpperCAmelCase = 'mask2former' __UpperCAmelCase = ['swin'] __UpperCAmelCase = {'hidden_size': 'hidden_dim'} def __init__( self : Union[str, Any] ,snake_case : Dict = None ,snake_case : List[str] = 256 ,snake_case : Optional[int] = 256 ,snake_case : Any = 256 ,snake_case : int = 1024 ,snake_case : Dict = "relu" ,snake_case : Dict = 6 ,snake_case : Dict = 10 ,snake_case : Optional[int] = 8 ,snake_case : List[str] = 0.0 ,snake_case : Union[str, Any] = 2048 ,snake_case : List[str] = False ,snake_case : str = False ,snake_case : str = 4 ,snake_case : Dict = 255 ,snake_case : Dict = 100 ,snake_case : Optional[int] = 0.1 ,snake_case : Optional[int] = 2.0 ,snake_case : List[str] = 5.0 ,snake_case : Tuple = 5.0 ,snake_case : int = 12544 ,snake_case : int = 3.0 ,snake_case : List[Any] = 0.75 ,snake_case : Union[str, Any] = 0.02 ,snake_case : Tuple = 1.0 ,snake_case : List[str] = True ,snake_case : int = [4, 8, 16, 32] ,snake_case : List[Any] = None ,**snake_case : List[str] ,): if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.' ) SCREAMING_SNAKE_CASE =CONFIG_MAPPING["""swin"""]( image_size=224 ,in_channels=3 ,patch_size=4 ,embed_dim=96 ,depths=[2, 2, 18, 2] ,num_heads=[3, 6, 12, 24] ,window_size=7 ,drop_path_rate=0.3 ,use_absolute_embeddings=_lowerCamelCase ,out_features=['stage1', 'stage2', 'stage3', 'stage4'] ,) if isinstance(_lowerCamelCase ,_lowerCamelCase ): SCREAMING_SNAKE_CASE =backbone_config.pop('model_type' ) SCREAMING_SNAKE_CASE =CONFIG_MAPPING[backbone_model_type] SCREAMING_SNAKE_CASE =config_class.from_dict(_lowerCamelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f'Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. ' f'Supported model types: {",".join(self.backbones_supported )}' ) SCREAMING_SNAKE_CASE =backbone_config SCREAMING_SNAKE_CASE =feature_size SCREAMING_SNAKE_CASE =mask_feature_size SCREAMING_SNAKE_CASE =hidden_dim SCREAMING_SNAKE_CASE =encoder_feedforward_dim SCREAMING_SNAKE_CASE =activation_function SCREAMING_SNAKE_CASE =encoder_layers SCREAMING_SNAKE_CASE =decoder_layers SCREAMING_SNAKE_CASE =num_attention_heads SCREAMING_SNAKE_CASE =dropout SCREAMING_SNAKE_CASE =dim_feedforward SCREAMING_SNAKE_CASE =pre_norm SCREAMING_SNAKE_CASE =enforce_input_projection SCREAMING_SNAKE_CASE =common_stride SCREAMING_SNAKE_CASE =ignore_value SCREAMING_SNAKE_CASE =num_queries SCREAMING_SNAKE_CASE =no_object_weight SCREAMING_SNAKE_CASE =class_weight SCREAMING_SNAKE_CASE =mask_weight SCREAMING_SNAKE_CASE =dice_weight SCREAMING_SNAKE_CASE =train_num_points SCREAMING_SNAKE_CASE =oversample_ratio SCREAMING_SNAKE_CASE =importance_sample_ratio SCREAMING_SNAKE_CASE =init_std SCREAMING_SNAKE_CASE =init_xavier_std SCREAMING_SNAKE_CASE =use_auxiliary_loss SCREAMING_SNAKE_CASE =feature_strides SCREAMING_SNAKE_CASE =output_auxiliary_logits SCREAMING_SNAKE_CASE =decoder_layers super().__init__(**_lowerCamelCase ) @classmethod def _lowerCAmelCase ( cls : Any ,snake_case : Union[str, Any] ,**snake_case : List[str] ): return cls( backbone_config=_lowerCamelCase ,**_lowerCamelCase ,) def _lowerCAmelCase ( self : Any ): SCREAMING_SNAKE_CASE =copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE =self.backbone_config.to_dict() SCREAMING_SNAKE_CASE =self.__class__.model_type return output
334
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ : Any = logging.get_logger(__name__) UpperCamelCase__ : Optional[int] = { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/config.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/config.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json' ), 'distilbert-base-uncased-finetuned-sst-2-english': ( 'https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json' ), } class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''distilbert''' lowerCamelCase = { '''hidden_size''': '''dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', } def __init__( self , _lowerCamelCase=3_0522 , _lowerCamelCase=512 , _lowerCamelCase=False , _lowerCamelCase=6 , _lowerCamelCase=12 , _lowerCamelCase=768 , _lowerCamelCase=4 * 768 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase="gelu" , _lowerCamelCase=0.02 , _lowerCamelCase=0.1 , _lowerCamelCase=0.2 , _lowerCamelCase=0 , **_lowerCamelCase , ) -> Optional[Any]: A_ : Tuple = vocab_size A_ : List[Any] = max_position_embeddings A_ : int = sinusoidal_pos_embds A_ : int = n_layers A_ : str = n_heads A_ : Optional[int] = dim A_ : int = hidden_dim A_ : Tuple = dropout A_ : List[Any] = attention_dropout A_ : int = activation A_ : Dict = initializer_range A_ : List[Any] = qa_dropout A_ : int = seq_classif_dropout super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": A_ : Union[str, Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A_ : int = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
344
0
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def _lowerCamelCase( lowercase__ , lowercase__=False ) -> Optional[Any]: '''simple docstring''' __lowercase= OmegaConf.load(a_ ) if display: print(yaml.dump(OmegaConf.to_container(a_ ) ) ) return config def _lowerCamelCase( lowercase__ , lowercase__=None , lowercase__=None ) -> Union[str, Any]: '''simple docstring''' if conf_path is None: __lowercase= """./model_checkpoints/vqgan_only.yaml""" __lowercase= load_config(a_ , display=a_ ) __lowercase= VQModel(**config.model.params ) if ckpt_path is None: __lowercase= """./model_checkpoints/vqgan_only.pt""" __lowercase= torch.load(a_ , map_location=a_ ) if ".ckpt" in ckpt_path: __lowercase= sd["""state_dict"""] model.load_state_dict(a_ , strict=a_ ) model.to(a_ ) del sd return model def _lowerCamelCase( lowercase__ , lowercase__ ) -> Tuple: '''simple docstring''' __lowercase= model.encode(a_ ) print(F'VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}' ) __lowercase= model.decode(a_ ) return xrec def _lowerCamelCase( lowercase__ , lowercase__=False ) -> Union[str, Any]: '''simple docstring''' __lowercase= string.rsplit('.' , 1 ) if reload: __lowercase= importlib.import_module(a_ ) importlib.reload(a_ ) return getattr(importlib.import_module(a_ , package=a_ ) , cls ) def _lowerCamelCase( lowercase__ ) -> str: '''simple docstring''' if "target" not in config: raise KeyError('Expected key `target` to instantiate.' ) return get_obj_from_str(config['target'] )(**config.get('params' , {} ) ) def _lowerCamelCase( lowercase__ , lowercase__ , lowercase__=True , lowercase__=True ) -> Optional[int]: '''simple docstring''' __lowercase= instantiate_from_config(a_ ) if sd is not None: model.load_state_dict(a_ ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def _lowerCamelCase( lowercase__ , lowercase__ , lowercase__ , lowercase__ ) -> Optional[Any]: '''simple docstring''' if ckpt: __lowercase= torch.load(a_ , map_location='cpu' ) __lowercase= pl_sd["""global_step"""] print(F'loaded model from global step {global_step}.' ) else: __lowercase= {"""state_dict""": None} __lowercase= None __lowercase= load_model_from_config(config.model , pl_sd['state_dict'] , gpu=a_ , eval_mode=a_ )["""model"""] return model, global_step
295
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ : int = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) UpperCamelCase__ : Any = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', f'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', f'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qpos_proj.weight', f'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kpos_proj.weight', f'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.weight', f'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', f'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', f'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kpos_proj.weight', f'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.weight', f'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', f'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', f'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', f'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_qpos_proj.bias', f'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_kpos_proj.bias', f'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.bias', f'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', f'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', f'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_kpos_proj.bias', f'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.bias', f'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', f'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ('transformer.decoder.ref_point_head.layers.0.weight', 'decoder.ref_point_head.layers.0.weight'), ('transformer.decoder.ref_point_head.layers.0.bias', 'decoder.ref_point_head.layers.0.bias'), ('transformer.decoder.ref_point_head.layers.1.weight', 'decoder.ref_point_head.layers.1.weight'), ('transformer.decoder.ref_point_head.layers.1.bias', 'decoder.ref_point_head.layers.1.bias'), ('transformer.decoder.query_scale.layers.0.weight', 'decoder.query_scale.layers.0.weight'), ('transformer.decoder.query_scale.layers.0.bias', 'decoder.query_scale.layers.0.bias'), ('transformer.decoder.query_scale.layers.1.weight', 'decoder.query_scale.layers.1.weight'), ('transformer.decoder.query_scale.layers.1.bias', 'decoder.query_scale.layers.1.bias'), ('transformer.decoder.layers.0.ca_qpos_proj.weight', 'decoder.layers.0.ca_qpos_proj.weight'), ('transformer.decoder.layers.0.ca_qpos_proj.bias', 'decoder.layers.0.ca_qpos_proj.bias'), ] ) def UpperCAmelCase ( a_ , a_ , a_ ) -> Optional[Any]: """simple docstring""" A_ : int = state_dict.pop(a_ ) A_ : Tuple = val def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : Union[str, Any] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: A_ : Optional[int] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) A_ : str = value else: A_ : int = value return new_state_dict def UpperCAmelCase ( a_ , a_=False ) -> Optional[int]: """simple docstring""" A_ : List[Any] = """""" if is_panoptic: A_ : Any = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) A_ : Optional[int] = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) A_ : str = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict A_ : Optional[Any] = in_proj_weight[:2_5_6, :] A_ : Tuple = in_proj_bias[:2_5_6] A_ : Dict = in_proj_weight[2_5_6:5_1_2, :] A_ : int = in_proj_bias[2_5_6:5_1_2] A_ : int = in_proj_weight[-2_5_6:, :] A_ : Optional[int] = in_proj_bias[-2_5_6:] def UpperCAmelCase ( ) -> Dict: """simple docstring""" A_ : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A_ : List[Any] = Image.open(requests.get(a_ , stream=a_ ).raw ) return im @torch.no_grad() def UpperCAmelCase ( a_ , a_ ) -> Dict: """simple docstring""" A_ : int = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: A_ : str = """resnet101""" if "dc5" in model_name: A_ : List[Any] = True A_ : str = """panoptic""" in model_name if is_panoptic: A_ : Dict = 2_5_0 else: A_ : Union[str, Any] = 9_1 A_ : str = """huggingface/label-files""" A_ : Union[str, Any] = """coco-detection-id2label.json""" A_ : Optional[Any] = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) A_ : str = {int(a_ ): v for k, v in idalabel.items()} A_ : Optional[int] = idalabel A_ : Tuple = {v: k for k, v in idalabel.items()} # load image processor A_ : List[Any] = """coco_panoptic""" if is_panoptic else """coco_detection""" A_ : Any = ConditionalDetrImageProcessor(format=a_ ) # prepare image A_ : Tuple = prepare_img() A_ : Any = image_processor(images=a_ , return_tensors="""pt""" ) A_ : Optional[int] = encoding["""pixel_values"""] logger.info(F"Converting model {model_name}..." ) # load original model from torch hub A_ : int = torch.hub.load("""DeppMeng/ConditionalDETR""" , a_ , pretrained=a_ ).eval() A_ : List[Any] = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: A_ : Union[str, Any] = """conditional_detr.""" + src rename_key(a_ , a_ , a_ ) A_ : Any = rename_backbone_keys(a_ ) # query, key and value matrices need special treatment read_in_q_k_v(a_ , is_panoptic=a_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them A_ : List[str] = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): A_ : Dict = state_dict.pop(a_ ) A_ : List[Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: A_ : str = state_dict.pop(a_ ) A_ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: A_ : Optional[int] = state_dict.pop(a_ ) A_ : str = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): A_ : Tuple = state_dict.pop(a_ ) A_ : Dict = val # finally, create HuggingFace model and load state dict A_ : Union[str, Any] = ConditionalDetrForSegmentation(a_ ) if is_panoptic else ConditionalDetrForObjectDetection(a_ ) model.load_state_dict(a_ ) model.eval() model.push_to_hub(repo_id=a_ , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion A_ : str = conditional_detr(a_ ) A_ : str = model(a_ ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(a_ ).mkdir(exist_ok=a_ ) model.save_pretrained(a_ ) image_processor.save_pretrained(a_ ) if __name__ == "__main__": UpperCamelCase__ : int = argparse.ArgumentParser() parser.add_argument( '--model_name', default='conditional_detr_resnet50', type=str, help='Name of the CONDITIONAL_DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) UpperCamelCase__ : Optional[Any] = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
344
0
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowercase : Optional[int] = get_tests_dir('fixtures/test_sentencepiece_with_bytefallback.model') @require_sentencepiece @require_tokenizers class A ( __A , unittest.TestCase ): __magic_name__ = GPTSwaTokenizer __magic_name__ = False __magic_name__ = True __magic_name__ = False def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing A : Tuple = GPTSwaTokenizer(_lowerCamelCase , eos_token='''<unk>''' , bos_token='''<unk>''' , pad_token='''<unk>''' ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" A : str = """This is a test""" A : Dict = """This is a test""" return input_text, output_text def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A : Optional[Any] = """<s>""" A : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCamelCase ) , _lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCamelCase ) , _lowerCamelCase ) def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" A : Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(_lowerCamelCase ) , 2000 ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2000 ) def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : Dict = GPTSwaTokenizer(_lowerCamelCase ) A : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowerCamelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [465, 287, 265, 631, 842] ) A : int = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) # fmt: off self.assertListEqual( _lowerCamelCase , ['''▁I''', '''▁was''', '''▁bor''', '''n''', '''▁in''', '''▁''', '''<0x39>''', '''2''', '''0''', '''0''', '''0''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁f''', '''al''', '''s''', '''<0xC3>''', '''<0xA9>''', '''.'''] , ) # fmt: on A : List[str] = tokenizer.convert_tokens_to_ids(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [262, 272, 1525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) A : Optional[Any] = tokenizer.convert_ids_to_tokens(_lowerCamelCase ) # fmt: off self.assertListEqual( _lowerCamelCase , ['''▁I''', '''▁was''', '''▁bor''', '''n''', '''▁in''', '''▁''', '''<0x39>''', '''2''', '''0''', '''0''', '''0''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁f''', '''al''', '''s''', '''<0xC3>''', '''<0xA9>''', '''.'''] ) # fmt: on def __lowerCAmelCase ( self ) -> int: """simple docstring""" A : Any = GPTSwaTokenizer(_lowerCamelCase ) A : Optional[int] = ["""This is a test""", """I was born in 92000, and this is falsé."""] A : Dict = [ [465, 287, 265, 631, 842], [262, 272, 1525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(_lowerCamelCase , _lowerCamelCase ): self.assertListEqual(tokenizer.encode_fast(_lowerCamelCase ) , _lowerCamelCase ) # Test that decode_fast returns the input text for text, token_ids in zip(_lowerCamelCase , _lowerCamelCase ): self.assertEqual(tokenizer.decode_fast(_lowerCamelCase ) , _lowerCamelCase ) @slow def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A : Dict = [ """<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')""", """Hey there, how are you doing this fine day?""", """This is a text with a trailing spaces followed by a dot .""", """Häj sväjs lillebrör! =)""", """Det är inget fel på Mr. Cool""", ] # fmt: off A : Dict = {"""input_ids""": [[63423, 5, 6811, 14954, 282, 816, 3821, 63466, 63425, 63462, 18, 63978, 678, 301, 1320, 63423, 63455, 63458, 18, 63982, 4246, 3940, 1901, 47789, 5547, 18994], [19630, 1100, 63446, 1342, 633, 544, 4488, 593, 5102, 2416, 63495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1652, 428, 268, 1936, 515, 268, 58593, 22413, 9106, 546, 268, 33213, 63979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55130, 63450, 924, 63449, 2249, 4062, 1558, 318, 63504, 21498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2827, 2559, 332, 6575, 63443, 26801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCamelCase , model_name='''AI-Sweden/gpt-sw3-126m''' , sequences=_lowerCamelCase , )
3
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = (UnCLIPScheduler,) def UpperCAmelCase_ ( self , **_lowerCamelCase ) -> List[Any]: A_ : Union[str, Any] = { """num_train_timesteps""": 1000, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**_lowerCamelCase ) return config def UpperCAmelCase_ ( self ) -> List[Any]: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample in [True, False]: self.check_over_configs(clip_sample=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[int]: for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=_lowerCamelCase , prev_timestep=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: A_ : Optional[int] = self.scheduler_classes[0] A_ : Any = self.get_scheduler_config(variance_type="""fixed_small_log""" ) A_ : List[Any] = scheduler_class(**_lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.054_9625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.999_4987 ) ) < 1e-5 def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : List[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config(variance_type="""learned_range""" ) A_ : Dict = scheduler_class(**_lowerCamelCase ) A_ : Dict = 0.5 assert scheduler._get_variance(1 , predicted_variance=_lowerCamelCase ) - -10.171_2790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=_lowerCamelCase ) - -5.799_8052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=_lowerCamelCase ) - -0.001_0011 < 1e-5 def UpperCAmelCase_ ( self ) -> Any: A_ : Optional[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config() A_ : Optional[Any] = scheduler_class(**_lowerCamelCase ) A_ : int = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : str = self.dummy_sample_deter A_ : Optional[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Any = model(_lowerCamelCase , _lowerCamelCase ) # 2. predict previous mean of sample x_t-1 A_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : List[Any] = pred_prev_sample A_ : Any = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Optional[Any] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 252.268_2495 ) < 1e-2 assert abs(result_mean.item() - 0.328_4743 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Dict: A_ : Union[str, Any] = self.scheduler_classes[0] A_ : Dict = self.get_scheduler_config() A_ : Tuple = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(25 ) A_ : List[str] = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : List[Any] = self.dummy_sample_deter A_ : List[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Optional[Any] = model(_lowerCamelCase , _lowerCamelCase ) if i + 1 == timesteps.shape[0]: A_ : List[str] = None else: A_ : Dict = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 A_ : str = scheduler.step( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , prev_timestep=_lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : Optional[Any] = pred_prev_sample A_ : Dict = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : List[str] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 258.204_4983 ) < 1e-2 assert abs(result_mean.item() - 0.336_2038 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> int: pass
344
0
from maths.prime_check import is_prime def __A ( __lowerCamelCase ) -> int: if not isinstance(a_ , a_ ): a = f'Input value of [number={number}] must be an integer' raise TypeError(a_ ) if is_prime(a_ ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
228
'''simple docstring''' import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=7 , _lowerCamelCase=3 , _lowerCamelCase=18 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=False , ) -> Optional[int]: A_ : Union[str, Any] = size if size is not None else {"""height""": 20, """width""": 20} A_ : Tuple = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} A_ : Optional[Any] = parent A_ : Optional[int] = batch_size A_ : Union[str, Any] = num_channels A_ : str = image_size A_ : Tuple = min_resolution A_ : Dict = max_resolution A_ : str = do_resize A_ : Tuple = size A_ : int = do_center_crop A_ : Dict = crop_size A_ : Tuple = do_normalize A_ : List[str] = image_mean A_ : Optional[Any] = image_std A_ : Any = do_reduce_labels def UpperCAmelCase_ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def UpperCAmelCase ( ) -> List[str]: """simple docstring""" A_ : Any = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(dataset[0]["""file"""] ) A_ : Dict = Image.open(dataset[1]["""file"""] ) return image, map def UpperCAmelCase ( ) -> Optional[int]: """simple docstring""" A_ : Tuple = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(ds[0]["""file"""] ) A_ : List[Any] = Image.open(ds[1]["""file"""] ) A_ : Any = Image.open(ds[2]["""file"""] ) A_ : str = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BeitImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : List[Any] = BeitImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) A_ : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=_lowerCamelCase ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> Dict: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> str: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : Union[str, Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> Optional[int]: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) A_ : Optional[int] = [] for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) A_ , A_ : List[Any] = prepare_semantic_single_inputs() A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) A_ , A_ : str = prepare_semantic_batch_inputs() A_ : Any = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image_processing A_ : Any = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 A_ , A_ : Tuple = prepare_semantic_single_inputs() A_ : str = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) A_ : str = True A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
344
0
'''simple docstring''' import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=7 , UpperCAmelCase=3 , UpperCAmelCase=18 , UpperCAmelCase=30 , UpperCAmelCase=400 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=True , UpperCAmelCase=[0.5, 0.5, 0.5] , UpperCAmelCase=[0.5, 0.5, 0.5] , UpperCAmelCase=False , ) -> Optional[int]: _snake_case = size if size is not None else {"""height""": 20, """width""": 20} _snake_case = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} _snake_case = parent _snake_case = batch_size _snake_case = num_channels _snake_case = image_size _snake_case = min_resolution _snake_case = max_resolution _snake_case = do_resize _snake_case = size _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_normalize _snake_case = image_mean _snake_case = image_std _snake_case = do_reduce_labels def lowercase (self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def __SCREAMING_SNAKE_CASE ( ): _snake_case = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) _snake_case = Image.open(dataset[0]["""file"""] ) _snake_case = Image.open(dataset[1]["""file"""] ) return image, map def __SCREAMING_SNAKE_CASE ( ): _snake_case = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) _snake_case = Image.open(ds[0]["""file"""] ) _snake_case = Image.open(ds[1]["""file"""] ) _snake_case = Image.open(ds[2]["""file"""] ) _snake_case = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _lowerCAmelCase ( __A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = BeitImageProcessor if is_vision_available() else None def lowercase (self ) -> Dict: _snake_case = BeitImageProcessingTester(self ) @property def lowercase (self ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def lowercase (self ) -> Optional[int]: _snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) def lowercase (self ) -> Optional[Any]: _snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) _snake_case = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=_lowerCamelCase ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) def lowercase (self ) -> Union[str, Any]: pass def lowercase (self ) -> Dict: # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _snake_case = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def lowercase (self ) -> List[str]: # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _snake_case = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def lowercase (self ) -> str: # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _snake_case = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def lowercase (self ) -> Optional[int]: # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) _snake_case = [] for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input _snake_case = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched _snake_case = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) _snake_case = prepare_semantic_single_inputs() _snake_case = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) _snake_case = prepare_semantic_batch_inputs() _snake_case = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def lowercase (self ) -> Tuple: # Initialize image_processing _snake_case = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 _snake_case = prepare_semantic_single_inputs() _snake_case = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) _snake_case = True _snake_case = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
341
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class _lowerCAmelCase ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None ) -> str: super().__init__() A_ : Optional[Any] = pad_token_id A_ : List[Any] = max_length A_ : str = vocab A_ : Union[str, Any] = merges A_ : List[Any] = BytePairTokenizer(_lowerCamelCase , _lowerCamelCase , sequence_length=_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> int: A_ : Tuple = [""" """.join(_lowerCamelCase ) for m in tokenizer.bpe_ranks.keys()] A_ : Dict = tokenizer.get_vocab() return cls(_lowerCamelCase , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> str: A_ : Tuple = GPTaTokenizer.from_pretrained(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) return cls.from_tokenizer(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase ) -> List[Any]: return cls(**_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ) -> Any: A_ : List[Any] = self.tf_tokenizer(_lowerCamelCase ) A_ : Any = tf.ones_like(_lowerCamelCase ) if self.pad_token_id is not None: # pad the tokens up to max length A_ : List[Any] = max_length if max_length is not None else self.max_length if max_length is not None: A_ , A_ : Tuple = pad_model_inputs( _lowerCamelCase , max_seq_length=_lowerCamelCase , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
344
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class lowerCAmelCase_ ( metaclass=__A ): __lowerCamelCase : Dict = ["torch", "scipy"] def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ) -> str: requires_backends(self , ["torch", "scipy"] ) @classmethod def _snake_case ( cls , *_lowerCAmelCase , **_lowerCAmelCase ) -> Dict: requires_backends(cls , ["torch", "scipy"] ) @classmethod def _snake_case ( cls , *_lowerCAmelCase , **_lowerCAmelCase ) -> Any: requires_backends(cls , ["torch", "scipy"] )
158
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase__ : Optional[int] = {'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : int = ['YolosFeatureExtractor'] UpperCamelCase__ : int = ['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Dict = [ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys UpperCamelCase__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
from manim import * class __A ( __A ): def lowercase__ ( self : List[str] ): lowerCAmelCase : Union[str, Any] = Rectangle(height=0.5 , width=0.5 ) lowerCAmelCase : List[Any] = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) lowerCAmelCase : str = [mem.copy() for i in range(6 )] lowerCAmelCase : Any = [mem.copy() for i in range(6 )] lowerCAmelCase : List[str] = VGroup(*_lowerCamelCase ).arrange(_lowerCamelCase , buff=0 ) lowerCAmelCase : Optional[Any] = VGroup(*_lowerCamelCase ).arrange(_lowerCamelCase , buff=0 ) lowerCAmelCase : Any = VGroup(_lowerCamelCase , _lowerCamelCase ).arrange(_lowerCamelCase , buff=0 ) lowerCAmelCase : int = Text('CPU' , font_size=24 ) lowerCAmelCase : Any = Group(_lowerCamelCase , _lowerCamelCase ).arrange(_lowerCamelCase , buff=0.5 , aligned_edge=_lowerCamelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_lowerCamelCase ) lowerCAmelCase : List[Any] = [mem.copy() for i in range(4 )] lowerCAmelCase : Dict = VGroup(*_lowerCamelCase ).arrange(_lowerCamelCase , buff=0 ) lowerCAmelCase : Dict = Text('GPU' , font_size=24 ) lowerCAmelCase : Optional[Any] = Group(_lowerCamelCase , _lowerCamelCase ).arrange(_lowerCamelCase , buff=0.5 , aligned_edge=_lowerCamelCase ) gpu.move_to([-1, -1, 0] ) self.add(_lowerCamelCase ) lowerCAmelCase : List[str] = [mem.copy() for i in range(6 )] lowerCAmelCase : List[str] = VGroup(*_lowerCamelCase ).arrange(_lowerCamelCase , buff=0 ) lowerCAmelCase : str = Text('Model' , font_size=24 ) lowerCAmelCase : str = Group(_lowerCamelCase , _lowerCamelCase ).arrange(_lowerCamelCase , buff=0.5 , aligned_edge=_lowerCamelCase ) model.move_to([3, -1.0, 0] ) self.add(_lowerCamelCase ) lowerCAmelCase : Optional[int] = [] for i, rect in enumerate(_lowerCamelCase ): rect.set_stroke(_lowerCamelCase ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) lowerCAmelCase : Optional[int] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(_lowerCamelCase , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=_lowerCamelCase ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=_lowerCamelCase , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=_lowerCamelCase , buff=0.0 ) self.add(_lowerCamelCase ) cpu_targs.append(_lowerCamelCase ) lowerCAmelCase : List[Any] = [mem.copy() for i in range(6 )] lowerCAmelCase : Optional[int] = VGroup(*_lowerCamelCase ).arrange(_lowerCamelCase , buff=0 ) lowerCAmelCase : int = Text('Loaded Checkpoint' , font_size=24 ) lowerCAmelCase : str = Group(_lowerCamelCase , _lowerCamelCase ).arrange(_lowerCamelCase , aligned_edge=_lowerCamelCase , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) lowerCAmelCase : List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowerCAmelCase : List[Any] = MarkupText( f"<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_lowerCamelCase , _lowerCamelCase ) lowerCAmelCase : List[Any] = MarkupText( f"<span fgcolor='{BLUE}'>●</span> Checkpoint" , font_size=18 , ) blue_text.next_to(_lowerCamelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) lowerCAmelCase : Any = MarkupText( f"Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_lowerCamelCase ) , Write(_lowerCamelCase ) ) self.play(Write(_lowerCamelCase , run_time=1 ) , Create(_lowerCamelCase , run_time=1 ) ) lowerCAmelCase : List[str] = [] lowerCAmelCase : str = [] for i, rect in enumerate(_lowerCamelCase ): lowerCAmelCase : Union[str, Any] = fill.copy().set_fill(_lowerCamelCase , opacity=0.7 ) target.move_to(_lowerCamelCase ) first_animations.append(GrowFromCenter(_lowerCamelCase , run_time=1 ) ) lowerCAmelCase : Optional[int] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(_lowerCamelCase , run_time=1.5 ) ) self.play(*_lowerCamelCase ) self.play(*_lowerCamelCase ) self.wait()
138
'''simple docstring''' class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Optional[Any] = name A_ : Dict = value A_ : Union[str, Any] = weight def __repr__( self ) -> List[str]: return F"{self.__class__.__name__}({self.name}, {self.value}, {self.weight})" def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.value def UpperCAmelCase_ ( self ) -> List[str]: return self.name def UpperCAmelCase_ ( self ) -> Tuple: return self.weight def UpperCAmelCase_ ( self ) -> Optional[int]: return self.value / self.weight def UpperCAmelCase ( a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Optional[int] = [] for i in range(len(a_ ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def UpperCAmelCase ( a_ , a_ , a_ ) -> List[Any]: """simple docstring""" A_ : Optional[Any] = sorted(a_ , key=a_ , reverse=a_ ) A_ : str = [] A_ , A_ : Dict = 0.0, 0.0 for i in range(len(a_ ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
344
0
import warnings from .state import AcceleratorState, GradientState warnings.filterwarnings('''ignore''', category=UserWarning, module='''torch.optim.lr_scheduler''') class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case = True , snake_case = False ): lowercase = scheduler lowercase = optimizers if isinstance(_lowerCamelCase , (list, tuple) ) else [optimizers] lowercase = split_batches lowercase = step_with_optimizer lowercase = GradientState() def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): if not self.step_with_optimizer: # No link between scheduler and optimizer -> just step self.scheduler.step(*_lowerCamelCase , **_lowerCamelCase ) return # Otherwise, first make sure the optimizer was stepped. if not self.gradient_state.sync_gradients: if self.gradient_state.adjust_scheduler: self.scheduler._step_count += 1 return for opt in self.optimizers: if opt.step_was_skipped: return if self.split_batches: # Split batches -> the training dataloader batch size is not changed so one step per training step self.scheduler.step(*_lowerCamelCase , **_lowerCamelCase ) else: # Otherwise the training dataloader batch size was multiplied by `num_processes`, so we need to do # num_processes steps per training step lowercase = AcceleratorState().num_processes for _ in range(_lowerCamelCase ): # Special case when using OneCycle and `drop_last` was not used if hasattr(self.scheduler , 'total_steps' ): if self.scheduler._step_count <= self.scheduler.total_steps: self.scheduler.step(*_lowerCamelCase , **_lowerCamelCase ) else: self.scheduler.step(*_lowerCamelCase , **_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): return self.scheduler.get_last_lr() def SCREAMING_SNAKE_CASE__ ( self ): return self.scheduler.state_dict() def SCREAMING_SNAKE_CASE__ ( self , snake_case ): self.scheduler.load_state_dict(_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): return self.scheduler.get_lr() def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.scheduler.print_lr(*_lowerCamelCase , **_lowerCamelCase )
195
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCAmelCase ( a_ , a_ ) -> tuple: """simple docstring""" if inductance <= 0: raise ValueError("""Inductance cannot be 0 or negative""" ) elif capacitance <= 0: raise ValueError("""Capacitance cannot be 0 or negative""" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
344
0
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, logging if is_torch_available(): import torch __UpperCAmelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __A ): UpperCAmelCase__ : Optional[Any] = ["pixel_values"] def __init__( self, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = PILImageResampling.BILINEAR, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = 1 / 255, SCREAMING_SNAKE_CASE_ = True, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> None: super().__init__(**_lowerCamelCase ) UpperCamelCase : int = size if size is not None else {"""shortest_edge""": 256} UpperCamelCase : Dict = get_size_dict(_lowerCamelCase, default_to_square=_lowerCamelCase ) UpperCamelCase : int = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} UpperCamelCase : Union[str, Any] = get_size_dict(_lowerCamelCase, param_name='crop_size' ) UpperCamelCase : int = do_resize UpperCamelCase : Any = size UpperCamelCase : Optional[int] = resample UpperCamelCase : List[Any] = do_center_crop UpperCamelCase : List[Any] = crop_size UpperCamelCase : List[Any] = do_rescale UpperCamelCase : Tuple = rescale_factor UpperCamelCase : Optional[Any] = do_normalize UpperCamelCase : Any = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCamelCase : Any = image_std if image_std is not None else IMAGENET_STANDARD_STD def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = PILImageResampling.BICUBIC, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: UpperCamelCase : List[Any] = get_size_dict(_lowerCamelCase, default_to_square=_lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCamelCase : Any = get_resize_output_image_size(_lowerCamelCase, size=size['shortest_edge'], default_to_square=_lowerCamelCase ) return resize(_lowerCamelCase, size=_lowerCamelCase, resample=_lowerCamelCase, data_format=_lowerCamelCase, **_lowerCamelCase ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: UpperCamelCase : Optional[int] = get_size_dict(_lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_lowerCamelCase, size=(size['height'], size['width']), data_format=_lowerCamelCase, **_lowerCamelCase ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_ ) -> np.ndarray: return rescale(_lowerCamelCase, scale=_lowerCamelCase, data_format=_lowerCamelCase, **_lowerCamelCase ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> np.ndarray: return normalize(_lowerCamelCase, mean=_lowerCamelCase, std=_lowerCamelCase, data_format=_lowerCamelCase, **_lowerCamelCase ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST, **SCREAMING_SNAKE_CASE_, ) -> Optional[Any]: UpperCamelCase : Dict = do_resize if do_resize is not None else self.do_resize UpperCamelCase : str = size if size is not None else self.size UpperCamelCase : str = get_size_dict(_lowerCamelCase, default_to_square=_lowerCamelCase ) UpperCamelCase : Optional[Any] = resample if resample is not None else self.resample UpperCamelCase : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase : int = crop_size if crop_size is not None else self.crop_size UpperCamelCase : Optional[Any] = get_size_dict(_lowerCamelCase, param_name='crop_size' ) UpperCamelCase : str = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase : Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase : List[Any] = image_mean if image_mean is not None else self.image_mean UpperCamelCase : Optional[Any] = image_std if image_std is not None else self.image_std UpperCamelCase : List[Any] = make_list_of_images(_lowerCamelCase ) if not valid_images(_lowerCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. UpperCamelCase : List[str] = [to_numpy_array(_lowerCamelCase ) for image in images] if do_resize: UpperCamelCase : Optional[Any] = [self.resize(image=_lowerCamelCase, size=_lowerCamelCase, resample=_lowerCamelCase ) for image in images] if do_center_crop: UpperCamelCase : Optional[int] = [self.center_crop(image=_lowerCamelCase, size=_lowerCamelCase ) for image in images] if do_rescale: UpperCamelCase : Tuple = [self.rescale(image=_lowerCamelCase, scale=_lowerCamelCase ) for image in images] if do_normalize: UpperCamelCase : Dict = [self.normalize(image=_lowerCamelCase, mean=_lowerCamelCase, std=_lowerCamelCase ) for image in images] UpperCamelCase : Optional[int] = [to_channel_dimension_format(_lowerCamelCase, _lowerCamelCase ) for image in images] UpperCamelCase : Any = {"""pixel_values""": images} return BatchFeature(data=_lowerCamelCase, tensor_type=_lowerCamelCase ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = None ) -> List[str]: UpperCamelCase : Optional[int] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError( 'Make sure that you pass in as many target sizes as the batch dimension of the logits' ) if is_torch_tensor(_lowerCamelCase ): UpperCamelCase : Union[str, Any] = target_sizes.numpy() UpperCamelCase : Any = [] for idx in range(len(_lowerCamelCase ) ): UpperCamelCase : Dict = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ), size=target_sizes[idx], mode='bilinear', align_corners=_lowerCamelCase ) UpperCamelCase : Tuple = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowerCamelCase ) else: UpperCamelCase : List[Any] = logits.argmax(dim=1 ) UpperCamelCase : Optional[int] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
119
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class _lowerCAmelCase ( __A ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase ) -> Any: A_ : List[Any] = parent A_ : int = config_class A_ : int = has_text_modality A_ : str = kwargs A_ : int = common_properties def UpperCAmelCase_ ( self ) -> str: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : Optional[int] = ( ["""hidden_size""", """num_attention_heads""", """num_hidden_layers"""] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["""vocab_size"""] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(_lowerCamelCase , _lowerCamelCase ) , msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(_lowerCamelCase ): try: setattr(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(_lowerCamelCase ): try: A_ : List[str] = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = self.config_class(**self.inputs_dict ) A_ : Optional[int] = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : str = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A_ : List[Any] = os.path.join(_lowerCamelCase , """config.json""" ) config_first.to_json_file(_lowerCamelCase ) A_ : Dict = self.config_class.from_json_file(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Any = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(_lowerCamelCase ) A_ : Union[str, Any] = self.config_class.from_pretrained(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : List[Any] = """test""" with tempfile.TemporaryDirectory() as tmpdirname: A_ : Any = os.path.join(_lowerCamelCase , _lowerCamelCase ) config_first.save_pretrained(_lowerCamelCase ) A_ : Any = self.config_class.from_pretrained(_lowerCamelCase , subfolder=_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Tuple = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) A_ : str = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def UpperCAmelCase_ ( self ) -> Optional[Any]: if self.config_class.is_composition: return A_ : Dict = self.config_class() self.parent.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : Any = copy.deepcopy(_lowerCamelCase ) A_ : Tuple = self.config_class(**_lowerCamelCase ) A_ : Optional[Any] = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("""torch_dtype""", config.torch_dtype, torch.floataa) ) elif getattr(_lowerCamelCase , _lowerCamelCase ) != value: wrong_values.append((key, getattr(_lowerCamelCase , _lowerCamelCase ), value) ) if len(_lowerCamelCase ) > 0: A_ : List[Any] = """\n""".join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def UpperCAmelCase_ ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
344
0
"""simple docstring""" from __future__ import annotations from random import choice def lowercase_ ( _UpperCAmelCase ): """simple docstring""" return choice(a_ ) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : int = random_pivot(a_ ) # partition based on pivot # linear time A_ : str = [e for e in lst if e < pivot] A_ : Dict = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(a_ ) == k - 1: return pivot # pivot is in elements bigger than k elif len(a_ ) < k - 1: return kth_number(a_ , k - len(a_ ) - 1 ) # pivot is in elements smaller than k else: return kth_number(a_ , a_ ) if __name__ == "__main__": import doctest doctest.testmod()
167
'''simple docstring''' from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" try: with open(a_ , """rb""" ) as flax_state_f: A_ : Tuple = from_bytes(a_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(a_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"Unable to convert {model_file} to Flax deserializable object. " ) return load_flax_weights_in_pytorch_model(a_ , a_ ) def UpperCAmelCase ( a_ , a_ ) -> Any: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights A_ : List[Any] = flatten_dict(jax.tree_util.tree_map(lambda a_ : x.dtype == jnp.bfloataa , a_ ) ).values() if any(a_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) A_ : str = jax.tree_util.tree_map( lambda a_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , a_ ) A_ : Any = """""" A_ : Optional[int] = flatten_dict(a_ , sep=""".""" ) A_ : List[str] = pt_model.state_dict() # keep track of unexpected & missing keys A_ : Union[str, Any] = [] A_ : Dict = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): A_ : List[Any] = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: A_ : Optional[Any] = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[Any] = jnp.transpose(a_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": A_ : int = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[int] = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": A_ : Any = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(a_ ): A_ : Tuple = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) A_ : Dict = """.""".join(a_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected " F"to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}." ) else: # add weight to pytorch dict A_ : Optional[Any] = np.asarray(a_ ) if not isinstance(a_ , np.ndarray ) else flax_tensor A_ : Tuple = torch.from_numpy(a_ ) # remove from missing keys missing_keys.remove(a_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(a_ ) pt_model.load_state_dict(a_ ) # re-transform missing_keys to list A_ : Dict = list(a_ ) if len(a_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing" F" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(a_ ) > 0: logger.warning( F"Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly" F" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to" """ use it for predictions and inference.""" ) return pt_model
344
0
from collections import deque from .hash_table import HashTable class __snake_case ( __A ): def __init__( self : Tuple , *A_ : Optional[Any] , **A_ : Tuple): super().__init__(*_lowerCamelCase , **_lowerCamelCase) def UpperCAmelCase__ ( self : Optional[int] , A_ : Dict , A_ : int): lowerCAmelCase_ : Optional[Any] = deque([]) if self.values[key] is None else self.values[key] self.values[key].appendleft(_lowerCamelCase) lowerCAmelCase_ : List[Any] = self.values[key] def UpperCAmelCase__ ( self : str): return ( sum(self.charge_factor - len(_lowerCamelCase) for slot in self.values) / self.size_table * self.charge_factor ) def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Optional[Any] , A_ : Tuple=None): if not ( len(self.values[key]) == self.charge_factor and self.values.count(_lowerCamelCase) == 0 ): return key return super()._collision_resolution(_lowerCamelCase , _lowerCamelCase)
103
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=3 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=10 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[1, 1, 2, 1] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="relu" , _lowerCamelCase=3 , _lowerCamelCase=None , ) -> List[str]: A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> Optional[Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : Any = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase_ ( self ) -> int: super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: pass def UpperCAmelCase_ ( self ) -> int: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase , _lowerCamelCase ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) def UpperCAmelCase_ ( self ) -> str: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> int: return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""tf""" ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 )
344
0
from __future__ import annotations def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, ): """simple docstring""" if (stress, tangential_force, area).count(0 ) != 1: raise ValueError('You cannot supply more or less than 2 values' ) elif stress < 0: raise ValueError('Stress cannot be negative' ) elif tangential_force < 0: raise ValueError('Tangential Force cannot be negative' ) elif area < 0: raise ValueError('Area cannot be negative' ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
334
'''simple docstring''' def UpperCAmelCase ( a_ = 1_0_0 ) -> int: """simple docstring""" A_ : Dict = n * (n + 1) * (2 * n + 1) / 6 A_ : Optional[int] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'{solution() = }')
344
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/config.json', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/config.json', 'funnel-transformer/medium-base': 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json', 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/config.json', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json', 'funnel-transformer/xlarge-base': 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json', } class A ( __A ): UpperCamelCase_ : Optional[int] ='''funnel''' UpperCamelCase_ : str ={ '''hidden_size''': '''d_model''', '''num_attention_heads''': '''n_head''', } def __init__(self , lowerCAmelCase=3_0_5_2_2 , lowerCAmelCase=[4, 4, 4] , lowerCAmelCase=None , lowerCAmelCase=2 , lowerCAmelCase=7_6_8 , lowerCAmelCase=1_2 , lowerCAmelCase=6_4 , lowerCAmelCase=3_0_7_2 , lowerCAmelCase="gelu_new" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=0.0 , lowerCAmelCase=0.1 , lowerCAmelCase=None , lowerCAmelCase=1E-9 , lowerCAmelCase="mean" , lowerCAmelCase="relative_shift" , lowerCAmelCase=True , lowerCAmelCase=True , lowerCAmelCase=True , **lowerCAmelCase , ): __lowercase= vocab_size __lowercase= block_sizes __lowercase= [1] * len(_lowerCamelCase ) if block_repeats is None else block_repeats assert len(_lowerCamelCase ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." __lowercase= num_decoder_layers __lowercase= d_model __lowercase= n_head __lowercase= d_head __lowercase= d_inner __lowercase= hidden_act __lowercase= hidden_dropout __lowercase= attention_dropout __lowercase= activation_dropout __lowercase= initializer_range __lowercase= initializer_std __lowercase= layer_norm_eps assert pooling_type in [ "mean", "max", ], f'Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.' __lowercase= pooling_type assert attention_type in [ "relative_shift", "factorized", ], f'Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.' __lowercase= attention_type __lowercase= separate_cls __lowercase= truncate_seq __lowercase= pool_q_only super().__init__(**_lowerCamelCase ) @property def _A (self ): return sum(self.block_sizes ) @num_hidden_layers.setter def _A (self , lowerCAmelCase ): raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.' ) @property def _A (self ): return len(self.block_sizes ) @num_blocks.setter def _A (self , lowerCAmelCase ): raise NotImplementedError('This model does not support the setting of `num_blocks`. Please set `block_sizes`.' )
295
'''simple docstring''' from typing import TYPE_CHECKING from ..utils import _LazyModule UpperCamelCase__ : int = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys UpperCamelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
'''simple docstring''' import pytest lowercase : Optional[int] = '__dummy_dataset1__' lowercase : List[Any] = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def lowerCAmelCase_ ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def lowerCAmelCase_ ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def lowerCAmelCase_ ( snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' A : Union[str, Any] = dataset_loading_script_name A : Dict = tmp_path / """datasets""" / script_name script_dir.mkdir(parents=a_ ) A : str = script_dir / F'{script_name}.py' with open(a_ , '''w''' ) as f: f.write(a_ ) return str(a_ )
3
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCAmelCase ( a_ ) -> Dict[str, torch.Tensor]: """simple docstring""" A_ : List[str] = [] A_ : Dict = [] A_ : List[Any] = [] for rt in rc.restypes: A_ : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) A_ : Union[str, Any] = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) A_ : Tuple = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : Optional[int] = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : List[Any] = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) A_ : Optional[int] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein A_ : Dict = restype_atomaa_to_atomaa[protein_aatype] A_ : Optional[Any] = restype_atomaa_mask[protein_aatype] A_ : Any = residx_atomaa_mask A_ : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back A_ : Tuple = restype_atomaa_to_atomaa[protein_aatype] A_ : Tuple = residx_atomaa_to_atomaa.long() # create the corresponding mask A_ : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): A_ : Optional[Any] = rc.restype_atoa[restype_letter] A_ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: A_ : Any = rc.atom_order[atom_name] A_ : Optional[int] = 1 A_ : Optional[int] = restype_atomaa_mask[protein_aatype] A_ : Dict = residx_atomaa_mask return protein def UpperCAmelCase ( a_ ) -> Dict[str, np.ndarray]: """simple docstring""" A_ : Union[str, Any] = tree_map(lambda a_ : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) A_ : Optional[int] = tensor_tree_map(lambda a_ : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
344
0
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowerCAmelCase ( __A , unittest.TestCase ): UpperCamelCase__ = KandinskyVaaControlnetImgaImgPipeline UpperCamelCase__ = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] UpperCamelCase__ = ['''image_embeds''', '''negative_image_embeds''', '''image''', '''hint'''] UpperCamelCase__ = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] UpperCamelCase__ = False @property def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' return 32 @property def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' return 32 @property def lowerCamelCase__ ( self :int ): '''simple docstring''' return self.time_input_dim @property def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase__ ( self :Any ): '''simple docstring''' return 100 @property def lowerCamelCase__ ( self :str ): '''simple docstring''' torch.manual_seed(0 ) a = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } a = UNetaDConditionModel(**_lowerCamelCase ) return model @property def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def lowerCamelCase__ ( self :int ): '''simple docstring''' torch.manual_seed(0 ) a = VQModel(**self.dummy_movq_kwargs ) return model def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = self.dummy_unet a = self.dummy_movq a = { """num_train_timesteps""": 1000, """beta_schedule""": """linear""", """beta_start""": 0.00085, """beta_end""": 0.012, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } a = DDIMScheduler(**_lowerCamelCase ) a = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def lowerCamelCase__ ( self :Dict , __magic_name__ :Optional[Any] , __magic_name__ :Dict=0 ): '''simple docstring''' a = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) a = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _lowerCamelCase ) # create init_image a = floats_tensor((1, 3, 64, 64) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) a = image.cpu().permute(0 , 2 , 3 , 1 )[0] a = Image.fromarray(np.uinta(_lowerCamelCase ) ).convert("""RGB""" ).resize((256, 256) ) # create hint a = floats_tensor((1, 3, 64, 64) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) if str(_lowerCamelCase ).startswith("""mps""" ): a = torch.manual_seed(_lowerCamelCase ) else: a = torch.Generator(device=_lowerCamelCase ).manual_seed(_lowerCamelCase ) a = { """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = """cpu""" a = self.get_dummy_components() a = self.pipeline_class(**_lowerCamelCase ) a = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) a = pipe(**self.get_dummy_inputs(_lowerCamelCase ) ) a = output.images a = pipe( **self.get_dummy_inputs(_lowerCamelCase ) , return_dict=_lowerCamelCase , )[0] a = image[0, -3:, -3:, -1] a = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) a = np.array( [0.54985034, 0.55509365, 0.52561504, 0.5570494, 0.5593818, 0.5263979, 0.50285643, 0.5069846, 0.51196736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy""" ) a = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) a = init_image.resize((512, 512) ) a = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) a = torch.from_numpy(np.array(_lowerCamelCase ) ).float() / 255.0 a = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) a = """A robot, 4k photo""" a = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCamelCase ) a = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" , torch_dtype=torch.floataa ) a = pipeline.to(_lowerCamelCase ) pipeline.set_progress_bar_config(disable=_lowerCamelCase ) a = torch.Generator(device="""cpu""" ).manual_seed(0 ) a = pipe_prior( _lowerCamelCase , image=_lowerCamelCase , strength=0.85 , generator=_lowerCamelCase , negative_prompt="""""" , ).to_tuple() a = pipeline( image=_lowerCamelCase , image_embeds=_lowerCamelCase , negative_image_embeds=_lowerCamelCase , hint=_lowerCamelCase , generator=_lowerCamelCase , num_inference_steps=100 , height=512 , width=512 , strength=0.5 , output_type="""np""" , ) a = output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(_lowerCamelCase , _lowerCamelCase )
228
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ) -> str: A_ : Optional[int] = parent A_ : Dict = batch_size A_ : List[Any] = image_size A_ : Optional[int] = patch_size A_ : List[str] = num_channels A_ : List[Any] = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_size A_ : str = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Union[str, Any] = intermediate_size A_ : Any = hidden_act A_ : Optional[Any] = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Dict = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : str = scope A_ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) A_ : Tuple = (image_size // patch_size) ** 2 A_ : Union[str, Any] = num_patches + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> int: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[str] = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : int = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A_ : Dict = 1 A_ : Optional[int] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Tuple = self.type_sequence_label_size A_ : Tuple = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : Dict = 1 A_ : Any = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : List[str] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = DeiTModelTester(self ) A_ : str = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: pass def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ , A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[str] = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Union[str, Any] = [*signature.parameters.keys()] A_ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: A_ : int = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase_ ( self ) -> Optional[Any]: if not self.model_tester.is_training: return A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[str] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : List[str] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> int: A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A_ : Any = False A_ : Union[str, Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue A_ : List[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Union[str, Any] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Tuple: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): A_ : Dict = problem_type["""title"""] A_ : List[Any] = problem_type["""num_labels"""] A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: A_ : Tuple = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) A_ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: A_ : List[str] = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Optional[Any]: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) A_ : Optional[int] = self.default_image_processor A_ : str = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Any = model(**_lowerCamelCase ) # verify the logits A_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : List[Any] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : int = self.default_image_processor A_ : List[str] = prepare_img() A_ : List[Any] = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) A_ : Union[str, Any] = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A_ : List[Any] = model(_lowerCamelCase )
344
0
'''simple docstring''' import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = JukeboxTokenizer lowerCAmelCase_ = { "artist": "Zac Brown Band", "genres": "Country", "lyrics": "I met a traveller from an antique land,\n Who said \"Two vast and trunkless legs of stone\n Stand in the desert. . . . Near them, on the sand,\n Half sunk a shattered visage lies, whose frown,\n And wrinkled lip, and sneer of cold command,\n Tell that its sculptor well those passions read\n Which yet survive, stamped on these lifeless things,\n The hand that mocked them, and the heart that fed;\n And on the pedestal, these words appear:\n My name is Ozymandias, King of Kings;\n Look on my Works, ye Mighty, and despair!\n Nothing beside remains. Round the decay\n Of that colossal Wreck, boundless and bare\n The lone and level sands stretch far away\n ", } @require_torch def lowercase (self ) -> List[Any]: import torch _snake_case = JukeboxTokenizer.from_pretrained("""openai/jukebox-1b-lyrics""" ) _snake_case = tokenizer(**self.metas )["""input_ids"""] # fmt: off _snake_case = [ torch.tensor([[ 0, 0, 0, 7169, 507, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowercase (self ) -> int: import torch _snake_case = JukeboxTokenizer.from_pretrained("""openai/jukebox-5b-lyrics""" ) _snake_case = tokenizer(**self.metas )["""input_ids"""] # fmt: off _snake_case = [ torch.tensor([[ 0, 0, 0, 1069, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
341
'''simple docstring''' import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 32 , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = True , _lowerCamelCase = [0.4814_5466, 0.457_8275, 0.4082_1073] , _lowerCamelCase = [0.2686_2954, 0.2613_0258, 0.2757_7711] , _lowerCamelCase = True , _lowerCamelCase=7 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=3 , ) -> Union[str, Any]: A_ : Optional[int] = parent A_ : Union[str, Any] = do_resize A_ : Optional[Any] = size if size is not None else {"""shortest_edge""": 288} A_ : Tuple = size_divisor A_ : List[Any] = do_rescale A_ : Dict = rescale_factor A_ : List[Any] = do_normalize A_ : Dict = do_center_crop A_ : Optional[Any] = image_mean A_ : List[str] = image_std A_ : str = do_pad A_ : Any = batch_size A_ : List[str] = num_channels A_ : List[str] = min_resolution A_ : Union[str, Any] = max_resolution def UpperCAmelCase_ ( self ) -> Any: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase=False ) -> Optional[int]: if not batched: A_ : Union[str, Any] = self.size["""shortest_edge"""] A_ : Dict = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image ): A_ , A_ : Optional[Any] = image.size else: A_ , A_ : int = image.shape[1], image.shape[2] A_ : Optional[int] = size / min(_lowerCamelCase , _lowerCamelCase ) if h < w: A_ , A_ : Optional[Any] = size, scale * w else: A_ , A_ : Dict = scale * h, size A_ : Union[str, Any] = int((1333 / 800) * size ) if max(_lowerCamelCase , _lowerCamelCase ) > max_size: A_ : str = max_size / max(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = newh * scale A_ : Dict = neww * scale A_ , A_ : str = int(newh + 0.5 ), int(neww + 0.5 ) A_ , A_ : Dict = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: A_ : Tuple = [] for image in image_inputs: A_ , A_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) A_ : List[Any] = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[0] )[0] A_ : Tuple = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : int = BridgeTowerImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size_divisor""" ) ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image processor A_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : int = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: # Initialize image processor A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image processor A_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
344
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowerCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=7 , _lowerCAmelCase=3 , _lowerCAmelCase=18 , _lowerCAmelCase=30 , _lowerCAmelCase=400 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , ) -> Tuple: _lowerCAmelCase = size if size is not None else {"""height""": 18, """width""": 18} _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = num_channels _lowerCAmelCase = image_size _lowerCAmelCase = min_resolution _lowerCAmelCase = max_resolution _lowerCAmelCase = do_resize _lowerCAmelCase = size _lowerCAmelCase = apply_ocr def _snake_case ( self ) -> List[Any]: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class lowerCAmelCase_ ( __A ,unittest.TestCase ): __lowerCamelCase : Dict = LayoutLMvaImageProcessor if is_pytesseract_available() else None def _snake_case ( self ) -> Tuple: _lowerCAmelCase = LayoutLMvaImageProcessingTester(self ) @property def _snake_case ( self ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def _snake_case ( self ) -> Tuple: _lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(_lowerCamelCase , "size" ) ) self.assertTrue(hasattr(_lowerCamelCase , "apply_ocr" ) ) def _snake_case ( self ) -> Dict: _lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) _lowerCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def _snake_case ( self ) -> Any: pass def _snake_case ( self ) -> Union[str, Any]: # Initialize image_processing _lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input _lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) self.assertIsInstance(encoding.words , _lowerCamelCase ) self.assertIsInstance(encoding.boxes , _lowerCamelCase ) # Test batched _lowerCAmelCase = image_processing(_lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def _snake_case ( self ) -> List[Any]: # Initialize image_processing _lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input _lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched _lowerCAmelCase = image_processing(_lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def _snake_case ( self ) -> List[str]: # Initialize image_processing _lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input _lowerCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched _lowerCAmelCase = image_processing(_lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def _snake_case ( self ) -> Optional[Any]: # with apply_OCR = True _lowerCAmelCase = LayoutLMvaImageProcessor() from datasets import load_dataset _lowerCAmelCase = load_dataset("hf-internal-testing/fixtures_docvqa" , split="test" ) _lowerCAmelCase = Image.open(ds[0]["file"] ).convert("RGB" ) _lowerCAmelCase = image_processing(_lowerCamelCase , return_tensors="pt" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 _lowerCAmelCase = [["""11:14""", """to""", """11:39""", """a.m""", """11:39""", """to""", """11:44""", """a.m.""", """11:44""", """a.m.""", """to""", """12:25""", """p.m.""", """12:25""", """to""", """12:58""", """p.m.""", """12:58""", """to""", """4:00""", """p.m.""", """2:00""", """to""", """5:00""", """p.m.""", """Coffee""", """Break""", """Coffee""", """will""", """be""", """served""", """for""", """men""", """and""", """women""", """in""", """the""", """lobby""", """adjacent""", """to""", """exhibit""", """area.""", """Please""", """move""", """into""", """exhibit""", """area.""", """(Exhibits""", """Open)""", """TRRF""", """GENERAL""", """SESSION""", """(PART""", """|)""", """Presiding:""", """Lee""", """A.""", """Waller""", """TRRF""", """Vice""", """President""", """“Introductory""", """Remarks”""", """Lee""", """A.""", """Waller,""", """TRRF""", """Vice""", """Presi-""", """dent""", """Individual""", """Interviews""", """with""", """TRRF""", """Public""", """Board""", """Members""", """and""", """Sci-""", """entific""", """Advisory""", """Council""", """Mem-""", """bers""", """Conducted""", """by""", """TRRF""", """Treasurer""", """Philip""", """G.""", """Kuehn""", """to""", """get""", """answers""", """which""", """the""", """public""", """refrigerated""", """warehousing""", """industry""", """is""", """looking""", """for.""", """Plus""", """questions""", """from""", """the""", """floor.""", """Dr.""", """Emil""", """M.""", """Mrak,""", """University""", """of""", """Cal-""", """ifornia,""", """Chairman,""", """TRRF""", """Board;""", """Sam""", """R.""", """Cecil,""", """University""", """of""", """Georgia""", """College""", """of""", """Agriculture;""", """Dr.""", """Stanley""", """Charm,""", """Tufts""", """University""", """School""", """of""", """Medicine;""", """Dr.""", """Robert""", """H.""", """Cotton,""", """ITT""", """Continental""", """Baking""", """Company;""", """Dr.""", """Owen""", """Fennema,""", """University""", """of""", """Wis-""", """consin;""", """Dr.""", """Robert""", """E.""", """Hardenburg,""", """USDA.""", """Questions""", """and""", """Answers""", """Exhibits""", """Open""", """Capt.""", """Jack""", """Stoney""", """Room""", """TRRF""", """Scientific""", """Advisory""", """Council""", """Meeting""", """Ballroom""", """Foyer"""]] # noqa: E231 _lowerCAmelCase = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , _lowerCamelCase ) self.assertListEqual(encoding.boxes , _lowerCamelCase ) # with apply_OCR = False _lowerCAmelCase = LayoutLMvaImageProcessor(apply_ocr=_lowerCamelCase ) _lowerCAmelCase = image_processing(_lowerCamelCase , return_tensors="pt" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
158
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(a_ ): for j in range(a_ ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) , end="""\t""" ) else: print("""INF""" , end="""\t""" ) print() def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" A_ : List[str] = [[float("""inf""" ) for _ in range(a_ )] for _ in range(a_ )] for i in range(a_ ): for j in range(a_ ): A_ : List[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(a_ ): # looping through rows of graph array for i in range(a_ ): # looping through columns of graph array for j in range(a_ ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): A_ : List[str] = dist[i][k] + dist[k][j] _print_dist(a_ , a_ ) return dist, v if __name__ == "__main__": UpperCamelCase__ : Tuple = int(input('Enter number of vertices: ')) UpperCamelCase__ : int = int(input('Enter number of edges: ')) UpperCamelCase__ : Dict = [[float('inf') for i in range(v)] for j in range(v)] for i in range(v): UpperCamelCase__ : Union[str, Any] = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print('\nEdge ', i + 1) UpperCamelCase__ : Union[str, Any] = int(input('Enter source:')) UpperCamelCase__ : int = int(input('Enter destination:')) UpperCamelCase__ : Optional[Any] = float(input('Enter weight:')) UpperCamelCase__ : Any = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
344
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> List[Any]: '''simple docstring''' lowerCAmelCase : Union[str, Any] = 384 if "tiny" in model_name: lowerCAmelCase : Union[str, Any] = [3, 3, 9, 3] lowerCAmelCase : int = [96, 192, 384, 768] if "small" in model_name: lowerCAmelCase : List[str] = [3, 3, 27, 3] lowerCAmelCase : str = [96, 192, 384, 768] if "base" in model_name: lowerCAmelCase : str = [3, 3, 27, 3] lowerCAmelCase : List[Any] = [128, 256, 512, 1_024] lowerCAmelCase : Optional[Any] = 512 if "large" in model_name: lowerCAmelCase : int = [3, 3, 27, 3] lowerCAmelCase : int = [192, 384, 768, 1_536] lowerCAmelCase : Union[str, Any] = 768 if "xlarge" in model_name: lowerCAmelCase : str = [3, 3, 27, 3] lowerCAmelCase : Optional[Any] = [256, 512, 1_024, 2_048] lowerCAmelCase : Union[str, Any] = 1_024 # set label information lowerCAmelCase : Tuple = 150 lowerCAmelCase : List[Any] = """huggingface/label-files""" lowerCAmelCase : Tuple = """ade20k-id2label.json""" lowerCAmelCase : Union[str, Any] = json.load(open(hf_hub_download(a_, a_, repo_type='dataset' ), 'r' ) ) lowerCAmelCase : Optional[Any] = {int(a_ ): v for k, v in idalabel.items()} lowerCAmelCase : List[str] = {v: k for k, v in idalabel.items()} lowerCAmelCase : List[Any] = ConvNextConfig( depths=a_, hidden_sizes=a_, out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) lowerCAmelCase : Any = UperNetConfig( backbone_config=a_, auxiliary_in_channels=a_, num_labels=a_, idalabel=a_, labelaid=a_, ) return config def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> int: '''simple docstring''' lowerCAmelCase : str = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"backbone.stages.{i}.{j}.gamma", f"backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter") ) rename_keys.append((f"backbone.stages.{i}.{j}.depthwise_conv.weight", f"backbone.encoder.stages.{i}.layers.{j}.dwconv.weight") ) rename_keys.append((f"backbone.stages.{i}.{j}.depthwise_conv.bias", f"backbone.encoder.stages.{i}.layers.{j}.dwconv.bias") ) rename_keys.append((f"backbone.stages.{i}.{j}.norm.weight", f"backbone.encoder.stages.{i}.layers.{j}.layernorm.weight") ) rename_keys.append((f"backbone.stages.{i}.{j}.norm.bias", f"backbone.encoder.stages.{i}.layers.{j}.layernorm.bias") ) rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv1.weight", f"backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight") ) rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv1.bias", f"backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias") ) rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv2.weight", f"backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight") ) rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv2.bias", f"backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias") ) if i > 0: rename_keys.append((f"backbone.downsample_layers.{i}.0.weight", f"backbone.encoder.stages.{i}.downsampling_layer.0.weight") ) rename_keys.append((f"backbone.downsample_layers.{i}.0.bias", f"backbone.encoder.stages.{i}.downsampling_layer.0.bias") ) rename_keys.append((f"backbone.downsample_layers.{i}.1.weight", f"backbone.encoder.stages.{i}.downsampling_layer.1.weight") ) rename_keys.append((f"backbone.downsample_layers.{i}.1.bias", f"backbone.encoder.stages.{i}.downsampling_layer.1.bias") ) rename_keys.append((f"backbone.norm{i}.weight", f"backbone.hidden_states_norms.stage{i+1}.weight") ) rename_keys.append((f"backbone.norm{i}.bias", f"backbone.hidden_states_norms.stage{i+1}.bias") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> Optional[int]: '''simple docstring''' lowerCAmelCase : Tuple = dct.pop(a_ ) lowerCAmelCase : Dict = val def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> List[str]: '''simple docstring''' lowerCAmelCase : List[Any] = { """upernet-convnext-tiny""": """https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth""", """upernet-convnext-small""": """https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth""", """upernet-convnext-base""": """https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth""", """upernet-convnext-large""": """https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth""", """upernet-convnext-xlarge""": """https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth""", } lowerCAmelCase : Dict = model_name_to_url[model_name] lowerCAmelCase : str = torch.hub.load_state_dict_from_url(a_, map_location='cpu' )["""state_dict"""] lowerCAmelCase : List[str] = get_upernet_config(a_ ) lowerCAmelCase : Union[str, Any] = UperNetForSemanticSegmentation(a_ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): lowerCAmelCase : Tuple = state_dict.pop(a_ ) if "bn" in key: lowerCAmelCase : List[Any] = key.replace('bn', 'batch_norm' ) lowerCAmelCase : int = val # rename keys lowerCAmelCase : int = create_rename_keys(a_ ) for src, dest in rename_keys: rename_key(a_, a_, a_ ) model.load_state_dict(a_ ) # verify on image lowerCAmelCase : List[str] = """https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg""" lowerCAmelCase : Union[str, Any] = Image.open(requests.get(a_, stream=a_ ).raw ).convert('RGB' ) lowerCAmelCase : Union[str, Any] = SegformerImageProcessor() lowerCAmelCase : List[str] = processor(a_, return_tensors='pt' ).pixel_values with torch.no_grad(): lowerCAmelCase : List[str] = model(a_ ) if model_name == "upernet-convnext-tiny": lowerCAmelCase : Union[str, Any] = torch.tensor( [[-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.7_7_4_6, -8.7_7_4_6, -8.6_1_3_0]] ) elif model_name == "upernet-convnext-small": lowerCAmelCase : Union[str, Any] = torch.tensor( [[-8.8_2_3_6, -8.8_2_3_6, -8.6_7_7_1], [-8.8_2_3_6, -8.8_2_3_6, -8.6_7_7_1], [-8.7_6_3_8, -8.7_6_3_8, -8.6_2_4_0]] ) elif model_name == "upernet-convnext-base": lowerCAmelCase : int = torch.tensor( [[-8.8_5_5_8, -8.8_5_5_8, -8.6_9_0_5], [-8.8_5_5_8, -8.8_5_5_8, -8.6_9_0_5], [-8.7_6_6_9, -8.7_6_6_9, -8.6_0_2_1]] ) elif model_name == "upernet-convnext-large": lowerCAmelCase : str = torch.tensor( [[-8.6_6_6_0, -8.6_6_6_0, -8.6_2_1_0], [-8.6_6_6_0, -8.6_6_6_0, -8.6_2_1_0], [-8.6_3_1_0, -8.6_3_1_0, -8.5_9_6_4]] ) elif model_name == "upernet-convnext-xlarge": lowerCAmelCase : Optional[int] = torch.tensor( [[-8.4_9_8_0, -8.4_9_8_0, -8.3_9_7_7], [-8.4_9_8_0, -8.4_9_8_0, -8.3_9_7_7], [-8.4_3_7_9, -8.4_3_7_9, -8.3_4_1_2]] ) print('Logits:', outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3], a_, atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(a_ ) print(f"Saving processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(a_ ) if push_to_hub: print(f"Pushing model and processor for {model_name} to hub" ) model.push_to_hub(f"openmmlab/{model_name}" ) processor.push_to_hub(f"openmmlab/{model_name}" ) if __name__ == "__main__": __A : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''upernet-convnext-tiny''', type=str, choices=[F'upernet-convnext-{size}' for size in ['''tiny''', '''small''', '''base''', '''large''', '''xlarge''']], help='''Name of the ConvNext UperNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __A : Any = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
138
'''simple docstring''' import datasets from .evaluate import evaluate UpperCamelCase__ : int = '\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n' UpperCamelCase__ : Any = '\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n' UpperCamelCase__ : Optional[Any] = '\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the SQuAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> squad_metric = datasets.load_metric("squad")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: A_ : Optional[Any] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} A_ : List[Any] = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] A_ : int = evaluate(dataset=_lowerCamelCase , predictions=_lowerCamelCase ) return score
344
0
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if not all(x.isalpha() for x in string ): raise ValueError('String must only contain alphabetic characters.' ) lowercase = sorted(string.lower() ) return len(a_ ) == len(set(a_ ) ) if __name__ == "__main__": UpperCAmelCase = input('''Enter a string ''').strip() UpperCAmelCase = is_isogram(input_str) print(F"""{input_str} is {"an" if isogram else "not an"} isogram.""")
195
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ : Any = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Optional[Any] = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] UpperCamelCase__ : List[str] = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] UpperCamelCase__ : str = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): UpperCamelCase__ : List[str] = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCamelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
def UpperCamelCase ( snake_case__ : Optional[Any] , snake_case__ : Optional[Any] ) -> int: while a != 0: UpperCamelCase : Union[str, Any] = b % a, a return b def UpperCamelCase ( snake_case__ : Any , snake_case__ : Tuple ) -> int: if gcd(a_ , a_ ) != 1: UpperCamelCase : Any = F"""mod inverse of {a!r} and {m!r} does not exist""" raise ValueError(a_ ) UpperCamelCase : str = 1, 0, a UpperCamelCase : Tuple = 0, 1, m while va != 0: UpperCamelCase : str = ua // va UpperCamelCase : Optional[Any] = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
119
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''vision-encoder-decoder''' lowerCamelCase = True def __init__( self , **_lowerCamelCase ) -> str: super().__init__(**_lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F"A configuraton of type {self.model_type} cannot be instantiated because " F"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) A_ : Optional[int] = kwargs.pop("""encoder""" ) A_ : List[str] = encoder_config.pop("""model_type""" ) A_ : str = kwargs.pop("""decoder""" ) A_ : Optional[Any] = decoder_config.pop("""model_type""" ) A_ : List[str] = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : str = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : Any = True @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A_ : int = True A_ : List[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = copy.deepcopy(self.__dict__ ) A_ : List[str] = self.encoder.to_dict() A_ : Union[str, Any] = self.decoder.to_dict() A_ : str = self.__class__.model_type return output class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> float: return 1e-4 @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: A_ : Optional[Any] = OrderedDict() A_ : Any = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : Optional[int] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = -1 , _lowerCamelCase = -1 , _lowerCamelCase = False , _lowerCamelCase = None , ) -> Mapping[str, Any]: import torch A_ : Optional[int] = OrderedDict() A_ : List[Any] = super().generate_dummy_inputs( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) A_ , A_ : str = dummy_input["""input_ids"""].shape A_ : Optional[int] = (batch, encoder_sequence, self._config.encoder_hidden_size) A_ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A_ : List[str] = dummy_input.pop("""attention_mask""" ) A_ : Optional[int] = torch.zeros(_lowerCamelCase ) return common_inputs class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> None: pass def UpperCAmelCase_ ( self , _lowerCamelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = "default" ) -> OnnxConfig: A_ : List[Any] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_lowerCamelCase , _lowerCamelCase )
344
0
"""simple docstring""" import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class lowercase ( unittest.TestCase): def a_ ( self : Tuple ): """simple docstring""" A_ : Optional[int] = 10 def a_ ( self : List[str] ): """simple docstring""" A_ : Tuple = [1, 2, 3, 4] A_ : Union[str, Any] = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(_lowerCamelCase , self.block_size , 0 ) , _lowerCamelCase ) def a_ ( self : str ): """simple docstring""" A_ : str = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] A_ : Union[str, Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_lowerCamelCase , self.block_size , 0 ) , _lowerCamelCase ) def a_ ( self : str ): """simple docstring""" A_ : Union[str, Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] A_ : Union[str, Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_lowerCamelCase , self.block_size , 0 ) , _lowerCamelCase ) def a_ ( self : Optional[Any] ): """simple docstring""" A_ : Union[str, Any] = """It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.""" A_ : Dict = process_story(_lowerCamelCase ) self.assertEqual(_lowerCamelCase , [] ) def a_ ( self : int ): """simple docstring""" A_ : Optional[Any] = """""" A_ : Dict = process_story(_lowerCamelCase ) self.assertEqual(_lowerCamelCase , [] ) self.assertEqual(_lowerCamelCase , [] ) def a_ ( self : str ): """simple docstring""" A_ : Dict = ( """It was the year of Our Lord one thousand seven hundred and """ """seventy-five\n\nSpiritual revelations were conceded to England """ """at that favoured period, as at this.\n@highlight\n\nIt was the best of times""" ) A_ : Optional[int] = process_story(_lowerCamelCase ) A_ : Optional[int] = [ """It was the year of Our Lord one thousand seven hundred and seventy-five.""", """Spiritual revelations were conceded to England at that favoured period, as at this.""", ] self.assertEqual(_lowerCamelCase , _lowerCamelCase ) A_ : Union[str, Any] = ["""It was the best of times."""] self.assertEqual(_lowerCamelCase , _lowerCamelCase ) def a_ ( self : List[Any] ): """simple docstring""" A_ : Optional[int] = torch.tensor([1, 2, 3, 4] ) A_ : Union[str, Any] = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(_lowerCamelCase , 0 ).numpy() , expected.numpy() ) def a_ ( self : Optional[Any] ): """simple docstring""" A_ : Tuple = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) A_ : List[str] = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(_lowerCamelCase , 23 ).numpy() , expected.numpy() ) def a_ ( self : int ): """simple docstring""" A_ : Union[str, Any] = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) A_ : Union[str, Any] = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(_lowerCamelCase , 1 ).numpy() , expected.numpy() ) def a_ ( self : List[Any] ): """simple docstring""" A_ : int = 1_01 A_ : List[str] = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 1_01, 5, 6], [1, 1_01, 3, 4, 1_01, 6]] ) A_ : int = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) A_ : Dict = compute_token_type_ids(_lowerCamelCase , _lowerCamelCase ) np.testing.assert_array_equal(_lowerCamelCase , _lowerCamelCase )
167
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCamelCase__ : Any = '\\n Text data.\n Second line of data.' UpperCamelCase__ : List[Any] = 'file' @pytest.fixture(scope="""session""" ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : int = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") A_ : int = bytes(a_ , """utf-8""" ) with zstd.open(a_ , """wb""" ) as f: f.write(a_ ) return path @pytest.fixture def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , a_ ) , """w""" ) as f: f.write(a_ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> Optional[int]: """simple docstring""" A_ : List[str] = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} A_ : Any = input_paths[compression_format] A_ : Tuple = tmp_path / """cache""" A_ : Tuple = DownloadConfig(cache_dir=a_ , extract_compressed_file=a_ ) A_ : Dict = cached_path(a_ , download_config=a_ ) with open(a_ ) as f: A_ : Optional[Any] = f.read() with open(a_ ) as f: A_ : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Union[str, Any] = """custom_cache""" A_ : List[str] = """custom_extracted_dir""" A_ : Optional[Any] = tmp_path / """custom_extracted_path""" if default_extracted: A_ : Any = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , a_ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(a_ ) ) A_ : Union[str, Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) A_ : List[Any] = xz_file A_ : Optional[int] = ( DownloadConfig(extract_compressed_file=a_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=a_ ) ) A_ : Union[str, Any] = cached_path(a_ , download_config=a_ ) assert Path(a_ ).parent.parts[-2:] == expected def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : str = str(Path(a_ ).resolve() ) assert cached_path(a_ ) == text_file # relative path A_ : List[str] = str(Path(a_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(a_ ) == text_file def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : Optional[Any] = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(a_ ): cached_path(a_ ) # relative path A_ : Tuple = """./__missing_file__.txt""" with pytest.raises(a_ ): cached_path(a_ ) def UpperCAmelCase ( a_ ) -> Tuple: """simple docstring""" A_ : Any = get_from_cache(F"tmp://{tmpfs_file}" ) with open(a_ ) as f: A_ : List[str] = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( ) -> List[str]: """simple docstring""" with pytest.raises(a_ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Union[str, Any]: """simple docstring""" A_ : List[str] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): http_get("""https://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): ftp_get("""ftp://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): fsspec_get("""s3://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): fsspec_head("""s3://huggingface.co""" )
344
0
import functools def UpperCamelCase( __UpperCamelCase : Union[str, Any] ,__UpperCamelCase : List[str] ): lowerCAmelCase_ : List[Any] = len(a_ ) lowerCAmelCase_ : List[str] = len(a_ ) @functools.cache def min_distance(__UpperCamelCase : Any ,__UpperCamelCase : Tuple ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa lowerCAmelCase_ : int = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 ,a_ ) ,1 + min_distance(a_ ,indexa + 1 ) ,diff + min_distance(indexa + 1 ,indexa + 1 ) ,) return min_distance(0 ,0 ) if __name__ == "__main__": import doctest doctest.testmod()
103
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) UpperCamelCase__ : int = {'processing_layoutxlm': ['LayoutXLMProcessor']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Tuple = ['LayoutXLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : List[Any] = ['LayoutXLMTokenizerFast'] if TYPE_CHECKING: from .processing_layoutxlm import LayoutXLMProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm import LayoutXLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm_fast import LayoutXLMTokenizerFast else: import sys UpperCamelCase__ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase ={'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =['YolosFeatureExtractor'] _lowerCamelCase =['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
334
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ : Any = logging.get_logger(__name__) UpperCamelCase__ : Optional[int] = { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/config.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/config.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json' ), 'distilbert-base-uncased-finetuned-sst-2-english': ( 'https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json' ), } class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''distilbert''' lowerCamelCase = { '''hidden_size''': '''dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', } def __init__( self , _lowerCamelCase=3_0522 , _lowerCamelCase=512 , _lowerCamelCase=False , _lowerCamelCase=6 , _lowerCamelCase=12 , _lowerCamelCase=768 , _lowerCamelCase=4 * 768 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase="gelu" , _lowerCamelCase=0.02 , _lowerCamelCase=0.1 , _lowerCamelCase=0.2 , _lowerCamelCase=0 , **_lowerCamelCase , ) -> Optional[Any]: A_ : Tuple = vocab_size A_ : List[Any] = max_position_embeddings A_ : int = sinusoidal_pos_embds A_ : int = n_layers A_ : str = n_heads A_ : Optional[int] = dim A_ : int = hidden_dim A_ : Tuple = dropout A_ : List[Any] = attention_dropout A_ : int = activation A_ : Dict = initializer_range A_ : List[Any] = qa_dropout A_ : int = seq_classif_dropout super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": A_ : Union[str, Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A_ : int = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
344
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=__A ) class A ( __A ): UpperCamelCase_ : Dict =field(default='''image-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) UpperCamelCase_ : List[str] =Features({'''image''': Image()} ) UpperCamelCase_ : Optional[Any] =Features({'''labels''': ClassLabel} ) UpperCamelCase_ : List[str] ='''image''' UpperCamelCase_ : Union[str, Any] ='''labels''' def _A (self , lowerCAmelCase ): if self.label_column not in features: raise ValueError(f'Column {self.label_column} is not present in features.' ) if not isinstance(features[self.label_column] , _lowerCamelCase ): raise ValueError(f'Column {self.label_column} is not a ClassLabel.' ) __lowercase= copy.deepcopy(self ) __lowercase= self.label_schema.copy() __lowercase= features[self.label_column] __lowercase= label_schema return task_template @property def _A (self ): return { self.image_column: "image", self.label_column: "labels", }
295
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ : int = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) UpperCamelCase__ : Any = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', f'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', f'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qpos_proj.weight', f'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kpos_proj.weight', f'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.weight', f'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', f'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', f'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kpos_proj.weight', f'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.weight', f'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', f'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', f'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', f'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_qpos_proj.bias', f'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_kpos_proj.bias', f'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.bias', f'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', f'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', f'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_kpos_proj.bias', f'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.bias', f'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', f'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ('transformer.decoder.ref_point_head.layers.0.weight', 'decoder.ref_point_head.layers.0.weight'), ('transformer.decoder.ref_point_head.layers.0.bias', 'decoder.ref_point_head.layers.0.bias'), ('transformer.decoder.ref_point_head.layers.1.weight', 'decoder.ref_point_head.layers.1.weight'), ('transformer.decoder.ref_point_head.layers.1.bias', 'decoder.ref_point_head.layers.1.bias'), ('transformer.decoder.query_scale.layers.0.weight', 'decoder.query_scale.layers.0.weight'), ('transformer.decoder.query_scale.layers.0.bias', 'decoder.query_scale.layers.0.bias'), ('transformer.decoder.query_scale.layers.1.weight', 'decoder.query_scale.layers.1.weight'), ('transformer.decoder.query_scale.layers.1.bias', 'decoder.query_scale.layers.1.bias'), ('transformer.decoder.layers.0.ca_qpos_proj.weight', 'decoder.layers.0.ca_qpos_proj.weight'), ('transformer.decoder.layers.0.ca_qpos_proj.bias', 'decoder.layers.0.ca_qpos_proj.bias'), ] ) def UpperCAmelCase ( a_ , a_ , a_ ) -> Optional[Any]: """simple docstring""" A_ : int = state_dict.pop(a_ ) A_ : Tuple = val def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : Union[str, Any] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: A_ : Optional[int] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) A_ : str = value else: A_ : int = value return new_state_dict def UpperCAmelCase ( a_ , a_=False ) -> Optional[int]: """simple docstring""" A_ : List[Any] = """""" if is_panoptic: A_ : Any = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) A_ : Optional[int] = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) A_ : str = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict A_ : Optional[Any] = in_proj_weight[:2_5_6, :] A_ : Tuple = in_proj_bias[:2_5_6] A_ : Dict = in_proj_weight[2_5_6:5_1_2, :] A_ : int = in_proj_bias[2_5_6:5_1_2] A_ : int = in_proj_weight[-2_5_6:, :] A_ : Optional[int] = in_proj_bias[-2_5_6:] def UpperCAmelCase ( ) -> Dict: """simple docstring""" A_ : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A_ : List[Any] = Image.open(requests.get(a_ , stream=a_ ).raw ) return im @torch.no_grad() def UpperCAmelCase ( a_ , a_ ) -> Dict: """simple docstring""" A_ : int = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: A_ : str = """resnet101""" if "dc5" in model_name: A_ : List[Any] = True A_ : str = """panoptic""" in model_name if is_panoptic: A_ : Dict = 2_5_0 else: A_ : Union[str, Any] = 9_1 A_ : str = """huggingface/label-files""" A_ : Union[str, Any] = """coco-detection-id2label.json""" A_ : Optional[Any] = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) A_ : str = {int(a_ ): v for k, v in idalabel.items()} A_ : Optional[int] = idalabel A_ : Tuple = {v: k for k, v in idalabel.items()} # load image processor A_ : List[Any] = """coco_panoptic""" if is_panoptic else """coco_detection""" A_ : Any = ConditionalDetrImageProcessor(format=a_ ) # prepare image A_ : Tuple = prepare_img() A_ : Any = image_processor(images=a_ , return_tensors="""pt""" ) A_ : Optional[int] = encoding["""pixel_values"""] logger.info(F"Converting model {model_name}..." ) # load original model from torch hub A_ : int = torch.hub.load("""DeppMeng/ConditionalDETR""" , a_ , pretrained=a_ ).eval() A_ : List[Any] = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: A_ : Union[str, Any] = """conditional_detr.""" + src rename_key(a_ , a_ , a_ ) A_ : Any = rename_backbone_keys(a_ ) # query, key and value matrices need special treatment read_in_q_k_v(a_ , is_panoptic=a_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them A_ : List[str] = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): A_ : Dict = state_dict.pop(a_ ) A_ : List[Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: A_ : str = state_dict.pop(a_ ) A_ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: A_ : Optional[int] = state_dict.pop(a_ ) A_ : str = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): A_ : Tuple = state_dict.pop(a_ ) A_ : Dict = val # finally, create HuggingFace model and load state dict A_ : Union[str, Any] = ConditionalDetrForSegmentation(a_ ) if is_panoptic else ConditionalDetrForObjectDetection(a_ ) model.load_state_dict(a_ ) model.eval() model.push_to_hub(repo_id=a_ , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion A_ : str = conditional_detr(a_ ) A_ : str = model(a_ ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(a_ ).mkdir(exist_ok=a_ ) model.save_pretrained(a_ ) image_processor.save_pretrained(a_ ) if __name__ == "__main__": UpperCamelCase__ : int = argparse.ArgumentParser() parser.add_argument( '--model_name', default='conditional_detr_resnet50', type=str, help='Name of the CONDITIONAL_DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) UpperCamelCase__ : Optional[Any] = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
344
0
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
3
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = (UnCLIPScheduler,) def UpperCAmelCase_ ( self , **_lowerCamelCase ) -> List[Any]: A_ : Union[str, Any] = { """num_train_timesteps""": 1000, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**_lowerCamelCase ) return config def UpperCAmelCase_ ( self ) -> List[Any]: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample in [True, False]: self.check_over_configs(clip_sample=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[int]: for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=_lowerCamelCase , prev_timestep=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: A_ : Optional[int] = self.scheduler_classes[0] A_ : Any = self.get_scheduler_config(variance_type="""fixed_small_log""" ) A_ : List[Any] = scheduler_class(**_lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.054_9625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.999_4987 ) ) < 1e-5 def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : List[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config(variance_type="""learned_range""" ) A_ : Dict = scheduler_class(**_lowerCamelCase ) A_ : Dict = 0.5 assert scheduler._get_variance(1 , predicted_variance=_lowerCamelCase ) - -10.171_2790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=_lowerCamelCase ) - -5.799_8052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=_lowerCamelCase ) - -0.001_0011 < 1e-5 def UpperCAmelCase_ ( self ) -> Any: A_ : Optional[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config() A_ : Optional[Any] = scheduler_class(**_lowerCamelCase ) A_ : int = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : str = self.dummy_sample_deter A_ : Optional[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Any = model(_lowerCamelCase , _lowerCamelCase ) # 2. predict previous mean of sample x_t-1 A_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : List[Any] = pred_prev_sample A_ : Any = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Optional[Any] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 252.268_2495 ) < 1e-2 assert abs(result_mean.item() - 0.328_4743 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Dict: A_ : Union[str, Any] = self.scheduler_classes[0] A_ : Dict = self.get_scheduler_config() A_ : Tuple = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(25 ) A_ : List[str] = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : List[Any] = self.dummy_sample_deter A_ : List[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Optional[Any] = model(_lowerCamelCase , _lowerCamelCase ) if i + 1 == timesteps.shape[0]: A_ : List[str] = None else: A_ : Dict = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 A_ : str = scheduler.step( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , prev_timestep=_lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : Optional[Any] = pred_prev_sample A_ : Dict = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : List[str] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 258.204_4983 ) < 1e-2 assert abs(result_mean.item() - 0.336_2038 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> int: pass
344
0
import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __UpperCamelCase : Optional[Any] = logging.get_logger(__name__) __UpperCamelCase : List[str] = { 'b0': efficientnet.EfficientNetBa, 'b1': efficientnet.EfficientNetBa, 'b2': efficientnet.EfficientNetBa, 'b3': efficientnet.EfficientNetBa, 'b4': efficientnet.EfficientNetBa, 'b5': efficientnet.EfficientNetBa, 'b6': efficientnet.EfficientNetBa, 'b7': efficientnet.EfficientNetBa, } __UpperCamelCase : Dict = { 'b0': { 'hidden_dim': 1_280, 'width_coef': 1.0, 'depth_coef': 1.0, 'image_size': 224, 'dropout_rate': 0.2, 'dw_padding': [], }, 'b1': { 'hidden_dim': 1_280, 'width_coef': 1.0, 'depth_coef': 1.1, 'image_size': 240, 'dropout_rate': 0.2, 'dw_padding': [16], }, 'b2': { 'hidden_dim': 1_408, 'width_coef': 1.1, 'depth_coef': 1.2, 'image_size': 260, 'dropout_rate': 0.3, 'dw_padding': [5, 8, 16], }, 'b3': { 'hidden_dim': 1_536, 'width_coef': 1.2, 'depth_coef': 1.4, 'image_size': 300, 'dropout_rate': 0.3, 'dw_padding': [5, 18], }, 'b4': { 'hidden_dim': 1_792, 'width_coef': 1.4, 'depth_coef': 1.8, 'image_size': 380, 'dropout_rate': 0.4, 'dw_padding': [6], }, 'b5': { 'hidden_dim': 2_048, 'width_coef': 1.6, 'depth_coef': 2.2, 'image_size': 456, 'dropout_rate': 0.4, 'dw_padding': [13, 27], }, 'b6': { 'hidden_dim': 2_304, 'width_coef': 1.8, 'depth_coef': 2.6, 'image_size': 528, 'dropout_rate': 0.5, 'dw_padding': [31], }, 'b7': { 'hidden_dim': 2_560, 'width_coef': 2.0, 'depth_coef': 3.1, 'image_size': 600, 'dropout_rate': 0.5, 'dw_padding': [18], }, } def __A ( __lowerCamelCase ) -> int: a = EfficientNetConfig() a = CONFIG_MAP[model_name]["""hidden_dim"""] a = CONFIG_MAP[model_name]["""width_coef"""] a = CONFIG_MAP[model_name]["""depth_coef"""] a = CONFIG_MAP[model_name]["""image_size"""] a = CONFIG_MAP[model_name]["""dropout_rate"""] a = CONFIG_MAP[model_name]["""dw_padding"""] a = """huggingface/label-files""" a = """imagenet-1k-id2label.json""" a = 1000 a = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) a = {int(a_ ): v for k, v in idalabel.items()} a = idalabel a = {v: k for k, v in idalabel.items()} return config def __A ( ) -> Dict: a = """http://images.cocodataset.org/val2017/000000039769.jpg""" a = Image.open(requests.get(a_ , stream=a_ ).raw ) return im def __A ( __lowerCamelCase ) -> Optional[Any]: a = CONFIG_MAP[model_name]["""image_size"""] a = EfficientNetImageProcessor( size={"""height""": size, """width""": size} , image_mean=[0.485, 0.456, 0.406] , image_std=[0.47853944, 0.4732864, 0.47434163] , do_center_crop=a_ , ) return preprocessor def __A ( __lowerCamelCase ) -> Optional[Any]: a = [v.split("""_""" )[0].split("""block""" )[1] for v in original_param_names if v.startswith("""block""" )] a = sorted(set(a_ ) ) a = len(a_ ) a = {b: str(a_ ) for b, i in zip(a_ , range(a_ ) )} a = [] rename_keys.append(("""stem_conv/kernel:0""", """embeddings.convolution.weight""") ) rename_keys.append(("""stem_bn/gamma:0""", """embeddings.batchnorm.weight""") ) rename_keys.append(("""stem_bn/beta:0""", """embeddings.batchnorm.bias""") ) rename_keys.append(("""stem_bn/moving_mean:0""", """embeddings.batchnorm.running_mean""") ) rename_keys.append(("""stem_bn/moving_variance:0""", """embeddings.batchnorm.running_var""") ) for b in block_names: a = block_name_mapping[b] rename_keys.append((f'block{b}_expand_conv/kernel:0', f'encoder.blocks.{hf_b}.expansion.expand_conv.weight') ) rename_keys.append((f'block{b}_expand_bn/gamma:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.weight') ) rename_keys.append((f'block{b}_expand_bn/beta:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.bias') ) rename_keys.append( (f'block{b}_expand_bn/moving_mean:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_mean') ) rename_keys.append( (f'block{b}_expand_bn/moving_variance:0', f'encoder.blocks.{hf_b}.expansion.expand_bn.running_var') ) rename_keys.append( (f'block{b}_dwconv/depthwise_kernel:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight') ) rename_keys.append((f'block{b}_bn/gamma:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight') ) rename_keys.append((f'block{b}_bn/beta:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias') ) rename_keys.append( (f'block{b}_bn/moving_mean:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean') ) rename_keys.append( (f'block{b}_bn/moving_variance:0', f'encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var') ) rename_keys.append((f'block{b}_se_reduce/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.weight') ) rename_keys.append((f'block{b}_se_reduce/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.reduce.bias') ) rename_keys.append((f'block{b}_se_expand/kernel:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.weight') ) rename_keys.append((f'block{b}_se_expand/bias:0', f'encoder.blocks.{hf_b}.squeeze_excite.expand.bias') ) rename_keys.append( (f'block{b}_project_conv/kernel:0', f'encoder.blocks.{hf_b}.projection.project_conv.weight') ) rename_keys.append((f'block{b}_project_bn/gamma:0', f'encoder.blocks.{hf_b}.projection.project_bn.weight') ) rename_keys.append((f'block{b}_project_bn/beta:0', f'encoder.blocks.{hf_b}.projection.project_bn.bias') ) rename_keys.append( (f'block{b}_project_bn/moving_mean:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_mean') ) rename_keys.append( (f'block{b}_project_bn/moving_variance:0', f'encoder.blocks.{hf_b}.projection.project_bn.running_var') ) rename_keys.append(("""top_conv/kernel:0""", """encoder.top_conv.weight""") ) rename_keys.append(("""top_bn/gamma:0""", """encoder.top_bn.weight""") ) rename_keys.append(("""top_bn/beta:0""", """encoder.top_bn.bias""") ) rename_keys.append(("""top_bn/moving_mean:0""", """encoder.top_bn.running_mean""") ) rename_keys.append(("""top_bn/moving_variance:0""", """encoder.top_bn.running_var""") ) a = {} for item in rename_keys: if item[0] in original_param_names: a = """efficientnet.""" + item[1] a = """classifier.weight""" a = """classifier.bias""" return key_mapping def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Union[str, Any]: for key, value in tf_params.items(): if "normalization" in key: continue a = key_mapping[key] if "_conv" in key and "kernel" in key: a = torch.from_numpy(a_ ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: a = torch.from_numpy(a_ ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: a = torch.from_numpy(np.transpose(a_ ) ) else: a = torch.from_numpy(a_ ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(a_ ) @torch.no_grad() def __A ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Optional[Any]: a = model_classes[model_name]( include_top=a_ , weights="""imagenet""" , input_tensor=a_ , input_shape=a_ , pooling=a_ , classes=1000 , classifier_activation="""softmax""" , ) a = original_model.trainable_variables a = original_model.non_trainable_variables a = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: a = param.numpy() a = list(tf_params.keys() ) # Load HuggingFace model a = get_efficientnet_config(a_ ) a = EfficientNetForImageClassification(a_ ).eval() a = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("""Converting parameters...""" ) a = rename_keys(a_ ) replace_params(a_ , a_ , a_ ) # Initialize preprocessor and preprocess input image a = convert_image_processor(a_ ) a = preprocessor(images=prepare_img() , return_tensors="""pt""" ) # HF model inference hf_model.eval() with torch.no_grad(): a = hf_model(**a_ ) a = outputs.logits.detach().numpy() # Original model inference a = False a = CONFIG_MAP[model_name]["""image_size"""] a = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) a = image.img_to_array(a_ ) a = np.expand_dims(a_ , axis=0 ) a = original_model.predict(a_ ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(a_ , a_ , atol=1E-3 ), "The predicted logits are not the same." print("""Model outputs match!""" ) if save_model: # Create folder to save model if not os.path.isdir(a_ ): os.mkdir(a_ ) # Save converted model and image processor hf_model.save_pretrained(a_ ) preprocessor.save_pretrained(a_ ) if push_to_hub: # Push model and image processor to hub print(f'Pushing converted {model_name} to the hub...' ) a = f'efficientnet-{model_name}' preprocessor.push_to_hub(a_ ) hf_model.push_to_hub(a_ ) if __name__ == "__main__": __UpperCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="b0", type=str, help="Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].", ) parser.add_argument( "--pytorch_dump_folder_path", default="hf_model", type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument("--save_model", action="store_true", help="Save model to local") parser.add_argument("--push_to_hub", action="store_true", help="Push model and image processor to the hub") __UpperCamelCase : Tuple = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
228
'''simple docstring''' import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=7 , _lowerCamelCase=3 , _lowerCamelCase=18 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=False , ) -> Optional[int]: A_ : Union[str, Any] = size if size is not None else {"""height""": 20, """width""": 20} A_ : Tuple = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} A_ : Optional[Any] = parent A_ : Optional[int] = batch_size A_ : Union[str, Any] = num_channels A_ : str = image_size A_ : Tuple = min_resolution A_ : Dict = max_resolution A_ : str = do_resize A_ : Tuple = size A_ : int = do_center_crop A_ : Dict = crop_size A_ : Tuple = do_normalize A_ : List[str] = image_mean A_ : Optional[Any] = image_std A_ : Any = do_reduce_labels def UpperCAmelCase_ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def UpperCAmelCase ( ) -> List[str]: """simple docstring""" A_ : Any = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(dataset[0]["""file"""] ) A_ : Dict = Image.open(dataset[1]["""file"""] ) return image, map def UpperCAmelCase ( ) -> Optional[int]: """simple docstring""" A_ : Tuple = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(ds[0]["""file"""] ) A_ : List[Any] = Image.open(ds[1]["""file"""] ) A_ : Any = Image.open(ds[2]["""file"""] ) A_ : str = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BeitImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : List[Any] = BeitImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) A_ : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=_lowerCamelCase ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> Dict: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> str: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : Union[str, Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> Optional[int]: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) A_ : Optional[int] = [] for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) A_ , A_ : List[Any] = prepare_semantic_single_inputs() A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) A_ , A_ : str = prepare_semantic_batch_inputs() A_ : Any = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image_processing A_ : Any = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 A_ , A_ : Tuple = prepare_semantic_single_inputs() A_ : str = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) A_ : str = True A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
344
0
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class _lowerCAmelCase ( __A ): '''simple docstring''' lowerCAmelCase_ = "mctct" def __init__(self , UpperCAmelCase=8065 , UpperCAmelCase=1536 , UpperCAmelCase=36 , UpperCAmelCase=6144 , UpperCAmelCase=4 , UpperCAmelCase=384 , UpperCAmelCase=920 , UpperCAmelCase=1e-5 , UpperCAmelCase=0.3 , UpperCAmelCase="relu" , UpperCAmelCase=0.02 , UpperCAmelCase=0.3 , UpperCAmelCase=0.3 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=2 , UpperCAmelCase=1 , UpperCAmelCase=0.3 , UpperCAmelCase=1 , UpperCAmelCase=(7,) , UpperCAmelCase=(3,) , UpperCAmelCase=80 , UpperCAmelCase=1 , UpperCAmelCase=None , UpperCAmelCase="sum" , UpperCAmelCase=False , **UpperCAmelCase , ) -> Optional[int]: super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase ) _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = intermediate_size _snake_case = num_attention_heads _snake_case = attention_head_dim _snake_case = max_position_embeddings _snake_case = layer_norm_eps _snake_case = layerdrop _snake_case = hidden_act _snake_case = initializer_range _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = pad_token_id _snake_case = bos_token_id _snake_case = eos_token_id _snake_case = conv_glu_dim _snake_case = conv_dropout _snake_case = num_conv_layers _snake_case = input_feat_per_channel _snake_case = input_channels _snake_case = conv_channels _snake_case = ctc_loss_reduction _snake_case = ctc_zero_infinity # prevents config testing fail with exporting to json _snake_case = list(_lowerCamelCase ) _snake_case = list(_lowerCamelCase ) if len(self.conv_kernel ) != self.num_conv_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.conv_kernel)` == `config.num_conv_layers` """ f"""but is `len(config.conv_kernel) = {len(self.conv_kernel )}`, """ f"""`config.num_conv_layers = {self.num_conv_layers}`.""" )
341
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class _lowerCAmelCase ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None ) -> str: super().__init__() A_ : Optional[Any] = pad_token_id A_ : List[Any] = max_length A_ : str = vocab A_ : Union[str, Any] = merges A_ : List[Any] = BytePairTokenizer(_lowerCamelCase , _lowerCamelCase , sequence_length=_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> int: A_ : Tuple = [""" """.join(_lowerCamelCase ) for m in tokenizer.bpe_ranks.keys()] A_ : Dict = tokenizer.get_vocab() return cls(_lowerCamelCase , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> str: A_ : Tuple = GPTaTokenizer.from_pretrained(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) return cls.from_tokenizer(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase ) -> List[Any]: return cls(**_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ) -> Any: A_ : List[Any] = self.tf_tokenizer(_lowerCamelCase ) A_ : Any = tf.ones_like(_lowerCamelCase ) if self.pad_token_id is not None: # pad the tokens up to max length A_ : List[Any] = max_length if max_length is not None else self.max_length if max_length is not None: A_ , A_ : Tuple = pad_model_inputs( _lowerCamelCase , max_seq_length=_lowerCamelCase , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
344
0
'''simple docstring''' from __future__ import annotations import typing from collections import Counter def __a(SCREAMING_SNAKE_CASE_ : Dict ): '''simple docstring''' _lowerCAmelCase = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(a_ , max_perimeter + 1 ): _lowerCAmelCase = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(a_ ): _lowerCAmelCase = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def __a(SCREAMING_SNAKE_CASE_ : List[Any] = 1000 ): '''simple docstring''' _lowerCAmelCase = pythagorean_triple(a_ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(f'''Perimeter {solution()} has maximum solutions''')
158
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase__ : Optional[int] = {'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : int = ['YolosFeatureExtractor'] UpperCamelCase__ : int = ['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Dict = [ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys UpperCamelCase__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
import argparse import os import re __A : int = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict __A : Optional[int] = re.compile(R'''[A-Z_]+_MAPPING(\s+|_[A-Z_]+\s+)=\s+OrderedDict''') # re pattern that matches identifiers in mappings __A : Optional[int] = re.compile(R'''\s*\(\s*"(\S[^"]+)"''') def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase = False ) -> List[str]: '''simple docstring''' with open(a_, 'r', encoding='utf-8' ) as f: lowerCAmelCase : Optional[int] = f.read() lowerCAmelCase : List[Any] = content.split('\n' ) lowerCAmelCase : List[Any] = [] lowerCAmelCase : Optional[int] = 0 while line_idx < len(a_ ): if _re_intro_mapping.search(lines[line_idx] ) is not None: lowerCAmelCase : Optional[int] = len(re.search(r'^(\s*)\S', lines[line_idx] ).groups()[0] ) + 8 # Start of a new mapping! while not lines[line_idx].startswith(' ' * indent + '(' ): new_lines.append(lines[line_idx] ) line_idx += 1 lowerCAmelCase : Any = [] while lines[line_idx].strip() != "]": # Blocks either fit in one line or not if lines[line_idx].strip() == "(": lowerCAmelCase : List[str] = line_idx while not lines[line_idx].startswith(' ' * indent + ')' ): line_idx += 1 blocks.append('\n'.join(lines[start_idx : line_idx + 1] ) ) else: blocks.append(lines[line_idx] ) line_idx += 1 # Sort blocks by their identifiers lowerCAmelCase : Optional[int] = sorted(a_, key=lambda _UpperCAmelCase : _re_identifier.search(a_ ).groups()[0] ) new_lines += blocks else: new_lines.append(lines[line_idx] ) line_idx += 1 if overwrite: with open(a_, 'w', encoding='utf-8' ) as f: f.write('\n'.join(a_ ) ) elif "\n".join(a_ ) != content: return True def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase = False ) -> List[Any]: '''simple docstring''' lowerCAmelCase : List[str] = [os.path.join(a_, a_ ) for f in os.listdir(a_ ) if f.endswith('.py' )] lowerCAmelCase : Union[str, Any] = [sort_auto_mapping(a_, overwrite=a_ ) for fname in fnames] if not overwrite and any(a_ ): lowerCAmelCase : Tuple = [f for f, d in zip(a_, a_ ) if d] raise ValueError( f"The following files have auto mappings that need sorting: {', '.join(a_ )}. Run `make style` to fix" ' this.' ) if __name__ == "__main__": __A : int = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') __A : List[Any] = parser.parse_args() sort_all_auto_mappings(not args.check_only)
138
'''simple docstring''' class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Optional[Any] = name A_ : Dict = value A_ : Union[str, Any] = weight def __repr__( self ) -> List[str]: return F"{self.__class__.__name__}({self.name}, {self.value}, {self.weight})" def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.value def UpperCAmelCase_ ( self ) -> List[str]: return self.name def UpperCAmelCase_ ( self ) -> Tuple: return self.weight def UpperCAmelCase_ ( self ) -> Optional[int]: return self.value / self.weight def UpperCAmelCase ( a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Optional[int] = [] for i in range(len(a_ ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def UpperCAmelCase ( a_ , a_ , a_ ) -> List[Any]: """simple docstring""" A_ : Optional[Any] = sorted(a_ , key=a_ , reverse=a_ ) A_ : str = [] A_ , A_ : Dict = 0.0, 0.0 for i in range(len(a_ ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
344
0
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCAmelCase = logging.get_logger(__name__) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): try: with open(a_ , 'rb' ) as flax_state_f: lowercase = from_bytes(a_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(a_ ) as f: if f.read().startswith('version' ): raise OSError( 'You seem to have cloned a repository without having git-lfs installed. Please' ' install git-lfs and run `git lfs install` followed by `git lfs pull` in the' ' folder you cloned.' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F'''Unable to convert {model_file} to Flax deserializable object. ''' ) return load_flax_weights_in_pytorch_model(a_ , a_ ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): try: import torch # noqa: F401 except ImportError: logger.error( 'Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights lowercase = flatten_dict(jax.tree_util.tree_map(lambda __SCREAMING_SNAKE_CASE : x.dtype == jnp.bfloataa , a_ ) ).values() if any(a_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) lowercase = jax.tree_util.tree_map( lambda __SCREAMING_SNAKE_CASE : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , a_ ) lowercase = """""" lowercase = flatten_dict(a_ , sep='.' ) lowercase = pt_model.state_dict() # keep track of unexpected & missing keys lowercase = [] lowercase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowercase = flax_key_tuple.split('.' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: lowercase = flax_key_tuple_array[:-1] + ["""weight"""] lowercase = jnp.transpose(a_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": lowercase = flax_key_tuple_array[:-1] + ["""weight"""] lowercase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": lowercase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(a_ ): lowercase = ( flax_key_tuple_string.replace('_0' , '.0' ) .replace('_1' , '.1' ) .replace('_2' , '.2' ) .replace('_3' , '.3' ) .replace('_4' , '.4' ) .replace('_5' , '.5' ) .replace('_6' , '.6' ) .replace('_7' , '.7' ) .replace('_8' , '.8' ) .replace('_9' , '.9' ) ) lowercase = """.""".join(a_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'''Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ''' F'''to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) else: # add weight to pytorch dict lowercase = np.asarray(a_ ) if not isinstance(a_ , np.ndarray ) else flax_tensor lowercase = torch.from_numpy(a_ ) # remove from missing keys missing_keys.remove(a_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(a_ ) pt_model.load_state_dict(a_ ) # re-transform missing_keys to list lowercase = list(a_ ) if len(a_ ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' F''' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing''' F''' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture''' ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' F''' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect''' ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) if len(a_ ) > 0: logger.warning( F'''Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly''' F''' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to''' ' use it for predictions and inference.' ) return pt_model
195
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCAmelCase ( a_ , a_ ) -> tuple: """simple docstring""" if inductance <= 0: raise ValueError("""Inductance cannot be 0 or negative""" ) elif capacitance <= 0: raise ValueError("""Capacitance cannot be 0 or negative""" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
344
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __UpperCAmelCase = { 'configuration_altclip': [ 'ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'AltCLIPConfig', 'AltCLIPTextConfig', 'AltCLIPVisionConfig', ], 'processing_altclip': ['AltCLIPProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'AltCLIPPreTrainedModel', 'AltCLIPModel', 'AltCLIPTextModel', 'AltCLIPVisionModel', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
119
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class _lowerCAmelCase ( __A ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase ) -> Any: A_ : List[Any] = parent A_ : int = config_class A_ : int = has_text_modality A_ : str = kwargs A_ : int = common_properties def UpperCAmelCase_ ( self ) -> str: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : Optional[int] = ( ["""hidden_size""", """num_attention_heads""", """num_hidden_layers"""] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["""vocab_size"""] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(_lowerCamelCase , _lowerCamelCase ) , msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(_lowerCamelCase ): try: setattr(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(_lowerCamelCase ): try: A_ : List[str] = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = self.config_class(**self.inputs_dict ) A_ : Optional[int] = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : str = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A_ : List[Any] = os.path.join(_lowerCamelCase , """config.json""" ) config_first.to_json_file(_lowerCamelCase ) A_ : Dict = self.config_class.from_json_file(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Any = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(_lowerCamelCase ) A_ : Union[str, Any] = self.config_class.from_pretrained(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : List[Any] = """test""" with tempfile.TemporaryDirectory() as tmpdirname: A_ : Any = os.path.join(_lowerCamelCase , _lowerCamelCase ) config_first.save_pretrained(_lowerCamelCase ) A_ : Any = self.config_class.from_pretrained(_lowerCamelCase , subfolder=_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Tuple = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) A_ : str = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def UpperCAmelCase_ ( self ) -> Optional[Any]: if self.config_class.is_composition: return A_ : Dict = self.config_class() self.parent.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : Any = copy.deepcopy(_lowerCamelCase ) A_ : Tuple = self.config_class(**_lowerCamelCase ) A_ : Optional[Any] = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("""torch_dtype""", config.torch_dtype, torch.floataa) ) elif getattr(_lowerCamelCase , _lowerCamelCase ) != value: wrong_values.append((key, getattr(_lowerCamelCase , _lowerCamelCase ), value) ) if len(_lowerCamelCase ) > 0: A_ : List[Any] = """\n""".join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def UpperCAmelCase_ ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
344
0
"""simple docstring""" from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowerCamelCase : Tuple = 10 def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" for i in range(a_ , a_ ): if array[i] == target: return i return -1 def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : Optional[Any] = 0 A_ : Optional[int] = len(a_ ) while left <= right: if right - left < precision: return lin_search(a_ , a_ , a_ , a_ ) A_ : str = (left + right) // 3 + 1 A_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: A_ : Optional[int] = one_third - 1 elif array[two_third] < target: A_ : Optional[int] = two_third + 1 else: A_ : Union[str, Any] = one_third + 1 A_ : List[Any] = two_third - 1 else: return -1 def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" if left < right: if right - left < precision: return lin_search(a_ , a_ , a_ , a_ ) A_ : Any = (left + right) // 3 + 1 A_ : Dict = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(a_ , one_third - 1 , a_ , a_ ) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , a_ , a_ , a_ ) else: return rec_ternary_search(one_third + 1 , two_third - 1 , a_ , a_ ) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowerCamelCase : Any = input('Enter numbers separated by comma:\n').strip() _lowerCamelCase : Optional[Any] = [int(item.strip()) for item in user_input.split(',')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowerCamelCase : Any = int(input('Enter the number to be found in the list:\n').strip()) _lowerCamelCase : Optional[int] = ite_ternary_search(collection, target) _lowerCamelCase : Any = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f'Iterative search: {target} found at positions: {resulta}') print(f'Recursive search: {target} found at positions: {resulta}') else: print('Not found')
167
'''simple docstring''' from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" try: with open(a_ , """rb""" ) as flax_state_f: A_ : Tuple = from_bytes(a_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(a_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"Unable to convert {model_file} to Flax deserializable object. " ) return load_flax_weights_in_pytorch_model(a_ , a_ ) def UpperCAmelCase ( a_ , a_ ) -> Any: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights A_ : List[Any] = flatten_dict(jax.tree_util.tree_map(lambda a_ : x.dtype == jnp.bfloataa , a_ ) ).values() if any(a_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) A_ : str = jax.tree_util.tree_map( lambda a_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , a_ ) A_ : Any = """""" A_ : Optional[int] = flatten_dict(a_ , sep=""".""" ) A_ : List[str] = pt_model.state_dict() # keep track of unexpected & missing keys A_ : Union[str, Any] = [] A_ : Dict = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): A_ : List[Any] = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: A_ : Optional[Any] = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[Any] = jnp.transpose(a_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": A_ : int = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[int] = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": A_ : Any = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(a_ ): A_ : Tuple = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) A_ : Dict = """.""".join(a_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected " F"to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}." ) else: # add weight to pytorch dict A_ : Optional[Any] = np.asarray(a_ ) if not isinstance(a_ , np.ndarray ) else flax_tensor A_ : Tuple = torch.from_numpy(a_ ) # remove from missing keys missing_keys.remove(a_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(a_ ) pt_model.load_state_dict(a_ ) # re-transform missing_keys to list A_ : Dict = list(a_ ) if len(a_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing" F" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(a_ ) > 0: logger.warning( F"Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly" F" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to" """ use it for predictions and inference.""" ) return pt_model
344
0
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class __snake_case ( __A ): _a = (DPMSolverSinglestepScheduler,) _a = (('''num_inference_steps''', 25),) def UpperCAmelCase__ ( self : str , **A_ : List[str]): lowerCAmelCase_ : Union[str, Any] = { """num_train_timesteps""": 1_0_0_0, """beta_start""": 0.0001, """beta_end""": 0.02, """beta_schedule""": """linear""", """solver_order""": 2, """prediction_type""": """epsilon""", """thresholding""": False, """sample_max_value""": 1.0, """algorithm_type""": """dpmsolver++""", """solver_type""": """midpoint""", """lambda_min_clipped""": -float('''inf'''), """variance_type""": None, } config.update(**_lowerCamelCase) return config def UpperCAmelCase__ ( self : Optional[Any] , A_ : Optional[Any]=0 , **A_ : List[str]): lowerCAmelCase_ : List[str] = dict(self.forward_default_kwargs) lowerCAmelCase_ : int = kwargs.pop('''num_inference_steps''' , _lowerCamelCase) lowerCAmelCase_ : List[str] = self.dummy_sample lowerCAmelCase_ : str = 0.1 * sample lowerCAmelCase_ : Dict = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCAmelCase_ : Tuple = self.get_scheduler_config(**_lowerCamelCase) lowerCAmelCase_ : List[Any] = scheduler_class(**_lowerCamelCase) scheduler.set_timesteps(_lowerCamelCase) # copy over dummy past residuals lowerCAmelCase_ : List[str] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_lowerCamelCase) lowerCAmelCase_ : List[str] = scheduler_class.from_pretrained(_lowerCamelCase) new_scheduler.set_timesteps(_lowerCamelCase) # copy over dummy past residuals lowerCAmelCase_ : Optional[Any] = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCAmelCase_ : Optional[int] = sample, sample for t in range(_lowerCamelCase , time_step + scheduler.config.solver_order + 1): lowerCAmelCase_ : List[Any] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase).prev_sample lowerCAmelCase_ : List[str] = new_scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def UpperCAmelCase__ ( self : Union[str, Any]): pass def UpperCAmelCase__ ( self : Union[str, Any] , A_ : List[Any]=0 , **A_ : Dict): lowerCAmelCase_ : Optional[int] = dict(self.forward_default_kwargs) lowerCAmelCase_ : Union[str, Any] = kwargs.pop('''num_inference_steps''' , _lowerCamelCase) lowerCAmelCase_ : Tuple = self.dummy_sample lowerCAmelCase_ : List[Any] = 0.1 * sample lowerCAmelCase_ : int = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCAmelCase_ : Union[str, Any] = self.get_scheduler_config() lowerCAmelCase_ : Optional[Any] = scheduler_class(**_lowerCamelCase) scheduler.set_timesteps(_lowerCamelCase) # copy over dummy past residuals (must be after setting timesteps) lowerCAmelCase_ : Tuple = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_lowerCamelCase) lowerCAmelCase_ : int = scheduler_class.from_pretrained(_lowerCamelCase) # copy over dummy past residuals new_scheduler.set_timesteps(_lowerCamelCase) # copy over dummy past residual (must be after setting timesteps) lowerCAmelCase_ : List[Any] = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCAmelCase_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase).prev_sample lowerCAmelCase_ : Optional[int] = new_scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def UpperCAmelCase__ ( self : Tuple , A_ : int=None , **A_ : int): if scheduler is None: lowerCAmelCase_ : List[Any] = self.scheduler_classes[0] lowerCAmelCase_ : Dict = self.get_scheduler_config(**_lowerCamelCase) lowerCAmelCase_ : Any = scheduler_class(**_lowerCamelCase) lowerCAmelCase_ : Optional[int] = self.scheduler_classes[0] lowerCAmelCase_ : Tuple = self.get_scheduler_config(**_lowerCamelCase) lowerCAmelCase_ : List[Any] = scheduler_class(**_lowerCamelCase) lowerCAmelCase_ : Tuple = 1_0 lowerCAmelCase_ : Tuple = self.dummy_model() lowerCAmelCase_ : str = self.dummy_sample_deter scheduler.set_timesteps(_lowerCamelCase) for i, t in enumerate(scheduler.timesteps): lowerCAmelCase_ : List[Any] = model(_lowerCamelCase , _lowerCamelCase) lowerCAmelCase_ : int = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase).prev_sample return sample def UpperCAmelCase__ ( self : Union[str, Any]): lowerCAmelCase_ : List[str] = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) lowerCAmelCase_ : List[Any] = 5_0 lowerCAmelCase_ : int = self.dummy_model() lowerCAmelCase_ : Any = self.dummy_sample_deter scheduler.set_timesteps(_lowerCamelCase) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): lowerCAmelCase_ : Union[str, Any] = model(_lowerCamelCase , _lowerCamelCase) lowerCAmelCase_ : int = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase).prev_sample lowerCAmelCase_ : Optional[int] = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.2574) < 1e-3 def UpperCAmelCase__ ( self : Tuple): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=_lowerCamelCase) def UpperCAmelCase__ ( self : Any): # make sure that iterating over schedulers with same config names gives same results # for defaults lowerCAmelCase_ : str = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) lowerCAmelCase_ : Dict = self.full_loop(scheduler=_lowerCamelCase) lowerCAmelCase_ : str = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.2791) < 1e-3 lowerCAmelCase_ : Any = DEISMultistepScheduler.from_config(scheduler.config) lowerCAmelCase_ : Union[str, Any] = DPMSolverMultistepScheduler.from_config(scheduler.config) lowerCAmelCase_ : Union[str, Any] = UniPCMultistepScheduler.from_config(scheduler.config) lowerCAmelCase_ : Optional[Any] = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowerCAmelCase_ : int = self.full_loop(scheduler=_lowerCamelCase) lowerCAmelCase_ : Optional[Any] = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.2791) < 1e-3 def UpperCAmelCase__ ( self : Union[str, Any]): self.check_over_configs(thresholding=_lowerCamelCase) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=_lowerCamelCase , prediction_type=_lowerCamelCase , sample_max_value=_lowerCamelCase , algorithm_type='''dpmsolver++''' , solver_order=_lowerCamelCase , solver_type=_lowerCamelCase , ) def UpperCAmelCase__ ( self : str): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_lowerCamelCase) def UpperCAmelCase__ ( self : Dict): for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=_lowerCamelCase , solver_type=_lowerCamelCase , prediction_type=_lowerCamelCase , algorithm_type=_lowerCamelCase , ) lowerCAmelCase_ : Optional[int] = self.full_loop( solver_order=_lowerCamelCase , solver_type=_lowerCamelCase , prediction_type=_lowerCamelCase , algorithm_type=_lowerCamelCase , ) assert not torch.isnan(_lowerCamelCase).any(), "Samples have nan numbers" def UpperCAmelCase__ ( self : str): self.check_over_configs(lower_order_final=_lowerCamelCase) self.check_over_configs(lower_order_final=_lowerCamelCase) def UpperCAmelCase__ ( self : Optional[Any]): self.check_over_configs(lambda_min_clipped=-float('''inf''')) self.check_over_configs(lambda_min_clipped=-5.1) def UpperCAmelCase__ ( self : Dict): self.check_over_configs(variance_type=_lowerCamelCase) self.check_over_configs(variance_type='''learned_range''') def UpperCAmelCase__ ( self : Optional[int]): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=_lowerCamelCase , time_step=0) def UpperCAmelCase__ ( self : Dict): lowerCAmelCase_ : Dict = self.full_loop() lowerCAmelCase_ : str = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.2791) < 1e-3 def UpperCAmelCase__ ( self : int): lowerCAmelCase_ : Any = self.full_loop(use_karras_sigmas=_lowerCamelCase) lowerCAmelCase_ : Union[str, Any] = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.2248) < 1e-3 def UpperCAmelCase__ ( self : Any): lowerCAmelCase_ : str = self.full_loop(prediction_type='''v_prediction''') lowerCAmelCase_ : List[Any] = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.1453) < 1e-3 def UpperCAmelCase__ ( self : Union[str, Any]): lowerCAmelCase_ : List[Any] = self.full_loop(prediction_type='''v_prediction''' , use_karras_sigmas=_lowerCamelCase) lowerCAmelCase_ : Tuple = torch.mean(torch.abs(_lowerCamelCase)) assert abs(result_mean.item() - 0.0649) < 1e-3 def UpperCAmelCase__ ( self : Tuple): lowerCAmelCase_ : int = self.scheduler_classes[0] lowerCAmelCase_ : List[str] = self.get_scheduler_config(thresholding=_lowerCamelCase , dynamic_thresholding_ratio=0) lowerCAmelCase_ : List[Any] = scheduler_class(**_lowerCamelCase) lowerCAmelCase_ : List[Any] = 1_0 lowerCAmelCase_ : Union[str, Any] = self.dummy_model() lowerCAmelCase_ : str = self.dummy_sample_deter.half() scheduler.set_timesteps(_lowerCamelCase) for i, t in enumerate(scheduler.timesteps): lowerCAmelCase_ : Dict = model(_lowerCamelCase , _lowerCamelCase) lowerCAmelCase_ : Optional[Any] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase).prev_sample assert sample.dtype == torch.floataa
103
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=3 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=10 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[1, 1, 2, 1] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="relu" , _lowerCamelCase=3 , _lowerCamelCase=None , ) -> List[str]: A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> Optional[Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : Any = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase_ ( self ) -> int: super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: pass def UpperCAmelCase_ ( self ) -> int: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase , _lowerCamelCase ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) def UpperCAmelCase_ ( self ) -> str: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> int: return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""tf""" ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 )
344
0
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class a_ ( __A ): """simple docstring""" def __init__( self : int ,snake_case : List[str] ,snake_case : int=13 ,snake_case : Union[str, Any]=7 ,snake_case : Union[str, Any]=True ,snake_case : Tuple=True ,snake_case : List[Any]=False ,snake_case : Any=True ,snake_case : List[Any]=99 ,snake_case : Tuple=32 ,snake_case : Union[str, Any]=5 ,snake_case : List[Any]=4 ,snake_case : Any=37 ,snake_case : Union[str, Any]="gelu" ,snake_case : Optional[int]=0.1 ,snake_case : List[str]=0.1 ,snake_case : Optional[int]=512 ,snake_case : Tuple=16 ,snake_case : Optional[Any]=2 ,snake_case : Optional[int]=0.02 ,snake_case : List[str]=3 ,snake_case : int=4 ,snake_case : Dict=None ,): SCREAMING_SNAKE_CASE =parent SCREAMING_SNAKE_CASE =batch_size SCREAMING_SNAKE_CASE =seq_length SCREAMING_SNAKE_CASE =is_training SCREAMING_SNAKE_CASE =use_input_mask SCREAMING_SNAKE_CASE =use_token_type_ids SCREAMING_SNAKE_CASE =use_labels SCREAMING_SNAKE_CASE =vocab_size SCREAMING_SNAKE_CASE =hidden_size SCREAMING_SNAKE_CASE =num_hidden_layers SCREAMING_SNAKE_CASE =num_attention_heads SCREAMING_SNAKE_CASE =intermediate_size SCREAMING_SNAKE_CASE =hidden_act SCREAMING_SNAKE_CASE =hidden_dropout_prob SCREAMING_SNAKE_CASE =attention_probs_dropout_prob SCREAMING_SNAKE_CASE =max_position_embeddings SCREAMING_SNAKE_CASE =type_vocab_size SCREAMING_SNAKE_CASE =type_sequence_label_size SCREAMING_SNAKE_CASE =initializer_range SCREAMING_SNAKE_CASE =num_labels SCREAMING_SNAKE_CASE =num_choices SCREAMING_SNAKE_CASE =scope def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) SCREAMING_SNAKE_CASE =None if self.use_input_mask: SCREAMING_SNAKE_CASE =random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE =None SCREAMING_SNAKE_CASE =None SCREAMING_SNAKE_CASE =None if self.use_labels: SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size] ,self.type_sequence_label_size ) SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size] ,self.num_choices ) SCREAMING_SNAKE_CASE =self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCAmelCase ( self : List[str] ): return DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,) def _lowerCAmelCase ( self : Union[str, Any] ,snake_case : Union[str, Any] ,snake_case : Dict ,snake_case : Any ,snake_case : Dict ,snake_case : int ,snake_case : Union[str, Any] ): SCREAMING_SNAKE_CASE =DistilBertModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE =model(_lowerCamelCase ,_lowerCamelCase ) SCREAMING_SNAKE_CASE =model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCAmelCase ( self : Optional[Any] ,snake_case : Tuple ,snake_case : str ,snake_case : str ,snake_case : Optional[Any] ,snake_case : Union[str, Any] ,snake_case : List[str] ): SCREAMING_SNAKE_CASE =DistilBertForMaskedLM(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE =model(_lowerCamelCase ,attention_mask=_lowerCamelCase ,labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCAmelCase ( self : Dict ,snake_case : str ,snake_case : Optional[Any] ,snake_case : Union[str, Any] ,snake_case : Optional[Any] ,snake_case : int ,snake_case : List[str] ): SCREAMING_SNAKE_CASE =DistilBertForQuestionAnswering(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE =model( _lowerCamelCase ,attention_mask=_lowerCamelCase ,start_positions=_lowerCamelCase ,end_positions=_lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def _lowerCAmelCase ( self : Any ,snake_case : Tuple ,snake_case : Union[str, Any] ,snake_case : str ,snake_case : List[Any] ,snake_case : Optional[int] ,snake_case : int ): SCREAMING_SNAKE_CASE =self.num_labels SCREAMING_SNAKE_CASE =DistilBertForSequenceClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE =model(_lowerCamelCase ,attention_mask=_lowerCamelCase ,labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def _lowerCAmelCase ( self : List[str] ,snake_case : str ,snake_case : Dict ,snake_case : Union[str, Any] ,snake_case : Tuple ,snake_case : Optional[Any] ,snake_case : int ): SCREAMING_SNAKE_CASE =self.num_labels SCREAMING_SNAKE_CASE =DistilBertForTokenClassification(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE =model(_lowerCamelCase ,attention_mask=_lowerCamelCase ,labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def _lowerCAmelCase ( self : List[Any] ,snake_case : List[str] ,snake_case : List[str] ,snake_case : Optional[int] ,snake_case : int ,snake_case : Optional[Any] ,snake_case : List[Any] ): SCREAMING_SNAKE_CASE =self.num_choices SCREAMING_SNAKE_CASE =DistilBertForMultipleChoice(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE =input_ids.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() SCREAMING_SNAKE_CASE =input_mask.unsqueeze(1 ).expand(-1 ,self.num_choices ,-1 ).contiguous() SCREAMING_SNAKE_CASE =model( _lowerCamelCase ,attention_mask=_lowerCamelCase ,labels=_lowerCamelCase ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def _lowerCAmelCase ( self : List[Any] ): SCREAMING_SNAKE_CASE =self.prepare_config_and_inputs() (SCREAMING_SNAKE_CASE) =config_and_inputs SCREAMING_SNAKE_CASE ={"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a_ ( __A , __A , unittest.TestCase ): """simple docstring""" __UpperCAmelCase = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) __UpperCAmelCase = ( { 'feature-extraction': DistilBertModel, 'fill-mask': DistilBertForMaskedLM, 'question-answering': DistilBertForQuestionAnswering, 'text-classification': DistilBertForSequenceClassification, 'token-classification': DistilBertForTokenClassification, 'zero-shot': DistilBertForSequenceClassification, } if is_torch_available() else {} ) __UpperCAmelCase = True __UpperCAmelCase = True __UpperCAmelCase = True __UpperCAmelCase = True def _lowerCAmelCase ( self : List[Any] ): SCREAMING_SNAKE_CASE =DistilBertModelTester(self ) SCREAMING_SNAKE_CASE =ConfigTester(self ,config_class=_lowerCamelCase ,dim=37 ) def _lowerCAmelCase ( self : int ): self.config_tester.run_common_tests() def _lowerCAmelCase ( self : List[str] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_lowerCamelCase ) def _lowerCAmelCase ( self : Any ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_lowerCamelCase ) def _lowerCAmelCase ( self : Tuple ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_lowerCamelCase ) def _lowerCAmelCase ( self : Any ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_lowerCamelCase ) def _lowerCAmelCase ( self : Dict ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_lowerCamelCase ) def _lowerCAmelCase ( self : Optional[int] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_lowerCamelCase ) @slow def _lowerCAmelCase ( self : Optional[Any] ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE =DistilBertModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) @slow @require_torch_gpu def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return SCREAMING_SNAKE_CASE =True SCREAMING_SNAKE_CASE =model_class(config=_lowerCamelCase ) SCREAMING_SNAKE_CASE =self._prepare_for_class(_lowerCamelCase ,_lowerCamelCase ) SCREAMING_SNAKE_CASE =torch.jit.trace( _lowerCamelCase ,(inputs_dict['input_ids'].to('cpu' ), inputs_dict['attention_mask'].to('cpu' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_lowerCamelCase ,os.path.join(_lowerCamelCase ,'traced_model.pt' ) ) SCREAMING_SNAKE_CASE =torch.jit.load(os.path.join(_lowerCamelCase ,'traced_model.pt' ) ,map_location=_lowerCamelCase ) loaded(inputs_dict['input_ids'].to(_lowerCamelCase ) ,inputs_dict['attention_mask'].to(_lowerCamelCase ) ) @require_torch class a_ ( unittest.TestCase ): """simple docstring""" @slow def _lowerCAmelCase ( self : Any ): SCREAMING_SNAKE_CASE =DistilBertModel.from_pretrained('distilbert-base-uncased' ) SCREAMING_SNAKE_CASE =torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) SCREAMING_SNAKE_CASE =torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): SCREAMING_SNAKE_CASE =model(_lowerCamelCase ,attention_mask=_lowerCamelCase )[0] SCREAMING_SNAKE_CASE =torch.Size((1, 11, 768) ) self.assertEqual(output.shape ,_lowerCamelCase ) SCREAMING_SNAKE_CASE =torch.tensor( [[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] ,_lowerCamelCase ,atol=1e-4 ) )
334
'''simple docstring''' def UpperCAmelCase ( a_ = 1_0_0 ) -> int: """simple docstring""" A_ : Dict = n * (n + 1) * (2 * n + 1) / 6 A_ : Optional[int] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'{solution() = }')
344
0
from math import factorial def _lowerCamelCase( lowercase__ = 1_0_0 ) -> int: '''simple docstring''' return sum(int(a_ ) for x in str(factorial(a_ ) ) ) if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
295
'''simple docstring''' from typing import TYPE_CHECKING from ..utils import _LazyModule UpperCamelCase__ : int = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys UpperCamelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
'''simple docstring''' from __future__ import annotations def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(a_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(a_ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
3
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCAmelCase ( a_ ) -> Dict[str, torch.Tensor]: """simple docstring""" A_ : List[str] = [] A_ : Dict = [] A_ : List[Any] = [] for rt in rc.restypes: A_ : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) A_ : Union[str, Any] = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) A_ : Tuple = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : Optional[int] = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : List[Any] = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) A_ : Optional[int] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein A_ : Dict = restype_atomaa_to_atomaa[protein_aatype] A_ : Optional[Any] = restype_atomaa_mask[protein_aatype] A_ : Any = residx_atomaa_mask A_ : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back A_ : Tuple = restype_atomaa_to_atomaa[protein_aatype] A_ : Tuple = residx_atomaa_to_atomaa.long() # create the corresponding mask A_ : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): A_ : Optional[Any] = rc.restype_atoa[restype_letter] A_ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: A_ : Any = rc.atom_order[atom_name] A_ : Optional[int] = 1 A_ : Optional[int] = restype_atomaa_mask[protein_aatype] A_ : Dict = residx_atomaa_mask return protein def UpperCAmelCase ( a_ ) -> Dict[str, np.ndarray]: """simple docstring""" A_ : Union[str, Any] = tree_map(lambda a_ : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) A_ : Optional[int] = tensor_tree_map(lambda a_ : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
344
0
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class __lowerCAmelCase : @property def lowerCamelCase__ ( self :Dict ): '''simple docstring''' return self.get_dummy_input() @property def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(F'\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.' ) def lowerCamelCase__ ( self :Optional[Any] , __magic_name__ :Union[str, Any]=True , __magic_name__ :Union[str, Any]=False , __magic_name__ :List[str]=False , __magic_name__ :List[Any]=False , ): '''simple docstring''' a = 4 a = 32 a = (32, 32) a = torch.manual_seed(0 ) a = torch.device(_lowerCamelCase ) a = (batch_size, num_channels) + sizes a = randn_tensor(_lowerCamelCase , generator=_lowerCamelCase , device=_lowerCamelCase ) a = {"""hidden_states""": hidden_states} if include_temb: a = 128 a = randn_tensor((batch_size, temb_channels) , generator=_lowerCamelCase , device=_lowerCamelCase ) if include_res_hidden_states_tuple: a = torch.manual_seed(1 ) a = (randn_tensor(_lowerCamelCase , generator=_lowerCamelCase , device=_lowerCamelCase ),) if include_encoder_hidden_states: a = floats_tensor((batch_size, 32, 32) ).to(_lowerCamelCase ) if include_skip_sample: a = randn_tensor(((batch_size, 3) + sizes) , generator=_lowerCamelCase , device=_lowerCamelCase ) return dummy_input def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' a = { """in_channels""": 32, """out_channels""": 32, """temb_channels""": 128, } if self.block_type == "up": a = 32 if self.block_type == "mid": init_dict.pop("""out_channels""" ) a = self.dummy_input return init_dict, inputs_dict def lowerCamelCase__ ( self :List[Any] , __magic_name__ :List[str] ): '''simple docstring''' a = self.prepare_init_args_and_inputs_for_common() a = self.block_class(**_lowerCamelCase ) unet_block.to(_lowerCamelCase ) unet_block.eval() with torch.no_grad(): a = unet_block(**_lowerCamelCase ) if isinstance(_lowerCamelCase , _lowerCamelCase ): a = output[0] self.assertEqual(output.shape , self.output_shape ) a = output[0, -1, -3:, -3:] a = torch.tensor(_lowerCamelCase ).to(_lowerCamelCase ) assert torch_all_close(output_slice.flatten() , _lowerCamelCase , atol=5E-3 ) @unittest.skipIf(torch_device == """mps""" , """Training is not supported in mps""" ) def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' a = self.prepare_init_args_and_inputs_for_common() a = self.block_class(**_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() a = model(**_lowerCamelCase ) if isinstance(_lowerCamelCase , _lowerCamelCase ): a = output[0] a = torch.device(_lowerCamelCase ) a = randn_tensor(output.shape , device=_lowerCamelCase ) a = torch.nn.functional.mse_loss(_lowerCamelCase , _lowerCamelCase ) loss.backward()
228
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ) -> str: A_ : Optional[int] = parent A_ : Dict = batch_size A_ : List[Any] = image_size A_ : Optional[int] = patch_size A_ : List[str] = num_channels A_ : List[Any] = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_size A_ : str = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Union[str, Any] = intermediate_size A_ : Any = hidden_act A_ : Optional[Any] = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Dict = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : str = scope A_ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) A_ : Tuple = (image_size // patch_size) ** 2 A_ : Union[str, Any] = num_patches + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> int: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[str] = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : int = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A_ : Dict = 1 A_ : Optional[int] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Tuple = self.type_sequence_label_size A_ : Tuple = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : Dict = 1 A_ : Any = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : List[str] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = DeiTModelTester(self ) A_ : str = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: pass def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ , A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[str] = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Union[str, Any] = [*signature.parameters.keys()] A_ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: A_ : int = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase_ ( self ) -> Optional[Any]: if not self.model_tester.is_training: return A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[str] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : List[str] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> int: A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A_ : Any = False A_ : Union[str, Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue A_ : List[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Union[str, Any] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Tuple: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): A_ : Dict = problem_type["""title"""] A_ : List[Any] = problem_type["""num_labels"""] A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: A_ : Tuple = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) A_ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: A_ : List[str] = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Optional[Any]: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) A_ : Optional[int] = self.default_image_processor A_ : str = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Any = model(**_lowerCamelCase ) # verify the logits A_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : List[Any] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : int = self.default_image_processor A_ : List[str] = prepare_img() A_ : List[Any] = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) A_ : Union[str, Any] = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A_ : List[Any] = model(_lowerCamelCase )
344
0
'''simple docstring''' import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin class _lowerCAmelCase ( __A , __A ): '''simple docstring''' lowerCAmelCase_ = 1 @register_to_config def __init__(self , UpperCAmelCase=2000 , UpperCAmelCase=0.1 , UpperCAmelCase=20 , UpperCAmelCase=1e-3 ) -> str: _snake_case = None _snake_case = None _snake_case = None def lowercase (self , UpperCAmelCase , UpperCAmelCase = None ) -> Any: _snake_case = torch.linspace(1 , self.config.sampling_eps , _lowerCamelCase , device=_lowerCamelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None ) -> Dict: if self.timesteps is None: raise ValueError( """`self.timesteps` is not set, you need to run 'set_timesteps' after creating the scheduler""" ) # TODO(Patrick) better comments + non-PyTorch # postprocess model score _snake_case = ( -0.25 * t**2 * (self.config.beta_max - self.config.beta_min) - 0.5 * t * self.config.beta_min ) _snake_case = torch.sqrt(1.0 - torch.exp(2.0 * log_mean_coeff ) ) _snake_case = std.flatten() while len(std.shape ) < len(score.shape ): _snake_case = std.unsqueeze(-1 ) _snake_case = -score / std # compute _snake_case = -1.0 / len(self.timesteps ) _snake_case = self.config.beta_min + t * (self.config.beta_max - self.config.beta_min) _snake_case = beta_t.flatten() while len(beta_t.shape ) < len(x.shape ): _snake_case = beta_t.unsqueeze(-1 ) _snake_case = -0.5 * beta_t * x _snake_case = torch.sqrt(_lowerCamelCase ) _snake_case = drift - diffusion**2 * score _snake_case = x + drift * dt # add noise _snake_case = randn_tensor(x.shape , layout=x.layout , generator=_lowerCamelCase , device=x.device , dtype=x.dtype ) _snake_case = x_mean + diffusion * math.sqrt(-dt ) * noise return x, x_mean def __len__(self ) -> str: return self.config.num_train_timesteps
341
'''simple docstring''' import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 32 , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = True , _lowerCamelCase = [0.4814_5466, 0.457_8275, 0.4082_1073] , _lowerCamelCase = [0.2686_2954, 0.2613_0258, 0.2757_7711] , _lowerCamelCase = True , _lowerCamelCase=7 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=3 , ) -> Union[str, Any]: A_ : Optional[int] = parent A_ : Union[str, Any] = do_resize A_ : Optional[Any] = size if size is not None else {"""shortest_edge""": 288} A_ : Tuple = size_divisor A_ : List[Any] = do_rescale A_ : Dict = rescale_factor A_ : List[Any] = do_normalize A_ : Dict = do_center_crop A_ : Optional[Any] = image_mean A_ : List[str] = image_std A_ : str = do_pad A_ : Any = batch_size A_ : List[str] = num_channels A_ : List[str] = min_resolution A_ : Union[str, Any] = max_resolution def UpperCAmelCase_ ( self ) -> Any: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase=False ) -> Optional[int]: if not batched: A_ : Union[str, Any] = self.size["""shortest_edge"""] A_ : Dict = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image ): A_ , A_ : Optional[Any] = image.size else: A_ , A_ : int = image.shape[1], image.shape[2] A_ : Optional[int] = size / min(_lowerCamelCase , _lowerCamelCase ) if h < w: A_ , A_ : Optional[Any] = size, scale * w else: A_ , A_ : Dict = scale * h, size A_ : Union[str, Any] = int((1333 / 800) * size ) if max(_lowerCamelCase , _lowerCamelCase ) > max_size: A_ : str = max_size / max(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = newh * scale A_ : Dict = neww * scale A_ , A_ : str = int(newh + 0.5 ), int(neww + 0.5 ) A_ , A_ : Dict = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: A_ : Tuple = [] for image in image_inputs: A_ , A_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) A_ : List[Any] = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[0] )[0] A_ : Tuple = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : int = BridgeTowerImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size_divisor""" ) ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image processor A_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : int = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: # Initialize image processor A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image processor A_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
344
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { 'microsoft/unispeech-large-1500h-cv': ( 'https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json' ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class lowerCAmelCase_ ( __A ): __lowerCamelCase : Tuple = "unispeech" def __init__( self , _lowerCAmelCase=32 , _lowerCAmelCase=768 , _lowerCAmelCase=12 , _lowerCAmelCase=12 , _lowerCAmelCase=3072 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.02 , _lowerCAmelCase=1E-5 , _lowerCAmelCase="group" , _lowerCAmelCase="gelu" , _lowerCAmelCase=(512, 512, 512, 512, 512, 512, 512) , _lowerCAmelCase=(5, 2, 2, 2, 2, 2, 2) , _lowerCAmelCase=(10, 3, 3, 3, 3, 2, 2) , _lowerCAmelCase=False , _lowerCAmelCase=128 , _lowerCAmelCase=16 , _lowerCAmelCase=False , _lowerCAmelCase=True , _lowerCAmelCase=0.05 , _lowerCAmelCase=10 , _lowerCAmelCase=2 , _lowerCAmelCase=0.0 , _lowerCAmelCase=10 , _lowerCAmelCase=0 , _lowerCAmelCase=320 , _lowerCAmelCase=2 , _lowerCAmelCase=0.1 , _lowerCAmelCase=100 , _lowerCAmelCase=256 , _lowerCAmelCase=256 , _lowerCAmelCase=0.1 , _lowerCAmelCase="mean" , _lowerCAmelCase=False , _lowerCAmelCase=False , _lowerCAmelCase=256 , _lowerCAmelCase=80 , _lowerCAmelCase=0 , _lowerCAmelCase=1 , _lowerCAmelCase=2 , _lowerCAmelCase=0.5 , **_lowerCAmelCase , ) -> Optional[Any]: super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase ) _lowerCAmelCase = hidden_size _lowerCAmelCase = feat_extract_norm _lowerCAmelCase = feat_extract_activation _lowerCAmelCase = list(_lowerCamelCase ) _lowerCAmelCase = list(_lowerCamelCase ) _lowerCAmelCase = list(_lowerCamelCase ) _lowerCAmelCase = conv_bias _lowerCAmelCase = num_conv_pos_embeddings _lowerCAmelCase = num_conv_pos_embedding_groups _lowerCAmelCase = len(self.conv_dim ) _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_act _lowerCAmelCase = num_attention_heads _lowerCAmelCase = hidden_dropout _lowerCAmelCase = attention_dropout _lowerCAmelCase = activation_dropout _lowerCAmelCase = feat_proj_dropout _lowerCAmelCase = final_dropout _lowerCAmelCase = layerdrop _lowerCAmelCase = layer_norm_eps _lowerCAmelCase = initializer_range _lowerCAmelCase = num_ctc_classes _lowerCAmelCase = vocab_size _lowerCAmelCase = do_stable_layer_norm _lowerCAmelCase = use_weighted_layer_sum _lowerCAmelCase = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _lowerCAmelCase = apply_spec_augment _lowerCAmelCase = mask_time_prob _lowerCAmelCase = mask_time_length _lowerCAmelCase = mask_time_min_masks _lowerCAmelCase = mask_feature_prob _lowerCAmelCase = mask_feature_length _lowerCAmelCase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations _lowerCAmelCase = num_codevectors_per_group _lowerCAmelCase = num_codevector_groups _lowerCAmelCase = contrastive_logits_temperature _lowerCAmelCase = feat_quantizer_dropout _lowerCAmelCase = num_negatives _lowerCAmelCase = codevector_dim _lowerCAmelCase = proj_codevector_dim _lowerCAmelCase = diversity_loss_weight # ctc loss _lowerCAmelCase = ctc_loss_reduction _lowerCAmelCase = ctc_zero_infinity # pretraining loss _lowerCAmelCase = replace_prob @property def _snake_case ( self ) -> int: return functools.reduce(operator.mul , self.conv_stride , 1 )
158
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(a_ ): for j in range(a_ ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) , end="""\t""" ) else: print("""INF""" , end="""\t""" ) print() def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" A_ : List[str] = [[float("""inf""" ) for _ in range(a_ )] for _ in range(a_ )] for i in range(a_ ): for j in range(a_ ): A_ : List[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(a_ ): # looping through rows of graph array for i in range(a_ ): # looping through columns of graph array for j in range(a_ ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): A_ : List[str] = dist[i][k] + dist[k][j] _print_dist(a_ , a_ ) return dist, v if __name__ == "__main__": UpperCamelCase__ : Tuple = int(input('Enter number of vertices: ')) UpperCamelCase__ : int = int(input('Enter number of edges: ')) UpperCamelCase__ : Dict = [[float('inf') for i in range(v)] for j in range(v)] for i in range(v): UpperCamelCase__ : Union[str, Any] = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print('\nEdge ', i + 1) UpperCamelCase__ : Union[str, Any] = int(input('Enter source:')) UpperCamelCase__ : int = int(input('Enter destination:')) UpperCamelCase__ : Optional[Any] = float(input('Enter weight:')) UpperCamelCase__ : Any = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
344
0
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class __A : def __init__( self : Any , UpperCAmelCase_ : List[Any] , ): lowerCAmelCase : Optional[Any] = parent lowerCAmelCase : List[Any] = 13 lowerCAmelCase : List[Any] = 7 lowerCAmelCase : Dict = True lowerCAmelCase : List[str] = True lowerCAmelCase : Dict = True lowerCAmelCase : List[str] = 99 lowerCAmelCase : Union[str, Any] = 32 lowerCAmelCase : List[Any] = 2 lowerCAmelCase : Optional[int] = 4 lowerCAmelCase : Any = 37 lowerCAmelCase : int = """gelu""" lowerCAmelCase : Dict = 0.1 lowerCAmelCase : int = 0.1 lowerCAmelCase : str = 512 lowerCAmelCase : Tuple = 16 lowerCAmelCase : Tuple = 2 lowerCAmelCase : Dict = 0.02 lowerCAmelCase : Dict = 3 lowerCAmelCase : Dict = 4 lowerCAmelCase : Optional[Any] = None def lowercase__ ( self : Optional[int] ): lowerCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase : List[str] = None if self.use_input_mask: lowerCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase : Any = None lowerCAmelCase : int = None lowerCAmelCase : Any = None if self.use_labels: lowerCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase : Optional[int] = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase__ ( self : Optional[int] ): ( lowerCAmelCase ) : Tuple = self.prepare_config_and_inputs() lowerCAmelCase : Union[str, Any] = True lowerCAmelCase : Dict = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def lowercase__ ( self : int , UpperCAmelCase_ : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Optional[int] ): lowerCAmelCase : Any = TFEsmModel(config=_lowerCamelCase ) lowerCAmelCase : List[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} lowerCAmelCase : Dict = model(_lowerCamelCase ) lowerCAmelCase : Optional[Any] = [input_ids, input_mask] lowerCAmelCase : Tuple = model(_lowerCamelCase ) lowerCAmelCase : Optional[Any] = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : str , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Tuple , ): lowerCAmelCase : str = True lowerCAmelCase : Dict = TFEsmModel(config=_lowerCamelCase ) lowerCAmelCase : Dict = { """input_ids""": input_ids, """attention_mask""": input_mask, """encoder_hidden_states""": encoder_hidden_states, """encoder_attention_mask""": encoder_attention_mask, } lowerCAmelCase : Union[str, Any] = model(_lowerCamelCase ) lowerCAmelCase : Union[str, Any] = [input_ids, input_mask] lowerCAmelCase : Optional[int] = model(_lowerCamelCase , encoder_hidden_states=_lowerCamelCase ) # Also check the case where encoder outputs are not passed lowerCAmelCase : Union[str, Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : Union[str, Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any] ): lowerCAmelCase : Dict = TFEsmForMaskedLM(config=_lowerCamelCase ) lowerCAmelCase : Optional[Any] = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase__ ( self : Optional[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Any , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Union[str, Any] ): lowerCAmelCase : str = self.num_labels lowerCAmelCase : Optional[Any] = TFEsmForTokenClassification(config=_lowerCamelCase ) lowerCAmelCase : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask} lowerCAmelCase : Optional[int] = model(_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase__ ( self : Dict ): lowerCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() ( lowerCAmelCase ) : str = config_and_inputs lowerCAmelCase : str = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __A ( __A , __A , unittest.TestCase ): lowerCAmelCase_ : List[str] = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase_ : Optional[Any] = ( { "feature-extraction": TFEsmModel, "fill-mask": TFEsmForMaskedLM, "text-classification": TFEsmForSequenceClassification, "token-classification": TFEsmForTokenClassification, "zero-shot": TFEsmForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase_ : Optional[Any] = False lowerCAmelCase_ : str = False def lowercase__ ( self : str ): lowerCAmelCase : List[str] = TFEsmModelTester(self ) lowerCAmelCase : str = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def lowercase__ ( self : Any ): self.config_tester.run_common_tests() def lowercase__ ( self : Tuple ): lowerCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def lowercase__ ( self : List[str] ): lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_lowerCamelCase ) def lowercase__ ( self : List[Any] ): lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCamelCase ) def lowercase__ ( self : Any ): lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCamelCase ) @slow def lowercase__ ( self : List[Any] ): for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase : Optional[int] = TFEsmModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) @unittest.skip('Protein models do not support embedding resizing.' ) def lowercase__ ( self : Dict ): pass @unittest.skip('Protein models do not support embedding resizing.' ) def lowercase__ ( self : List[Any] ): pass def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase : int = model_class(_lowerCamelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer lowerCAmelCase : Any = model.get_bias() assert isinstance(_lowerCamelCase , _lowerCamelCase ) for k, v in name.items(): assert isinstance(_lowerCamelCase , tf.Variable ) else: lowerCAmelCase : Any = model.get_output_embeddings() assert x is None lowerCAmelCase : int = model.get_bias() assert name is None @require_tf class __A ( unittest.TestCase ): @slow def lowercase__ ( self : Optional[int] ): lowerCAmelCase : Tuple = TFEsmForMaskedLM.from_pretrained('facebook/esm2_t6_8M_UR50D' ) lowerCAmelCase : int = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase : str = model(_lowerCamelCase )[0] lowerCAmelCase : Optional[int] = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , _lowerCamelCase ) # compare the actual values for a slice. lowerCAmelCase : int = tf.constant( [ [ [8.92_15_18, -10.58_98_14, -6.4_67_13_07], [-6.3_96_71_56, -13.91_13_77, -1.1_21_19_15], [-7.78_12_47, -13.95_15_57, -3.74_05_92], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-2 ) ) @slow def lowercase__ ( self : List[Any] ): lowerCAmelCase : List[Any] = TFEsmModel.from_pretrained('facebook/esm2_t6_8M_UR50D' ) lowerCAmelCase : List[Any] = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) lowerCAmelCase : Optional[Any] = model(_lowerCamelCase )[0] # compare the actual values for a slice. lowerCAmelCase : Union[str, Any] = tf.constant( [ [ [0.14_44_30_92, 0.54_12_53_27, 0.3_24_77_39], [0.30_34_04_84, 0.00_52_66_76, 0.31_07_77_22], [0.32_27_80_43, -0.24_98_70_96, 0.3_41_46_28], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
138
'''simple docstring''' import datasets from .evaluate import evaluate UpperCamelCase__ : int = '\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n' UpperCamelCase__ : Any = '\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n' UpperCamelCase__ : Optional[Any] = '\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the SQuAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> squad_metric = datasets.load_metric("squad")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: A_ : Optional[Any] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} A_ : List[Any] = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] A_ : int = evaluate(dataset=_lowerCamelCase , predictions=_lowerCamelCase ) return score
344
0
import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .tokenization_wavaveca import WavaVecaCTCTokenizer class A_ ( __A ): '''simple docstring''' _UpperCamelCase : Tuple = """Wav2Vec2FeatureExtractor""" _UpperCamelCase : List[str] = """AutoTokenizer""" def __init__( self , snake_case , snake_case ): super().__init__(_lowerCamelCase , _lowerCamelCase ) lowercase = self.feature_extractor lowercase = False @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case , **snake_case ): try: return super().from_pretrained(_lowerCamelCase , **_lowerCamelCase ) except OSError: warnings.warn( F'''Loading a tokenizer inside {cls.__name__} from a config that does not''' ' include a `tokenizer_class` attribute is deprecated and will be ' 'removed in v5. Please add `\'tokenizer_class\': \'Wav2Vec2CTCTokenizer\'`' ' attribute to either your `config.json` or `tokenizer_config.json` ' 'file to suppress this warning: ' , _lowerCamelCase , ) lowercase = WavaVecaFeatureExtractor.from_pretrained(_lowerCamelCase , **_lowerCamelCase ) lowercase = WavaVecaCTCTokenizer.from_pretrained(_lowerCamelCase , **_lowerCamelCase ) return cls(feature_extractor=_lowerCamelCase , tokenizer=_lowerCamelCase ) def __call__( self , *snake_case , **snake_case ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_lowerCamelCase , **_lowerCamelCase ) if "raw_speech" in kwargs: warnings.warn('Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.' ) lowercase = kwargs.pop('raw_speech' ) else: lowercase = kwargs.pop('audio' , _lowerCamelCase ) lowercase = kwargs.pop('sampling_rate' , _lowerCamelCase ) lowercase = kwargs.pop('text' , _lowerCamelCase ) if len(_lowerCamelCase ) > 0: lowercase = args[0] lowercase = args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if audio is not None: lowercase = self.feature_extractor(_lowerCamelCase , *_lowerCamelCase , sampling_rate=_lowerCamelCase , **_lowerCamelCase ) if text is not None: lowercase = self.tokenizer(_lowerCamelCase , **_lowerCamelCase ) if text is None: return inputs elif audio is None: return encodings else: lowercase = encodings["""input_ids"""] return inputs def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*_lowerCamelCase , **_lowerCamelCase ) lowercase = kwargs.pop('input_features' , _lowerCamelCase ) lowercase = kwargs.pop('labels' , _lowerCamelCase ) if len(_lowerCamelCase ) > 0: lowercase = args[0] lowercase = args[1:] if input_features is not None: lowercase = self.feature_extractor.pad(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) if labels is not None: lowercase = self.tokenizer.pad(_lowerCamelCase , **_lowerCamelCase ) if labels is None: return input_features elif input_features is None: return labels else: lowercase = labels["""input_ids"""] return input_features def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.tokenizer.batch_decode(*_lowerCamelCase , **_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.tokenizer.decode(*_lowerCamelCase , **_lowerCamelCase ) @contextmanager def SCREAMING_SNAKE_CASE__ ( self ): warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your audio inputs, or in a separate call.' ) lowercase = True lowercase = self.tokenizer yield lowercase = self.feature_extractor lowercase = False
195
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ : Any = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Optional[Any] = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] UpperCamelCase__ : List[str] = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] UpperCamelCase__ : str = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): UpperCamelCase__ : List[str] = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCamelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCamelCase : Tuple = None UpperCamelCase : List[Any] = None UpperCamelCase : List[str] = graph self._normalize_graph(_lowerCamelCase, _lowerCamelCase ) UpperCamelCase : str = len(_lowerCamelCase ) UpperCamelCase : str = None def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> List[str]: if sources is int: UpperCamelCase : Tuple = [sources] if sinks is int: UpperCamelCase : List[str] = [sinks] if len(_lowerCamelCase ) == 0 or len(_lowerCamelCase ) == 0: return UpperCamelCase : Optional[Any] = sources[0] UpperCamelCase : List[str] = sinks[0] # make fake vertex if there are more # than one source or sink if len(_lowerCamelCase ) > 1 or len(_lowerCamelCase ) > 1: UpperCamelCase : List[str] = 0 for i in sources: max_input_flow += sum(self.graph[i] ) UpperCamelCase : Union[str, Any] = len(self.graph ) + 1 for room in self.graph: room.insert(0, 0 ) self.graph.insert(0, [0] * size ) for i in sources: UpperCamelCase : int = max_input_flow UpperCamelCase : str = 0 UpperCamelCase : List[str] = len(self.graph ) + 1 for room in self.graph: room.append(0 ) self.graph.append([0] * size ) for i in sinks: UpperCamelCase : Optional[Any] = max_input_flow UpperCamelCase : str = size - 1 def snake_case_ ( self ) -> str: if self.maximum_flow_algorithm is None: raise Exception('You need to set maximum flow algorithm before.' ) if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> Dict: UpperCamelCase : Union[str, Any] = algorithm(self ) class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_ ) -> Tuple: UpperCamelCase : List[Any] = flow_network UpperCamelCase : Union[str, Any] = flow_network.verticesCount UpperCamelCase : List[Any] = flow_network.sourceIndex UpperCamelCase : Dict = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that UpperCamelCase : Optional[int] = flow_network.graph UpperCamelCase : Optional[int] = False def snake_case_ ( self ) -> Optional[int]: if not self.executed: self._algorithm() UpperCamelCase : Optional[Any] = True def snake_case_ ( self ) -> List[str]: pass class lowerCAmelCase_ ( __A ): def __init__( self, SCREAMING_SNAKE_CASE_ ) -> List[Any]: super().__init__(_lowerCamelCase ) # use this to save your result UpperCamelCase : Dict = -1 def snake_case_ ( self ) -> Any: if not self.executed: raise Exception('You should execute algorithm before using its result!' ) return self.maximum_flow class lowerCAmelCase_ ( __A ): def __init__( self, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: super().__init__(_lowerCamelCase ) UpperCamelCase : Tuple = [[0] * self.verticies_count for i in range(self.verticies_count )] UpperCamelCase : List[Any] = [0] * self.verticies_count UpperCamelCase : Union[str, Any] = [0] * self.verticies_count def snake_case_ ( self ) -> Union[str, Any]: UpperCamelCase : List[Any] = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index] ): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule UpperCamelCase : Optional[int] = [ i for i in range(self.verticies_count ) if i != self.source_index and i != self.sink_index ] # move through list UpperCamelCase : List[Any] = 0 while i < len(_lowerCamelCase ): UpperCamelCase : str = vertices_list[i] UpperCamelCase : Optional[Any] = self.heights[vertex_index] self.process_vertex(_lowerCamelCase ) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0, vertices_list.pop(_lowerCamelCase ) ) UpperCamelCase : Union[str, Any] = 0 else: i += 1 UpperCamelCase : Any = sum(self.preflow[self.source_index] ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count ): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(_lowerCamelCase, _lowerCamelCase ) self.relabel(_lowerCamelCase ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCamelCase : Union[str, Any] = min( self.excesses[from_index], self.graph[from_index][to_index] - self.preflow[from_index][to_index], ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> List[str]: UpperCamelCase : Dict = None for to_index in range(self.verticies_count ): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): UpperCamelCase : List[Any] = self.heights[to_index] if min_height is not None: UpperCamelCase : Optional[Any] = min_height + 1 if __name__ == "__main__": __UpperCAmelCase = [0] __UpperCAmelCase = [3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] __UpperCAmelCase = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network __UpperCAmelCase = FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate __UpperCAmelCase = flow_network.find_maximum_flow() print(F"""maximum flow is {maximum_flow}""")
119
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''vision-encoder-decoder''' lowerCamelCase = True def __init__( self , **_lowerCamelCase ) -> str: super().__init__(**_lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F"A configuraton of type {self.model_type} cannot be instantiated because " F"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) A_ : Optional[int] = kwargs.pop("""encoder""" ) A_ : List[str] = encoder_config.pop("""model_type""" ) A_ : str = kwargs.pop("""decoder""" ) A_ : Optional[Any] = decoder_config.pop("""model_type""" ) A_ : List[str] = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : str = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : Any = True @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A_ : int = True A_ : List[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = copy.deepcopy(self.__dict__ ) A_ : List[str] = self.encoder.to_dict() A_ : Union[str, Any] = self.decoder.to_dict() A_ : str = self.__class__.model_type return output class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> float: return 1e-4 @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: A_ : Optional[Any] = OrderedDict() A_ : Any = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : Optional[int] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = -1 , _lowerCamelCase = -1 , _lowerCamelCase = False , _lowerCamelCase = None , ) -> Mapping[str, Any]: import torch A_ : Optional[int] = OrderedDict() A_ : List[Any] = super().generate_dummy_inputs( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) A_ , A_ : str = dummy_input["""input_ids"""].shape A_ : Optional[int] = (batch, encoder_sequence, self._config.encoder_hidden_size) A_ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A_ : List[str] = dummy_input.pop("""attention_mask""" ) A_ : Optional[int] = torch.zeros(_lowerCamelCase ) return common_inputs class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> None: pass def UpperCAmelCase_ ( self , _lowerCamelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = "default" ) -> OnnxConfig: A_ : List[Any] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_lowerCamelCase , _lowerCamelCase )
344
0
"""simple docstring""" import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def lowercase_ ( ): """simple docstring""" A_ : Tuple = ArgumentParser( description=( '''PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes''' ) ) # Optional arguments for the launch helper parser.add_argument('''--num_cores''' , type=a_ , default=1 , help='''Number of TPU cores to use (1 or 8).''' ) # positional parser.add_argument( '''training_script''' , type=a_ , help=( '''The full path to the single TPU training ''' '''program/script to be launched in parallel, ''' '''followed by all the arguments for the ''' '''training script''' ) , ) # rest from the training program parser.add_argument('''training_script_args''' , nargs=a_ ) return parser.parse_args() def lowercase_ ( ): """simple docstring""" A_ : List[Any] = parse_args() # Import training_script as a module. A_ : Tuple = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) A_ : Optional[Any] = script_fpath.stem A_ : int = importlib.import_module(a_ ) # Patch sys.argv A_ : Any = [args.training_script] + args.training_script_args + ["""--tpu_num_cores""", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
167
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCamelCase__ : Any = '\\n Text data.\n Second line of data.' UpperCamelCase__ : List[Any] = 'file' @pytest.fixture(scope="""session""" ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : int = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") A_ : int = bytes(a_ , """utf-8""" ) with zstd.open(a_ , """wb""" ) as f: f.write(a_ ) return path @pytest.fixture def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , a_ ) , """w""" ) as f: f.write(a_ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> Optional[int]: """simple docstring""" A_ : List[str] = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} A_ : Any = input_paths[compression_format] A_ : Tuple = tmp_path / """cache""" A_ : Tuple = DownloadConfig(cache_dir=a_ , extract_compressed_file=a_ ) A_ : Dict = cached_path(a_ , download_config=a_ ) with open(a_ ) as f: A_ : Optional[Any] = f.read() with open(a_ ) as f: A_ : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Union[str, Any] = """custom_cache""" A_ : List[str] = """custom_extracted_dir""" A_ : Optional[Any] = tmp_path / """custom_extracted_path""" if default_extracted: A_ : Any = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , a_ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(a_ ) ) A_ : Union[str, Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) A_ : List[Any] = xz_file A_ : Optional[int] = ( DownloadConfig(extract_compressed_file=a_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=a_ ) ) A_ : Union[str, Any] = cached_path(a_ , download_config=a_ ) assert Path(a_ ).parent.parts[-2:] == expected def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : str = str(Path(a_ ).resolve() ) assert cached_path(a_ ) == text_file # relative path A_ : List[str] = str(Path(a_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(a_ ) == text_file def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : Optional[Any] = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(a_ ): cached_path(a_ ) # relative path A_ : Tuple = """./__missing_file__.txt""" with pytest.raises(a_ ): cached_path(a_ ) def UpperCAmelCase ( a_ ) -> Tuple: """simple docstring""" A_ : Any = get_from_cache(F"tmp://{tmpfs_file}" ) with open(a_ ) as f: A_ : List[str] = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( ) -> List[str]: """simple docstring""" with pytest.raises(a_ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Union[str, Any]: """simple docstring""" A_ : List[str] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): http_get("""https://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): ftp_get("""ftp://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): fsspec_get("""s3://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): fsspec_head("""s3://huggingface.co""" )
344
0
def UpperCamelCase( __UpperCamelCase : str ): return "".join([hex(a_ )[2:].zfill(2 ).upper() for byte in list(a_ )] ) def UpperCamelCase( __UpperCamelCase : Tuple ): if (len(a_ ) % 2) != 0: raise ValueError( '''Base16 encoded data is invalid: Data does not have an even number of hex digits.''' ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(a_ ) <= set('''0123456789ABCDEF''' ): raise ValueError( '''Base16 encoded data is invalid: Data is not uppercase hex or it contains invalid characters.''' ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] ,16 ) for i in range(0 ,len(a_ ) ,2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
103
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) UpperCamelCase__ : int = {'processing_layoutxlm': ['LayoutXLMProcessor']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Tuple = ['LayoutXLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : List[Any] = ['LayoutXLMTokenizerFast'] if TYPE_CHECKING: from .processing_layoutxlm import LayoutXLMProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm import LayoutXLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm_fast import LayoutXLMTokenizerFast else: import sys UpperCamelCase__ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
class a_ : """simple docstring""" def __init__( self : int ): SCREAMING_SNAKE_CASE =0 SCREAMING_SNAKE_CASE =0 SCREAMING_SNAKE_CASE ={} def _lowerCAmelCase ( self : str ,snake_case : Optional[int] ): if vertex not in self.adjacency: SCREAMING_SNAKE_CASE ={} self.num_vertices += 1 def _lowerCAmelCase ( self : int ,snake_case : Dict ,snake_case : Any ,snake_case : str ): self.add_vertex(_lowerCamelCase ) self.add_vertex(_lowerCamelCase ) if head == tail: return SCREAMING_SNAKE_CASE =weight SCREAMING_SNAKE_CASE =weight def _lowerCAmelCase ( self : Dict ): SCREAMING_SNAKE_CASE =self.get_edges() for edge in edges: SCREAMING_SNAKE_CASE =edge edges.remove((tail, head, weight) ) for i in range(len(_lowerCamelCase ) ): SCREAMING_SNAKE_CASE =list(edges[i] ) edges.sort(key=lambda snake_case : e[2] ) for i in range(len(_lowerCamelCase ) - 1 ): if edges[i][2] >= edges[i + 1][2]: SCREAMING_SNAKE_CASE =edges[i][2] + 1 for edge in edges: SCREAMING_SNAKE_CASE =edge SCREAMING_SNAKE_CASE =weight SCREAMING_SNAKE_CASE =weight def __str__( self : List[str] ): SCREAMING_SNAKE_CASE ="""""" for tail in self.adjacency: for head in self.adjacency[tail]: SCREAMING_SNAKE_CASE =self.adjacency[head][tail] string += f'{head} -> {tail} == {weight}\n' return string.rstrip('\n' ) def _lowerCAmelCase ( self : Optional[Any] ): SCREAMING_SNAKE_CASE =[] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def _lowerCAmelCase ( self : List[Any] ): return self.adjacency.keys() @staticmethod def _lowerCAmelCase ( snake_case : List[Any]=None ,snake_case : Dict=None ): SCREAMING_SNAKE_CASE =Graph() if vertices is None: SCREAMING_SNAKE_CASE =[] if edges is None: SCREAMING_SNAKE_CASE =[] for vertex in vertices: g.add_vertex(_lowerCamelCase ) for edge in edges: g.add_edge(*_lowerCamelCase ) return g class a_ : """simple docstring""" def __init__( self : Optional[int] ): SCREAMING_SNAKE_CASE ={} SCREAMING_SNAKE_CASE ={} def __len__( self : List[str] ): return len(self.parent ) def _lowerCAmelCase ( self : Dict ,snake_case : Dict ): if item in self.parent: return self.find(_lowerCamelCase ) SCREAMING_SNAKE_CASE =item SCREAMING_SNAKE_CASE =0 return item def _lowerCAmelCase ( self : Dict ,snake_case : List[Any] ): if item not in self.parent: return self.make_set(_lowerCamelCase ) if item != self.parent[item]: SCREAMING_SNAKE_CASE =self.find(self.parent[item] ) return self.parent[item] def _lowerCAmelCase ( self : List[Any] ,snake_case : List[str] ,snake_case : Any ): SCREAMING_SNAKE_CASE =self.find(_lowerCamelCase ) SCREAMING_SNAKE_CASE =self.find(_lowerCamelCase ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: SCREAMING_SNAKE_CASE =roota return roota if self.rank[roota] < self.rank[roota]: SCREAMING_SNAKE_CASE =roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 SCREAMING_SNAKE_CASE =roota return roota return None @staticmethod def _lowerCAmelCase ( snake_case : Any ): SCREAMING_SNAKE_CASE =graph.num_vertices SCREAMING_SNAKE_CASE =Graph.UnionFind() SCREAMING_SNAKE_CASE =[] while num_components > 1: SCREAMING_SNAKE_CASE ={} for vertex in graph.get_vertices(): SCREAMING_SNAKE_CASE =-1 SCREAMING_SNAKE_CASE =graph.get_edges() for edge in edges: SCREAMING_SNAKE_CASE =edge edges.remove((tail, head, weight) ) for edge in edges: SCREAMING_SNAKE_CASE =edge SCREAMING_SNAKE_CASE =union_find.find(_lowerCamelCase ) SCREAMING_SNAKE_CASE =union_find.find(_lowerCamelCase ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: SCREAMING_SNAKE_CASE =[head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: SCREAMING_SNAKE_CASE =[head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: SCREAMING_SNAKE_CASE =cheap_edge[vertex] if union_find.find(_lowerCamelCase ) != union_find.find(_lowerCamelCase ): union_find.union(_lowerCamelCase ,_lowerCamelCase ) mst_edges.append(cheap_edge[vertex] ) SCREAMING_SNAKE_CASE =num_components - 1 SCREAMING_SNAKE_CASE =Graph.build(edges=_lowerCamelCase ) return mst
334
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ : Any = logging.get_logger(__name__) UpperCamelCase__ : Optional[int] = { 'distilbert-base-uncased': 'https://huggingface.co/distilbert-base-uncased/resolve/main/config.json', 'distilbert-base-uncased-distilled-squad': ( 'https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-cased': 'https://huggingface.co/distilbert-base-cased/resolve/main/config.json', 'distilbert-base-cased-distilled-squad': ( 'https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json' ), 'distilbert-base-german-cased': 'https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json', 'distilbert-base-multilingual-cased': ( 'https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json' ), 'distilbert-base-uncased-finetuned-sst-2-english': ( 'https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json' ), } class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''distilbert''' lowerCamelCase = { '''hidden_size''': '''dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', } def __init__( self , _lowerCamelCase=3_0522 , _lowerCamelCase=512 , _lowerCamelCase=False , _lowerCamelCase=6 , _lowerCamelCase=12 , _lowerCamelCase=768 , _lowerCamelCase=4 * 768 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase="gelu" , _lowerCamelCase=0.02 , _lowerCamelCase=0.1 , _lowerCamelCase=0.2 , _lowerCamelCase=0 , **_lowerCamelCase , ) -> Optional[Any]: A_ : Tuple = vocab_size A_ : List[Any] = max_position_embeddings A_ : int = sinusoidal_pos_embds A_ : int = n_layers A_ : str = n_heads A_ : Optional[int] = dim A_ : int = hidden_dim A_ : Tuple = dropout A_ : List[Any] = attention_dropout A_ : int = activation A_ : Dict = initializer_range A_ : List[Any] = qa_dropout A_ : int = seq_classif_dropout super().__init__(**_lowerCamelCase , pad_token_id=_lowerCamelCase ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": A_ : Union[str, Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A_ : int = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
344
0
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class A ( __A ): UpperCamelCase_ : Dict =42 @flax_register_to_config class A ( nn.Module , __A , __A ): UpperCamelCase_ : str =32 UpperCamelCase_ : List[str] =4 UpperCamelCase_ : Any =4 UpperCamelCase_ : Any =( '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''', '''DownBlock2D''', ) UpperCamelCase_ : Optional[Any] =('''UpBlock2D''', '''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''') UpperCamelCase_ : str =False UpperCamelCase_ : Dict =(320, 640, 1_280, 1_280) UpperCamelCase_ : List[Any] =2 UpperCamelCase_ : Union[str, Any] =8 UpperCamelCase_ : Tuple =None UpperCamelCase_ : List[Any] =1_280 UpperCamelCase_ : str =0.0 UpperCamelCase_ : List[Any] =False UpperCamelCase_ : List[str] =jnp.floataa UpperCamelCase_ : Optional[int] =True UpperCamelCase_ : int =0 UpperCamelCase_ : int =False def _A (self , lowerCAmelCase ): # init input tensors __lowercase= (1, self.in_channels, self.sample_size, self.sample_size) __lowercase= jnp.zeros(_lowerCamelCase , dtype=jnp.floataa ) __lowercase= jnp.ones((1,) , dtype=jnp.intaa ) __lowercase= jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) __lowercase= jax.random.split(_lowerCamelCase ) __lowercase= {"""params""": params_rng, """dropout""": dropout_rng} return self.init(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase )["params"] def _A (self ): __lowercase= self.block_out_channels __lowercase= block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( 'At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.' ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. __lowercase= self.num_attention_heads or self.attention_head_dim # input __lowercase= nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time __lowercase= FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) __lowercase= FlaxTimestepEmbedding(_lowerCamelCase , dtype=self.dtype ) __lowercase= self.only_cross_attention if isinstance(_lowerCamelCase , _lowerCamelCase ): __lowercase= (only_cross_attention,) * len(self.down_block_types ) if isinstance(_lowerCamelCase , _lowerCamelCase ): __lowercase= (num_attention_heads,) * len(self.down_block_types ) # down __lowercase= [] __lowercase= block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): __lowercase= output_channel __lowercase= block_out_channels[i] __lowercase= i == len(_lowerCamelCase ) - 1 if down_block_type == "CrossAttnDownBlock2D": __lowercase= FlaxCrossAttnDownBlockaD( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: __lowercase= FlaxDownBlockaD( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(_lowerCamelCase ) __lowercase= down_blocks # mid __lowercase= FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up __lowercase= [] __lowercase= list(reversed(_lowerCamelCase ) ) __lowercase= list(reversed(_lowerCamelCase ) ) __lowercase= list(reversed(_lowerCamelCase ) ) __lowercase= reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): __lowercase= output_channel __lowercase= reversed_block_out_channels[i] __lowercase= reversed_block_out_channels[min(i + 1 , len(_lowerCamelCase ) - 1 )] __lowercase= i == len(_lowerCamelCase ) - 1 if up_block_type == "CrossAttnUpBlock2D": __lowercase= FlaxCrossAttnUpBlockaD( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , prev_output_channel=_lowerCamelCase , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: __lowercase= FlaxUpBlockaD( in_channels=_lowerCamelCase , out_channels=_lowerCamelCase , prev_output_channel=_lowerCamelCase , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(_lowerCamelCase ) __lowercase= output_channel __lowercase= up_blocks # out __lowercase= nn.GroupNorm(num_groups=3_2 , epsilon=1E-5 ) __lowercase= nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__(self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase = True , lowerCAmelCase = False , ): # 1. time if not isinstance(_lowerCamelCase , jnp.ndarray ): __lowercase= jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(_lowerCamelCase , jnp.ndarray ) and len(timesteps.shape ) == 0: __lowercase= timesteps.astype(dtype=jnp.floataa ) __lowercase= jnp.expand_dims(_lowerCamelCase , 0 ) __lowercase= self.time_proj(_lowerCamelCase ) __lowercase= self.time_embedding(_lowerCamelCase ) # 2. pre-process __lowercase= jnp.transpose(_lowerCamelCase , (0, 2, 3, 1) ) __lowercase= self.conv_in(_lowerCamelCase ) # 3. down __lowercase= (sample,) for down_block in self.down_blocks: if isinstance(_lowerCamelCase , _lowerCamelCase ): __lowercase= down_block(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , deterministic=not train ) else: __lowercase= down_block(_lowerCamelCase , _lowerCamelCase , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: __lowercase= () for down_block_res_sample, down_block_additional_residual in zip( _lowerCamelCase , _lowerCamelCase ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) __lowercase= new_down_block_res_samples # 4. mid __lowercase= self.mid_block(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: __lowercase= down_block_res_samples[-(self.layers_per_block + 1) :] __lowercase= down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(_lowerCamelCase , _lowerCamelCase ): __lowercase= up_block( _lowerCamelCase , temb=_lowerCamelCase , encoder_hidden_states=_lowerCamelCase , res_hidden_states_tuple=_lowerCamelCase , deterministic=not train , ) else: __lowercase= up_block(_lowerCamelCase , temb=_lowerCamelCase , res_hidden_states_tuple=_lowerCamelCase , deterministic=not train ) # 6. post-process __lowercase= self.conv_norm_out(_lowerCamelCase ) __lowercase= nn.silu(_lowerCamelCase ) __lowercase= self.conv_out(_lowerCamelCase ) __lowercase= jnp.transpose(_lowerCamelCase , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=_lowerCamelCase )
295
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ : int = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) UpperCamelCase__ : Any = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.cross_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.weight', f'decoder.layers.{i}.sa_qcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.weight', f'decoder.layers.{i}.sa_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qpos_proj.weight', f'decoder.layers.{i}.sa_qpos_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kpos_proj.weight', f'decoder.layers.{i}.sa_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.weight', f'decoder.layers.{i}.sa_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.weight', f'decoder.layers.{i}.ca_qcontent_proj.weight') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.weight', f'decoder.layers.{i}.ca_kcontent_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kpos_proj.weight', f'decoder.layers.{i}.ca_kpos_proj.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.weight', f'decoder.layers.{i}.ca_v_proj.weight')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight', f'decoder.layers.{i}.ca_qpos_sine_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_qcontent_proj.bias', f'decoder.layers.{i}.sa_qcontent_proj.bias') ) rename_keys.append( (f'transformer.decoder.layers.{i}.sa_kcontent_proj.bias', f'decoder.layers.{i}.sa_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.sa_qpos_proj.bias', f'decoder.layers.{i}.sa_qpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_kpos_proj.bias', f'decoder.layers.{i}.sa_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.sa_v_proj.bias', f'decoder.layers.{i}.sa_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qcontent_proj.bias', f'decoder.layers.{i}.ca_qcontent_proj.bias') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_kcontent_proj.bias', f'decoder.layers.{i}.ca_kcontent_proj.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.ca_kpos_proj.bias', f'decoder.layers.{i}.ca_kpos_proj.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.ca_v_proj.bias', f'decoder.layers.{i}.ca_v_proj.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias', f'decoder.layers.{i}.ca_qpos_sine_proj.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ('transformer.decoder.ref_point_head.layers.0.weight', 'decoder.ref_point_head.layers.0.weight'), ('transformer.decoder.ref_point_head.layers.0.bias', 'decoder.ref_point_head.layers.0.bias'), ('transformer.decoder.ref_point_head.layers.1.weight', 'decoder.ref_point_head.layers.1.weight'), ('transformer.decoder.ref_point_head.layers.1.bias', 'decoder.ref_point_head.layers.1.bias'), ('transformer.decoder.query_scale.layers.0.weight', 'decoder.query_scale.layers.0.weight'), ('transformer.decoder.query_scale.layers.0.bias', 'decoder.query_scale.layers.0.bias'), ('transformer.decoder.query_scale.layers.1.weight', 'decoder.query_scale.layers.1.weight'), ('transformer.decoder.query_scale.layers.1.bias', 'decoder.query_scale.layers.1.bias'), ('transformer.decoder.layers.0.ca_qpos_proj.weight', 'decoder.layers.0.ca_qpos_proj.weight'), ('transformer.decoder.layers.0.ca_qpos_proj.bias', 'decoder.layers.0.ca_qpos_proj.bias'), ] ) def UpperCAmelCase ( a_ , a_ , a_ ) -> Optional[Any]: """simple docstring""" A_ : int = state_dict.pop(a_ ) A_ : Tuple = val def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : Union[str, Any] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: A_ : Optional[int] = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) A_ : str = value else: A_ : int = value return new_state_dict def UpperCAmelCase ( a_ , a_=False ) -> Optional[int]: """simple docstring""" A_ : List[Any] = """""" if is_panoptic: A_ : Any = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) A_ : Optional[int] = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) A_ : str = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict A_ : Optional[Any] = in_proj_weight[:2_5_6, :] A_ : Tuple = in_proj_bias[:2_5_6] A_ : Dict = in_proj_weight[2_5_6:5_1_2, :] A_ : int = in_proj_bias[2_5_6:5_1_2] A_ : int = in_proj_weight[-2_5_6:, :] A_ : Optional[int] = in_proj_bias[-2_5_6:] def UpperCAmelCase ( ) -> Dict: """simple docstring""" A_ : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A_ : List[Any] = Image.open(requests.get(a_ , stream=a_ ).raw ) return im @torch.no_grad() def UpperCAmelCase ( a_ , a_ ) -> Dict: """simple docstring""" A_ : int = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: A_ : str = """resnet101""" if "dc5" in model_name: A_ : List[Any] = True A_ : str = """panoptic""" in model_name if is_panoptic: A_ : Dict = 2_5_0 else: A_ : Union[str, Any] = 9_1 A_ : str = """huggingface/label-files""" A_ : Union[str, Any] = """coco-detection-id2label.json""" A_ : Optional[Any] = json.load(open(hf_hub_download(a_ , a_ , repo_type="""dataset""" ) , """r""" ) ) A_ : str = {int(a_ ): v for k, v in idalabel.items()} A_ : Optional[int] = idalabel A_ : Tuple = {v: k for k, v in idalabel.items()} # load image processor A_ : List[Any] = """coco_panoptic""" if is_panoptic else """coco_detection""" A_ : Any = ConditionalDetrImageProcessor(format=a_ ) # prepare image A_ : Tuple = prepare_img() A_ : Any = image_processor(images=a_ , return_tensors="""pt""" ) A_ : Optional[int] = encoding["""pixel_values"""] logger.info(F"Converting model {model_name}..." ) # load original model from torch hub A_ : int = torch.hub.load("""DeppMeng/ConditionalDETR""" , a_ , pretrained=a_ ).eval() A_ : List[Any] = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: A_ : Union[str, Any] = """conditional_detr.""" + src rename_key(a_ , a_ , a_ ) A_ : Any = rename_backbone_keys(a_ ) # query, key and value matrices need special treatment read_in_q_k_v(a_ , is_panoptic=a_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them A_ : List[str] = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): A_ : Dict = state_dict.pop(a_ ) A_ : List[Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: A_ : str = state_dict.pop(a_ ) A_ : Any = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: A_ : Optional[int] = state_dict.pop(a_ ) A_ : str = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): A_ : Tuple = state_dict.pop(a_ ) A_ : Dict = val # finally, create HuggingFace model and load state dict A_ : Union[str, Any] = ConditionalDetrForSegmentation(a_ ) if is_panoptic else ConditionalDetrForObjectDetection(a_ ) model.load_state_dict(a_ ) model.eval() model.push_to_hub(repo_id=a_ , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion A_ : str = conditional_detr(a_ ) A_ : str = model(a_ ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(a_ ).mkdir(exist_ok=a_ ) model.save_pretrained(a_ ) image_processor.save_pretrained(a_ ) if __name__ == "__main__": UpperCamelCase__ : int = argparse.ArgumentParser() parser.add_argument( '--model_name', default='conditional_detr_resnet50', type=str, help='Name of the CONDITIONAL_DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) UpperCamelCase__ : Optional[Any] = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
344
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase : int = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
3
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = (UnCLIPScheduler,) def UpperCAmelCase_ ( self , **_lowerCamelCase ) -> List[Any]: A_ : Union[str, Any] = { """num_train_timesteps""": 1000, """variance_type""": """fixed_small_log""", """clip_sample""": True, """clip_sample_range""": 1.0, """prediction_type""": """epsilon""", } config.update(**_lowerCamelCase ) return config def UpperCAmelCase_ ( self ) -> List[Any]: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample in [True, False]: self.check_over_configs(clip_sample=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[int]: for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=_lowerCamelCase , prev_timestep=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[Any]: A_ : Optional[int] = self.scheduler_classes[0] A_ : Any = self.get_scheduler_config(variance_type="""fixed_small_log""" ) A_ : List[Any] = scheduler_class(**_lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.054_9625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.999_4987 ) ) < 1e-5 def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : List[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config(variance_type="""learned_range""" ) A_ : Dict = scheduler_class(**_lowerCamelCase ) A_ : Dict = 0.5 assert scheduler._get_variance(1 , predicted_variance=_lowerCamelCase ) - -10.171_2790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=_lowerCamelCase ) - -5.799_8052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=_lowerCamelCase ) - -0.001_0011 < 1e-5 def UpperCAmelCase_ ( self ) -> Any: A_ : Optional[Any] = self.scheduler_classes[0] A_ : Tuple = self.get_scheduler_config() A_ : Optional[Any] = scheduler_class(**_lowerCamelCase ) A_ : int = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : str = self.dummy_sample_deter A_ : Optional[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Any = model(_lowerCamelCase , _lowerCamelCase ) # 2. predict previous mean of sample x_t-1 A_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : List[Any] = pred_prev_sample A_ : Any = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Optional[Any] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 252.268_2495 ) < 1e-2 assert abs(result_mean.item() - 0.328_4743 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Dict: A_ : Union[str, Any] = self.scheduler_classes[0] A_ : Dict = self.get_scheduler_config() A_ : Tuple = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(25 ) A_ : List[str] = scheduler.timesteps A_ : List[Any] = self.dummy_model() A_ : List[Any] = self.dummy_sample_deter A_ : List[Any] = torch.manual_seed(0 ) for i, t in enumerate(_lowerCamelCase ): # 1. predict noise residual A_ : Optional[Any] = model(_lowerCamelCase , _lowerCamelCase ) if i + 1 == timesteps.shape[0]: A_ : List[str] = None else: A_ : Dict = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 A_ : str = scheduler.step( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , prev_timestep=_lowerCamelCase , generator=_lowerCamelCase ).prev_sample A_ : Optional[Any] = pred_prev_sample A_ : Dict = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : List[str] = torch.mean(torch.abs(_lowerCamelCase ) ) assert abs(result_sum.item() - 258.204_4983 ) < 1e-2 assert abs(result_mean.item() - 0.336_2038 ) < 1e-3 def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> int: pass
344
0
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __A ( ) -> Any: raise RuntimeError("""CUDA out of memory.""" ) class __lowerCAmelCase ( nn.Module ): def __init__( self :List[str] ): '''simple docstring''' super().__init__() a = nn.Linear(3 , 4 ) a = nn.BatchNormad(4 ) a = nn.Linear(4 , 5 ) def lowerCamelCase__ ( self :str , __magic_name__ :List[Any] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(_lowerCamelCase ) ) ) class __lowerCAmelCase ( unittest.TestCase ): def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(__magic_name__ :Tuple ): nonlocal batch_sizes batch_sizes.append(_lowerCamelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(_lowerCamelCase , [128, 64, 32, 16, 8] ) def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' a = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(__magic_name__ :Optional[int] , __magic_name__ :str ): nonlocal batch_sizes batch_sizes.append(_lowerCamelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga a = mock_training_loop_function("""hello""" ) self.assertListEqual(_lowerCamelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, """hello"""] ) def lowerCamelCase__ ( self :Any ): '''simple docstring''' @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(__magic_name__ :Dict ): pass with self.assertRaises(_lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("""No executable batch size found, reached zero.""" , cm.exception.args[0] ) def lowerCamelCase__ ( self :Dict ): '''simple docstring''' @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(__magic_name__ :Union[str, Any] ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(_lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("""No executable batch size found, reached zero.""" , cm.exception.args[0] ) def lowerCamelCase__ ( self :Any ): '''simple docstring''' @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(__magic_name__ :Optional[Any] , __magic_name__ :List[Any] , __magic_name__ :Tuple ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(_lowerCamelCase ) as cm: mock_training_loop_function(128 , """hello""" , """world""" ) self.assertIn("""Batch size was passed into `f`""" , cm.exception.args[0] ) self.assertIn("""`f(arg1='hello', arg2='world')""" , cm.exception.args[0] ) def lowerCamelCase__ ( self :str ): '''simple docstring''' @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(__magic_name__ :Union[str, Any] ): raise ValueError("""Oops, we had an error!""" ) with self.assertRaises(_lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("""Oops, we had an error!""" , cm.exception.args[0] ) @require_cuda def lowerCamelCase__ ( self :int ): '''simple docstring''' a = torch.cuda.memory_allocated() a = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , _lowerCamelCase ) a = release_memory(_lowerCamelCase ) self.assertEqual(torch.cuda.memory_allocated() , _lowerCamelCase )
228
'''simple docstring''' import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=7 , _lowerCamelCase=3 , _lowerCamelCase=18 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=[0.5, 0.5, 0.5] , _lowerCamelCase=False , ) -> Optional[int]: A_ : Union[str, Any] = size if size is not None else {"""height""": 20, """width""": 20} A_ : Tuple = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} A_ : Optional[Any] = parent A_ : Optional[int] = batch_size A_ : Union[str, Any] = num_channels A_ : str = image_size A_ : Tuple = min_resolution A_ : Dict = max_resolution A_ : str = do_resize A_ : Tuple = size A_ : int = do_center_crop A_ : Dict = crop_size A_ : Tuple = do_normalize A_ : List[str] = image_mean A_ : Optional[Any] = image_std A_ : Any = do_reduce_labels def UpperCAmelCase_ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def UpperCAmelCase ( ) -> List[str]: """simple docstring""" A_ : Any = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(dataset[0]["""file"""] ) A_ : Dict = Image.open(dataset[1]["""file"""] ) return image, map def UpperCAmelCase ( ) -> Optional[int]: """simple docstring""" A_ : Tuple = load_dataset("""hf-internal-testing/fixtures_ade20k""" , split="""test""" ) A_ : Tuple = Image.open(ds[0]["""file"""] ) A_ : List[Any] = Image.open(ds[1]["""file"""] ) A_ : Any = Image.open(ds[2]["""file"""] ) A_ : str = Image.open(ds[3]["""file"""] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BeitImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : List[Any] = BeitImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[int]: A_ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """center_crop""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 20, """width""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) A_ : int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=_lowerCamelCase ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) self.assertEqual(image_processor.do_reduce_labels , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> Dict: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image_processing A_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> str: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A_ : Union[str, Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def UpperCAmelCase_ ( self ) -> Optional[int]: # Initialize image_processing A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) A_ : Optional[int] = [] for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input A_ : Union[str, Any] = image_processing(image_inputs[0] , maps[0] , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test not batched input (PIL images) A_ , A_ : List[Any] = prepare_semantic_single_inputs() A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 1, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) # Test batched input (PIL images) A_ , A_ : str = prepare_semantic_batch_inputs() A_ : Any = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertEqual( encoding["""pixel_values"""].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual( encoding["""labels"""].shape , ( 2, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) self.assertEqual(encoding["""labels"""].dtype , torch.long ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image_processing A_ : Any = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 A_ , A_ : Tuple = prepare_semantic_single_inputs() A_ : str = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 150 ) A_ : str = True A_ : Union[str, Any] = image_processing(_lowerCamelCase , _lowerCamelCase , return_tensors="""pt""" ) self.assertTrue(encoding["""labels"""].min().item() >= 0 ) self.assertTrue(encoding["""labels"""].max().item() <= 255 )
344
0
'''simple docstring''' import fire from utils import calculate_rouge, save_json def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE ): _snake_case = [x.strip() for x in open(a_ ).readlines()] _snake_case = [x.strip() for x in open(a_ ).readlines()][: len(a_ )] _snake_case = calculate_rouge(a_ , a_ , **a_ ) if save_path is not None: save_json(a_ , a_ , indent=a_ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
341
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class _lowerCAmelCase ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None ) -> str: super().__init__() A_ : Optional[Any] = pad_token_id A_ : List[Any] = max_length A_ : str = vocab A_ : Union[str, Any] = merges A_ : List[Any] = BytePairTokenizer(_lowerCamelCase , _lowerCamelCase , sequence_length=_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> int: A_ : Tuple = [""" """.join(_lowerCamelCase ) for m in tokenizer.bpe_ranks.keys()] A_ : Dict = tokenizer.get_vocab() return cls(_lowerCamelCase , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) -> str: A_ : Tuple = GPTaTokenizer.from_pretrained(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) return cls.from_tokenizer(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase ) -> List[Any]: return cls(**_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ) -> Any: A_ : List[Any] = self.tf_tokenizer(_lowerCamelCase ) A_ : Any = tf.ones_like(_lowerCamelCase ) if self.pad_token_id is not None: # pad the tokens up to max length A_ : List[Any] = max_length if max_length is not None else self.max_length if max_length is not None: A_ , A_ : Tuple = pad_model_inputs( _lowerCamelCase , max_seq_length=_lowerCamelCase , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
344
0
'''simple docstring''' from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class lowerCAmelCase_ : __lowerCamelCase : str = 42 __lowerCamelCase : int = 42 class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase ) -> int: _lowerCAmelCase = [[] for _ in range(_lowerCamelCase )] _lowerCAmelCase = size def __getitem__( self , _lowerCAmelCase ) -> Iterator[Edge]: return iter(self._graph[vertex] ) @property def _snake_case ( self ) -> Tuple: return self._size def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: if weight not in (0, 1): raise ValueError("Edge weight must be either 0 or 1." ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError("Vertex indexes must be in [0; size)." ) self._graph[from_vertex].append(Edge(_lowerCamelCase , _lowerCamelCase ) ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase ) -> int | None: _lowerCAmelCase = deque([start_vertex] ) _lowerCAmelCase = [None] * self.size _lowerCAmelCase = 0 while queue: _lowerCAmelCase = queue.popleft() _lowerCAmelCase = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: _lowerCAmelCase = current_distance + edge.weight _lowerCAmelCase = distances[edge.destination_vertex] if ( isinstance(_lowerCamelCase , _lowerCamelCase ) and new_distance >= dest_vertex_distance ): continue _lowerCAmelCase = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError("No path from start_vertex to finish_vertex." ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
158
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase__ : Optional[int] = {'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : int = ['YolosFeatureExtractor'] UpperCamelCase__ : int = ['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Dict = [ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys UpperCamelCase__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
from __future__ import annotations __A : List[str] = list[list[int]] # assigning initial values to the grid __A : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution __A : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> bool: '''simple docstring''' for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> tuple[int, int] | None: '''simple docstring''' for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> Matrix | None: '''simple docstring''' if location := find_empty_location(a_ ): lowerCAmelCase : Optional[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1, 10 ): if is_safe(a_, a_, a_, a_ ): lowerCAmelCase : str = digit if sudoku(a_ ) is not None: return grid lowerCAmelCase : Optional[int] = 0 return None def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> None: '''simple docstring''' for row in grid: for cell in row: print(a_, end=' ' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('''\nExample grid:\n''' + '''=''' * 20) print_solution(example_grid) print('''\nExample grid solution:''') __A : Any = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('''Cannot find a solution.''')
138
'''simple docstring''' class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Optional[Any] = name A_ : Dict = value A_ : Union[str, Any] = weight def __repr__( self ) -> List[str]: return F"{self.__class__.__name__}({self.name}, {self.value}, {self.weight})" def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.value def UpperCAmelCase_ ( self ) -> List[str]: return self.name def UpperCAmelCase_ ( self ) -> Tuple: return self.weight def UpperCAmelCase_ ( self ) -> Optional[int]: return self.value / self.weight def UpperCAmelCase ( a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Optional[int] = [] for i in range(len(a_ ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def UpperCAmelCase ( a_ , a_ , a_ ) -> List[Any]: """simple docstring""" A_ : Optional[Any] = sorted(a_ , key=a_ , reverse=a_ ) A_ : str = [] A_ , A_ : Dict = 0.0, 0.0 for i in range(len(a_ ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
344
0
UpperCAmelCase = '0.18.2' from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_note_seq_available, is_onnx_available, is_scipy_available, is_torch_available, is_torchsde_available, is_transformers_available, is_transformers_version, is_unidecode_available, logging, ) try: if not is_onnx_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_onnx_objects import * # noqa F403 else: from .pipelines import OnnxRuntimeModel try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_pt_objects import * # noqa F403 else: from .models import ( AutoencoderKL, ControlNetModel, ModelMixin, PriorTransformer, TaFilmDecoder, TransformeraDModel, UNetaDModel, UNetaDConditionModel, UNetaDModel, UNetaDConditionModel, VQModel, ) from .optimization import ( get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, get_scheduler, ) from .pipelines import ( AudioPipelineOutput, ConsistencyModelPipeline, DanceDiffusionPipeline, DDIMPipeline, DDPMPipeline, DiffusionPipeline, DiTPipeline, ImagePipelineOutput, KarrasVePipeline, LDMPipeline, LDMSuperResolutionPipeline, PNDMPipeline, RePaintPipeline, ScoreSdeVePipeline, ) from .schedulers import ( CMStochasticIterativeScheduler, DDIMInverseScheduler, DDIMParallelScheduler, DDIMScheduler, DDPMParallelScheduler, DDPMScheduler, DEISMultistepScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, HeunDiscreteScheduler, IPNDMScheduler, KarrasVeScheduler, KDPMaAncestralDiscreteScheduler, KDPMaDiscreteScheduler, PNDMScheduler, RePaintScheduler, SchedulerMixin, ScoreSdeVeScheduler, UnCLIPScheduler, UniPCMultistepScheduler, VQDiffusionScheduler, ) from .training_utils import EMAModel try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .schedulers import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .schedulers import DPMSolverSDEScheduler try: if not (is_torch_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipelines import ( AltDiffusionImgaImgPipeline, AltDiffusionPipeline, AudioLDMPipeline, CycleDiffusionPipeline, IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ImageTextPipelineOutput, KandinskyImgaImgPipeline, KandinskyInpaintPipeline, KandinskyPipeline, KandinskyPriorPipeline, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaControlnetPipeline, KandinskyVaaImgaImgPipeline, KandinskyVaaInpaintPipeline, KandinskyVaaPipeline, KandinskyVaaPriorEmbaEmbPipeline, KandinskyVaaPriorPipeline, LDMTextToImagePipeline, PaintByExamplePipeline, SemanticStableDiffusionPipeline, ShapEImgaImgPipeline, ShapEPipeline, StableDiffusionAttendAndExcitePipeline, StableDiffusionControlNetImgaImgPipeline, StableDiffusionControlNetInpaintPipeline, StableDiffusionControlNetPipeline, StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionImageVariationPipeline, StableDiffusionImgaImgPipeline, StableDiffusionInpaintPipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionInstructPixaPixPipeline, StableDiffusionLatentUpscalePipeline, StableDiffusionLDMaDPipeline, StableDiffusionModelEditingPipeline, StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPipeline, StableDiffusionPipelineSafe, StableDiffusionPixaPixZeroPipeline, StableDiffusionSAGPipeline, StableDiffusionUpscalePipeline, StableUnCLIPImgaImgPipeline, StableUnCLIPPipeline, TextToVideoSDPipeline, TextToVideoZeroPipeline, UnCLIPImageVariationPipeline, UnCLIPPipeline, UniDiffuserModel, UniDiffuserPipeline, UniDiffuserTextDecoder, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, VideoToVideoSDPipeline, VQDiffusionPipeline, ) try: if not (is_torch_available() and is_transformers_available() and is_invisible_watermark_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_invisible_watermark_objects import * # noqa F403 else: from .pipelines import StableDiffusionXLImgaImgPipeline, StableDiffusionXLPipeline try: if not (is_torch_available() and is_transformers_available() and is_k_diffusion_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipelines import StableDiffusionKDiffusionPipeline try: if not (is_torch_available() and is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_onnx_objects import * # noqa F403 else: from .pipelines import ( OnnxStableDiffusionImgaImgPipeline, OnnxStableDiffusionInpaintPipeline, OnnxStableDiffusionInpaintPipelineLegacy, OnnxStableDiffusionPipeline, OnnxStableDiffusionUpscalePipeline, StableDiffusionOnnxPipeline, ) try: if not (is_torch_available() and is_librosa_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_librosa_objects import * # noqa F403 else: from .pipelines import AudioDiffusionPipeline, Mel try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .pipelines import SpectrogramDiffusionPipeline try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_objects import * # noqa F403 else: from .models.controlnet_flax import FlaxControlNetModel from .models.modeling_flax_utils import FlaxModelMixin from .models.unet_ad_condition_flax import FlaxUNetaDConditionModel from .models.vae_flax import FlaxAutoencoderKL from .pipelines import FlaxDiffusionPipeline from .schedulers import ( FlaxDDIMScheduler, FlaxDDPMScheduler, FlaxDPMSolverMultistepScheduler, FlaxKarrasVeScheduler, FlaxLMSDiscreteScheduler, FlaxPNDMScheduler, FlaxSchedulerMixin, FlaxScoreSdeVeScheduler, ) try: if not (is_flax_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_and_transformers_objects import * # noqa F403 else: from .pipelines import ( FlaxStableDiffusionControlNetPipeline, FlaxStableDiffusionImgaImgPipeline, FlaxStableDiffusionInpaintPipeline, FlaxStableDiffusionPipeline, ) try: if not (is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_note_seq_objects import * # noqa F403 else: from .pipelines import MidiProcessor
195
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCAmelCase ( a_ , a_ ) -> tuple: """simple docstring""" if inductance <= 0: raise ValueError("""Inductance cannot be 0 or negative""" ) elif capacitance <= 0: raise ValueError("""Capacitance cannot be 0 or negative""" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
344
0
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) __UpperCAmelCase = '\\n Text data.\n Second line of data.' __UpperCAmelCase = 'file' @pytest.fixture(scope='session' ) def UpperCamelCase ( snake_case__ : List[Any] ) -> Optional[int]: UpperCamelCase : int = tmp_path_factory.mktemp('data' ) / (FILE_PATH + """.zstd""") UpperCamelCase : int = bytes(a_ , 'utf-8' ) with zstd.open(a_ , 'wb' ) as f: f.write(a_ ) return path @pytest.fixture def UpperCamelCase ( snake_case__ : Dict ) -> Optional[int]: with open(os.path.join(tmpfs.local_root_dir , a_ ) , 'w' ) as f: f.write(a_ ) return FILE_PATH @pytest.mark.parametrize('compression_format' , ['gzip', 'xz', 'zstd'] ) def UpperCamelCase ( snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : int , snake_case__ : Optional[Any] , snake_case__ : List[Any] , snake_case__ : Optional[int] ) -> Optional[int]: UpperCamelCase : List[str] = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} UpperCamelCase : Any = input_paths[compression_format] UpperCamelCase : Tuple = tmp_path / """cache""" UpperCamelCase : Tuple = DownloadConfig(cache_dir=a_ , extract_compressed_file=a_ ) UpperCamelCase : Dict = cached_path(a_ , download_config=a_ ) with open(a_ ) as f: UpperCamelCase : Optional[Any] = f.read() with open(a_ ) as f: UpperCamelCase : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('default_extracted' , [True, False] ) @pytest.mark.parametrize('default_cache_dir' , [True, False] ) def UpperCamelCase ( snake_case__ : Optional[int] , snake_case__ : Any , snake_case__ : List[str] , snake_case__ : Tuple , snake_case__ : Optional[int] ) -> str: UpperCamelCase : Union[str, Any] = """custom_cache""" UpperCamelCase : List[str] = """custom_extracted_dir""" UpperCamelCase : Optional[Any] = tmp_path / """custom_extracted_path""" if default_extracted: UpperCamelCase : Any = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_DIR' , a_ ) monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(a_ ) ) UpperCamelCase : Union[str, Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) UpperCamelCase : List[Any] = xz_file UpperCamelCase : Optional[int] = ( DownloadConfig(extract_compressed_file=a_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=a_ ) ) UpperCamelCase : Union[str, Any] = cached_path(a_ , download_config=a_ ) assert Path(a_ ).parent.parts[-2:] == expected def UpperCamelCase ( snake_case__ : Union[str, Any] ) -> Dict: UpperCamelCase : str = str(Path(a_ ).resolve() ) assert cached_path(a_ ) == text_file # relative path UpperCamelCase : List[str] = str(Path(a_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(a_ ) == text_file def UpperCamelCase ( snake_case__ : Any ) -> int: UpperCamelCase : Optional[Any] = str(tmp_path.resolve() / '__missing_file__.txt' ) with pytest.raises(a_ ): cached_path(a_ ) # relative path UpperCamelCase : Tuple = """./__missing_file__.txt""" with pytest.raises(a_ ): cached_path(a_ ) def UpperCamelCase ( snake_case__ : List[Any] ) -> Tuple: UpperCamelCase : Any = get_from_cache(F"""tmp://{tmpfs_file}""" ) with open(a_ ) as f: UpperCamelCase : List[str] = f.read() assert output_file_content == FILE_CONTENT @patch('datasets.config.HF_DATASETS_OFFLINE' , a_ ) def UpperCamelCase ( ) -> List[str]: with pytest.raises(a_ ): cached_path('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , a_ ) def UpperCamelCase ( snake_case__ : str ) -> Union[str, Any]: UpperCamelCase : List[str] = tmp_path_factory.mktemp('data' ) / """file.html""" with pytest.raises(a_ ): http_get('https://huggingface.co' , temp_file=a_ ) with pytest.raises(a_ ): http_head('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , a_ ) def UpperCamelCase ( snake_case__ : Any ) -> int: UpperCamelCase : List[Any] = tmp_path_factory.mktemp('data' ) / """file.html""" with pytest.raises(a_ ): ftp_get('ftp://huggingface.co' , temp_file=a_ ) with pytest.raises(a_ ): ftp_head('ftp://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , a_ ) def UpperCamelCase ( snake_case__ : List[Any] ) -> Optional[int]: UpperCamelCase : Optional[int] = tmp_path_factory.mktemp('data' ) / """file.html""" with pytest.raises(a_ ): fsspec_get('s3://huggingface.co' , temp_file=a_ ) with pytest.raises(a_ ): fsspec_head('s3://huggingface.co' )
119
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class _lowerCAmelCase ( __A ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase ) -> Any: A_ : List[Any] = parent A_ : int = config_class A_ : int = has_text_modality A_ : str = kwargs A_ : int = common_properties def UpperCAmelCase_ ( self ) -> str: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : Optional[int] = ( ["""hidden_size""", """num_attention_heads""", """num_hidden_layers"""] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["""vocab_size"""] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(_lowerCamelCase , _lowerCamelCase ) , msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(_lowerCamelCase ): try: setattr(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(_lowerCamelCase ): try: A_ : List[str] = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(_lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase , msg=F"`{name} value {idx} expected, but was {getattr(_lowerCamelCase , _lowerCamelCase )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = self.config_class(**self.inputs_dict ) A_ : Optional[int] = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : str = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A_ : List[Any] = os.path.join(_lowerCamelCase , """config.json""" ) config_first.to_json_file(_lowerCamelCase ) A_ : Dict = self.config_class.from_json_file(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Any = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(_lowerCamelCase ) A_ : Union[str, Any] = self.config_class.from_pretrained(_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Optional[int] = self.config_class(**self.inputs_dict ) A_ : List[Any] = """test""" with tempfile.TemporaryDirectory() as tmpdirname: A_ : Any = os.path.join(_lowerCamelCase , _lowerCamelCase ) config_first.save_pretrained(_lowerCamelCase ) A_ : Any = self.config_class.from_pretrained(_lowerCamelCase , subfolder=_lowerCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Tuple = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) A_ : str = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def UpperCAmelCase_ ( self ) -> Optional[Any]: if self.config_class.is_composition: return A_ : Dict = self.config_class() self.parent.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ : Any = copy.deepcopy(_lowerCamelCase ) A_ : Tuple = self.config_class(**_lowerCamelCase ) A_ : Optional[Any] = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("""torch_dtype""", config.torch_dtype, torch.floataa) ) elif getattr(_lowerCamelCase , _lowerCamelCase ) != value: wrong_values.append((key, getattr(_lowerCamelCase , _lowerCamelCase ), value) ) if len(_lowerCamelCase ) > 0: A_ : List[Any] = """\n""".join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def UpperCAmelCase_ ( self ) -> Optional[int]: self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
344
0
"""simple docstring""" from __future__ import annotations def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : list[list[int]] = [] create_all_state(1 , a_ , a_ , [] , a_ ) return result def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): """simple docstring""" if level == 0: total_list.append(current_list[:] ) return for i in range(a_ , total_number - level + 2 ): current_list.append(a_ ) create_all_state(i + 1 , a_ , level - 1 , a_ , a_ ) current_list.pop() def lowercase_ ( _UpperCAmelCase ): """simple docstring""" for i in total_list: print(*a_ ) if __name__ == "__main__": _lowerCamelCase : List[Any] = 4 _lowerCamelCase : str = 2 _lowerCamelCase : Tuple = generate_all_combinations(n, k) print_all_state(total_list)
167
'''simple docstring''' from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" try: with open(a_ , """rb""" ) as flax_state_f: A_ : Tuple = from_bytes(a_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(a_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"Unable to convert {model_file} to Flax deserializable object. " ) return load_flax_weights_in_pytorch_model(a_ , a_ ) def UpperCAmelCase ( a_ , a_ ) -> Any: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights A_ : List[Any] = flatten_dict(jax.tree_util.tree_map(lambda a_ : x.dtype == jnp.bfloataa , a_ ) ).values() if any(a_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) A_ : str = jax.tree_util.tree_map( lambda a_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , a_ ) A_ : Any = """""" A_ : Optional[int] = flatten_dict(a_ , sep=""".""" ) A_ : List[str] = pt_model.state_dict() # keep track of unexpected & missing keys A_ : Union[str, Any] = [] A_ : Dict = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): A_ : List[Any] = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: A_ : Optional[Any] = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[Any] = jnp.transpose(a_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": A_ : int = flax_key_tuple_array[:-1] + ["""weight"""] A_ : Optional[int] = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": A_ : Any = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(a_ ): A_ : Tuple = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) A_ : Dict = """.""".join(a_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected " F"to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}." ) else: # add weight to pytorch dict A_ : Optional[Any] = np.asarray(a_ ) if not isinstance(a_ , np.ndarray ) else flax_tensor A_ : Tuple = torch.from_numpy(a_ ) # remove from missing keys missing_keys.remove(a_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(a_ ) pt_model.load_state_dict(a_ ) # re-transform missing_keys to list A_ : Dict = list(a_ ) if len(a_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing" F" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(a_ ) > 0: logger.warning( F"Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly" F" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to" """ use it for predictions and inference.""" ) return pt_model
344
0
A__ : List[Any] = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def UpperCamelCase( __UpperCamelCase : Optional[Any] ,__UpperCamelCase : str ,__UpperCamelCase : Optional[Any] ,__UpperCamelCase : Union[str, Any] ): lowerCAmelCase_ : Any = [False] * len(a_ ) lowerCAmelCase_ : List[str] = [s] lowerCAmelCase_ : int = True while queue: lowerCAmelCase_ : Tuple = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(a_ ) lowerCAmelCase_ : Union[str, Any] = True lowerCAmelCase_ : List[str] = u return visited[t] def UpperCamelCase( __UpperCamelCase : Dict ,__UpperCamelCase : Dict ,__UpperCamelCase : Dict ): lowerCAmelCase_ : Optional[Any] = [-1] * (len(a_ )) lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : List[Any] = [i[:] for i in graph] # Record original cut, copy. while bfs(a_ ,a_ ,a_ ,a_ ): lowerCAmelCase_ : Any = float('''Inf''' ) lowerCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path lowerCAmelCase_ : Optional[int] = min(a_ ,graph[parent[s]][s] ) lowerCAmelCase_ : Tuple = parent[s] max_flow += path_flow lowerCAmelCase_ : str = sink while v != source: lowerCAmelCase_ : List[Any] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow lowerCAmelCase_ : str = parent[v] for i in range(len(a_ ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
103
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=3 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=10 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[1, 1, 2, 1] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="relu" , _lowerCamelCase=3 , _lowerCamelCase=None , ) -> List[str]: A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> Optional[Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : Any = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase_ ( self ) -> int: super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: pass def UpperCAmelCase_ ( self ) -> int: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase , _lowerCamelCase ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) def UpperCAmelCase_ ( self ) -> str: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> int: return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""tf""" ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 )
344
0
import itertools import string from collections.abc import Generator, Iterable def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE =iter(a_ ) while True: SCREAMING_SNAKE_CASE =tuple(itertools.islice(a_, a_ ) ) if not chunk: return yield chunk def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE ="""""".join([c.upper() for c in dirty if c in string.ascii_letters] ) SCREAMING_SNAKE_CASE ="""""" if len(a_ ) < 2: return dirty for i in range(len(a_ ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(a_ ) & 1: clean += "X" return clean def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE ="""ABCDEFGHIKLMNOPQRSTUVWXYZ""" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler SCREAMING_SNAKE_CASE =[] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(a_ ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(a_ ) return table def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE =generate_table(a_ ) SCREAMING_SNAKE_CASE =prepare_input(a_ ) SCREAMING_SNAKE_CASE ="""""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(a_, 2 ): SCREAMING_SNAKE_CASE =divmod(table.index(a_ ), 5 ) SCREAMING_SNAKE_CASE =divmod(table.index(a_ ), 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def snake_case__ ( lowerCAmelCase_, lowerCAmelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE =generate_table(a_ ) SCREAMING_SNAKE_CASE ="""""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(a_, 2 ): SCREAMING_SNAKE_CASE =divmod(table.index(a_ ), 5 ) SCREAMING_SNAKE_CASE =divmod(table.index(a_ ), 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
334
'''simple docstring''' def UpperCAmelCase ( a_ = 1_0_0 ) -> int: """simple docstring""" A_ : Dict = n * (n + 1) * (2 * n + 1) / 6 A_ : Optional[int] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'{solution() = }')
344
0
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class A ( __A ): UpperCamelCase_ : Tuple ='''EncodecFeatureExtractor''' UpperCamelCase_ : Dict =('''T5Tokenizer''', '''T5TokenizerFast''') def __init__(self , lowerCAmelCase , lowerCAmelCase ): super().__init__(_lowerCamelCase , _lowerCamelCase ) __lowercase= self.feature_extractor __lowercase= False def _A (self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase=True ): return self.tokenizer.get_decoder_prompt_ids(task=_lowerCamelCase , language=_lowerCamelCase , no_timestamps=_lowerCamelCase ) def __call__(self , *lowerCAmelCase , **lowerCAmelCase ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_lowerCamelCase , **_lowerCamelCase ) __lowercase= kwargs.pop('audio' , _lowerCamelCase ) __lowercase= kwargs.pop('sampling_rate' , _lowerCamelCase ) __lowercase= kwargs.pop('text' , _lowerCamelCase ) if len(_lowerCamelCase ) > 0: __lowercase= args[0] __lowercase= args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if text is not None: __lowercase= self.tokenizer(_lowerCamelCase , **_lowerCamelCase ) if audio is not None: __lowercase= self.feature_extractor(_lowerCamelCase , *_lowerCamelCase , sampling_rate=_lowerCamelCase , **_lowerCamelCase ) if audio is None: return inputs elif text is None: return audio_inputs else: __lowercase= audio_inputs["""input_values"""] if "padding_mask" in audio_inputs: __lowercase= audio_inputs["""padding_mask"""] return inputs def _A (self , *lowerCAmelCase , **lowerCAmelCase ): __lowercase= kwargs.pop('audio' , _lowerCamelCase ) __lowercase= kwargs.pop('padding_mask' , _lowerCamelCase ) if len(_lowerCamelCase ) > 0: __lowercase= args[0] __lowercase= args[1:] if audio_values is not None: return self._decode_audio(_lowerCamelCase , padding_mask=_lowerCamelCase ) else: return self.tokenizer.batch_decode(*_lowerCamelCase , **_lowerCamelCase ) def _A (self , *lowerCAmelCase , **lowerCAmelCase ): return self.tokenizer.decode(*_lowerCamelCase , **_lowerCamelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase = None ): __lowercase= to_numpy(_lowerCamelCase ) __lowercase= audio_values.shape if padding_mask is None: return list(_lowerCamelCase ) __lowercase= to_numpy(_lowerCamelCase ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) __lowercase= seq_len - padding_mask.shape[-1] __lowercase= 1 - self.feature_extractor.padding_value __lowercase= np.pad(_lowerCamelCase , ((0, 0), (0, difference)) , 'constant' , constant_values=_lowerCamelCase ) __lowercase= audio_values.tolist() for i in range(_lowerCamelCase ): __lowercase= np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] __lowercase= sliced_audio.reshape(_lowerCamelCase , -1 ) return audio_values
295
'''simple docstring''' from typing import TYPE_CHECKING from ..utils import _LazyModule UpperCamelCase__ : int = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys UpperCamelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('Program to check whether a number is a Perfect number or not...') lowercase : Tuple = int(input('Enter number: ').strip()) print(f'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
3
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def UpperCAmelCase ( a_ ) -> Dict[str, torch.Tensor]: """simple docstring""" A_ : List[str] = [] A_ : Dict = [] A_ : List[Any] = [] for rt in rc.restypes: A_ : Tuple = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) A_ : Union[str, Any] = {name: i for i, name in enumerate(a_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 1_4 ) restype_atomaa_to_atomaa_list.append([0] * 3_7 ) restype_atomaa_mask_list.append([0.0] * 1_4 ) A_ : Tuple = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : Optional[int] = torch.tensor( a_ , dtype=torch.intaa , device=protein["""aatype"""].device , ) A_ : List[Any] = torch.tensor( a_ , dtype=torch.floataa , device=protein["""aatype"""].device , ) A_ : Optional[int] = protein["""aatype"""].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein A_ : Dict = restype_atomaa_to_atomaa[protein_aatype] A_ : Optional[Any] = restype_atomaa_mask[protein_aatype] A_ : Any = residx_atomaa_mask A_ : List[str] = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back A_ : Tuple = restype_atomaa_to_atomaa[protein_aatype] A_ : Tuple = residx_atomaa_to_atomaa.long() # create the corresponding mask A_ : Optional[Any] = torch.zeros([2_1, 3_7] , dtype=torch.floataa , device=protein["""aatype"""].device ) for restype, restype_letter in enumerate(rc.restypes ): A_ : Optional[Any] = rc.restype_atoa[restype_letter] A_ : Any = rc.residue_atoms[restype_name] for atom_name in atom_names: A_ : Any = rc.atom_order[atom_name] A_ : Optional[int] = 1 A_ : Optional[int] = restype_atomaa_mask[protein_aatype] A_ : Dict = residx_atomaa_mask return protein def UpperCAmelCase ( a_ ) -> Dict[str, np.ndarray]: """simple docstring""" A_ : Union[str, Any] = tree_map(lambda a_ : torch.tensor(a_ , device=batch["""aatype"""].device ) , a_ , np.ndarray ) A_ : Optional[int] = tensor_tree_map(lambda a_ : np.array(a_ ) , make_atomaa_masks(a_ ) ) return out
344
0
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __lowerCAmelCase ( __A ): def lowerCamelCase__ ( self :int ): '''simple docstring''' a = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_lowerCamelCase , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(_lowerCamelCase , """num_attention_heads""" ) ) self.parent.assertTrue(hasattr(_lowerCamelCase , """num_encoder_blocks""" ) ) class __lowerCAmelCase : def __init__( self :Union[str, Any] , __magic_name__ :Optional[Any] , __magic_name__ :Optional[Any]=13 , __magic_name__ :Any=64 , __magic_name__ :List[Any]=3 , __magic_name__ :List[str]=4 , __magic_name__ :Dict=[2, 2, 2, 2] , __magic_name__ :Tuple=[8, 4, 2, 1] , __magic_name__ :Optional[int]=[16, 32, 64, 128] , __magic_name__ :Dict=[1, 4, 8, 16] , __magic_name__ :Union[str, Any]=[1, 2, 4, 8] , __magic_name__ :str=True , __magic_name__ :Union[str, Any]=True , __magic_name__ :Optional[int]="gelu" , __magic_name__ :Tuple=0.1 , __magic_name__ :Dict=0.1 , __magic_name__ :Optional[int]=0.02 , __magic_name__ :Optional[Any]=3 , __magic_name__ :Any=None , ): '''simple docstring''' a = parent a = batch_size a = image_size a = num_channels a = num_encoder_blocks a = sr_ratios a = depths a = hidden_sizes a = downsampling_rates a = num_attention_heads a = is_training a = use_labels a = hidden_act a = hidden_dropout_prob a = attention_probs_dropout_prob a = initializer_range a = num_labels a = scope def lowerCamelCase__ ( self :Dict ): '''simple docstring''' a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a = None if self.use_labels: a = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) a = self.get_config() return config, pixel_values, labels def lowerCamelCase__ ( self :Any ): '''simple docstring''' return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase__ ( self :int , __magic_name__ :Optional[int] , __magic_name__ :Optional[int] , __magic_name__ :List[Any] ): '''simple docstring''' a = SegformerModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() a = model(_lowerCamelCase ) a = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def lowerCamelCase__ ( self :int , __magic_name__ :List[Any] , __magic_name__ :Optional[Any] , __magic_name__ :Tuple ): '''simple docstring''' a = self.num_labels a = SegformerForSemanticSegmentation(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() a = model(_lowerCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) a = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def lowerCamelCase__ ( self :str , __magic_name__ :List[Any] , __magic_name__ :Optional[Any] , __magic_name__ :Optional[Any] ): '''simple docstring''' a = 1 a = SegformerForSemanticSegmentation(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() a = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(_lowerCamelCase ) a = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertGreater(result.loss , 0.0 ) def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' a = self.prepare_config_and_inputs() a = config_and_inputs a = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( __A , __A , unittest.TestCase ): UpperCamelCase__ = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) UpperCamelCase__ = ( { '''feature-extraction''': SegformerModel, '''image-classification''': SegformerForImageClassification, '''image-segmentation''': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def lowerCamelCase__ ( self :Any ): '''simple docstring''' a = SegformerModelTester(self ) a = SegformerConfigTester(self , config_class=_lowerCamelCase ) def lowerCamelCase__ ( self :Any ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*_lowerCamelCase ) def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*_lowerCamelCase ) @unittest.skip("""SegFormer does not use inputs_embeds""" ) def lowerCamelCase__ ( self :Dict ): '''simple docstring''' pass @unittest.skip("""SegFormer does not have get_input_embeddings method and get_output_embeddings methods""" ) def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' pass def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a = model_class(_lowerCamelCase ) a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a = [*signature.parameters.keys()] a = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs_for_common() a = True for model_class in self.all_model_classes: a = True a = False a = True a = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): a = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) a = outputs.attentions a = sum(self.model_tester.depths ) self.assertEqual(len(_lowerCamelCase ) , _lowerCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] a = True a = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): a = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) a = outputs.attentions self.assertEqual(len(_lowerCamelCase ) , _lowerCamelCase ) # verify the first attentions (first block, first layer) a = (self.model_tester.image_size // 4) ** 2 a = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) a = (self.model_tester.image_size // 32) ** 2 a = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) a = len(_lowerCamelCase ) # Check attention is always last and order is fine a = True a = True a = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): a = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) self.assertEqual(out_len + 1 , len(_lowerCamelCase ) ) a = outputs.attentions self.assertEqual(len(_lowerCamelCase ) , _lowerCamelCase ) # verify the first attentions (first block, first layer) a = (self.model_tester.image_size // 4) ** 2 a = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' def check_hidden_states_output(__magic_name__ :Optional[int] , __magic_name__ :str , __magic_name__ :str ): a = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() with torch.no_grad(): a = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) ) a = outputs.hidden_states a = self.model_tester.num_encoder_blocks self.assertEqual(len(_lowerCamelCase ) , _lowerCamelCase ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' if not self.model_tester.is_training: return a = self.model_tester.prepare_config_and_inputs_for_common() a = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ): continue a = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() a = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) a = model(**_lowerCamelCase ).loss loss.backward() @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' pass @slow def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a = SegformerModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def __A ( ) -> Tuple: a = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' a = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCamelCase , align=_lowerCamelCase , do_random_crop=_lowerCamelCase ) a = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( _lowerCamelCase ) a = prepare_img() a = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) a = encoded_inputs.pixel_values.to(_lowerCamelCase ) with torch.no_grad(): a = model(_lowerCamelCase ) a = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) a = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _lowerCamelCase , atol=1E-4 ) ) @slow def lowerCamelCase__ ( self :int ): '''simple docstring''' a = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCamelCase , align=_lowerCamelCase , do_random_crop=_lowerCamelCase ) a = SegformerForSemanticSegmentation.from_pretrained( """nvidia/segformer-b1-finetuned-cityscapes-1024-1024""" ).to(_lowerCamelCase ) a = prepare_img() a = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) a = encoded_inputs.pixel_values.to(_lowerCamelCase ) with torch.no_grad(): a = model(_lowerCamelCase ) a = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) a = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _lowerCamelCase , atol=1E-1 ) ) @slow def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCamelCase , align=_lowerCamelCase , do_random_crop=_lowerCamelCase ) a = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( _lowerCamelCase ) a = prepare_img() a = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) a = encoded_inputs.pixel_values.to(_lowerCamelCase ) with torch.no_grad(): a = model(_lowerCamelCase ) a = outputs.logits.detach().cpu() a = image_processor.post_process_semantic_segmentation(outputs=_lowerCamelCase , target_sizes=[(500, 300)] ) a = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , _lowerCamelCase ) a = image_processor.post_process_semantic_segmentation(outputs=_lowerCamelCase ) a = torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , _lowerCamelCase )
228
'''simple docstring''' import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ) -> str: A_ : Optional[int] = parent A_ : Dict = batch_size A_ : List[Any] = image_size A_ : Optional[int] = patch_size A_ : List[str] = num_channels A_ : List[Any] = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_size A_ : str = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Union[str, Any] = intermediate_size A_ : Any = hidden_act A_ : Optional[Any] = hidden_dropout_prob A_ : List[Any] = attention_probs_dropout_prob A_ : Dict = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : str = scope A_ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) A_ : Tuple = (image_size // patch_size) ** 2 A_ : Union[str, Any] = num_patches + 2 def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> int: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[str] = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : int = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A_ : Dict = 1 A_ : Optional[int] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : int = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : Tuple = self.type_sequence_label_size A_ : Tuple = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : int = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : Dict = 1 A_ : Any = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : List[str] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': DeiTModel, '''image-classification''': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : int = DeiTModelTester(self ) A_ : str = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> List[str]: self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Optional[int]: pass def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ , A_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A_ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : List[str] = model_class(_lowerCamelCase ) A_ : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Union[str, Any] = [*signature.parameters.keys()] A_ : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ) -> Union[str, Any]: A_ : int = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase_ ( self ) -> Optional[Any]: if not self.model_tester.is_training: return A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[str] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : List[str] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> int: A_ , A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return A_ : Any = False A_ : Union[str, Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue A_ : List[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Union[str, Any] = model(**_lowerCamelCase ).loss loss.backward() def UpperCAmelCase_ ( self ) -> Tuple: A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() A_ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"Testing {model_class} with {problem_type['title']}" ): A_ : Dict = problem_type["""title"""] A_ : List[Any] = problem_type["""num_labels"""] A_ : List[str] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() A_ : List[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: A_ : Tuple = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) A_ : Union[str, Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: A_ : List[str] = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"Something is going wrong in the regression problem: intercepted {w.message}" ) loss.backward() @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> Optional[Any]: return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : Any = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) A_ : Optional[int] = self.default_image_processor A_ : str = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): A_ : Any = model(**_lowerCamelCase ) # verify the logits A_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : List[Any] = torch.tensor([-1.0266, 0.1912, -1.2861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) A_ : int = self.default_image_processor A_ : List[str] = prepare_img() A_ : List[Any] = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) A_ : Union[str, Any] = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A_ : List[Any] = model(_lowerCamelCase )
344
0
'''simple docstring''' import os def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = "input.txt" ): with open(os.path.join(os.path.dirname(a_ ) , a_ ) ) as input_file: _snake_case = [ [int(a_ ) for element in line.split(""",""" )] for line in input_file.readlines() ] _snake_case = len(a_ ) _snake_case = len(matrix[0] ) _snake_case = [[-1 for _ in range(a_ )] for _ in range(a_ )] for i in range(a_ ): _snake_case = matrix[i][0] for j in range(1 , a_ ): for i in range(a_ ): _snake_case = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , a_ ): _snake_case = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): _snake_case = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(f'''{solution() = }''')
341
'''simple docstring''' import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 32 , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = True , _lowerCamelCase = [0.4814_5466, 0.457_8275, 0.4082_1073] , _lowerCamelCase = [0.2686_2954, 0.2613_0258, 0.2757_7711] , _lowerCamelCase = True , _lowerCamelCase=7 , _lowerCamelCase=30 , _lowerCamelCase=400 , _lowerCamelCase=3 , ) -> Union[str, Any]: A_ : Optional[int] = parent A_ : Union[str, Any] = do_resize A_ : Optional[Any] = size if size is not None else {"""shortest_edge""": 288} A_ : Tuple = size_divisor A_ : List[Any] = do_rescale A_ : Dict = rescale_factor A_ : List[Any] = do_normalize A_ : Dict = do_center_crop A_ : Optional[Any] = image_mean A_ : List[str] = image_std A_ : str = do_pad A_ : Any = batch_size A_ : List[str] = num_channels A_ : List[str] = min_resolution A_ : Union[str, Any] = max_resolution def UpperCAmelCase_ ( self ) -> Any: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase=False ) -> Optional[int]: if not batched: A_ : Union[str, Any] = self.size["""shortest_edge"""] A_ : Dict = image_inputs[0] if isinstance(_lowerCamelCase , Image.Image ): A_ , A_ : Optional[Any] = image.size else: A_ , A_ : int = image.shape[1], image.shape[2] A_ : Optional[int] = size / min(_lowerCamelCase , _lowerCamelCase ) if h < w: A_ , A_ : Optional[Any] = size, scale * w else: A_ , A_ : Dict = scale * h, size A_ : Union[str, Any] = int((1333 / 800) * size ) if max(_lowerCamelCase , _lowerCamelCase ) > max_size: A_ : str = max_size / max(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = newh * scale A_ : Dict = neww * scale A_ , A_ : str = int(newh + 0.5 ), int(neww + 0.5 ) A_ , A_ : Dict = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: A_ : Tuple = [] for image in image_inputs: A_ , A_ : Tuple = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) A_ : List[Any] = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[0] )[0] A_ : Tuple = max(_lowerCamelCase , key=lambda _lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _lowerCAmelCase ( __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> Dict: A_ : int = BridgeTowerImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(_lowerCamelCase , """size_divisor""" ) ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: pass def UpperCAmelCase_ ( self ) -> List[str]: # Initialize image processor A_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A_ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , Image.Image ) # Test not batched input A_ : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : Optional[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : int = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Union[str, Any]: # Initialize image processor A_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , numpify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , np.ndarray ) # Test not batched input A_ : Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : int = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image processor A_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCamelCase , torchify=_lowerCamelCase ) for image in image_inputs: self.assertIsInstance(_lowerCamelCase , torch.Tensor ) # Test not batched input A_ : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values A_ , A_ : Tuple = self.image_processor_tester.get_expected_values(_lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched A_ : List[Any] = image_processing(_lowerCamelCase , return_tensors="""pt""" ).pixel_values A_ , A_ : List[str] = self.image_processor_tester.get_expected_values(_lowerCamelCase , batched=_lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
344
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {'configuration_opt': ['OPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'OPTConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'OPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'OPTForCausalLM', 'OPTModel', 'OPTPreTrainedModel', 'OPTForSequenceClassification', 'OPTForQuestionAnswering', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ['TFOPTForCausalLM', 'TFOPTModel', 'TFOPTPreTrainedModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ 'FlaxOPTForCausalLM', 'FlaxOPTModel', 'FlaxOPTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
158
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> Optional[int]: """simple docstring""" print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(a_ ): for j in range(a_ ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ) , end="""\t""" ) else: print("""INF""" , end="""\t""" ) print() def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" A_ : List[str] = [[float("""inf""" ) for _ in range(a_ )] for _ in range(a_ )] for i in range(a_ ): for j in range(a_ ): A_ : List[Any] = graph[i][j] # check vertex k against all other vertices (i, j) for k in range(a_ ): # looping through rows of graph array for i in range(a_ ): # looping through columns of graph array for j in range(a_ ): if ( dist[i][k] != float("""inf""" ) and dist[k][j] != float("""inf""" ) and dist[i][k] + dist[k][j] < dist[i][j] ): A_ : List[str] = dist[i][k] + dist[k][j] _print_dist(a_ , a_ ) return dist, v if __name__ == "__main__": UpperCamelCase__ : Tuple = int(input('Enter number of vertices: ')) UpperCamelCase__ : int = int(input('Enter number of edges: ')) UpperCamelCase__ : Dict = [[float('inf') for i in range(v)] for j in range(v)] for i in range(v): UpperCamelCase__ : Union[str, Any] = 0.0 # src and dst are indices that must be within the array size graph[e][v] # failure to follow this will result in an error for i in range(e): print('\nEdge ', i + 1) UpperCamelCase__ : Union[str, Any] = int(input('Enter source:')) UpperCamelCase__ : int = int(input('Enter destination:')) UpperCamelCase__ : Optional[Any] = float(input('Enter weight:')) UpperCamelCase__ : Any = weight floyd_warshall(graph, v) # Example Input # Enter number of vertices: 3 # Enter number of edges: 2 # # generated graph from vertex and edge inputs # [[inf, inf, inf], [inf, inf, inf], [inf, inf, inf]] # [[0.0, inf, inf], [inf, 0.0, inf], [inf, inf, 0.0]] # specify source, destination and weight for edge #1 # Edge 1 # Enter source:1 # Enter destination:2 # Enter weight:2 # specify source, destination and weight for edge #2 # Edge 2 # Enter source:2 # Enter destination:1 # Enter weight:1 # # Expected Output from the vertice, edge and src, dst, weight inputs!! # 0 INF INF # INF 0 2 # INF 1 0
344
0
from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings @maybe_allow_in_graph class __A ( nn.Module ): def __init__( self : List[str] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Any=0.0 , UpperCAmelCase_ : Tuple = None , UpperCAmelCase_ : Optional[Any] = "geglu" , UpperCAmelCase_ : Union[str, Any] = None , UpperCAmelCase_ : Any = False , UpperCAmelCase_ : Union[str, Any] = False , UpperCAmelCase_ : Dict = False , UpperCAmelCase_ : Dict = False , UpperCAmelCase_ : Optional[int] = True , UpperCAmelCase_ : Optional[int] = "layer_norm" , UpperCAmelCase_ : Any = False , ): super().__init__() lowerCAmelCase : Tuple = only_cross_attention lowerCAmelCase : Dict = (num_embeds_ada_norm is not None) and norm_type == """ada_norm_zero""" lowerCAmelCase : Tuple = (num_embeds_ada_norm is not None) and norm_type == """ada_norm""" if norm_type in ("ada_norm", "ada_norm_zero") and num_embeds_ada_norm is None: raise ValueError( f"`norm_type` is set to {norm_type}, but `num_embeds_ada_norm` is not defined. Please make sure to" f" define `num_embeds_ada_norm` if setting `norm_type` to {norm_type}." ) # Define 3 blocks. Each block has its own normalization layer. # 1. Self-Attn if self.use_ada_layer_norm: lowerCAmelCase : Union[str, Any] = AdaLayerNorm(_lowerCamelCase , _lowerCamelCase ) elif self.use_ada_layer_norm_zero: lowerCAmelCase : Optional[Any] = AdaLayerNormZero(_lowerCamelCase , _lowerCamelCase ) else: lowerCAmelCase : Dict = nn.LayerNorm(_lowerCamelCase , elementwise_affine=_lowerCamelCase ) lowerCAmelCase : List[str] = Attention( query_dim=_lowerCamelCase , heads=_lowerCamelCase , dim_head=_lowerCamelCase , dropout=_lowerCamelCase , bias=_lowerCamelCase , cross_attention_dim=cross_attention_dim if only_cross_attention else None , upcast_attention=_lowerCamelCase , ) # 2. Cross-Attn if cross_attention_dim is not None or double_self_attention: # We currently only use AdaLayerNormZero for self attention where there will only be one attention block. # I.e. the number of returned modulation chunks from AdaLayerZero would not make sense if returned during # the second cross attention block. lowerCAmelCase : List[Any] = ( AdaLayerNorm(_lowerCamelCase , _lowerCamelCase ) if self.use_ada_layer_norm else nn.LayerNorm(_lowerCamelCase , elementwise_affine=_lowerCamelCase ) ) lowerCAmelCase : Optional[int] = Attention( query_dim=_lowerCamelCase , cross_attention_dim=cross_attention_dim if not double_self_attention else None , heads=_lowerCamelCase , dim_head=_lowerCamelCase , dropout=_lowerCamelCase , bias=_lowerCamelCase , upcast_attention=_lowerCamelCase , ) # is self-attn if encoder_hidden_states is none else: lowerCAmelCase : Tuple = None lowerCAmelCase : Any = None # 3. Feed-forward lowerCAmelCase : str = nn.LayerNorm(_lowerCamelCase , elementwise_affine=_lowerCamelCase ) lowerCAmelCase : Any = FeedForward(_lowerCamelCase , dropout=_lowerCamelCase , activation_fn=_lowerCamelCase , final_dropout=_lowerCamelCase ) # let chunk size default to None lowerCAmelCase : Optional[Any] = None lowerCAmelCase : int = 0 def lowercase__ ( self : List[Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[str] ): # Sets chunk feed-forward lowerCAmelCase : Union[str, Any] = chunk_size lowerCAmelCase : List[str] = dim def lowercase__ ( self : Union[str, Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Any = None , UpperCAmelCase_ : str = None , UpperCAmelCase_ : Optional[Any] = None , UpperCAmelCase_ : Any = None , UpperCAmelCase_ : List[Any] = None , UpperCAmelCase_ : Optional[Any] = None , ): # Notice that normalization is always applied before the real computation in the following blocks. # 1. Self-Attention if self.use_ada_layer_norm: lowerCAmelCase : List[Any] = self.norma(_lowerCamelCase , _lowerCamelCase ) elif self.use_ada_layer_norm_zero: lowerCAmelCase : Dict = self.norma( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , hidden_dtype=hidden_states.dtype ) else: lowerCAmelCase : List[Any] = self.norma(_lowerCamelCase ) lowerCAmelCase : List[Any] = cross_attention_kwargs if cross_attention_kwargs is not None else {} lowerCAmelCase : int = self.attna( _lowerCamelCase , encoder_hidden_states=encoder_hidden_states if self.only_cross_attention else None , attention_mask=_lowerCamelCase , **_lowerCamelCase , ) if self.use_ada_layer_norm_zero: lowerCAmelCase : Optional[Any] = gate_msa.unsqueeze(1 ) * attn_output lowerCAmelCase : int = attn_output + hidden_states # 2. Cross-Attention if self.attna is not None: lowerCAmelCase : Optional[Any] = ( self.norma(_lowerCamelCase , _lowerCamelCase ) if self.use_ada_layer_norm else self.norma(_lowerCamelCase ) ) lowerCAmelCase : Optional[int] = self.attna( _lowerCamelCase , encoder_hidden_states=_lowerCamelCase , attention_mask=_lowerCamelCase , **_lowerCamelCase , ) lowerCAmelCase : Tuple = attn_output + hidden_states # 3. Feed-forward lowerCAmelCase : List[str] = self.norma(_lowerCamelCase ) if self.use_ada_layer_norm_zero: lowerCAmelCase : str = norm_hidden_states * (1 + scale_mlp[:, None]) + shift_mlp[:, None] if self._chunk_size is not None: # "feed_forward_chunk_size" can be used to save memory if norm_hidden_states.shape[self._chunk_dim] % self._chunk_size != 0: raise ValueError( f"`hidden_states` dimension to be chunked: {norm_hidden_states.shape[self._chunk_dim]} has to be divisible by chunk size: {self._chunk_size}. Make sure to set an appropriate `chunk_size` when calling `unet.enable_forward_chunking`." ) lowerCAmelCase : List[Any] = norm_hidden_states.shape[self._chunk_dim] // self._chunk_size lowerCAmelCase : List[str] = torch.cat( [self.ff(_lowerCamelCase ) for hid_slice in norm_hidden_states.chunk(_lowerCamelCase , dim=self._chunk_dim )] , dim=self._chunk_dim , ) else: lowerCAmelCase : Optional[int] = self.ff(_lowerCamelCase ) if self.use_ada_layer_norm_zero: lowerCAmelCase : List[Any] = gate_mlp.unsqueeze(1 ) * ff_output lowerCAmelCase : Tuple = ff_output + hidden_states return hidden_states class __A ( nn.Module ): def __init__( self : str , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple = None , UpperCAmelCase_ : Any = 4 , UpperCAmelCase_ : int = 0.0 , UpperCAmelCase_ : Dict = "geglu" , UpperCAmelCase_ : List[Any] = False , ): super().__init__() lowerCAmelCase : List[str] = int(dim * mult ) lowerCAmelCase : List[Any] = dim_out if dim_out is not None else dim if activation_fn == "gelu": lowerCAmelCase : Any = GELU(_lowerCamelCase , _lowerCamelCase ) if activation_fn == "gelu-approximate": lowerCAmelCase : Dict = GELU(_lowerCamelCase , _lowerCamelCase , approximate='tanh' ) elif activation_fn == "geglu": lowerCAmelCase : int = GEGLU(_lowerCamelCase , _lowerCamelCase ) elif activation_fn == "geglu-approximate": lowerCAmelCase : Tuple = ApproximateGELU(_lowerCamelCase , _lowerCamelCase ) lowerCAmelCase : Dict = nn.ModuleList([] ) # project in self.net.append(_lowerCamelCase ) # project dropout self.net.append(nn.Dropout(_lowerCamelCase ) ) # project out self.net.append(nn.Linear(_lowerCamelCase , _lowerCamelCase ) ) # FF as used in Vision Transformer, MLP-Mixer, etc. have a final dropout if final_dropout: self.net.append(nn.Dropout(_lowerCamelCase ) ) def lowercase__ ( self : int , UpperCAmelCase_ : List[str] ): for module in self.net: lowerCAmelCase : Dict = module(_lowerCamelCase ) return hidden_states class __A ( nn.Module ): def __init__( self : int , UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Dict = "none" ): super().__init__() lowerCAmelCase : int = nn.Linear(_lowerCamelCase , _lowerCamelCase ) lowerCAmelCase : Union[str, Any] = approximate def lowercase__ ( self : List[Any] , UpperCAmelCase_ : Optional[int] ): if gate.device.type != "mps": return F.gelu(_lowerCamelCase , approximate=self.approximate ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ) , approximate=self.approximate ).to(dtype=gate.dtype ) def lowercase__ ( self : List[str] , UpperCAmelCase_ : Tuple ): lowerCAmelCase : Tuple = self.proj(_lowerCamelCase ) lowerCAmelCase : Any = self.gelu(_lowerCamelCase ) return hidden_states class __A ( nn.Module ): def __init__( self : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Any ): super().__init__() lowerCAmelCase : Any = nn.Linear(_lowerCamelCase , dim_out * 2 ) def lowercase__ ( self : Any , UpperCAmelCase_ : Optional[Any] ): if gate.device.type != "mps": return F.gelu(_lowerCamelCase ) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa ) ).to(dtype=gate.dtype ) def lowercase__ ( self : Any , UpperCAmelCase_ : Tuple ): lowerCAmelCase : List[str] = self.proj(_lowerCamelCase ).chunk(2 , dim=-1 ) return hidden_states * self.gelu(_lowerCamelCase ) class __A ( nn.Module ): def __init__( self : str , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : int ): super().__init__() lowerCAmelCase : List[Any] = nn.Linear(_lowerCamelCase , _lowerCamelCase ) def lowercase__ ( self : Optional[int] , UpperCAmelCase_ : str ): lowerCAmelCase : str = self.proj(_lowerCamelCase ) return x * torch.sigmoid(1.7_02 * x ) class __A ( nn.Module ): def __init__( self : Tuple , UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any] ): super().__init__() lowerCAmelCase : Dict = nn.Embedding(_lowerCamelCase , _lowerCamelCase ) lowerCAmelCase : int = nn.SiLU() lowerCAmelCase : int = nn.Linear(_lowerCamelCase , embedding_dim * 2 ) lowerCAmelCase : str = nn.LayerNorm(_lowerCamelCase , elementwise_affine=_lowerCamelCase ) def lowercase__ ( self : List[Any] , UpperCAmelCase_ : Any , UpperCAmelCase_ : Tuple ): lowerCAmelCase : Dict = self.linear(self.silu(self.emb(_lowerCamelCase ) ) ) lowerCAmelCase : Any = torch.chunk(_lowerCamelCase , 2 ) lowerCAmelCase : Union[str, Any] = self.norm(_lowerCamelCase ) * (1 + scale) + shift return x class __A ( nn.Module ): def __init__( self : Tuple , UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[Any] ): super().__init__() lowerCAmelCase : Tuple = CombinedTimestepLabelEmbeddings(_lowerCamelCase , _lowerCamelCase ) lowerCAmelCase : Optional[Any] = nn.SiLU() lowerCAmelCase : Union[str, Any] = nn.Linear(_lowerCamelCase , 6 * embedding_dim , bias=_lowerCamelCase ) lowerCAmelCase : Any = nn.LayerNorm(_lowerCamelCase , elementwise_affine=_lowerCamelCase , eps=1E-6 ) def lowercase__ ( self : Optional[Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str=None ): lowerCAmelCase : Optional[Any] = self.linear(self.silu(self.emb(_lowerCamelCase , _lowerCamelCase , hidden_dtype=_lowerCamelCase ) ) ) lowerCAmelCase : int = emb.chunk(6 , dim=1 ) lowerCAmelCase : List[Any] = self.norm(_lowerCamelCase ) * (1 + scale_msa[:, None]) + shift_msa[:, None] return x, gate_msa, shift_mlp, scale_mlp, gate_mlp class __A ( nn.Module ): def __init__( self : Dict , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] = None , UpperCAmelCase_ : List[Any] = 1E-5 ): super().__init__() lowerCAmelCase : Tuple = num_groups lowerCAmelCase : Union[str, Any] = eps if act_fn is None: lowerCAmelCase : List[str] = None else: lowerCAmelCase : Dict = get_activation(_lowerCamelCase ) lowerCAmelCase : Optional[Any] = nn.Linear(_lowerCamelCase , out_dim * 2 ) def lowercase__ ( self : List[str] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[Any] ): if self.act: lowerCAmelCase : Dict = self.act(_lowerCamelCase ) lowerCAmelCase : Union[str, Any] = self.linear(_lowerCamelCase ) lowerCAmelCase : Tuple = emb[:, :, None, None] lowerCAmelCase : int = emb.chunk(2 , dim=1 ) lowerCAmelCase : Union[str, Any] = F.group_norm(_lowerCamelCase , self.num_groups , eps=self.eps ) lowerCAmelCase : Any = x * (1 + scale) + shift return x
138
'''simple docstring''' import datasets from .evaluate import evaluate UpperCamelCase__ : int = '\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n' UpperCamelCase__ : Any = '\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n' UpperCamelCase__ : Optional[Any] = '\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the SQuAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> squad_metric = datasets.load_metric("squad")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase_ ( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: A_ : Optional[Any] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} A_ : List[Any] = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] A_ : int = evaluate(dataset=_lowerCamelCase , predictions=_lowerCamelCase ) return score
344
0
from collections.abc import Sequence def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return sum(c * (x**i) for i, c in enumerate(a_ ) ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = 0.0 for coeff in reversed(a_ ): lowercase = result * x + coeff return result if __name__ == "__main__": UpperCAmelCase = (0.0, 0.0, 5.0, 9.3, 7.0) UpperCAmelCase = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
195
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase__ : Any = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Optional[Any] = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] UpperCamelCase__ : List[str] = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] UpperCamelCase__ : str = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): UpperCamelCase__ : List[str] = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCamelCase__ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
344
0
import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def UpperCamelCase ( snake_case__ : Union[str, Any] , snake_case__ : List[str] ) -> int: assert isinstance(a_ , a_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('keep_in_memory' , [False, True] ) def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : str , snake_case__ : int ) -> Tuple: UpperCamelCase : str = tmp_path / """cache""" UpperCamelCase : int = {"""text""": """string"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase : Optional[Any] = TextDatasetReader(a_ , cache_dir=a_ , keep_in_memory=a_ ).read() _check_text_dataset(a_ , a_ ) @pytest.mark.parametrize( 'features' , [ None, {'text': 'string'}, {'text': 'int32'}, {'text': 'float32'}, ] , ) def UpperCamelCase ( snake_case__ : Any , snake_case__ : Tuple , snake_case__ : Union[str, Any] ) -> Optional[int]: UpperCamelCase : Any = tmp_path / """cache""" UpperCamelCase : Dict = {"""text""": """string"""} UpperCamelCase : Any = features.copy() if features else default_expected_features UpperCamelCase : Union[str, Any] = ( Features({feature: Value(a_ ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase : List[str] = TextDatasetReader(a_ , features=a_ , cache_dir=a_ ).read() _check_text_dataset(a_ , a_ ) @pytest.mark.parametrize('split' , [None, NamedSplit('train' ), 'train', 'test'] ) def UpperCamelCase ( snake_case__ : Union[str, Any] , snake_case__ : List[str] , snake_case__ : Tuple ) -> Union[str, Any]: UpperCamelCase : List[Any] = tmp_path / """cache""" UpperCamelCase : Optional[int] = {"""text""": """string"""} UpperCamelCase : List[str] = TextDatasetReader(a_ , cache_dir=a_ , split=a_ ).read() _check_text_dataset(a_ , a_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('path_type' , [str, list] ) def UpperCamelCase ( snake_case__ : str , snake_case__ : Any , snake_case__ : str ) -> Tuple: if issubclass(a_ , a_ ): UpperCamelCase : Optional[Any] = text_path elif issubclass(a_ , a_ ): UpperCamelCase : Tuple = [text_path] UpperCamelCase : str = tmp_path / """cache""" UpperCamelCase : Optional[Any] = {"""text""": """string"""} UpperCamelCase : Union[str, Any] = TextDatasetReader(a_ , cache_dir=a_ ).read() _check_text_dataset(a_ , a_ ) def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Optional[Any] , snake_case__ : Optional[Any]=("train",) ) -> Optional[int]: assert isinstance(a_ , a_ ) for split in splits: UpperCamelCase : Optional[int] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('keep_in_memory' , [False, True] ) def UpperCamelCase ( snake_case__ : List[Any] , snake_case__ : Dict , snake_case__ : Dict ) -> Optional[Any]: UpperCamelCase : Optional[int] = tmp_path / """cache""" UpperCamelCase : List[str] = {"""text""": """string"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCamelCase : int = TextDatasetReader({'train': text_path} , cache_dir=a_ , keep_in_memory=a_ ).read() _check_text_datasetdict(a_ , a_ ) @pytest.mark.parametrize( 'features' , [ None, {'text': 'string'}, {'text': 'int32'}, {'text': 'float32'}, ] , ) def UpperCamelCase ( snake_case__ : int , snake_case__ : List[str] , snake_case__ : Optional[int] ) -> Union[str, Any]: UpperCamelCase : Union[str, Any] = tmp_path / """cache""" # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" UpperCamelCase : Tuple = {"""text""": """string"""} UpperCamelCase : Optional[Any] = features.copy() if features else default_expected_features UpperCamelCase : Dict = ( Features({feature: Value(a_ ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCamelCase : List[str] = TextDatasetReader({'train': text_path} , features=a_ , cache_dir=a_ ).read() _check_text_datasetdict(a_ , a_ ) @pytest.mark.parametrize('split' , [None, NamedSplit('train' ), 'train', 'test'] ) def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Any , snake_case__ : Tuple ) -> Optional[Any]: if split: UpperCamelCase : int = {split: text_path} else: UpperCamelCase : str = """train""" UpperCamelCase : Any = {"""train""": text_path, """test""": text_path} UpperCamelCase : List[Any] = tmp_path / """cache""" UpperCamelCase : Dict = {"""text""": """string"""} UpperCamelCase : Optional[Any] = TextDatasetReader(a_ , cache_dir=a_ ).read() _check_text_datasetdict(a_ , a_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
119
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType UpperCamelCase__ : Optional[Any] = logging.get_logger(__name__) class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = '''vision-encoder-decoder''' lowerCamelCase = True def __init__( self , **_lowerCamelCase ) -> str: super().__init__(**_lowerCamelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F"A configuraton of type {self.model_type} cannot be instantiated because " F"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}" ) A_ : Optional[int] = kwargs.pop("""encoder""" ) A_ : List[str] = encoder_config.pop("""model_type""" ) A_ : str = kwargs.pop("""decoder""" ) A_ : Optional[Any] = decoder_config.pop("""model_type""" ) A_ : List[str] = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : str = AutoConfig.for_model(_lowerCamelCase , **_lowerCamelCase ) A_ : Any = True @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A_ : int = True A_ : List[Any] = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Any: A_ : Dict = copy.deepcopy(self.__dict__ ) A_ : List[str] = self.encoder.to_dict() A_ : Union[str, Any] = self.decoder.to_dict() A_ : str = self.__class__.model_type return output class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = version.parse('''1.11''' ) @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def UpperCAmelCase_ ( self ) -> float: return 1e-4 @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: A_ : Optional[Any] = OrderedDict() A_ : Any = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A_ : Optional[int] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = -1 , _lowerCamelCase = -1 , _lowerCamelCase = False , _lowerCamelCase = None , ) -> Mapping[str, Any]: import torch A_ : Optional[int] = OrderedDict() A_ : List[Any] = super().generate_dummy_inputs( _lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) A_ , A_ : str = dummy_input["""input_ids"""].shape A_ : Optional[int] = (batch, encoder_sequence, self._config.encoder_hidden_size) A_ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A_ : List[str] = dummy_input.pop("""attention_mask""" ) A_ : Optional[int] = torch.zeros(_lowerCamelCase ) return common_inputs class _lowerCAmelCase ( __A ): """simple docstring""" @property def UpperCAmelCase_ ( self ) -> None: pass def UpperCAmelCase_ ( self , _lowerCamelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = "default" ) -> OnnxConfig: A_ : List[Any] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(_lowerCamelCase , _lowerCamelCase )
344
0
"""simple docstring""" import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : str = logging.get_logger(__name__) def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" A_ : int = os.path.abspath(a_ ) logger.info(f"""Converting TensorFlow checkpoint from {tf_path}""" ) # Load weights from TF model A_ : Any = tf.train.list_variables(a_ ) A_ : str = [] A_ : List[Any] = [] A_ : Union[str, Any] = [] for full_name, shape in init_vars: # logger.info(f"Loading TF weight {name} with shape {shape}") A_ : List[Any] = full_name.split('''/''' ) if full_name == "_CHECKPOINTABLE_OBJECT_GRAPH" or name[0] in ["global_step", "save_counter"]: logger.info(f"""Skipping non-model layer {full_name}""" ) continue if "optimizer" in full_name: logger.info(f"""Skipping optimization layer {full_name}""" ) continue if name[0] == "model": # ignore initial 'model' A_ : int = name[1:] # figure out how many levels deep the name is A_ : Any = 0 for _name in name: if _name.startswith('''layer_with_weights''' ): depth += 1 else: break layer_depth.append(a_ ) # read data A_ : int = tf.train.load_variable(a_ , a_ ) names.append('''/'''.join(a_ ) ) arrays.append(a_ ) logger.info(f"""Read a total of {len(a_ ):,} layers""" ) # Sanity check if len(set(a_ ) ) != 1: raise ValueError(f"""Found layer names with different depths (layer depth {list(set(a_ ) )})""" ) A_ : Dict = list(set(a_ ) )[0] if layer_depth != 1: raise ValueError( '''The model contains more than just the embedding/encoder layers. This script does not handle MLM/NSP''' ''' heads.''' ) # convert layers logger.info('''Converting weights...''' ) for full_name, array in zip(a_ , a_ ): A_ : List[Any] = full_name.split('''/''' ) A_ : Dict = model A_ : Optional[Any] = [] for i, m_name in enumerate(a_ ): if m_name == ".ATTRIBUTES": # variable names end with .ATTRIBUTES/VARIABLE_VALUE break if m_name.startswith('''layer_with_weights''' ): A_ : List[Any] = int(m_name.split('''-''' )[-1] ) if layer_num <= 2: # embedding layers # layer_num 0: word_embeddings # layer_num 1: position_embeddings # layer_num 2: token_type_embeddings continue elif layer_num == 3: # embedding LayerNorm trace.extend(['''embeddings''', '''LayerNorm'''] ) A_ : Dict = getattr(a_ , '''embeddings''' ) A_ : List[str] = getattr(a_ , '''LayerNorm''' ) elif layer_num > 3 and layer_num < config.num_hidden_layers + 4: # encoder layers trace.extend(['''encoder''', '''layer''', str(layer_num - 4 )] ) A_ : List[Any] = getattr(a_ , '''encoder''' ) A_ : str = getattr(a_ , '''layer''' ) A_ : Dict = pointer[layer_num - 4] elif layer_num == config.num_hidden_layers + 4: # pooler layer trace.extend(['''pooler''', '''dense'''] ) A_ : Tuple = getattr(a_ , '''pooler''' ) A_ : List[str] = getattr(a_ , '''dense''' ) elif m_name == "embeddings": trace.append('''embeddings''' ) A_ : Any = getattr(a_ , '''embeddings''' ) if layer_num == 0: trace.append('''word_embeddings''' ) A_ : Optional[int] = getattr(a_ , '''word_embeddings''' ) elif layer_num == 1: trace.append('''position_embeddings''' ) A_ : str = getattr(a_ , '''position_embeddings''' ) elif layer_num == 2: trace.append('''token_type_embeddings''' ) A_ : List[Any] = getattr(a_ , '''token_type_embeddings''' ) else: raise ValueError(f"""Unknown embedding layer with name {full_name}""" ) trace.append('''weight''' ) A_ : List[str] = getattr(a_ , '''weight''' ) elif m_name == "_attention_layer": # self-attention layer trace.extend(['''attention''', '''self'''] ) A_ : str = getattr(a_ , '''attention''' ) A_ : Optional[Any] = getattr(a_ , '''self''' ) elif m_name == "_attention_layer_norm": # output attention norm trace.extend(['''attention''', '''output''', '''LayerNorm'''] ) A_ : Union[str, Any] = getattr(a_ , '''attention''' ) A_ : Union[str, Any] = getattr(a_ , '''output''' ) A_ : Union[str, Any] = getattr(a_ , '''LayerNorm''' ) elif m_name == "_attention_output_dense": # output attention dense trace.extend(['''attention''', '''output''', '''dense'''] ) A_ : str = getattr(a_ , '''attention''' ) A_ : List[str] = getattr(a_ , '''output''' ) A_ : Union[str, Any] = getattr(a_ , '''dense''' ) elif m_name == "_output_dense": # output dense trace.extend(['''output''', '''dense'''] ) A_ : List[str] = getattr(a_ , '''output''' ) A_ : Optional[int] = getattr(a_ , '''dense''' ) elif m_name == "_output_layer_norm": # output dense trace.extend(['''output''', '''LayerNorm'''] ) A_ : List[Any] = getattr(a_ , '''output''' ) A_ : Dict = getattr(a_ , '''LayerNorm''' ) elif m_name == "_key_dense": # attention key trace.append('''key''' ) A_ : Optional[Any] = getattr(a_ , '''key''' ) elif m_name == "_query_dense": # attention query trace.append('''query''' ) A_ : Optional[int] = getattr(a_ , '''query''' ) elif m_name == "_value_dense": # attention value trace.append('''value''' ) A_ : Optional[Any] = getattr(a_ , '''value''' ) elif m_name == "_intermediate_dense": # attention intermediate dense trace.extend(['''intermediate''', '''dense'''] ) A_ : List[Any] = getattr(a_ , '''intermediate''' ) A_ : List[str] = getattr(a_ , '''dense''' ) elif m_name == "_output_layer_norm": # output layer norm trace.append('''output''' ) A_ : Optional[Any] = getattr(a_ , '''output''' ) # weights & biases elif m_name in ["bias", "beta"]: trace.append('''bias''' ) A_ : List[str] = getattr(a_ , '''bias''' ) elif m_name in ["kernel", "gamma"]: trace.append('''weight''' ) A_ : Any = getattr(a_ , '''weight''' ) else: logger.warning(f"""Ignored {m_name}""" ) # for certain layers reshape is necessary A_ : Optional[Any] = """.""".join(a_ ) if re.match(r'''(\S+)\.attention\.self\.(key|value|query)\.(bias|weight)''' , a_ ) or re.match( r'''(\S+)\.attention\.output\.dense\.weight''' , a_ ): A_ : List[Any] = array.reshape(pointer.data.shape ) if "kernel" in full_name: A_ : Optional[Any] = array.transpose() if pointer.shape == array.shape: A_ : Optional[int] = torch.from_numpy(a_ ) else: raise ValueError( f"""Shape mismatch in layer {full_name}: Model expects shape {pointer.shape} but layer contains shape:""" f""" {array.shape}""" ) logger.info(f"""Successfully set variable {full_name} to PyTorch layer {trace}""" ) return model def lowercase_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): """simple docstring""" logger.info(f"""Loading model based on config from {config_path}...""" ) A_ : Union[str, Any] = BertConfig.from_json_file(a_ ) A_ : Any = BertModel(a_ ) # Load weights from checkpoint logger.info(f"""Loading weights from checkpoint {tf_checkpoint_path}...""" ) load_tfa_weights_in_bert(a_ , a_ , a_ ) # Save pytorch-model logger.info(f"""Saving PyTorch model to {pytorch_dump_path}...""" ) torch.save(model.state_dict() , a_ ) if __name__ == "__main__": _lowerCamelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--tf_checkpoint_path', type=str, required=True, help='Path to the TensorFlow 2.x checkpoint path.' ) parser.add_argument( '--bert_config_file', type=str, required=True, help='The config json file corresponding to the BERT model. This specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', type=str, required=True, help='Path to the output PyTorch model (must include filename).', ) _lowerCamelCase : Dict = parser.parse_args() convert_tfa_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
167
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) UpperCamelCase__ : Any = '\\n Text data.\n Second line of data.' UpperCamelCase__ : List[Any] = 'file' @pytest.fixture(scope="""session""" ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : int = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") A_ : int = bytes(a_ , """utf-8""" ) with zstd.open(a_ , """wb""" ) as f: f.write(a_ ) return path @pytest.fixture def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , a_ ) , """w""" ) as f: f.write(a_ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> Optional[int]: """simple docstring""" A_ : List[str] = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} A_ : Any = input_paths[compression_format] A_ : Tuple = tmp_path / """cache""" A_ : Tuple = DownloadConfig(cache_dir=a_ , extract_compressed_file=a_ ) A_ : Dict = cached_path(a_ , download_config=a_ ) with open(a_ ) as f: A_ : Optional[Any] = f.read() with open(a_ ) as f: A_ : List[str] = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ ) -> str: """simple docstring""" A_ : Union[str, Any] = """custom_cache""" A_ : List[str] = """custom_extracted_dir""" A_ : Optional[Any] = tmp_path / """custom_extracted_path""" if default_extracted: A_ : Any = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , a_ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(a_ ) ) A_ : Union[str, Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) A_ : List[Any] = xz_file A_ : Optional[int] = ( DownloadConfig(extract_compressed_file=a_ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=a_ ) ) A_ : Union[str, Any] = cached_path(a_ , download_config=a_ ) assert Path(a_ ).parent.parts[-2:] == expected def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" A_ : str = str(Path(a_ ).resolve() ) assert cached_path(a_ ) == text_file # relative path A_ : List[str] = str(Path(a_ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(a_ ) == text_file def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : Optional[Any] = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(a_ ): cached_path(a_ ) # relative path A_ : Tuple = """./__missing_file__.txt""" with pytest.raises(a_ ): cached_path(a_ ) def UpperCAmelCase ( a_ ) -> Tuple: """simple docstring""" A_ : Any = get_from_cache(F"tmp://{tmpfs_file}" ) with open(a_ ) as f: A_ : List[str] = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( ) -> List[str]: """simple docstring""" with pytest.raises(a_ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Union[str, Any]: """simple docstring""" A_ : List[str] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): http_get("""https://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" A_ : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): ftp_get("""ftp://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , a_ ) def UpperCAmelCase ( a_ ) -> Optional[int]: """simple docstring""" A_ : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(a_ ): fsspec_get("""s3://huggingface.co""" , temp_file=a_ ) with pytest.raises(a_ ): fsspec_head("""s3://huggingface.co""" )
344
0
import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase_ = logging.get_logger(__name__) def lowerCamelCase_ ( _a : Tuple ): '''simple docstring''' UpperCAmelCase_ : str = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: UpperCAmelCase_ : Optional[Any] = 128 elif "12-12" in model_name: UpperCAmelCase_ : List[str] = 12 UpperCAmelCase_ : Optional[int] = 12 elif "14-14" in model_name: UpperCAmelCase_ : Any = 14 UpperCAmelCase_ : List[Any] = 14 elif "16-16" in model_name: UpperCAmelCase_ : Dict = 16 UpperCAmelCase_ : Optional[int] = 16 else: raise ValueError("""Model not supported""" ) UpperCAmelCase_ : str = """huggingface/label-files""" if "speech-commands" in model_name: UpperCAmelCase_ : Union[str, Any] = 35 UpperCAmelCase_ : Tuple = """speech-commands-v2-id2label.json""" else: UpperCAmelCase_ : int = 527 UpperCAmelCase_ : Tuple = """audioset-id2label.json""" UpperCAmelCase_ : int = json.load(open(hf_hub_download(_a , _a , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase_ : Tuple = {int(_a ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} return config def lowerCamelCase_ ( _a : Union[str, Any] ): '''simple docstring''' if "module.v" in name: UpperCAmelCase_ : List[str] = name.replace("""module.v""" , """audio_spectrogram_transformer""" ) if "cls_token" in name: UpperCAmelCase_ : List[Any] = name.replace("""cls_token""" , """embeddings.cls_token""" ) if "dist_token" in name: UpperCAmelCase_ : Dict = name.replace("""dist_token""" , """embeddings.distillation_token""" ) if "pos_embed" in name: UpperCAmelCase_ : Any = name.replace("""pos_embed""" , """embeddings.position_embeddings""" ) if "patch_embed.proj" in name: UpperCAmelCase_ : Dict = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) # transformer blocks if "blocks" in name: UpperCAmelCase_ : int = name.replace("""blocks""" , """encoder.layer""" ) if "attn.proj" in name: UpperCAmelCase_ : Any = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: UpperCAmelCase_ : Union[str, Any] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: UpperCAmelCase_ : Optional[Any] = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: UpperCAmelCase_ : Dict = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: UpperCAmelCase_ : str = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: UpperCAmelCase_ : Tuple = name.replace("""mlp.fc2""" , """output.dense""" ) # final layernorm if "audio_spectrogram_transformer.norm" in name: UpperCAmelCase_ : str = name.replace("""audio_spectrogram_transformer.norm""" , """audio_spectrogram_transformer.layernorm""" ) # classifier head if "module.mlp_head.0" in name: UpperCAmelCase_ : List[str] = name.replace("""module.mlp_head.0""" , """classifier.layernorm""" ) if "module.mlp_head.1" in name: UpperCAmelCase_ : Dict = name.replace("""module.mlp_head.1""" , """classifier.dense""" ) return name def lowerCamelCase_ ( _a : Optional[int] , _a : List[str] ): '''simple docstring''' for key in orig_state_dict.copy().keys(): UpperCAmelCase_ : Tuple = orig_state_dict.pop(_a ) if "qkv" in key: UpperCAmelCase_ : Optional[Any] = key.split(""".""" ) UpperCAmelCase_ : Union[str, Any] = int(key_split[3] ) UpperCAmelCase_ : str = config.hidden_size if "weight" in key: UpperCAmelCase_ : Any = val[:dim, :] UpperCAmelCase_ : Optional[Any] = val[dim : dim * 2, :] UpperCAmelCase_ : Optional[Any] = val[-dim:, :] else: UpperCAmelCase_ : str = val[:dim] UpperCAmelCase_ : str = val[dim : dim * 2] UpperCAmelCase_ : Dict = val[-dim:] else: UpperCAmelCase_ : Tuple = val return orig_state_dict def lowerCamelCase_ ( _a : Tuple ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = [ """module.v.head.weight""", """module.v.head.bias""", """module.v.head_dist.weight""", """module.v.head_dist.bias""", ] for k in ignore_keys: state_dict.pop(_a , _a ) @torch.no_grad() def lowerCamelCase_ ( _a : str , _a : Any , _a : List[str]=False ): '''simple docstring''' UpperCAmelCase_ : Dict = get_audio_spectrogram_transformer_config(_a ) UpperCAmelCase_ : List[str] = { """ast-finetuned-audioset-10-10-0.4593""": ( """https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1""" ), """ast-finetuned-audioset-10-10-0.450""": ( """https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1""" ), """ast-finetuned-audioset-10-10-0.448""": ( """https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1""" ), """ast-finetuned-audioset-10-10-0.448-v2""": ( """https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1""" ), """ast-finetuned-audioset-12-12-0.447""": ( """https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1""" ), """ast-finetuned-audioset-14-14-0.443""": ( """https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1""" ), """ast-finetuned-audioset-16-16-0.442""": ( """https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1""" ), """ast-finetuned-speech-commands-v2""": ( """https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1""" ), } # load original state_dict UpperCAmelCase_ : Optional[Any] = model_name_to_url[model_name] UpperCAmelCase_ : Tuple = torch.hub.load_state_dict_from_url(_a , map_location="""cpu""" ) # remove some keys remove_keys(_a ) # rename some keys UpperCAmelCase_ : List[str] = convert_state_dict(_a , _a ) # load 🤗 model UpperCAmelCase_ : Optional[Any] = ASTForAudioClassification(_a ) model.eval() model.load_state_dict(_a ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 UpperCAmelCase_ : List[str] = -4.2_6_7_7_3_9_3 if """speech-commands""" not in model_name else -6.8_4_5_9_7_8 UpperCAmelCase_ : Union[str, Any] = 4.5_6_8_9_9_7_4 if """speech-commands""" not in model_name else 5.5_6_5_4_5_2_6 UpperCAmelCase_ : Any = 1024 if """speech-commands""" not in model_name else 128 UpperCAmelCase_ : Optional[int] = ASTFeatureExtractor(mean=_a , std=_a , max_length=_a ) if "speech-commands" in model_name: UpperCAmelCase_ : str = load_dataset("""speech_commands""" , """v0.02""" , split="""validation""" ) UpperCAmelCase_ : List[str] = dataset[0]["""audio"""]["""array"""] else: UpperCAmelCase_ : Dict = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""" , filename="""sample_audio.flac""" , repo_type="""dataset""" , ) UpperCAmelCase_ , UpperCAmelCase_ : Any = torchaudio.load(_a ) UpperCAmelCase_ : str = waveform.squeeze().numpy() UpperCAmelCase_ : Any = feature_extractor(_a , sampling_rate=1_6000 , return_tensors="""pt""" ) # forward pass UpperCAmelCase_ : Union[str, Any] = model(**_a ) UpperCAmelCase_ : Union[str, Any] = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": UpperCAmelCase_ : Any = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": UpperCAmelCase_ : List[Any] = torch.tensor([-1.1_9_8_6, -7.0_9_0_3, -8.2_7_1_8] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": UpperCAmelCase_ : Union[str, Any] = torch.tensor([-2.6_1_2_8, -8.0_0_8_0, -9.4_3_4_4] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": UpperCAmelCase_ : Dict = torch.tensor([-1.5_0_8_0, -7.4_5_3_4, -8.8_9_1_7] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": UpperCAmelCase_ : Tuple = torch.tensor([-0.5_0_5_0, -6.5_8_3_3, -8.0_8_4_3] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": UpperCAmelCase_ : Union[str, Any] = torch.tensor([-0.3_8_2_6, -7.0_3_3_6, -8.2_4_1_3] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": UpperCAmelCase_ : str = torch.tensor([-1.2_1_1_3, -6.9_1_0_1, -8.3_4_7_0] ) elif model_name == "ast-finetuned-speech-commands-v2": UpperCAmelCase_ : List[Any] = torch.tensor([6.1_5_8_9, -8.0_5_6_6, -8.7_9_8_4] ) else: raise ValueError("""Unknown model name""" ) if not torch.allclose(logits[0, :3] , _a , atol=1E-4 ): raise ValueError("""Logits don't match""" ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(_a ).mkdir(exist_ok=_a ) print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_a ) print(F'''Saving feature extractor to {pytorch_dump_folder_path}''' ) feature_extractor.save_pretrained(_a ) if push_to_hub: print("""Pushing model and feature extractor to the hub...""" ) model.push_to_hub(F'''MIT/{model_name}''' ) feature_extractor.push_to_hub(F'''MIT/{model_name}''' ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''ast-finetuned-audioset-10-10-0.4593''', type=str, help='''Name of the Audio Spectrogram Transformer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) UpperCamelCase_ = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
345
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : '''simple docstring''' def __init__( self: Tuple ,lowerCamelCase_: List[str] ,lowerCamelCase_: int=13 ,lowerCamelCase_: int=32 ,lowerCamelCase_: Optional[int]=2 ,lowerCamelCase_: Any=3 ,lowerCamelCase_: str=16 ,lowerCamelCase_: Optional[Any]=[1, 2, 1] ,lowerCamelCase_: Tuple=[2, 2, 4] ,lowerCamelCase_: int=2 ,lowerCamelCase_: List[Any]=2.0 ,lowerCamelCase_: str=True ,lowerCamelCase_: Optional[int]=0.0 ,lowerCamelCase_: List[Any]=0.0 ,lowerCamelCase_: List[str]=0.1 ,lowerCamelCase_: Tuple="gelu" ,lowerCamelCase_: Union[str, Any]=False ,lowerCamelCase_: Union[str, Any]=True ,lowerCamelCase_: Optional[int]=0.0_2 ,lowerCamelCase_: int=1e-5 ,lowerCamelCase_: Optional[int]=True ,lowerCamelCase_: Union[str, Any]=None ,lowerCamelCase_: Union[str, Any]=True ,lowerCamelCase_: Optional[int]=10 ,lowerCamelCase_: Tuple=8 ,) -> List[Any]: UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = batch_size UpperCAmelCase_ : int = image_size UpperCAmelCase_ : Union[str, Any] = patch_size UpperCAmelCase_ : Optional[Any] = num_channels UpperCAmelCase_ : int = embed_dim UpperCAmelCase_ : Union[str, Any] = depths UpperCAmelCase_ : List[str] = num_heads UpperCAmelCase_ : int = window_size UpperCAmelCase_ : List[str] = mlp_ratio UpperCAmelCase_ : Tuple = qkv_bias UpperCAmelCase_ : Tuple = hidden_dropout_prob UpperCAmelCase_ : str = attention_probs_dropout_prob UpperCAmelCase_ : Tuple = drop_path_rate UpperCAmelCase_ : List[str] = hidden_act UpperCAmelCase_ : int = use_absolute_embeddings UpperCAmelCase_ : Any = patch_norm UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : Tuple = initializer_range UpperCAmelCase_ : Optional[Any] = is_training UpperCAmelCase_ : Dict = scope UpperCAmelCase_ : int = use_labels UpperCAmelCase_ : Optional[Any] = type_sequence_label_size UpperCAmelCase_ : List[str] = encoder_stride def A__ ( self: Any ) -> int: UpperCAmelCase_ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : List[Any] = None if self.use_labels: UpperCAmelCase_ : Optional[int] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) UpperCAmelCase_ : str = self.get_config() return config, pixel_values, labels def A__ ( self: List[Any] ) -> Union[str, Any]: return SwinvaConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,depths=self.depths ,num_heads=self.num_heads ,window_size=self.window_size ,mlp_ratio=self.mlp_ratio ,qkv_bias=self.qkv_bias ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,drop_path_rate=self.drop_path_rate ,hidden_act=self.hidden_act ,use_absolute_embeddings=self.use_absolute_embeddings ,path_norm=self.patch_norm ,layer_norm_eps=self.layer_norm_eps ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def A__ ( self: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Union[str, Any] ,lowerCamelCase_: List[str] ) -> str: UpperCAmelCase_ : str = SwinvaModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[Any] = model(lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCAmelCase_ : List[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, expected_seq_len, expected_dim) ) def A__ ( self: List[Any] ,lowerCamelCase_: List[Any] ,lowerCamelCase_: int ,lowerCamelCase_: int ) -> int: UpperCAmelCase_ : Any = SwinvaForMaskedImageModeling(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Union[str, Any] = model(lowerCamelCase_ ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase_ : str = 1 UpperCAmelCase_ : Optional[Any] = SwinvaForMaskedImageModeling(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : int = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def A__ ( self: int ,lowerCamelCase_: int ,lowerCamelCase_: List[Any] ,lowerCamelCase_: Optional[Any] ) -> int: UpperCAmelCase_ : Union[str, Any] = self.type_sequence_label_size UpperCAmelCase_ : int = SwinvaForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() UpperCAmelCase_ : Optional[int] = model(lowerCamelCase_ ,labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def A__ ( self: str ) -> Union[str, Any]: UpperCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = config_and_inputs UpperCAmelCase_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _snake_case ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' A__ : Tuple = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) A__ : Optional[Any] = ( {"feature-extraction": SwinvaModel, "image-classification": SwinvaForImageClassification} if is_torch_available() else {} ) A__ : List[Any] = False A__ : Tuple = False A__ : int = False A__ : Union[str, Any] = False def A__ ( self: List[str] ) -> Optional[Any]: UpperCAmelCase_ : Any = SwinvaModelTester(self ) UpperCAmelCase_ : str = ConfigTester(self ,config_class=lowerCamelCase_ ,embed_dim=37 ) def A__ ( self: Optional[int] ) -> List[Any]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self: Any ) -> Dict: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) @unittest.skip(reason="""Got `CUDA error: misaligned address` with PyTorch 2.0.0.""" ) def A__ ( self: int ) -> Dict: pass @unittest.skip(reason="""Swinv2 does not use inputs_embeds""" ) def A__ ( self: Tuple ) -> List[str]: pass def A__ ( self: str ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : int = model_class(lowerCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) UpperCAmelCase_ : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase_ ,nn.Linear ) ) def A__ ( self: Optional[Any] ) -> Optional[int]: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Dict = model_class(lowerCamelCase_ ) UpperCAmelCase_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : int = [*signature.parameters.keys()] UpperCAmelCase_ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,lowerCamelCase_ ) def A__ ( self: Union[str, Any] ) -> Optional[Any]: UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Any = True for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[Any] = True UpperCAmelCase_ : Union[str, Any] = False UpperCAmelCase_ : str = True UpperCAmelCase_ : List[Any] = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Optional[int] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : Optional[Any] = outputs.attentions UpperCAmelCase_ : List[str] = len(self.model_tester.depths ) self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase_ : str = True UpperCAmelCase_ : Optional[Any] = config.window_size**2 UpperCAmelCase_ : Optional[int] = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : List[Any] = outputs.attentions self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) UpperCAmelCase_ : Optional[Any] = len(lowerCamelCase_ ) # Check attention is always last and order is fine UpperCAmelCase_ : Tuple = True UpperCAmelCase_ : List[Any] = True UpperCAmelCase_ : Tuple = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Union[str, Any] = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) if hasattr(self.model_tester ,"""num_hidden_states_types""" ): UpperCAmelCase_ : List[Any] = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states UpperCAmelCase_ : List[str] = 2 self.assertEqual(out_len + added_hidden_states ,len(lowerCamelCase_ ) ) UpperCAmelCase_ : Any = outputs.attentions self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) def A__ ( self: List[str] ,lowerCamelCase_: Dict ,lowerCamelCase_: Tuple ,lowerCamelCase_: Optional[Any] ,lowerCamelCase_: Optional[int] ) -> List[Any]: UpperCAmelCase_ : str = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : int = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) UpperCAmelCase_ : List[str] = outputs.hidden_states UpperCAmelCase_ : Optional[Any] = getattr( self.model_tester ,"""expected_num_hidden_layers""" ,len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) # Swinv2 has a different seq_length UpperCAmelCase_ : Optional[Any] = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCAmelCase_ : int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) UpperCAmelCase_ : Optional[int] = outputs.reshaped_hidden_states self.assertEqual(len(lowerCamelCase_ ) ,lowerCamelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = reshaped_hidden_states[0].shape UpperCAmelCase_ : Optional[Any] = ( reshaped_hidden_states[0].view(lowerCamelCase_ ,lowerCamelCase_ ,height * width ).permute(0 ,2 ,1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) def A__ ( self: Any ) -> int: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Dict = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: UpperCAmelCase_ : Any = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : str = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) def A__ ( self: List[str] ) -> Dict: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Union[str, Any] = 3 UpperCAmelCase_ : Optional[int] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCAmelCase_ : List[str] = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCAmelCase_ : List[Any] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCAmelCase_ : Optional[Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: UpperCAmelCase_ : Optional[Any] = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,(padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : List[str] = True self.check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,(padded_height, padded_width) ) def A__ ( self: Optional[int] ) -> str: UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase_ ) def A__ ( self: Union[str, Any] ) -> Dict: UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @slow def A__ ( self: str ) -> Tuple: for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Dict = SwinvaModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def A__ ( self: Any ) -> int: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : List[str] = _config_zero_init(lowerCamelCase_ ) for model_class in self.all_model_classes: UpperCAmelCase_ : int = model_class(config=lowerCamelCase_ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' ,) @require_vision @require_torch class _snake_case ( unittest.TestCase ): '''simple docstring''' @cached_property def A__ ( self: Dict ) -> Optional[Any]: return ( AutoImageProcessor.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ) if is_vision_available() else None ) @slow def A__ ( self: str ) -> List[Any]: UpperCAmelCase_ : Tuple = SwinvaForImageClassification.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ).to( lowerCamelCase_ ) UpperCAmelCase_ : Any = self.default_image_processor UpperCAmelCase_ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) UpperCAmelCase_ : Optional[int] = image_processor(images=lowerCamelCase_ ,return_tensors="""pt""" ).to(lowerCamelCase_ ) # forward pass with torch.no_grad(): UpperCAmelCase_ : Optional[Any] = model(**lowerCamelCase_ ) # verify the logits UpperCAmelCase_ : Dict = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase_ ) UpperCAmelCase_ : Any = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(lowerCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCamelCase_ ,atol=1e-4 ) )
345
1