code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _A : Dict =logging.get_logger(__name__) _A : Any ={ '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class _lowercase ( _lowercase ): a = """table-transformer""" a = ["""past_key_values"""] a = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self: Dict , UpperCamelCase__: Any=True , UpperCamelCase__: str=None , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: Any=100 , UpperCamelCase__: str=6 , UpperCamelCase__: Tuple=2_048 , UpperCamelCase__: Any=8 , UpperCamelCase__: List[str]=6 , UpperCamelCase__: Union[str, Any]=2_048 , UpperCamelCase__: Dict=8 , UpperCamelCase__: List[Any]=0.0 , UpperCamelCase__: Optional[int]=0.0 , UpperCamelCase__: Dict=True , UpperCamelCase__: List[str]="relu" , UpperCamelCase__: Optional[Any]=256 , UpperCamelCase__: Dict=0.1 , UpperCamelCase__: Optional[Any]=0.0 , UpperCamelCase__: str=0.0 , UpperCamelCase__: List[str]=0.02 , UpperCamelCase__: Union[str, Any]=1.0 , UpperCamelCase__: Dict=False , UpperCamelCase__: Dict="sine" , UpperCamelCase__: str="resnet50" , UpperCamelCase__: List[str]=True , UpperCamelCase__: Tuple=False , UpperCamelCase__: List[str]=1 , UpperCamelCase__: Tuple=5 , UpperCamelCase__: Any=2 , UpperCamelCase__: Optional[int]=1 , UpperCamelCase__: Tuple=1 , UpperCamelCase__: List[Any]=5 , UpperCamelCase__: Optional[int]=2 , UpperCamelCase__: int=0.1 , **UpperCamelCase__: Optional[int] , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) lowerCamelCase__ : str = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : Optional[int] = backbone_config.get("""model_type""" ) lowerCamelCase__ : Tuple = CONFIG_MAPPING[backbone_model_type] lowerCamelCase__ : str = config_class.from_dict(UpperCamelCase__ ) # set timm attributes to None lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[int] = None, None, None lowerCamelCase__ : Union[str, Any] = use_timm_backbone lowerCamelCase__ : Optional[Any] = backbone_config lowerCamelCase__ : Any = num_channels lowerCamelCase__ : Tuple = num_queries lowerCamelCase__ : Dict = d_model lowerCamelCase__ : List[Any] = encoder_ffn_dim lowerCamelCase__ : int = encoder_layers lowerCamelCase__ : Union[str, Any] = encoder_attention_heads lowerCamelCase__ : Dict = decoder_ffn_dim lowerCamelCase__ : List[Any] = decoder_layers lowerCamelCase__ : Any = decoder_attention_heads lowerCamelCase__ : Any = dropout lowerCamelCase__ : List[Any] = attention_dropout lowerCamelCase__ : List[str] = activation_dropout lowerCamelCase__ : List[Any] = activation_function lowerCamelCase__ : Tuple = init_std lowerCamelCase__ : List[str] = init_xavier_std lowerCamelCase__ : Optional[int] = encoder_layerdrop lowerCamelCase__ : Any = decoder_layerdrop lowerCamelCase__ : int = encoder_layers lowerCamelCase__ : Tuple = auxiliary_loss lowerCamelCase__ : Tuple = position_embedding_type lowerCamelCase__ : List[Any] = backbone lowerCamelCase__ : Optional[Any] = use_pretrained_backbone lowerCamelCase__ : Any = dilation # Hungarian matcher lowerCamelCase__ : Dict = class_cost lowerCamelCase__ : Union[str, Any] = bbox_cost lowerCamelCase__ : Any = giou_cost # Loss coefficients lowerCamelCase__ : Any = mask_loss_coefficient lowerCamelCase__ : Dict = dice_loss_coefficient lowerCamelCase__ : Union[str, Any] = bbox_loss_coefficient lowerCamelCase__ : List[Any] = giou_loss_coefficient lowerCamelCase__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=UpperCamelCase__ , **UpperCamelCase__ ) @property def lowerCamelCase_ ( self: List[Any] ): return self.encoder_attention_heads @property def lowerCamelCase_ ( self: Optional[Any] ): return self.d_model class _lowercase ( _lowercase ): a = version.parse("""1.11""" ) @property def lowerCamelCase_ ( self: Any ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def lowerCamelCase_ ( self: List[Any] ): return 1e-5 @property def lowerCamelCase_ ( self: Tuple ): return 12
41
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=4 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[2, 2, 3, 2] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase=37 , UpperCAmelCase="gelu" , UpperCAmelCase=10 , UpperCAmelCase=0.02 , UpperCAmelCase=["stage2", "stage3", "stage4"] , UpperCAmelCase=3 , UpperCAmelCase=None , ) -> List[Any]: _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = num_stages _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = intermediate_size _snake_case = hidden_act _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = out_features _snake_case = num_labels _snake_case = scope _snake_case = num_stages def lowercase (self ) -> List[Any]: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def lowercase (self ) -> Tuple: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def lowercase (self ) -> Any: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCAmelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCAmelCase , loss_ignore_index=255 , num_labels=self.num_labels , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: _snake_case = UperNetForSemanticSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowercase (self ) -> Tuple: _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ), ( _snake_case ), ( _snake_case ), ) = config_and_inputs _snake_case = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCAmelCase_ = {"image-segmentation": UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> Optional[Any]: _snake_case = UperNetModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=37 ) def lowercase (self ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase (self ) -> Union[str, Any]: return def lowercase (self ) -> Union[str, Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def lowercase (self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> str: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> int: pass def lowercase (self ) -> List[str]: def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = _config_zero_init(UpperCAmelCase ) _snake_case = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _snake_case = model_class(config=UpperCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def lowercase (self ) -> Optional[Any]: pass @slow def lowercase (self ) -> Tuple: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = UperNetForSemanticSegmentation.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) _snake_case = Image.open(_SCREAMING_SNAKE_CASE ).convert("""RGB""" ) return image @require_torch @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) ) def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
341
0
'''simple docstring''' import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel lowercase : List[str] = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> Tuple: # save results if os.path.exists(__A ): if os.path.exists(os.path.join(__A , 'config.json' ) ) and os.path.isfile( os.path.join(__A , 'config.json' ) ): os.remove(os.path.join(__A , 'config.json' ) ) if os.path.exists(os.path.join(__A , 'pytorch_model.bin' ) ) and os.path.isfile( os.path.join(__A , 'pytorch_model.bin' ) ): os.remove(os.path.join(__A , 'pytorch_model.bin' ) ) else: os.makedirs(__A ) model.save_pretrained(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A=False ) -> List[Any]: _snake_case = 2 if unlogit: _snake_case = torch.pow(__A , __A ) _snake_case = p * torch.log(__A ) _snake_case = 0 return -plogp.sum(dim=-1 ) def SCREAMING_SNAKE_CASE__ ( __A ) -> Any: logger.info('lv, h >\t' + '\t'.join(F'{x + 1}' for x in range(len(__A ) ) ) ) for row in range(len(__A ) ): if tensor.dtype != torch.long: logger.info(F'layer {row + 1}:\t' + '\t'.join(F'{x:.5f}' for x in tensor[row].cpu().data ) ) else: logger.info(F'layer {row + 1}:\t' + '\t'.join(F'{x:d}' for x in tensor[row].cpu().data ) ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=True , __A=True , __A=None , __A=False ) -> Dict: _snake_case , _snake_case = model.config.num_hidden_layers, model.config.num_attention_heads _snake_case = torch.zeros(__A , __A ).to(args.device ) _snake_case = torch.zeros(__A , __A ).to(args.device ) if head_mask is None: _snake_case = torch.ones(__A , __A ).to(args.device ) head_mask.requires_grad_(requires_grad=__A ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: _snake_case = None _snake_case = 0.0 _snake_case = 0.0 for step, inputs in enumerate(tqdm(__A , desc='Iteration' , disable=args.local_rank not in [-1, 0] ) ): _snake_case = tuple(t.to(args.device ) for t in inputs ) ((_snake_case) , ) = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) _snake_case = model(__A , labels=__A , head_mask=__A ) # (loss), lm_logits, presents, (all hidden_states), (attentions) _snake_case , _snake_case , _snake_case = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(__A ): _snake_case = entropy(attn.detach() , __A ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(__A ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: _snake_case = 2 _snake_case = torch.pow(torch.pow(__A , __A ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-20 if not args.dont_normalize_global_importance: _snake_case = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info('Attention entropies' ) print_ad_tensor(__A ) if compute_importance: logger.info('Head importance scores' ) print_ad_tensor(__A ) logger.info('Head ranked by importance scores' ) _snake_case = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) _snake_case = torch.arange( head_importance.numel() , device=args.device ) _snake_case = head_ranks.view_as(__A ) print_ad_tensor(__A ) return attn_entropy, head_importance, total_loss def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> List[str]: _snake_case , _snake_case , _snake_case = compute_heads_importance(__A , __A , __A , compute_entropy=__A ) _snake_case = 1 / loss # instead of downsteam score use the LM loss logger.info('Pruning: original score: %f, threshold: %f' , __A , original_score * args.masking_threshold ) _snake_case = torch.ones_like(__A ) _snake_case = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) _snake_case = original_score while current_score >= original_score * args.masking_threshold: _snake_case = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads _snake_case = float('Inf' ) _snake_case = head_importance.view(-1 ).sort()[1] if len(__A ) <= num_to_mask: print('BREAK BY num_to_mask' ) break # mask heads _snake_case = current_heads_to_mask[:num_to_mask] logger.info('Heads to mask: %s' , str(current_heads_to_mask.tolist() ) ) _snake_case = new_head_mask.view(-1 ) _snake_case = 0.0 _snake_case = new_head_mask.view_as(__A ) _snake_case = new_head_mask.clone().detach() print_ad_tensor(__A ) # Compute metric and head importance again _snake_case , _snake_case , _snake_case = compute_heads_importance( __A , __A , __A , compute_entropy=__A , head_mask=__A ) _snake_case = 1 / loss logger.info( 'Masking: current score: %f, remaining heads %d (%.1f percents)' , __A , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info('Final head mask' ) print_ad_tensor(__A ) np.save(os.path.join(args.output_dir , 'head_mask.npy' ) , head_mask.detach().cpu().numpy() ) return head_mask def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A ) -> Optional[int]: _snake_case = datetime.now() _snake_case , _snake_case , _snake_case = compute_heads_importance( __A , __A , __A , compute_entropy=__A , compute_importance=__A , head_mask=__A ) _snake_case = 1 / loss _snake_case = datetime.now() - before_time _snake_case = sum(p.numel() for p in model.parameters() ) _snake_case = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(__A ) ) } for k, v in heads_to_prune.items(): if isinstance(__A , __A ): _snake_case = [ v, ] assert sum(len(__A ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(__A ) _snake_case = sum(p.numel() for p in model.parameters() ) _snake_case = datetime.now() _snake_case , _snake_case , _snake_case = compute_heads_importance( __A , __A , __A , compute_entropy=__A , compute_importance=__A , head_mask=__A , actually_pruned=__A , ) _snake_case = 1 / loss _snake_case = datetime.now() - before_time logger.info( 'Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)' , __A , __A , pruned_num_params / original_num_params * 100 , ) logger.info('Pruning: score with masking: %f score with pruning: %f' , __A , __A ) logger.info('Pruning: speed ratio (original timing / new timing): %f percents' , original_time / new_time * 100 ) save_model(__A , args.output_dir ) def SCREAMING_SNAKE_CASE__ ( ) -> Any: _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--data_dir' , default=__A , type=__A , required=__A , help='The input data dir. Should contain the .tsv files (or other data files) for the task.' , ) parser.add_argument( '--model_name_or_path' , default=__A , type=__A , required=__A , help='Path to pretrained model or model identifier from huggingface.co/models' , ) parser.add_argument( '--output_dir' , default=__A , type=__A , required=__A , help='The output directory where the model predictions and checkpoints will be written.' , ) # Other parameters parser.add_argument( '--config_name' , default='' , type=__A , help='Pretrained config name or path if not the same as model_name_or_path' , ) parser.add_argument( '--tokenizer_name' , default='' , type=__A , help='Pretrained tokenizer name or path if not the same as model_name_or_path' , ) parser.add_argument( '--cache_dir' , default=__A , type=__A , help='Where do you want to store the pre-trained models downloaded from s3' , ) parser.add_argument( '--data_subset' , type=__A , default=-1 , help='If > 0: limit the data to a subset of data_subset instances.' ) parser.add_argument( '--overwrite_output_dir' , action='store_true' , help='Whether to overwrite data in output directory' ) parser.add_argument( '--overwrite_cache' , action='store_true' , help='Overwrite the cached training and evaluation sets' ) parser.add_argument( '--dont_normalize_importance_by_layer' , action='store_true' , help='Don\'t normalize importance score by layers' ) parser.add_argument( '--dont_normalize_global_importance' , action='store_true' , help='Don\'t normalize all importance scores between 0 and 1' , ) parser.add_argument( '--try_masking' , action='store_true' , help='Whether to try to mask head until a threshold of accuracy.' ) parser.add_argument( '--masking_threshold' , default=0.9 , type=__A , help='masking threshold in term of metrics (stop masking when metric < threshold * original metric value).' , ) parser.add_argument( '--masking_amount' , default=0.1 , type=__A , help='Amount to heads to masking at each masking step.' ) parser.add_argument('--metric_name' , default='acc' , type=__A , help='Metric to use for head masking.' ) parser.add_argument( '--max_seq_length' , default=128 , type=__A , help=( 'The maximum total input sequence length after WordPiece tokenization. \n' 'Sequences longer than this will be truncated, sequences shorter padded.' ) , ) parser.add_argument('--batch_size' , default=1 , type=__A , help='Batch size.' ) parser.add_argument('--seed' , type=__A , default=42 ) parser.add_argument('--local_rank' , type=__A , default=-1 , help='local_rank for distributed training on gpus' ) parser.add_argument('--no_cuda' , action='store_true' , help='Whether not to use CUDA when available' ) parser.add_argument('--server_ip' , type=__A , default='' , help='Can be used for distant debugging.' ) parser.add_argument('--server_port' , type=__A , default='' , help='Can be used for distant debugging.' ) _snake_case = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('Waiting for debugger attach' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=__A ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: _snake_case = torch.device('cuda' if torch.cuda.is_available() and not args.no_cuda else 'cpu' ) _snake_case = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) _snake_case = torch.device('cuda' , args.local_rank ) _snake_case = 1 torch.distributed.init_process_group(backend='nccl' ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info('device: {} n_gpu: {}, distributed: {}'.format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) _snake_case = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: _snake_case = nn.parallel.DistributedDataParallel( __A , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=__A ) elif args.n_gpu > 1: _snake_case = nn.DataParallel(__A ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=__A ) torch.save(__A , os.path.join(args.output_dir , 'run_args.bin' ) ) logger.info('Training/evaluation parameters %s' , __A ) # Prepare dataset _snake_case = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) _snake_case = (torch.from_numpy(__A ),) _snake_case = TensorDataset(*__A ) _snake_case = RandomSampler(__A ) _snake_case = DataLoader(__A , sampler=__A , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(__A , __A , __A ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: _snake_case = mask_heads(__A , __A , __A ) prune_heads(__A , __A , __A , __A ) if __name__ == "__main__": main()
42
'''simple docstring''' import argparse from collections import defaultdict def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = f"""class {class_name}(""" _snake_case = f"""{4 * " "}def {test_name}(""" _snake_case = f"""{8 * " "}{correct_line.split()[0]}""" _snake_case = f"""{16 * " "}{correct_line.split()[0]}""" _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = 0 _snake_case = 0 _snake_case = [] for line in lines: if line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and in_func and (line.startswith(_SCREAMING_SNAKE_CASE ) or line.startswith(_SCREAMING_SNAKE_CASE )): _snake_case = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: _snake_case = True if in_class and in_func and in_line: if ")" not in line: continue else: _snake_case = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) _snake_case = _snake_case = _snake_case = _snake_case = False else: new_lines.append(_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: for line in new_lines: f.write(_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): if fail is not None: with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = {l.strip() for l in f.readlines()} else: _snake_case = None with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = defaultdict(_SCREAMING_SNAKE_CASE ) for line in correct_lines: _snake_case, _snake_case, _snake_case, _snake_case = line.split(""";""" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) __lowerCAmelCase = parser.parse_args() main(args.correct_filename, args.fail_filename)
341
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer __lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name __lowercase = ''' Examples: ```py >>> from PIL import Image >>> import torch >>> from diffusers import DiffusionPipeline >>> from diffusers.utils import export_to_gif, load_image >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu") >>> repo = "openai/shap-e-img2img" >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16) >>> pipe = pipe.to(device) >>> guidance_scale = 3.0 >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png" >>> image = load_image(image_url).convert("RGB") >>> images = pipe( ... image, ... guidance_scale=guidance_scale, ... num_inference_steps=64, ... frame_size=256, ... ).images >>> gif_path = export_to_gif(images[0], "corgi_3d.gif") ``` ''' @dataclass class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Union[PIL.Image.Image, np.ndarray] class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Optional[Any]: super().__init__() self.register_modules( prior=__lowercase , image_encoder=__lowercase , image_processor=__lowercase , scheduler=__lowercase , renderer=__lowercase , ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) -> List[Any]: if latents is None: __UpperCamelCase :Any = randn_tensor(__lowercase , generator=__lowercase , device=__lowercase , dtype=__lowercase) else: if latents.shape != shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {shape}""") __UpperCamelCase :List[Any] = latents.to(__lowercase) __UpperCamelCase :int = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , __lowercase=0) -> Any: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''') __UpperCamelCase :Tuple = torch.device(f"""cuda:{gpu_id}""") __UpperCamelCase :str = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__lowercase , __lowercase) @property def UpperCamelCase__ ( self) -> Tuple: if self.device != torch.device('''meta''') or not hasattr(self.image_encoder , '''_hf_hook'''): return self.device for module in self.image_encoder.modules(): if ( hasattr(__lowercase , '''_hf_hook''') and hasattr(module._hf_hook , '''execution_device''') and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device) return self.device def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , ) -> str: if isinstance(__lowercase , __lowercase) and isinstance(image[0] , torch.Tensor): __UpperCamelCase :Union[str, Any] = torch.cat(__lowercase , axis=0) if image[0].ndim == 4 else torch.stack(__lowercase , axis=0) if not isinstance(__lowercase , torch.Tensor): __UpperCamelCase :Any = self.image_processor(__lowercase , return_tensors='''pt''').pixel_values[0].unsqueeze(0) __UpperCamelCase :Optional[Any] = image.to(dtype=self.image_encoder.dtype , device=__lowercase) __UpperCamelCase :Union[str, Any] = self.image_encoder(__lowercase)['''last_hidden_state'''] __UpperCamelCase :Any = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 __UpperCamelCase :Any = image_embeds.repeat_interleave(__lowercase , dim=0) if do_classifier_free_guidance: __UpperCamelCase :Dict = torch.zeros_like(__lowercase) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __UpperCamelCase :Any = torch.cat([negative_image_embeds, image_embeds]) return image_embeds @torch.no_grad() @replace_example_docstring(__lowercase) def __call__( self , __lowercase , __lowercase = 1 , __lowercase = 25 , __lowercase = None , __lowercase = None , __lowercase = 4.0 , __lowercase = 64 , __lowercase = "pil" , __lowercase = True , ) -> List[Any]: if isinstance(__lowercase , PIL.Image.Image): __UpperCamelCase :List[Any] = 1 elif isinstance(__lowercase , torch.Tensor): __UpperCamelCase :str = image.shape[0] elif isinstance(__lowercase , __lowercase) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image)): __UpperCamelCase :Dict = len(__lowercase) else: raise ValueError( f"""`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(__lowercase)}""") __UpperCamelCase :Tuple = self._execution_device __UpperCamelCase :List[Any] = batch_size * num_images_per_prompt __UpperCamelCase :List[Any] = guidance_scale > 1.0 __UpperCamelCase :str = self._encode_image(__lowercase , __lowercase , __lowercase , __lowercase) # prior self.scheduler.set_timesteps(__lowercase , device=__lowercase) __UpperCamelCase :str = self.scheduler.timesteps __UpperCamelCase :str = self.prior.config.num_embeddings __UpperCamelCase :Optional[Any] = self.prior.config.embedding_dim __UpperCamelCase :List[str] = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , __lowercase , __lowercase , __lowercase , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim __UpperCamelCase :List[Any] = latents.reshape(latents.shape[0] , __lowercase , __lowercase) for i, t in enumerate(self.progress_bar(__lowercase)): # expand the latents if we are doing classifier free guidance __UpperCamelCase :int = torch.cat([latents] * 2) if do_classifier_free_guidance else latents __UpperCamelCase :Dict = self.scheduler.scale_model_input(__lowercase , __lowercase) __UpperCamelCase :List[Any] = self.prior( __lowercase , timestep=__lowercase , proj_embedding=__lowercase , ).predicted_image_embedding # remove the variance __UpperCamelCase , __UpperCamelCase :Dict = noise_pred.split( scaled_model_input.shape[2] , dim=2) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: __UpperCamelCase , __UpperCamelCase :Union[str, Any] = noise_pred.chunk(2) __UpperCamelCase :Optional[Any] = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) __UpperCamelCase :Optional[int] = self.scheduler.step( __lowercase , timestep=__lowercase , sample=__lowercase , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=__lowercase) __UpperCamelCase :List[Any] = [] for i, latent in enumerate(__lowercase): print() __UpperCamelCase :Any = self.renderer.decode( latent[None, :] , __lowercase , size=__lowercase , ray_batch_size=4_096 , n_coarse_samples=64 , n_fine_samples=128 , ) images.append(__lowercase) __UpperCamelCase :List[str] = torch.stack(__lowercase) if output_type not in ["np", "pil"]: raise ValueError(f"""Only the output types `pil` and `np` are supported not output_type={output_type}""") __UpperCamelCase :Optional[int] = images.cpu().numpy() if output_type == "pil": __UpperCamelCase :Optional[Any] = [self.numpy_to_pil(__lowercase) for image in images] # Offload last model to CPU if hasattr(self , '''final_offload_hook''') and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=__lowercase)
43
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _a : Union[str, Any] = logging.get_logger(__name__) _a : Any = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Union[str, Any] = "bart" _UpperCamelCase : Tuple = ["past_key_values"] _UpperCamelCase : Optional[int] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self , a__=50265 , a__=1024 , a__=12 , a__=4096 , a__=16 , a__=12 , a__=4096 , a__=16 , a__=0.0 , a__=0.0 , a__="gelu" , a__=1024 , a__=0.1 , a__=0.0 , a__=0.0 , a__=0.0_2 , a__=0.0 , a__=False , a__=True , a__=3 , a__=1 , a__=0 , a__=2 , a__=True , a__=2 , a__=2 , **a__ , ): _lowerCAmelCase : Optional[int] = vocab_size _lowerCAmelCase : Union[str, Any] = max_position_embeddings _lowerCAmelCase : Dict = d_model _lowerCAmelCase : List[Any] = encoder_ffn_dim _lowerCAmelCase : Tuple = encoder_layers _lowerCAmelCase : Any = encoder_attention_heads _lowerCAmelCase : Dict = decoder_ffn_dim _lowerCAmelCase : Any = decoder_layers _lowerCAmelCase : Dict = decoder_attention_heads _lowerCAmelCase : Union[str, Any] = dropout _lowerCAmelCase : List[Any] = attention_dropout _lowerCAmelCase : Union[str, Any] = activation_dropout _lowerCAmelCase : Optional[int] = activation_function _lowerCAmelCase : Optional[Any] = init_std _lowerCAmelCase : int = encoder_layerdrop _lowerCAmelCase : List[Any] = decoder_layerdrop _lowerCAmelCase : int = classifier_dropout _lowerCAmelCase : Optional[int] = use_cache _lowerCAmelCase : str = encoder_layers _lowerCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=a__ , pad_token_id=a__ , bos_token_id=a__ , eos_token_id=a__ , is_encoder_decoder=a__ , decoder_start_token_id=a__ , forced_eos_token_id=a__ , **a__ , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , a__ ): _lowerCAmelCase : Union[str, Any] = self.bos_token_id warnings.warn( F"Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. " """The config can simply be saved and uploaded again to be fixed.""" ) class __A ( SCREAMING_SNAKE_CASE_ ): @property def __A ( self ): if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase : Dict = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: _lowerCAmelCase : Optional[Any] = {0: """batch"""} _lowerCAmelCase : Optional[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: _lowerCAmelCase : str = {0: """batch""", 1: """decoder_sequence"""} _lowerCAmelCase : Dict = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(a__ , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. _lowerCAmelCase : str = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = self.num_layers for i in range(a__ ): _lowerCAmelCase : int = {0: """batch""", 2: """past_sequence + sequence"""} _lowerCAmelCase : int = {0: """batch""", 2: """past_sequence + sequence"""} else: _lowerCAmelCase : int = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def __A ( self ): if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase : List[Any] = super().outputs else: _lowerCAmelCase : List[str] = super(a__ , self ).outputs if self.use_past: _lowerCAmelCase , _lowerCAmelCase : Any = self.num_layers for i in range(a__ ): _lowerCAmelCase : str = {0: """batch""", 2: """past_sequence + sequence"""} _lowerCAmelCase : Optional[int] = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def __A ( self , a__ , a__ = -1 , a__ = -1 , a__ = False , a__ = None , ): _lowerCAmelCase : Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( a__ , a__ , a__ , a__ , a__ ) # Generate decoder inputs _lowerCAmelCase : List[Any] = seq_length if not self.use_past else 1 _lowerCAmelCase : Any = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( a__ , a__ , a__ , a__ , a__ ) _lowerCAmelCase : List[str] = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} _lowerCAmelCase : Optional[int] = dict(**a__ , **a__ ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch _lowerCAmelCase , _lowerCAmelCase : List[str] = common_inputs["""input_ids"""].shape _lowerCAmelCase : Tuple = common_inputs["""decoder_input_ids"""].shape[1] _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.num_attention_heads _lowerCAmelCase : List[Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) _lowerCAmelCase : List[Any] = decoder_seq_length + 3 _lowerCAmelCase : Tuple = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) _lowerCAmelCase : Union[str, Any] = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(a__ , a__ )] , dim=1 ) _lowerCAmelCase : str = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered _lowerCAmelCase , _lowerCAmelCase : List[str] = self.num_layers _lowerCAmelCase : Dict = min(a__ , a__ ) _lowerCAmelCase : List[Any] = max(a__ , a__ ) - min_num_layers _lowerCAmelCase : Union[str, Any] = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(a__ ): common_inputs["past_key_values"].append( ( torch.zeros(a__ ), torch.zeros(a__ ), torch.zeros(a__ ), torch.zeros(a__ ), ) ) # TODO: test this. _lowerCAmelCase : List[str] = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(a__ , a__ ): common_inputs["past_key_values"].append((torch.zeros(a__ ), torch.zeros(a__ )) ) return common_inputs def __A ( self , a__ , a__ = -1 , a__ = -1 , a__ = False , a__ = None , ): _lowerCAmelCase : Any = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( a__ , a__ , a__ , a__ , a__ ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values _lowerCAmelCase : Optional[int] = seqlen + 2 _lowerCAmelCase , _lowerCAmelCase : Dict = self.num_layers _lowerCAmelCase , _lowerCAmelCase : int = self.num_attention_heads _lowerCAmelCase : Tuple = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) _lowerCAmelCase : Optional[Any] = common_inputs["""attention_mask"""].dtype _lowerCAmelCase : Any = torch.cat( [common_inputs["""attention_mask"""], torch.ones(a__ , a__ , dtype=a__ )] , dim=1 ) _lowerCAmelCase : List[Any] = [ (torch.zeros(a__ ), torch.zeros(a__ )) for _ in range(a__ ) ] return common_inputs def __A ( self , a__ , a__ = -1 , a__ = -1 , a__ = False , a__ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _lowerCAmelCase : Tuple = compute_effective_axis_dimension( a__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _lowerCAmelCase : Union[str, Any] = tokenizer.num_special_tokens_to_add(a__ ) _lowerCAmelCase : Any = compute_effective_axis_dimension( a__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=a__ ) # Generate dummy inputs according to compute batch and sequence _lowerCAmelCase : int = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size _lowerCAmelCase : str = dict(tokenizer(a__ , return_tensors=a__ ) ) return common_inputs def __A ( self , a__ , a__ = -1 , a__ = -1 , a__ = False , a__ = None , ): if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase : Optional[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( a__ , batch_size=a__ , seq_length=a__ , is_pair=a__ , framework=a__ ) elif self.task == "causal-lm": _lowerCAmelCase : str = self._generate_dummy_inputs_for_causal_lm( a__ , batch_size=a__ , seq_length=a__ , is_pair=a__ , framework=a__ ) else: _lowerCAmelCase : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( a__ , batch_size=a__ , seq_length=a__ , is_pair=a__ , framework=a__ ) return common_inputs def __A ( self , a__ , a__ , a__ , a__ ): if self.task in ["default", "seq2seq-lm"]: _lowerCAmelCase : List[str] = super()._flatten_past_key_values_(a__ , a__ , a__ , a__ ) else: _lowerCAmelCase : Tuple = super(a__ , self )._flatten_past_key_values_( a__ , a__ , a__ , a__ )
44
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = ["pixel_values"] def __init__(self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: super().__init__(**UpperCAmelCase ) _snake_case = size if size is not None else {"""height""": 256, """width""": 256} _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( UpperCAmelCase , size=(size["""height"""], size["""width"""]) , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(UpperCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> List[Any]: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> PIL.Image.Image: _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = size if size is not None else self.size _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=UpperCAmelCase , size=UpperCAmelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] _snake_case = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _snake_case = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
341
0
"""simple docstring""" import pprint import requests lowercase_ = "https://zenquotes.io/api" def lowercase ( ) -> list: return requests.get(API_ENDPOINT_URL + '''/today''' ).json() def lowercase ( ) -> list: return requests.get(API_ENDPOINT_URL + '''/random''' ).json() if __name__ == "__main__": lowercase_ = random_quotes() pprint.pprint(response)
45
'''simple docstring''' __lowerCAmelCase = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure the supplied data is a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_SCREAMING_SNAKE_CASE ) _snake_case = """""".join(bin(_SCREAMING_SNAKE_CASE )[2:].zfill(8 ) for byte in data ) _snake_case = len(_SCREAMING_SNAKE_CASE ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b"""=""" * ((6 - len(_SCREAMING_SNAKE_CASE ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_SCREAMING_SNAKE_CASE ) % 6) else: _snake_case = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 6 ) ).encode() + padding ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = ( """argument should be a bytes-like object or ASCII string, """ f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_SCREAMING_SNAKE_CASE ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): try: _snake_case = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) _snake_case = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_SCREAMING_SNAKE_CASE ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 8 ) ] return bytes(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json" ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class lowercase ( _UpperCAmelCase ): _SCREAMING_SNAKE_CASE = 'roformer' def __init__( self , lowercase=50_000 , lowercase=None , lowercase=768 , lowercase=12 , lowercase=12 , lowercase=3_072 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=1_536 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=0 , lowercase=False , lowercase=True , **lowercase , ) -> Dict: super().__init__(pad_token_id=lowercase , **lowercase ) lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size if embedding_size is None else embedding_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = hidden_act lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = initializer_range lowerCAmelCase = layer_norm_eps lowerCAmelCase = rotary_value lowerCAmelCase = use_cache class lowercase ( _UpperCAmelCase ): @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowerCAmelCase = {0: """batch""", 1: """sequence"""} lowerCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
46
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def _lowerCAmelCase ( _UpperCamelCase : Optional[int] ) -> List[str]: """simple docstring""" _SCREAMING_SNAKE_CASE =botoa.client('iam' ) _SCREAMING_SNAKE_CASE ={ 'Version': '2012-10-17', 'Statement': [ {'Effect': 'Allow', 'Principal': {'Service': 'sagemaker.amazonaws.com'}, 'Action': 'sts:AssumeRole'} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=_UpperCamelCase , AssumeRolePolicyDocument=json.dumps(_UpperCamelCase , indent=2 ) ) _SCREAMING_SNAKE_CASE ={ 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 'sagemaker:*', 'ecr:GetDownloadUrlForLayer', 'ecr:BatchGetImage', 'ecr:BatchCheckLayerAvailability', 'ecr:GetAuthorizationToken', 'cloudwatch:PutMetricData', 'cloudwatch:GetMetricData', 'cloudwatch:GetMetricStatistics', 'cloudwatch:ListMetrics', 'logs:CreateLogGroup', 'logs:CreateLogStream', 'logs:DescribeLogStreams', 'logs:PutLogEvents', 'logs:GetLogEvents', 's3:CreateBucket', 's3:ListBucket', 's3:GetBucketLocation', 's3:GetObject', 's3:PutObject', ], 'Resource': '*', } ], } # attach policy to role iam_client.put_role_policy( RoleName=_UpperCamelCase , PolicyName=f"{role_name}_policy_permission" , PolicyDocument=json.dumps(_UpperCamelCase , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(f"role {role_name} already exists. Using existing one" ) def _lowerCAmelCase ( _UpperCamelCase : List[str] ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =botoa.client('iam' ) return iam_client.get_role(RoleName=_UpperCamelCase )["Role"]["Arn"] def _lowerCAmelCase ( ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE =_ask_options( 'How do you want to authorize?' , ['AWS Profile', 'Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) '] , _UpperCamelCase , ) _SCREAMING_SNAKE_CASE =None if credentials_configuration == 0: _SCREAMING_SNAKE_CASE =_ask_field('Enter your AWS Profile name: [default] ' , default='default' ) _SCREAMING_SNAKE_CASE =aws_profile else: print( 'Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,' '`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`' ) _SCREAMING_SNAKE_CASE =_ask_field('AWS Access Key ID: ' ) _SCREAMING_SNAKE_CASE =aws_access_key_id _SCREAMING_SNAKE_CASE =_ask_field('AWS Secret Access Key: ' ) _SCREAMING_SNAKE_CASE =aws_secret_access_key _SCREAMING_SNAKE_CASE =_ask_field('Enter your AWS Region: [us-east-1]' , default='us-east-1' ) _SCREAMING_SNAKE_CASE =aws_region _SCREAMING_SNAKE_CASE =_ask_options( 'Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?' , ['Provide IAM Role name', 'Create new IAM role using credentials'] , _UpperCamelCase , ) if role_management == 0: _SCREAMING_SNAKE_CASE =_ask_field('Enter your IAM role name: ' ) else: _SCREAMING_SNAKE_CASE ='accelerate_sagemaker_execution_role' print(f"Accelerate will create an iam role \"{iam_role_name}\" using the provided credentials" ) _create_iam_role_for_sagemaker(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to use custom Docker image? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =None if is_custom_docker_image: _SCREAMING_SNAKE_CASE =_ask_field('Enter your Docker image: ' , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to provide SageMaker input channels with data locations? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =None if is_sagemaker_inputs_enabled: _SCREAMING_SNAKE_CASE =_ask_field( 'Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): ' , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to enable SageMaker metrics? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =None if is_sagemaker_metrics_enabled: _SCREAMING_SNAKE_CASE =_ask_field( 'Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): ' , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() , ) _SCREAMING_SNAKE_CASE =_ask_options( 'What is the distributed mode?' , ['No distributed training', 'Data parallelism'] , _convert_sagemaker_distributed_mode , ) _SCREAMING_SNAKE_CASE ={} _SCREAMING_SNAKE_CASE =_ask_field( 'Do you wish to optimize your script with torch dynamo?[yes/NO]:' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) if use_dynamo: _SCREAMING_SNAKE_CASE ='dynamo_' _SCREAMING_SNAKE_CASE =_ask_options( 'Which dynamo backend would you like to use?' , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to customize the defaults sent to torch.compile? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) if use_custom_options: _SCREAMING_SNAKE_CASE =_ask_options( 'Which mode do you want to use?' , _UpperCamelCase , lambda _UpperCamelCase : TORCH_DYNAMO_MODES[int(_UpperCamelCase )] , default='default' , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE =_ask_field( 'Do you want to enable dynamic shape tracing? [yes/NO]: ' , _convert_yes_no_to_bool , default=_UpperCamelCase , error_message='Please enter yes or no.' , ) _SCREAMING_SNAKE_CASE ='Which EC2 instance type you want to use for your training?' if distributed_type != SageMakerDistributedType.NO: _SCREAMING_SNAKE_CASE =_ask_options( _UpperCamelCase , _UpperCamelCase , lambda _UpperCamelCase : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(_UpperCamelCase )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" _SCREAMING_SNAKE_CASE =_ask_field(_UpperCamelCase , lambda _UpperCamelCase : str(_UpperCamelCase ).lower() , default='ml.p3.2xlarge' ) _SCREAMING_SNAKE_CASE =1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): _SCREAMING_SNAKE_CASE =_ask_field( 'How many machines do you want use? [1]: ' , _UpperCamelCase , default=1 , ) _SCREAMING_SNAKE_CASE =_ask_options( 'Do you wish to use FP16 or BF16 (mixed precision)?' , ['no', 'fp16', 'bf16', 'fp8'] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( 'Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.' ) return SageMakerConfig( image_uri=_UpperCamelCase , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=_UpperCamelCase , use_cpu=_UpperCamelCase , dynamo_config=_UpperCamelCase , eca_instance_type=_UpperCamelCase , profile=_UpperCamelCase , region=_UpperCamelCase , iam_role_name=_UpperCamelCase , mixed_precision=_UpperCamelCase , num_machines=_UpperCamelCase , sagemaker_inputs_file=_UpperCamelCase , sagemaker_metrics_file=_UpperCamelCase , )
47
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Dict = {'vocab_file': 'vocab.txt'} SCREAMING_SNAKE_CASE__ : Tuple = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } SCREAMING_SNAKE_CASE__ : Dict = { 'YituTech/conv-bert-base': 512, 'YituTech/conv-bert-medium-small': 512, 'YituTech/conv-bert-small': 512, } SCREAMING_SNAKE_CASE__ : int = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : Optional[Any] = VOCAB_FILES_NAMES lowerCamelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ : Optional[int] = PRETRAINED_INIT_CONFIGURATION lowerCamelCase_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ : Dict = ConvBertTokenizer def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=True , UpperCamelCase__="[UNK]" , UpperCamelCase__="[SEP]" , UpperCamelCase__="[PAD]" , UpperCamelCase__="[CLS]" , UpperCamelCase__="[MASK]" , UpperCamelCase__=True , UpperCamelCase__=None , **UpperCamelCase__ , ) -> List[str]: super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , do_lower_case=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , tokenize_chinese_chars=UpperCamelCase__ , strip_accents=UpperCamelCase__ , **UpperCamelCase__ , ) lowerCamelCase : int = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , UpperCamelCase__ ) != do_lower_case or normalizer_state.get("strip_accents" , UpperCamelCase__ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , UpperCamelCase__ ) != tokenize_chinese_chars ): lowerCamelCase : Dict = getattr(UpperCamelCase__ , normalizer_state.pop("type" ) ) lowerCamelCase : Any = do_lower_case lowerCamelCase : Optional[Any] = strip_accents lowerCamelCase : Optional[int] = tokenize_chinese_chars lowerCamelCase : Optional[int] = normalizer_class(**UpperCamelCase__ ) lowerCamelCase : int = do_lower_case def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__=None ) -> int: lowerCamelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> List[int]: lowerCamelCase : Any = [self.sep_token_id] lowerCamelCase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> Tuple[str]: lowerCamelCase : Optional[Any] = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__ ) return tuple(UpperCamelCase__ )
48
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __lowerCAmelCase = TypeVar('T') __lowerCAmelCase = Union[List[T], Tuple[T, ...]] __lowerCAmelCase = Union[T, List[T], Dict[str, T]] __lowerCAmelCase = Union[str, bytes, os.PathLike]
341
0
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _A ( __UpperCAmelCase ,unittest.TestCase ): UpperCamelCase__ : Tuple = CodeGenTokenizer UpperCamelCase__ : Tuple = CodeGenTokenizerFast UpperCamelCase__ : int = True UpperCamelCase__ : List[str] = {'''add_prefix_space''': True} UpperCamelCase__ : str = False def _lowerCamelCase ( self : Optional[int]): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __a = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] __a = dict(zip(__SCREAMING_SNAKE_CASE , range(len(__SCREAMING_SNAKE_CASE)))) __a = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __a = {'''unk_token''': '''<unk>'''} __a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) __a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as fp: fp.write(json.dumps(__SCREAMING_SNAKE_CASE) + '''\n''') with open(self.merges_file , '''w''' , encoding='''utf-8''') as fp: fp.write('''\n'''.join(__SCREAMING_SNAKE_CASE)) def _lowerCamelCase ( self : Any , **__SCREAMING_SNAKE_CASE : Optional[int]): '''simple docstring''' kwargs.update(self.special_tokens_map) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Tuple , **__SCREAMING_SNAKE_CASE : Optional[int]): '''simple docstring''' kwargs.update(self.special_tokens_map) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : str): '''simple docstring''' __a = '''lower newer''' __a = '''lower newer''' return input_text, output_text def _lowerCamelCase ( self : List[str]): '''simple docstring''' __a = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) __a = '''lower newer''' __a = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] __a = tokenizer.tokenize(__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) __a = tokens + [tokenizer.unk_token] __a = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : int): '''simple docstring''' if not self.test_rust_tokenizer: return __a = self.get_tokenizer() __a = self.get_rust_tokenizer(add_prefix_space=__SCREAMING_SNAKE_CASE) __a = '''lower newer''' # Testing tokenization __a = tokenizer.tokenize(__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE) __a = rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # Testing conversion to ids without special tokens __a = tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE) __a = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # Testing conversion to ids with special tokens __a = self.get_rust_tokenizer(add_prefix_space=__SCREAMING_SNAKE_CASE) __a = tokenizer.encode(__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE) __a = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) # Testing the unknown token __a = tokens + [rust_tokenizer.unk_token] __a = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Union[str, Any] , *__SCREAMING_SNAKE_CASE : Optional[int] , **__SCREAMING_SNAKE_CASE : Dict): '''simple docstring''' pass def _lowerCamelCase ( self : str , __SCREAMING_SNAKE_CASE : str=15): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})'): __a = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE) # Simple input __a = '''This is a simple input''' __a = ['''This is a simple input 1''', '''This is a simple input 2'''] __a = ('''This is a simple input''', '''This is a pair''') __a = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(__SCREAMING_SNAKE_CASE , tokenizer_r.encode , __SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , padding='''max_length''') # Simple input self.assertRaises(__SCREAMING_SNAKE_CASE , tokenizer_r.encode_plus , __SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , padding='''max_length''') # Simple input self.assertRaises( __SCREAMING_SNAKE_CASE , tokenizer_r.batch_encode_plus , __SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , padding='''max_length''' , ) # Pair input self.assertRaises(__SCREAMING_SNAKE_CASE , tokenizer_r.encode , __SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , padding='''max_length''') # Pair input self.assertRaises(__SCREAMING_SNAKE_CASE , tokenizer_r.encode_plus , __SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , padding='''max_length''') # Pair input self.assertRaises( __SCREAMING_SNAKE_CASE , tokenizer_r.batch_encode_plus , __SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , padding='''max_length''' , ) def _lowerCamelCase ( self : Any): '''simple docstring''' __a = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''') # Simple input __a = '''This is a simple input''' __a = ['''This is a simple input looooooooong''', '''This is a simple input'''] __a = ('''This is a simple input''', '''This is a pair''') __a = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] __a = tokenizer.pad_token_id __a = tokenizer(__SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=30 , return_tensors='''np''') __a = tokenizer(__SCREAMING_SNAKE_CASE , padding=__SCREAMING_SNAKE_CASE , truncate=__SCREAMING_SNAKE_CASE , return_tensors='''np''') __a = tokenizer(*__SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=60 , return_tensors='''np''') __a = tokenizer(__SCREAMING_SNAKE_CASE , padding=__SCREAMING_SNAKE_CASE , truncate=__SCREAMING_SNAKE_CASE , return_tensors='''np''') # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30) self.assertTrue(pad_token_id in out_s['''input_ids''']) self.assertTrue(0 in out_s['''attention_mask''']) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0]) self.assertFalse(0 in out_sa['''attention_mask'''][0]) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1]) self.assertTrue(0 in out_sa['''attention_mask'''][1]) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60) self.assertTrue(pad_token_id in out_p['''input_ids''']) self.assertTrue(0 in out_p['''attention_mask''']) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0]) self.assertFalse(0 in out_pa['''attention_mask'''][0]) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1]) self.assertTrue(0 in out_pa['''attention_mask'''][1]) def _lowerCamelCase ( self : List[str]): '''simple docstring''' __a = '''$$$''' __a = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=__SCREAMING_SNAKE_CASE , add_bos_token=__SCREAMING_SNAKE_CASE) __a = '''This is a simple input''' __a = ['''This is a simple input 1''', '''This is a simple input 2'''] __a = tokenizer.bos_token_id __a = tokenizer(__SCREAMING_SNAKE_CASE) __a = tokenizer(__SCREAMING_SNAKE_CASE) self.assertEqual(out_s.input_ids[0] , __SCREAMING_SNAKE_CASE) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids)) __a = tokenizer.decode(out_s.input_ids) __a = tokenizer.batch_decode(out_sa.input_ids) self.assertEqual(decode_s.split()[0] , __SCREAMING_SNAKE_CASE) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa)) @slow def _lowerCamelCase ( self : Tuple): '''simple docstring''' __a = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''') __a = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' __a = '''\nif len_a > len_b: result = a\nelse: result = b''' __a = tokenizer.encode(__SCREAMING_SNAKE_CASE) __a = ['''^#''', re.escape('''<|endoftext|>'''), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] __a = tokenizer.decode(__SCREAMING_SNAKE_CASE , truncate_before_pattern=__SCREAMING_SNAKE_CASE) self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' pass
49
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: _snake_case = data _snake_case = previous _snake_case = next_node def __str__(self ) -> str: return f"""{self.data}""" def lowercase (self ) -> int: return self.data def lowercase (self ) -> Dict: return self.next def lowercase (self ) -> Union[str, Any]: return self.previous class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase ) -> List[str]: _snake_case = head def __iter__(self ) -> Optional[Any]: return self def lowercase (self ) -> str: if not self.current: raise StopIteration else: _snake_case = self.current.get_data() _snake_case = self.current.get_next() return value class _lowerCAmelCase : '''simple docstring''' def __init__(self ) -> Optional[int]: _snake_case = None # First node in list _snake_case = None # Last node in list def __str__(self ) -> Optional[int]: _snake_case = self.head _snake_case = [] while current is not None: nodes.append(current.get_data() ) _snake_case = current.get_next() return " ".join(str(UpperCAmelCase ) for node in nodes ) def __contains__(self , UpperCAmelCase ) -> int: _snake_case = self.head while current: if current.get_data() == value: return True _snake_case = current.get_next() return False def __iter__(self ) -> Union[str, Any]: return LinkedListIterator(self.head ) def lowercase (self ) -> str: if self.head: return self.head.get_data() return None def lowercase (self ) -> List[Any]: if self.tail: return self.tail.get_data() return None def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: _snake_case = node _snake_case = node else: self.insert_before_node(self.head , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: self.set_head(UpperCAmelCase ) else: self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: _snake_case = Node(UpperCAmelCase ) if self.head is None: self.set_head(UpperCAmelCase ) else: self.set_tail(UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.previous if node.get_previous() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.next if node.get_next() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = 1 _snake_case = Node(UpperCAmelCase ) _snake_case = self.head while node: if current_position == position: self.insert_before_node(UpperCAmelCase , UpperCAmelCase ) return current_position += 1 _snake_case = node.next self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> Node: _snake_case = self.head while node: if node.get_data() == item: return node _snake_case = node.get_next() raise Exception("""Node not found""" ) def lowercase (self , UpperCAmelCase ) -> Optional[int]: if (node := self.get_node(UpperCAmelCase )) is not None: if node == self.head: _snake_case = self.head.get_next() if node == self.tail: _snake_case = self.tail.get_previous() self.remove_node_pointers(UpperCAmelCase ) @staticmethod def lowercase (UpperCAmelCase ) -> None: if node.get_next(): _snake_case = node.previous if node.get_previous(): _snake_case = node.next _snake_case = None _snake_case = None def lowercase (self ) -> Dict: return self.head is None def __SCREAMING_SNAKE_CASE ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
341
0
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _UpperCAmelCase : int = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple[int, int]: def constraint_to_multiple_of(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=0 , _UpperCAmelCase=None ): lowerCamelCase__ : int = round(val / multiple ) * multiple if max_val is not None and x > max_val: lowerCamelCase__ : Tuple = math.floor(val / multiple ) * multiple if x < min_val: lowerCamelCase__ : Tuple = math.ceil(val / multiple ) * multiple return x lowerCamelCase__ : Any = (output_size, output_size) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else output_size lowerCamelCase__ , lowerCamelCase__ : Dict = get_image_size(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = output_size # determine new height and width lowerCamelCase__ : Dict = output_height / input_height lowerCamelCase__ : Optional[int] = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width lowerCamelCase__ : Union[str, Any] = scale_width else: # fit height lowerCamelCase__ : Tuple = scale_height lowerCamelCase__ : Optional[int] = constraint_to_multiple_of(scale_height * input_height , multiple=_UpperCAmelCase ) lowerCamelCase__ : Dict = constraint_to_multiple_of(scale_width * input_width , multiple=_UpperCAmelCase ) return (new_height, new_width) class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = ["""pixel_values"""] def __init__( self : Optional[int] , UpperCAmelCase : bool = True , UpperCAmelCase : Dict[str, int] = None , UpperCAmelCase : PILImageResampling = PILImageResampling.BILINEAR , UpperCAmelCase : bool = False , UpperCAmelCase : int = 1 , UpperCAmelCase : bool = True , UpperCAmelCase : Union[int, float] = 1 / 255 , UpperCAmelCase : bool = True , UpperCAmelCase : Optional[Union[float, List[float]]] = None , UpperCAmelCase : Optional[Union[float, List[float]]] = None , **UpperCAmelCase : Optional[Any] , ) -> None: super().__init__(**UpperCAmelCase ) lowerCamelCase__ : Optional[int] = size if size is not None else {'height': 384, 'width': 384} lowerCamelCase__ : str = get_size_dict(UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = do_resize lowerCamelCase__ : List[str] = size lowerCamelCase__ : Optional[int] = keep_aspect_ratio lowerCamelCase__ : Union[str, Any] = ensure_multiple_of lowerCamelCase__ : int = resample lowerCamelCase__ : str = do_rescale lowerCamelCase__ : List[Any] = rescale_factor lowerCamelCase__ : List[Any] = do_normalize lowerCamelCase__ : str = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase__ : Tuple = image_std if image_std is not None else IMAGENET_STANDARD_STD def A_ ( self : List[str] , UpperCAmelCase : np.ndarray , UpperCAmelCase : Dict[str, int] , UpperCAmelCase : bool = False , UpperCAmelCase : int = 1 , UpperCAmelCase : PILImageResampling = PILImageResampling.BICUBIC , UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase : str , ) -> np.ndarray: lowerCamelCase__ : Optional[int] = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) lowerCamelCase__ : Union[str, Any] = get_resize_output_image_size( UpperCAmelCase , output_size=(size['height'], size['width']) , keep_aspect_ratio=UpperCAmelCase , multiple=UpperCAmelCase , ) return resize(UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : List[str] , UpperCAmelCase : np.ndarray , UpperCAmelCase : Union[int, float] , UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase : int , ) -> str: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Optional[int] , UpperCAmelCase : np.ndarray , UpperCAmelCase : Union[float, List[float]] , UpperCAmelCase : Union[float, List[float]] , UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCAmelCase : str , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def A_ ( self : Union[str, Any] , UpperCAmelCase : ImageInput , UpperCAmelCase : bool = None , UpperCAmelCase : int = None , UpperCAmelCase : bool = None , UpperCAmelCase : int = None , UpperCAmelCase : PILImageResampling = None , UpperCAmelCase : bool = None , UpperCAmelCase : float = None , UpperCAmelCase : bool = None , UpperCAmelCase : Optional[Union[float, List[float]]] = None , UpperCAmelCase : Optional[Union[float, List[float]]] = None , UpperCAmelCase : Optional[Union[str, TensorType]] = None , UpperCAmelCase : ChannelDimension = ChannelDimension.FIRST , **UpperCAmelCase : List[str] , ) -> PIL.Image.Image: lowerCamelCase__ : Optional[int] = do_resize if do_resize is not None else self.do_resize lowerCamelCase__ : str = size if size is not None else self.size lowerCamelCase__ : int = get_size_dict(UpperCAmelCase ) lowerCamelCase__ : List[str] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio lowerCamelCase__ : Union[str, Any] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of lowerCamelCase__ : Any = resample if resample is not None else self.resample lowerCamelCase__ : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase__ : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase__ : Dict = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase__ : Dict = image_mean if image_mean is not None else self.image_mean lowerCamelCase__ : List[Any] = image_std if image_std is not None else self.image_std lowerCamelCase__ : Optional[Any] = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. lowerCamelCase__ : Dict = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: lowerCamelCase__ : Tuple = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_rescale: lowerCamelCase__ : Tuple = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: lowerCamelCase__ : List[Any] = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] lowerCamelCase__ : List[str] = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] lowerCamelCase__ : Tuple = {'pixel_values': images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase ) def A_ ( self : Dict , UpperCAmelCase : Union[str, Any] , UpperCAmelCase : List[Tuple] = None ) -> List[Any]: lowerCamelCase__ : Dict = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(UpperCAmelCase ) != len(UpperCAmelCase ): raise ValueError( 'Make sure that you pass in as many target sizes as the batch dimension of the logits' ) if is_torch_tensor(UpperCAmelCase ): lowerCamelCase__ : Dict = target_sizes.numpy() lowerCamelCase__ : int = [] for idx in range(len(UpperCAmelCase ) ): lowerCamelCase__ : Optional[int] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='bilinear' , align_corners=UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(UpperCAmelCase ) else: lowerCamelCase__ : Optional[Any] = logits.argmax(dim=1 ) lowerCamelCase__ : Dict = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
50
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput __lowerCAmelCase = 8 def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x * 255).int().clamp(0 , 255 ) _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c h w -> b c 1 h w""" ) _snake_case = ((x & mask) != 0).float() _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c d h w -> b (c d) h w""" ) _snake_case = bits * 2 - 1 return bits def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x > 0).int() _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE , dtype=torch.intaa ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b (c d) h w -> b c d h w""" , d=8 ) _snake_case = reduce(x * mask , """b c d h w -> b c h w""" , """sum""" ) return (dec / 255).clamp(0.0 , 1.0 ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _snake_case = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _snake_case = self.alphas_cumprod[timestep] _snake_case = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _snake_case = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _snake_case = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _snake_case = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _snake_case = model_output.device if torch.is_tensor(_SCREAMING_SNAKE_CASE ) else """cpu""" _snake_case = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise _snake_case = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="epsilon" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): _snake_case = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _snake_case, _snake_case = torch.split(_SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: _snake_case = None # 1. compute alphas, betas _snake_case = self.alphas_cumprod[t] _snake_case = self.alphas_cumprod[t - 1] if t > 0 else self.one _snake_case = 1 - alpha_prod_t _snake_case = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _snake_case = model_output else: raise ValueError(f"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _snake_case = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _snake_case = 0 if t > 0: _snake_case = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_SCREAMING_SNAKE_CASE ).to(model_output.device ) _snake_case = (self._get_variance(_SCREAMING_SNAKE_CASE , predicted_variance=_SCREAMING_SNAKE_CASE ) ** 0.5) * noise _snake_case = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 1.0 , ) -> Tuple: super().__init__() _snake_case = bit_scale _snake_case = ( ddim_bit_scheduler_step if isinstance(UpperCAmelCase , UpperCAmelCase ) else ddpm_bit_scheduler_step ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) @torch.no_grad() def __call__(self , UpperCAmelCase = 256 , UpperCAmelCase = 256 , UpperCAmelCase = 50 , UpperCAmelCase = None , UpperCAmelCase = 1 , UpperCAmelCase = "pil" , UpperCAmelCase = True , **UpperCAmelCase , ) -> Union[Tuple, ImagePipelineOutput]: _snake_case = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=UpperCAmelCase , ) _snake_case = decimal_to_bits(UpperCAmelCase ) * self.bit_scale _snake_case = latents.to(self.device ) self.scheduler.set_timesteps(UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _snake_case = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _snake_case = self.scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ).prev_sample _snake_case = bits_to_decimal(UpperCAmelCase ) if output_type == "pil": _snake_case = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase )
341
0
def A (__A : float , __A : float , __A : int ) -> float: """simple docstring""" if principal <= 0: raise Exception('''Principal borrowed must be > 0''' ) if rate_per_annum < 0: raise Exception('''Rate of interest must be >= 0''' ) if years_to_repay <= 0 or not isinstance(__A , __A ): raise Exception('''Years to repay must be an integer > 0''' ) # Yearly rate is divided by 12 to get monthly rate UpperCAmelCase_ = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly UpperCAmelCase_ = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
51
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 10**9 ): _snake_case = 1 _snake_case = 2 _snake_case = 0 _snake_case = 0 _snake_case = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value _snake_case = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f'''{solution() = }''')
341
0
import pytest import datasets # Import fixture modules as plugins __lowerCamelCase : Any = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ["integration", "unit"] ): continue item.add_marker(pytest.mark.unit ) def A_ ( _lowerCAmelCase ) -> Optional[int]: config.addinivalue_line("markers" , "torchaudio_latest: mark test to run with torchaudio>=0.12" ) @pytest.fixture(autouse=_lowerCAmelCase ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[int]: # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? UpperCamelCase : List[Any] = tmp_path_factory.getbasetemp() / "cache" UpperCamelCase : Optional[int] = test_hf_cache_home / "datasets" UpperCamelCase : str = test_hf_cache_home / "metrics" UpperCamelCase : Union[str, Any] = test_hf_cache_home / "modules" monkeypatch.setattr("datasets.config.HF_DATASETS_CACHE" , str(_lowerCAmelCase ) ) monkeypatch.setattr("datasets.config.HF_METRICS_CACHE" , str(_lowerCAmelCase ) ) monkeypatch.setattr("datasets.config.HF_MODULES_CACHE" , str(_lowerCAmelCase ) ) UpperCamelCase : Dict = test_hf_datasets_cache / "downloads" monkeypatch.setattr("datasets.config.DOWNLOADED_DATASETS_PATH" , str(_lowerCAmelCase ) ) UpperCamelCase : Tuple = test_hf_datasets_cache / "downloads" / "extracted" monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(_lowerCAmelCase ) ) @pytest.fixture(autouse=_lowerCAmelCase , scope="session" ) def A_ ( ) -> Union[str, Any]: datasets.disable_progress_bar() @pytest.fixture(autouse=_lowerCAmelCase ) def A_ ( _lowerCAmelCase ) -> Union[str, Any]: # don't take tests into account when counting downloads monkeypatch.setattr("datasets.config.HF_UPDATE_DOWNLOAD_COUNTS" , _lowerCAmelCase ) @pytest.fixture def A_ ( _lowerCAmelCase ) -> Optional[int]: # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr("sqlalchemy.util.deprecations.SILENCE_UBER_WARNING" , _lowerCAmelCase )
52
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "deberta-v2" def __init__(self , UpperCAmelCase=128100 , UpperCAmelCase=1536 , UpperCAmelCase=24 , UpperCAmelCase=24 , UpperCAmelCase=6144 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-7 , UpperCAmelCase=False , UpperCAmelCase=-1 , UpperCAmelCase=0 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=0 , UpperCAmelCase="gelu" , **UpperCAmelCase , ) -> List[str]: super().__init__(**UpperCAmelCase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = relative_attention _snake_case = max_relative_positions _snake_case = pad_token_id _snake_case = position_biased_input # Backwards compatibility if type(UpperCAmelCase ) == str: _snake_case = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case = pos_att_type _snake_case = vocab_size _snake_case = layer_norm_eps _snake_case = kwargs.get("""pooler_hidden_size""" , UpperCAmelCase ) _snake_case = pooler_dropout _snake_case = pooler_hidden_act class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def lowercase (self ) -> int: return 12 def lowercase (self , UpperCAmelCase , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = 3 , UpperCAmelCase = 40 , UpperCAmelCase = 40 , UpperCAmelCase = None , ) -> Mapping[str, Any]: _snake_case = super().generate_dummy_inputs(preprocessor=UpperCAmelCase , framework=UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
0
'''simple docstring''' import os from collections import deque import torch from torch.utils.data import Dataset class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : Union[str, Any] , __A : List[str]="" , __A : List[Any]="train" ): assert os.path.isdir(__A ) __UpperCamelCase = [] __UpperCamelCase = os.listdir(__A ) for story_filename in story_filenames_list: if "summary" in story_filename: continue __UpperCamelCase = os.path.join(__A , __A ) if not os.path.isfile(__A ): continue self.documents.append(__A ) def __len__( self : Dict ): return len(self.documents ) def __getitem__( self : Optional[int] , __A : Optional[int] ): __UpperCamelCase = self.documents[idx] __UpperCamelCase = document_path.split('/' )[-1] with open(__A , encoding='utf-8' ) as source: __UpperCamelCase = source.read() __UpperCamelCase , __UpperCamelCase = process_story(__A ) return document_name, story_lines, summary_lines def lowercase__ ( __lowercase : Optional[int] ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = list(filter(lambda __lowercase : len(__lowercase ) != 0 , [line.strip() for line in raw_story.split('\n' )] ) ) # for some unknown reason some lines miss a period, add it __UpperCamelCase = [_add_missing_period(__lowercase ) for line in nonempty_lines] # gather article lines __UpperCamelCase = [] __UpperCamelCase = deque(__lowercase ) while True: try: __UpperCamelCase = lines.popleft() if element.startswith('@highlight' ): break story_lines.append(__lowercase ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines __UpperCamelCase = list(filter(lambda __lowercase : not t.startswith('@highlight' ) , __lowercase ) ) return story_lines, summary_lines def lowercase__ ( __lowercase : int ) -> List[str]: """simple docstring""" __UpperCamelCase = ['.', '!', '?', '...', '\'', '`', '"', '\u2019', '\u2019', ')'] if line.startswith('@highlight' ): return line if line[-1] in END_TOKENS: return line return line + "." def lowercase__ ( __lowercase : str , __lowercase : List[Any] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" if len(__lowercase ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(__lowercase )) ) return sequence def lowercase__ ( __lowercase : List[Any] , __lowercase : int ) -> Optional[Any]: """simple docstring""" __UpperCamelCase = torch.ones_like(__lowercase ) __UpperCamelCase = sequence == pad_token_id __UpperCamelCase = 0 return mask def lowercase__ ( __lowercase : Optional[int] , __lowercase : int , __lowercase : Union[str, Any] ) -> Optional[int]: """simple docstring""" __UpperCamelCase = [tokenizer.encode(__lowercase ) for line in story_lines] __UpperCamelCase = [token for sentence in story_lines_token_ids for token in sentence] __UpperCamelCase = [tokenizer.encode(__lowercase ) for line in summary_lines] __UpperCamelCase = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def lowercase__ ( __lowercase : Dict , __lowercase : Union[str, Any] ) -> str: """simple docstring""" __UpperCamelCase = [] for sequence in batch: __UpperCamelCase = -1 __UpperCamelCase = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(__lowercase ) return torch.tensor(__lowercase )
53
'''simple docstring''' __lowerCAmelCase = [ (1_000, 'M'), (900, 'CM'), (500, 'D'), (400, 'CD'), (100, 'C'), (90, 'XC'), (50, 'L'), (40, 'XL'), (10, 'X'), (9, 'IX'), (5, 'V'), (4, 'IV'), (1, 'I'), ] def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1000} _snake_case = 0 _snake_case = 0 while place < len(_SCREAMING_SNAKE_CASE ): if (place + 1 < len(_SCREAMING_SNAKE_CASE )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = [] for arabic, roman in ROMAN: ((_snake_case), (_snake_case)) = divmod(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) result.append(roman * factor ) if number == 0: break return "".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import copy import importlib.metadata import json import os from dataclasses import dataclass from typing import Any, Dict, Union from packaging import version from ..utils import is_torch_available, logging if is_torch_available(): import torch a__ : Union[str, Any] = logging.get_logger(__name__) @dataclass class UpperCamelCase_ : """simple docstring""" def __init__( self : Union[str, Any] , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : List[Any]=False , UpperCAmelCase__ : List[Any]=6.0 , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : List[str]=False , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : Union[str, Any]="fp4" , UpperCAmelCase__ : Optional[int]=False , **UpperCAmelCase__ : List[str] , ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = load_in_abit __SCREAMING_SNAKE_CASE = load_in_abit __SCREAMING_SNAKE_CASE = llm_inta_threshold __SCREAMING_SNAKE_CASE = llm_inta_skip_modules __SCREAMING_SNAKE_CASE = llm_inta_enable_fpaa_cpu_offload __SCREAMING_SNAKE_CASE = llm_inta_has_fpaa_weight __SCREAMING_SNAKE_CASE = bnb_abit_quant_type __SCREAMING_SNAKE_CASE = bnb_abit_use_double_quant if bnb_abit_compute_dtype is None: __SCREAMING_SNAKE_CASE = torch.floataa elif isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): __SCREAMING_SNAKE_CASE = getattr(UpperCAmelCase__ , UpperCAmelCase__ ) elif isinstance(UpperCAmelCase__ , torch.dtype ): __SCREAMING_SNAKE_CASE = bnb_abit_compute_dtype else: raise ValueError("bnb_4bit_compute_dtype must be a string or a torch.dtype" ) self.post_init() def UpperCAmelCase_ ( self : Optional[Any] ) -> List[Any]: if not isinstance(self.llm_inta_threshold , UpperCAmelCase__ ): raise ValueError("llm_int8_threshold must be a float" ) if self.llm_inta_skip_modules is not None and not isinstance(self.llm_inta_skip_modules , UpperCAmelCase__ ): raise ValueError("llm_int8_skip_modules must be a list of strings" ) if not isinstance(self.llm_inta_enable_fpaa_cpu_offload , UpperCAmelCase__ ): raise ValueError("llm_int8_enable_fp32_cpu_offload must be a boolean" ) if not isinstance(self.llm_inta_has_fpaa_weight , UpperCAmelCase__ ): raise ValueError("llm_int8_has_fp16_weight must be a boolean" ) if self.bnb_abit_compute_dtype is not None and not isinstance(self.bnb_abit_compute_dtype , torch.dtype ): raise ValueError("bnb_4bit_compute_dtype must be torch.dtype" ) if not isinstance(self.bnb_abit_quant_type , UpperCAmelCase__ ): raise ValueError("bnb_4bit_quant_type must be a string" ) if not isinstance(self.bnb_abit_use_double_quant , UpperCAmelCase__ ): raise ValueError("bnb_4bit_use_double_quant must be a boolean" ) if self.load_in_abit and not version.parse(importlib.metadata.version("bitsandbytes" ) ) >= version.parse( "0.39.0" ): raise ValueError( "4 bit quantization requires bitsandbytes>=0.39.0 - please upgrade your bitsandbytes version" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Union[str, Any]: return self.load_in_abit or self.load_in_abit def UpperCAmelCase_ ( self : List[str] ) -> int: if self.load_in_abit: return "llm_int8" elif self.load_in_abit and self.bnb_abit_quant_type == "fp4": return "fp4" elif self.load_in_abit and self.bnb_abit_quant_type == "nf4": return "nf4" else: return None @classmethod def UpperCAmelCase_ ( cls : Dict , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : List[Any] ) -> List[str]: __SCREAMING_SNAKE_CASE = cls(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = [] for key, value in kwargs.items(): if hasattr(UpperCAmelCase__ , UpperCAmelCase__ ): setattr(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) to_remove.append(UpperCAmelCase__ ) for key in to_remove: kwargs.pop(UpperCAmelCase__ , UpperCAmelCase__ ) if return_unused_kwargs: return config, kwargs else: return config def UpperCAmelCase_ ( self : Union[str, Any] , UpperCAmelCase__ : Union[str, os.PathLike] ) -> List[Any]: with open(UpperCAmelCase__ , "w" , encoding="utf-8" ) as writer: __SCREAMING_SNAKE_CASE = self.to_dict() __SCREAMING_SNAKE_CASE = json.dumps(UpperCAmelCase__ , indent=2 , sort_keys=UpperCAmelCase__ ) + "\n" writer.write(UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Dict[str, Any]: __SCREAMING_SNAKE_CASE = copy.deepcopy(self.__dict__ ) __SCREAMING_SNAKE_CASE = str(output["bnb_4bit_compute_dtype"] ).split("." )[1] return output def __repr__( self : Optional[int] ) -> List[Any]: return F"""{self.__class__.__name__} {self.to_json_string()}""" def UpperCAmelCase_ ( self : List[str] , UpperCAmelCase__ : bool = True ) -> str: if use_diff is True: __SCREAMING_SNAKE_CASE = self.to_diff_dict() else: __SCREAMING_SNAKE_CASE = self.to_dict() return json.dumps(UpperCAmelCase__ , indent=2 , sort_keys=UpperCAmelCase__ ) + "\n" def UpperCAmelCase_ ( self : Optional[Any] ) -> Dict[str, Any]: __SCREAMING_SNAKE_CASE = self.to_dict() # get the default config dict __SCREAMING_SNAKE_CASE = BitsAndBytesConfig().to_dict() __SCREAMING_SNAKE_CASE = {} # only serialize values that differ from the default config for key, value in config_dict.items(): if value != default_config_dict[key]: __SCREAMING_SNAKE_CASE = value return serializable_config_dict
54
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __lowerCAmelCase = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['PerceiverFeatureExtractor'] __lowerCAmelCase = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
'''simple docstring''' import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError("""At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training""") # TF training parameters a_ : Optional[Any] = False a_ : int = False def __snake_case ( UpperCAmelCase_ : Namespace ): return TrainCommand(UpperCAmelCase_ ) class snake_case ( lowercase ): """simple docstring""" @staticmethod def snake_case ( UpperCamelCase ): """simple docstring""" lowerCamelCase_ = parser.add_parser("train" , help="CLI tool to train a model on a task." ) train_parser.add_argument( "--train_data" , type=UpperCamelCase , required=UpperCamelCase , help="path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences." , ) train_parser.add_argument( "--column_label" , type=UpperCamelCase , default=0 , help="Column of the dataset csv file with example labels." ) train_parser.add_argument( "--column_text" , type=UpperCamelCase , default=1 , help="Column of the dataset csv file with example texts." ) train_parser.add_argument( "--column_id" , type=UpperCamelCase , default=2 , help="Column of the dataset csv file with example ids." ) train_parser.add_argument( "--skip_first_row" , action="store_true" , help="Skip the first row of the csv file (headers)." ) train_parser.add_argument("--validation_data" , type=UpperCamelCase , default="" , help="path to validation dataset." ) train_parser.add_argument( "--validation_split" , type=UpperCamelCase , default=0.1 , help="if validation dataset is not provided, fraction of train dataset to use as validation dataset." , ) train_parser.add_argument("--output" , type=UpperCamelCase , default="./" , help="path to saved the trained model." ) train_parser.add_argument( "--task" , type=UpperCamelCase , default="text_classification" , help="Task to train the model on." ) train_parser.add_argument( "--model" , type=UpperCamelCase , default="bert-base-uncased" , help="Model's name or path to stored model." ) train_parser.add_argument("--train_batch_size" , type=UpperCamelCase , default=32 , help="Batch size for training." ) train_parser.add_argument("--valid_batch_size" , type=UpperCamelCase , default=64 , help="Batch size for validation." ) train_parser.add_argument("--learning_rate" , type=UpperCamelCase , default=3e-5 , help="Learning rate." ) train_parser.add_argument("--adam_epsilon" , type=UpperCamelCase , default=1e-08 , help="Epsilon for Adam optimizer." ) train_parser.set_defaults(func=UpperCamelCase ) def __init__( self , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = logging.get_logger("transformers-cli/training" ) lowerCamelCase_ = "tf" if is_tf_available() else "torch" os.makedirs(args.output , exist_ok=UpperCamelCase ) lowerCamelCase_ = args.output lowerCamelCase_ = args.column_label lowerCamelCase_ = args.column_text lowerCamelCase_ = args.column_id self.logger.info(f'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": lowerCamelCase_ = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f'''Loading dataset from {args.train_data}''' ) lowerCamelCase_ = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) lowerCamelCase_ = None if args.validation_data: self.logger.info(f'''Loading validation dataset from {args.validation_data}''' ) lowerCamelCase_ = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) lowerCamelCase_ = args.validation_split lowerCamelCase_ = args.train_batch_size lowerCamelCase_ = args.valid_batch_size lowerCamelCase_ = args.learning_rate lowerCamelCase_ = args.adam_epsilon def snake_case ( self ): """simple docstring""" if self.framework == "tf": return self.run_tf() return self.run_torch() def snake_case ( self ): """simple docstring""" raise NotImplementedError def snake_case ( self ): """simple docstring""" self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
55
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __lowerCAmelCase = 'platform' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ): if attention_mask is None: _snake_case = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _snake_case = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _snake_case = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=7 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=99 , UpperCAmelCase=16 , UpperCAmelCase=2 , UpperCAmelCase=4 , UpperCAmelCase=4 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , ) -> Union[str, Any]: _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id _snake_case = initializer_range def lowercase (self ) -> str: _snake_case = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _snake_case = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=UpperCAmelCase , ) _snake_case = prepare_blenderbot_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, inputs_dict def lowercase (self ) -> Dict: _snake_case, _snake_case = self.prepare_config_and_inputs() return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) @require_flax class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = 99 def lowercase (self ) -> Any: _snake_case = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _snake_case = input_ids.shape[0] _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case = self._get_config_and_data() _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = lm_model(input_ids=UpperCAmelCase ) _snake_case = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _snake_case = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _snake_case = lm_model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) _snake_case = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(UpperCAmelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class _lowerCAmelCase ( __snake_case , unittest.TestCase , __snake_case ): '''simple docstring''' lowerCAmelCase_ = True lowerCAmelCase_ = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCAmelCase_ = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def lowercase (self ) -> Any: _snake_case = FlaxBlenderbotModelTester(self ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _snake_case = model_class(UpperCAmelCase ) @jax.jit def encode_jitted(UpperCAmelCase , UpperCAmelCase=None , **UpperCAmelCase ): return model.encode(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase ) with self.subTest("""JIT Enabled""" ): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = model_class(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""] ) _snake_case = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): return model.decode( decoder_input_ids=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , encoder_outputs=UpperCAmelCase , ) with self.subTest("""JIT Enabled""" ): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def lowercase (self ) -> Any: for model_class_name in self.all_model_classes: _snake_case = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _snake_case = np.ones((1, 1) ) * model.config.eos_token_id _snake_case = model(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) @unittest.skipUnless(jax_device != """cpu""" , """3B test too slow on CPU.""" ) @slow def lowercase (self ) -> Dict: _snake_case = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 15, """max_length""": 25} _snake_case = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} _snake_case = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""" , from_pt=UpperCAmelCase ) _snake_case = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) _snake_case = ["""Sam"""] _snake_case = tokenizer(UpperCAmelCase , return_tensors="""jax""" ) _snake_case = model.generate(**UpperCAmelCase , **UpperCAmelCase ) _snake_case = """Sam is a great name. It means \"sun\" in Gaelic.""" _snake_case = tokenizer.batch_decode(UpperCAmelCase , **UpperCAmelCase ) assert generated_txt[0].strip() == tgt_text
341
0
'''simple docstring''' import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase = None, ) -> str: '''simple docstring''' snake_case_ = {} if train_file is not None: snake_case_ = [train_file] if eval_file is not None: snake_case_ = [eval_file] if test_file is not None: snake_case_ = [test_file] snake_case_ = datasets.load_dataset('''csv''', data_files=__UpperCAmelCase ) snake_case_ = list(ds[list(files.keys() )[0]].features.keys() ) snake_case_ = features_name.pop(__UpperCAmelCase ) snake_case_ = list(set(ds[list(files.keys() )[0]][label_name] ) ) snake_case_ = {label: i for i, label in enumerate(__UpperCAmelCase )} snake_case_ = tokenizer.model_input_names snake_case_ = {} if len(__UpperCAmelCase ) == 1: for k in files.keys(): snake_case_ = ds[k].map( lambda __UpperCAmelCase : tokenizer.batch_encode_plus( example[features_name[0]], truncation=__UpperCAmelCase, max_length=__UpperCAmelCase, padding='''max_length''' ), batched=__UpperCAmelCase, ) elif len(__UpperCAmelCase ) == 2: for k in files.keys(): snake_case_ = ds[k].map( lambda __UpperCAmelCase : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]), truncation=__UpperCAmelCase, max_length=__UpperCAmelCase, padding='''max_length''', ), batched=__UpperCAmelCase, ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: snake_case_ = {k: v for k, v in ex.items() if k in input_names} snake_case_ = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: snake_case_ = {k: v for k, v in ex.items() if k in input_names} snake_case_ = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: snake_case_ = {k: v for k, v in ex.items() if k in input_names} snake_case_ = labelaid[ex[label_name]] yield (d, label) snake_case_ = ( tf.data.Dataset.from_generator( __UpperCAmelCase, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: snake_case_ = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) snake_case_ = ( tf.data.Dataset.from_generator( __UpperCAmelCase, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: snake_case_ = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) snake_case_ = ( tf.data.Dataset.from_generator( __UpperCAmelCase, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: snake_case_ = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid a : Tuple = logging.getLogger(__name__) @dataclass class a : snake_case_ = field(metadata={"help": "Which column contains the label"} ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "The path of the training file"} ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "The path of the development file"} ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "The path of the test file"} ) snake_case_ = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class a : snake_case_ = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) def __magic_name__ ( ) -> Tuple: '''simple docstring''' snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) snake_case_ ,snake_case_ ,snake_case_ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO, ) logger.info( F"n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, " F"16-bits training: {training_args.fpaa}" ) logger.info(F"Training/evaluation parameters {training_args}" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path, cache_dir=model_args.cache_dir, ) snake_case_ ,snake_case_ ,snake_case_ ,snake_case_ = get_tfds( train_file=data_args.train_file, eval_file=data_args.dev_file, test_file=data_args.test_file, tokenizer=__UpperCAmelCase, label_column_id=data_args.label_column_id, max_seq_length=data_args.max_seq_length, ) snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path, num_labels=len(__UpperCAmelCase ), labelaid=__UpperCAmelCase, idalabel={id: label for label, id in labelaid.items()}, finetuning_task='''text-classification''', cache_dir=model_args.cache_dir, ) with training_args.strategy.scope(): snake_case_ = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path, from_pt=bool('''.bin''' in model_args.model_name_or_path ), config=__UpperCAmelCase, cache_dir=model_args.cache_dir, ) def compute_metrics(__UpperCAmelCase ) -> Dict: snake_case_ = np.argmax(p.predictions, axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer snake_case_ = TFTrainer( model=__UpperCAmelCase, args=__UpperCAmelCase, train_dataset=__UpperCAmelCase, eval_dataset=__UpperCAmelCase, compute_metrics=__UpperCAmelCase, ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation snake_case_ = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case_ = trainer.evaluate() snake_case_ = os.path.join(training_args.output_dir, '''eval_results.txt''' ) with open(__UpperCAmelCase, '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(F" {key} = {value}" ) writer.write(F"{key} = {value}\n" ) results.update(__UpperCAmelCase ) return results if __name__ == "__main__": main()
56
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=2 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=10 , UpperCAmelCase=3 , UpperCAmelCase=32 * 4 , UpperCAmelCase=32 * 6 , UpperCAmelCase=4 , UpperCAmelCase=32 , ) -> Optional[Any]: _snake_case = parent _snake_case = batch_size _snake_case = is_training _snake_case = use_auxiliary_loss _snake_case = num_queries _snake_case = num_channels _snake_case = min_size _snake_case = max_size _snake_case = num_labels _snake_case = mask_feature_size def lowercase (self ) -> str: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( UpperCAmelCase ) _snake_case = torch.ones([self.batch_size, self.min_size, self.max_size] , device=UpperCAmelCase ) _snake_case = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=UpperCAmelCase ) > 0.5 ).float() _snake_case = (torch.rand((self.batch_size, self.num_labels) , device=UpperCAmelCase ) > 0.5).long() _snake_case = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowercase (self ) -> Tuple: return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.prepare_config_and_inputs() _snake_case = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> int: _snake_case = output.encoder_hidden_states _snake_case = output.pixel_decoder_hidden_states _snake_case = output.transformer_decoder_hidden_states self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , config.decoder_config.decoder_layers ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Union[str, Any]: with torch.no_grad(): _snake_case = MaskFormerModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase , output_hidden_states=UpperCAmelCase ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(UpperCAmelCase , UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: _snake_case = MaskFormerForInstanceSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() def comm_check_on_output(UpperCAmelCase ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) _snake_case = model( pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowerCAmelCase_ = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> int: _snake_case = MaskFormerModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def lowercase (self ) -> int: self.config_tester.run_common_tests() def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> Any: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*UpperCAmelCase ) @unittest.skip(reason="""MaskFormer does not use inputs_embeds""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip(reason="""MaskFormer does not have a get_input_embeddings method""" ) def lowercase (self ) -> Optional[int]: pass @unittest.skip(reason="""MaskFormer is not a generative model""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""MaskFormer does not use token embeddings""" ) def lowercase (self ) -> Optional[int]: pass @require_torch_multi_gpu @unittest.skip( reason="""MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> Tuple: pass def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @slow def lowercase (self ) -> int: for model_name in ["facebook/maskformer-swin-small-coco"]: _snake_case = MaskFormerModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = (self.model_tester.min_size,) * 2 _snake_case = { """pixel_values""": torch.randn((2, 3, *size) , device=UpperCAmelCase ), """mask_labels""": torch.randn((2, 10, *size) , device=UpperCAmelCase ), """class_labels""": torch.zeros(2 , 10 , device=UpperCAmelCase ).long(), } _snake_case = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertTrue(outputs.attentions is not None ) def lowercase (self ) -> Tuple: if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ).loss loss.backward() def lowercase (self ) -> List[str]: # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = True _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) _snake_case = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _snake_case = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _snake_case = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _snake_case = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=UpperCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1E-4 def __SCREAMING_SNAKE_CASE ( ): _snake_case = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase (self ) -> Optional[int]: return ( MaskFormerImageProcessor.from_pretrained("""facebook/maskformer-swin-small-coco""" ) if is_vision_available() else None ) def lowercase (self ) -> str: _snake_case = MaskFormerModel.from_pretrained("""facebook/maskformer-swin-small-coco""" ).to(UpperCAmelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[str]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [ [1.6_5_1_2e0_0, -5.2_5_7_2e0_0, -3.3_5_1_9e0_0], [3.6_1_6_9e-0_2, -5.9_0_2_5e0_0, -2.9_3_1_3e0_0], [1.0_7_6_6e-0_4, -7.7_6_3_0e0_0, -5.1_2_6_3e0_0], ] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[Any]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-resnet101-coco-stuff""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> Tuple: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) _snake_case = inputs["""pixel_values"""].to(UpperCAmelCase ) _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""mask_labels"""]] _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""class_labels"""]] with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None )
341
0
"""simple docstring""" import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 A : int = { "return_dict": False, "output_hidden_states": True, "output_attentions": True, "torchscript": True, "torch_dtype": "float16", "use_bfloat16": True, "tf_legacy_loss": True, "pruned_heads": {"a": 1}, "tie_word_embeddings": False, "is_decoder": True, "cross_attention_hidden_size": 1_2_8, "add_cross_attention": True, "tie_encoder_decoder": True, "max_length": 5_0, "min_length": 3, "do_sample": True, "early_stopping": True, "num_beams": 3, "num_beam_groups": 3, "diversity_penalty": 0.5, "temperature": 2.0, "top_k": 1_0, "top_p": 0.7, "typical_p": 0.2, "repetition_penalty": 0.8, "length_penalty": 0.8, "no_repeat_ngram_size": 5, "encoder_no_repeat_ngram_size": 5, "bad_words_ids": [1, 2, 3], "num_return_sequences": 3, "chunk_size_feed_forward": 5, "output_scores": True, "return_dict_in_generate": True, "forced_bos_token_id": 2, "forced_eos_token_id": 3, "remove_invalid_values": True, "architectures": ["BertModel"], "finetuning_task": "translation", "id2label": {0: "label"}, "label2id": {"label": "0"}, "tokenizer_class": "BertTokenizerFast", "prefix": "prefix", "bos_token_id": 6, "pad_token_id": 7, "eos_token_id": 8, "sep_token_id": 9, "decoder_start_token_id": 1_0, "exponential_decay_length_penalty": (5, 1.01), "suppress_tokens": [0, 1], "begin_suppress_tokens": 2, "task_specific_params": {"translation": "some_params"}, "problem_type": "regression", } @is_staging_test class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @classmethod def snake_case ( cls ): __lowerCAmelCase = TOKEN HfFolder.save_token(__a ) @classmethod def snake_case ( cls ): try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def snake_case ( self ): __lowerCAmelCase = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) __lowerCAmelCase = BertConfig.from_pretrained(f"{USER}/test-config" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__a , getattr(__a , __a ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__a , repo_id="test-config" , push_to_hub=__a , use_auth_token=self._token ) __lowerCAmelCase = BertConfig.from_pretrained(f"{USER}/test-config" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__a , getattr(__a , __a ) ) def snake_case ( self ): __lowerCAmelCase = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) __lowerCAmelCase = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__a , getattr(__a , __a ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( __a , repo_id="valid_org/test-config-org" , push_to_hub=__a , use_auth_token=self._token ) __lowerCAmelCase = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(__a , getattr(__a , __a ) ) def snake_case ( self ): CustomConfig.register_for_auto_class() __lowerCAmelCase = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) __lowerCAmelCase = AutoConfig.from_pretrained(f"{USER}/test-dynamic-config" , trust_remote_code=__a ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self ): __lowerCAmelCase = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated __lowerCAmelCase = c.n_embd + 1 # int __lowerCAmelCase = c.resid_pdrop + 1.0 # float __lowerCAmelCase = not c.scale_attn_weights # bool __lowerCAmelCase = c.summary_type + "foo" # str c.update_from_string( f"n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}" ) self.assertEqual(__a , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(__a , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(__a , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(__a , c.summary_type , "mismatch for key: summary_type" ) def snake_case ( self ): __lowerCAmelCase = PretrainedConfig() __lowerCAmelCase = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( __a , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) __lowerCAmelCase = [key for key, value in config_common_kwargs.items() if value == getattr(__a , __a )] if len(__a ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" f" {', '.join(__a )}." ) def snake_case ( self ): with self.assertRaises(__a ): # config is in subfolder, the following should not work without specifying the subfolder __lowerCAmelCase = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) __lowerCAmelCase = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(__a ) def snake_case ( self ): # A mock response for an HTTP head request to emulate server down __lowerCAmelCase = mock.Mock() __lowerCAmelCase = 5_00 __lowerCAmelCase = {} __lowerCAmelCase = HTTPError __lowerCAmelCase = {} # Download this model to make sure it's in the cache. __lowerCAmelCase = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=__a ) as mock_head: __lowerCAmelCase = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def snake_case ( self ): # This test is for deprecated behavior and can be removed in v5 __lowerCAmelCase = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def snake_case ( self ): __lowerCAmelCase = AutoConfig.from_pretrained("bert-base-cased" ) __lowerCAmelCase = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(__a ) __lowerCAmelCase = 2 json.dump(configuration.to_dict() , open(os.path.join(__a , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 __lowerCAmelCase = AutoConfig.from_pretrained(__a ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 __lowerCAmelCase = ["config.42.0.0.json"] __lowerCAmelCase = 7_68 configuration.save_pretrained(__a ) shutil.move(os.path.join(__a , "config.4.0.0.json" ) , os.path.join(__a , "config.42.0.0.json" ) ) __lowerCAmelCase = AutoConfig.from_pretrained(__a ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def snake_case ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. __lowerCAmelCase = "hf-internal-testing/test-two-configs" import transformers as new_transformers __lowerCAmelCase = "v4.0.0" __lowerCAmelCase , __lowerCAmelCase = new_transformers.models.auto.AutoConfig.from_pretrained( __a , return_unused_kwargs=__a ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(__a , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers __lowerCAmelCase = "v3.0.0" __lowerCAmelCase = old_transformers.models.auto.AutoConfig.from_pretrained(__a ) self.assertEqual(old_configuration.hidden_size , 7_68 )
57
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self , UpperCAmelCase ) -> Union[str, Any]: for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): _snake_case = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sgugger/tiny-distilbert-classification""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , only_pretrain_model=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , torchscript=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , fpaa=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) # set architectures equal to `None` _snake_case = None _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def lowercase (self ) -> Tuple: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Any: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> int: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> str: _snake_case = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , save_to_csv=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCAmelCase , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(UpperCAmelCase , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(UpperCAmelCase , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(UpperCAmelCase , """train_time.csv""" ) , env_info_csv_file=os.path.join(UpperCAmelCase , """env.csv""" ) , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """env.csv""" ) ).exists() ) def lowercase (self ) -> int: _snake_case = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCAmelCase ): self.assertTrue(hasattr(UpperCAmelCase , """sequential""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """cumulative""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """current""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCAmelCase , """log.txt""" ) , log_print=UpperCAmelCase , trace_memory_line_by_line=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """log.txt""" ) ).exists() )
341
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = """▁""" lowercase_ = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} lowercase_ = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } lowercase_ = {"""vinai/bartpho-syllable""": 1_024} class a_ ( snake_case_ ): '''simple docstring''' UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self , A , A , A="<s>" , A="</s>" , A="</s>" , A="<s>" , A="<unk>" , A="<pad>" , A="<mask>" , A = None , **A , ) -> None: # Mask token behave like a normal word, i.e. include the space before it _SCREAMING_SNAKE_CASE = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else mask_token _SCREAMING_SNAKE_CASE = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A , eos_token=A , unk_token=A , sep_token=A , cls_token=A , pad_token=A , mask_token=A , sp_model_kwargs=self.sp_model_kwargs , **A , ) _SCREAMING_SNAKE_CASE = vocab_file _SCREAMING_SNAKE_CASE = monolingual_vocab_file _SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(A ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(A ) not in self.fairseq_tokens_to_ids: _SCREAMING_SNAKE_CASE = cnt cnt += 1 with open(A , """r""" , encoding="""utf-8""" ) as f: for line in f.readlines(): _SCREAMING_SNAKE_CASE = line.strip().split()[0] _SCREAMING_SNAKE_CASE = len(self.fairseq_tokens_to_ids ) if str(A ) not in self.fairseq_tokens_to_ids: _SCREAMING_SNAKE_CASE = len(self.fairseq_tokens_to_ids ) _SCREAMING_SNAKE_CASE = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ) -> Dict: _SCREAMING_SNAKE_CASE = self.__dict__.copy() _SCREAMING_SNAKE_CASE = None _SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() return state def __setstate__( self , A ) -> Optional[int]: _SCREAMING_SNAKE_CASE = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def snake_case_( self , A , A = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] _SCREAMING_SNAKE_CASE = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def snake_case_( self , A , A = None , A = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A , token_ids_a=A , already_has_special_tokens=A ) if token_ids_a is None: return [1] + ([0] * len(A )) + [1] return [1] + ([0] * len(A )) + [1, 1] + ([0] * len(A )) + [1] def snake_case_( self , A , A = None ) -> List[int]: _SCREAMING_SNAKE_CASE = [self.sep_token_id] _SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def snake_case_( self ) -> int: return len(self.fairseq_ids_to_tokens ) def snake_case_( self ) -> Dict: _SCREAMING_SNAKE_CASE = {self.convert_ids_to_tokens(A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def snake_case_( self , A ) -> List[str]: return self.sp_model.encode(A , out_type=A ) def snake_case_( self , A ) -> Dict: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def snake_case_( self , A ) -> Optional[int]: return self.fairseq_ids_to_tokens[index] def snake_case_( self , A ) -> Any: _SCREAMING_SNAKE_CASE = """""".join(A ).replace(A , """ """ ).strip() return out_string def snake_case_( self , A , A = None ) -> Tuple[str]: if not os.path.isdir(A ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return _SCREAMING_SNAKE_CASE = os.path.join( A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) _SCREAMING_SNAKE_CASE = os.path.join( A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""monolingual_vocab_file"""] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A ) elif not os.path.isfile(self.vocab_file ): with open(A , """wb""" ) as fi: _SCREAMING_SNAKE_CASE = self.sp_model.serialized_model_proto() fi.write(A ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( A ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , A ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(A , """w""" , encoding="""utf-8""" ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(f'{str(A )} \n' ) return out_vocab_file, out_monolingual_vocab_file
58
'''simple docstring''' from __future__ import annotations def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): if len(_SCREAMING_SNAKE_CASE ) == 0: return [] _snake_case, _snake_case = min(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) _snake_case = int(max_value - min_value ) + 1 _snake_case = [[] for _ in range(_SCREAMING_SNAKE_CASE )] for i in my_list: buckets[int(i - min_value )].append(_SCREAMING_SNAKE_CASE ) return [v for bucket in buckets for v in sorted(_SCREAMING_SNAKE_CASE )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
341
0
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCAmelCase ( A_ ,unittest.TestCase ): A__ : Optional[int] = DiTPipeline A__ : Any = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS A__ : Optional[Any] = PipelineTesterMixin.required_optional_params - { "latents", "num_images_per_prompt", "callback", "callback_steps", } A__ : List[Any] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS A__ : Union[str, Any] = False def _SCREAMING_SNAKE_CASE (self : Any ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) snake_case : Optional[int] = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=snake_case__ , activation_fn="gelu-approximate" , num_embeds_ada_norm=10_00 , norm_type="ada_norm_zero" , norm_elementwise_affine=snake_case__ , ) snake_case : List[Any] = AutoencoderKL() snake_case : Dict = DDIMScheduler() snake_case : Optional[Any] = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def _SCREAMING_SNAKE_CASE (self : Optional[Any] , snake_case__ : Dict , snake_case__ : Dict=0 ) -> int: '''simple docstring''' if str(snake_case__ ).startswith("mps" ): snake_case : str = torch.manual_seed(snake_case__ ) else: snake_case : Tuple = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) snake_case : int = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def _SCREAMING_SNAKE_CASE (self : List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case : Optional[Any] = "cpu" snake_case : str = self.get_dummy_components() snake_case : Any = self.pipeline_class(**snake_case__ ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) snake_case : Union[str, Any] = self.get_dummy_inputs(snake_case__ ) snake_case : Optional[int] = pipe(**snake_case__ ).images snake_case : List[str] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) snake_case : Union[str, Any] = np.array([0.2946, 0.6601, 0.4329, 0.3296, 0.4144, 0.5319, 0.7273, 0.5013, 0.4457] ) snake_case : Tuple = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case__ , 1e-3 ) def _SCREAMING_SNAKE_CASE (self : Optional[int] ) -> int: '''simple docstring''' self._test_inference_batch_single_identical(relax_max_difference=snake_case__ , expected_max_diff=1e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _SCREAMING_SNAKE_CASE (self : Tuple ) -> int: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @require_torch_gpu @slow class UpperCAmelCase ( unittest.TestCase ): def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE (self : Any ) -> Dict: '''simple docstring''' snake_case : str = torch.manual_seed(0 ) snake_case : Dict = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) snake_case : List[Any] = ["vase", "umbrella", "white shark", "white wolf"] snake_case : Dict = pipe.get_label_ids(snake_case__ ) snake_case : Optional[Any] = pipe(snake_case__ , generator=snake_case__ , num_inference_steps=40 , output_type="np" ).images for word, image in zip(snake_case__ , snake_case__ ): snake_case : Union[str, Any] = load_numpy( f"""https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy""" ) assert np.abs((expected_image - image).max() ) < 1e-2 def _SCREAMING_SNAKE_CASE (self : int ) -> List[str]: '''simple docstring''' snake_case : Dict = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) snake_case : List[str] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) snake_case : Optional[Any] = ["vase", "umbrella"] snake_case : Optional[Any] = pipe.get_label_ids(snake_case__ ) snake_case : List[str] = torch.manual_seed(0 ) snake_case : int = pipe(snake_case__ , generator=snake_case__ , num_inference_steps=25 , output_type="np" ).images for word, image in zip(snake_case__ , snake_case__ ): snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" f"""/dit/{word}_512.npy""" ) assert np.abs((expected_image - image).max() ) < 1e-1
59
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: if not conversation_id: _snake_case = uuid.uuida() if past_user_inputs is None: _snake_case = [] if generated_responses is None: _snake_case = [] _snake_case = conversation_id _snake_case = past_user_inputs _snake_case = generated_responses _snake_case = text def __eq__(self , UpperCAmelCase ) -> Dict: if not isinstance(UpperCAmelCase , UpperCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = False ) -> int: if self.new_user_input: if overwrite: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ f"""with: \"{text}\".""" ) _snake_case = text else: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ f"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: _snake_case = text def lowercase (self ) -> int: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _snake_case = None def lowercase (self , UpperCAmelCase ) -> Any: self.generated_responses.append(UpperCAmelCase ) def lowercase (self ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__(self ) -> Optional[int]: _snake_case = f"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): _snake_case = """user""" if is_user else """bot""" output += f"""{name} >> {text} \n""" return output @add_end_docstrings( __snake_case , r"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , *UpperCAmelCase , **UpperCAmelCase ) -> Union[str, Any]: super().__init__(*UpperCAmelCase , **UpperCAmelCase ) if self.tokenizer.pad_token_id is None: _snake_case = self.tokenizer.eos_token def lowercase (self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase ) -> Dict: _snake_case = {} _snake_case = {} _snake_case = {} if min_length_for_response is not None: _snake_case = min_length_for_response if minimum_tokens is not None: _snake_case = minimum_tokens if "max_length" in generate_kwargs: _snake_case = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _snake_case = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__(self , UpperCAmelCase , UpperCAmelCase=0 , **UpperCAmelCase ) -> Union[str, Any]: _snake_case = super().__call__(UpperCAmelCase , num_workers=UpperCAmelCase , **UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) and len(UpperCAmelCase ) == 1: return outputs[0] return outputs def lowercase (self , UpperCAmelCase , UpperCAmelCase=32 ) -> Dict[str, Any]: if not isinstance(UpperCAmelCase , UpperCAmelCase ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( f"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): _snake_case = self.tokenizer._build_conversation_input_ids(UpperCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _snake_case = self._legacy_parse_and_tokenize(UpperCAmelCase ) if self.framework == "pt": _snake_case = torch.LongTensor([input_ids] ) elif self.framework == "tf": _snake_case = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=10 , **UpperCAmelCase ) -> Optional[int]: _snake_case = generate_kwargs.get("""max_length""" , self.model.config.max_length ) _snake_case = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(f"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) _snake_case = max_length - minimum_tokens _snake_case = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: _snake_case = model_inputs["""attention_mask"""][:, -trim:] _snake_case = model_inputs.pop("""conversation""" ) _snake_case = max_length _snake_case = self.model.generate(**UpperCAmelCase , **UpperCAmelCase ) if self.model.config.is_encoder_decoder: _snake_case = 1 else: _snake_case = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=True ) -> List[str]: _snake_case = model_outputs["""output_ids"""] _snake_case = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCAmelCase , clean_up_tokenization_spaces=UpperCAmelCase , ) _snake_case = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCAmelCase ) return conversation def lowercase (self , UpperCAmelCase ) -> Dict: _snake_case = self.tokenizer.eos_token_id _snake_case = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) ) if len(UpperCAmelCase ) > self.tokenizer.model_max_length: _snake_case = input_ids[-self.tokenizer.model_max_length :] return input_ids
341
0
"""simple docstring""" from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES snake_case__ : int = logging.get_logger(__name__) snake_case__ : int = OrderedDict( [ # Base model mapping ('''albert''', '''FlaxAlbertModel'''), ('''bart''', '''FlaxBartModel'''), ('''beit''', '''FlaxBeitModel'''), ('''bert''', '''FlaxBertModel'''), ('''big_bird''', '''FlaxBigBirdModel'''), ('''blenderbot''', '''FlaxBlenderbotModel'''), ('''blenderbot-small''', '''FlaxBlenderbotSmallModel'''), ('''clip''', '''FlaxCLIPModel'''), ('''distilbert''', '''FlaxDistilBertModel'''), ('''electra''', '''FlaxElectraModel'''), ('''gpt-sw3''', '''FlaxGPT2Model'''), ('''gpt2''', '''FlaxGPT2Model'''), ('''gpt_neo''', '''FlaxGPTNeoModel'''), ('''gptj''', '''FlaxGPTJModel'''), ('''longt5''', '''FlaxLongT5Model'''), ('''marian''', '''FlaxMarianModel'''), ('''mbart''', '''FlaxMBartModel'''), ('''mt5''', '''FlaxMT5Model'''), ('''opt''', '''FlaxOPTModel'''), ('''pegasus''', '''FlaxPegasusModel'''), ('''regnet''', '''FlaxRegNetModel'''), ('''resnet''', '''FlaxResNetModel'''), ('''roberta''', '''FlaxRobertaModel'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormModel'''), ('''roformer''', '''FlaxRoFormerModel'''), ('''t5''', '''FlaxT5Model'''), ('''vision-text-dual-encoder''', '''FlaxVisionTextDualEncoderModel'''), ('''vit''', '''FlaxViTModel'''), ('''wav2vec2''', '''FlaxWav2Vec2Model'''), ('''whisper''', '''FlaxWhisperModel'''), ('''xglm''', '''FlaxXGLMModel'''), ('''xlm-roberta''', '''FlaxXLMRobertaModel'''), ] ) snake_case__ : Dict = OrderedDict( [ # Model for pre-training mapping ('''albert''', '''FlaxAlbertForPreTraining'''), ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''bert''', '''FlaxBertForPreTraining'''), ('''big_bird''', '''FlaxBigBirdForPreTraining'''), ('''electra''', '''FlaxElectraForPreTraining'''), ('''longt5''', '''FlaxLongT5ForConditionalGeneration'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''mt5''', '''FlaxMT5ForConditionalGeneration'''), ('''roberta''', '''FlaxRobertaForMaskedLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMaskedLM'''), ('''roformer''', '''FlaxRoFormerForMaskedLM'''), ('''t5''', '''FlaxT5ForConditionalGeneration'''), ('''wav2vec2''', '''FlaxWav2Vec2ForPreTraining'''), ('''whisper''', '''FlaxWhisperForConditionalGeneration'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMaskedLM'''), ] ) snake_case__ : Dict = OrderedDict( [ # Model for Masked LM mapping ('''albert''', '''FlaxAlbertForMaskedLM'''), ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''bert''', '''FlaxBertForMaskedLM'''), ('''big_bird''', '''FlaxBigBirdForMaskedLM'''), ('''distilbert''', '''FlaxDistilBertForMaskedLM'''), ('''electra''', '''FlaxElectraForMaskedLM'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''roberta''', '''FlaxRobertaForMaskedLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMaskedLM'''), ('''roformer''', '''FlaxRoFormerForMaskedLM'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMaskedLM'''), ] ) snake_case__ : Tuple = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''blenderbot''', '''FlaxBlenderbotForConditionalGeneration'''), ('''blenderbot-small''', '''FlaxBlenderbotSmallForConditionalGeneration'''), ('''encoder-decoder''', '''FlaxEncoderDecoderModel'''), ('''longt5''', '''FlaxLongT5ForConditionalGeneration'''), ('''marian''', '''FlaxMarianMTModel'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''mt5''', '''FlaxMT5ForConditionalGeneration'''), ('''pegasus''', '''FlaxPegasusForConditionalGeneration'''), ('''t5''', '''FlaxT5ForConditionalGeneration'''), ] ) snake_case__ : int = OrderedDict( [ # Model for Image-classsification ('''beit''', '''FlaxBeitForImageClassification'''), ('''regnet''', '''FlaxRegNetForImageClassification'''), ('''resnet''', '''FlaxResNetForImageClassification'''), ('''vit''', '''FlaxViTForImageClassification'''), ] ) snake_case__ : List[str] = OrderedDict( [ ('''vision-encoder-decoder''', '''FlaxVisionEncoderDecoderModel'''), ] ) snake_case__ : Any = OrderedDict( [ # Model for Causal LM mapping ('''bart''', '''FlaxBartForCausalLM'''), ('''bert''', '''FlaxBertForCausalLM'''), ('''big_bird''', '''FlaxBigBirdForCausalLM'''), ('''electra''', '''FlaxElectraForCausalLM'''), ('''gpt-sw3''', '''FlaxGPT2LMHeadModel'''), ('''gpt2''', '''FlaxGPT2LMHeadModel'''), ('''gpt_neo''', '''FlaxGPTNeoForCausalLM'''), ('''gptj''', '''FlaxGPTJForCausalLM'''), ('''opt''', '''FlaxOPTForCausalLM'''), ('''roberta''', '''FlaxRobertaForCausalLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForCausalLM'''), ('''xglm''', '''FlaxXGLMForCausalLM'''), ('''xlm-roberta''', '''FlaxXLMRobertaForCausalLM'''), ] ) snake_case__ : str = OrderedDict( [ # Model for Sequence Classification mapping ('''albert''', '''FlaxAlbertForSequenceClassification'''), ('''bart''', '''FlaxBartForSequenceClassification'''), ('''bert''', '''FlaxBertForSequenceClassification'''), ('''big_bird''', '''FlaxBigBirdForSequenceClassification'''), ('''distilbert''', '''FlaxDistilBertForSequenceClassification'''), ('''electra''', '''FlaxElectraForSequenceClassification'''), ('''mbart''', '''FlaxMBartForSequenceClassification'''), ('''roberta''', '''FlaxRobertaForSequenceClassification'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForSequenceClassification'''), ('''roformer''', '''FlaxRoFormerForSequenceClassification'''), ('''xlm-roberta''', '''FlaxXLMRobertaForSequenceClassification'''), ] ) snake_case__ : List[str] = OrderedDict( [ # Model for Question Answering mapping ('''albert''', '''FlaxAlbertForQuestionAnswering'''), ('''bart''', '''FlaxBartForQuestionAnswering'''), ('''bert''', '''FlaxBertForQuestionAnswering'''), ('''big_bird''', '''FlaxBigBirdForQuestionAnswering'''), ('''distilbert''', '''FlaxDistilBertForQuestionAnswering'''), ('''electra''', '''FlaxElectraForQuestionAnswering'''), ('''mbart''', '''FlaxMBartForQuestionAnswering'''), ('''roberta''', '''FlaxRobertaForQuestionAnswering'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForQuestionAnswering'''), ('''roformer''', '''FlaxRoFormerForQuestionAnswering'''), ('''xlm-roberta''', '''FlaxXLMRobertaForQuestionAnswering'''), ] ) snake_case__ : Optional[Any] = OrderedDict( [ # Model for Token Classification mapping ('''albert''', '''FlaxAlbertForTokenClassification'''), ('''bert''', '''FlaxBertForTokenClassification'''), ('''big_bird''', '''FlaxBigBirdForTokenClassification'''), ('''distilbert''', '''FlaxDistilBertForTokenClassification'''), ('''electra''', '''FlaxElectraForTokenClassification'''), ('''roberta''', '''FlaxRobertaForTokenClassification'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForTokenClassification'''), ('''roformer''', '''FlaxRoFormerForTokenClassification'''), ('''xlm-roberta''', '''FlaxXLMRobertaForTokenClassification'''), ] ) snake_case__ : Union[str, Any] = OrderedDict( [ # Model for Multiple Choice mapping ('''albert''', '''FlaxAlbertForMultipleChoice'''), ('''bert''', '''FlaxBertForMultipleChoice'''), ('''big_bird''', '''FlaxBigBirdForMultipleChoice'''), ('''distilbert''', '''FlaxDistilBertForMultipleChoice'''), ('''electra''', '''FlaxElectraForMultipleChoice'''), ('''roberta''', '''FlaxRobertaForMultipleChoice'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMultipleChoice'''), ('''roformer''', '''FlaxRoFormerForMultipleChoice'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMultipleChoice'''), ] ) snake_case__ : str = OrderedDict( [ ('''bert''', '''FlaxBertForNextSentencePrediction'''), ] ) snake_case__ : List[Any] = OrderedDict( [ ('''speech-encoder-decoder''', '''FlaxSpeechEncoderDecoderModel'''), ('''whisper''', '''FlaxWhisperForConditionalGeneration'''), ] ) snake_case__ : List[Any] = OrderedDict( [ ('''whisper''', '''FlaxWhisperForAudioClassification'''), ] ) snake_case__ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) snake_case__ : Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) snake_case__ : Union[str, Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) snake_case__ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) snake_case__ : str = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) snake_case__ : Tuple = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) snake_case__ : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) snake_case__ : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) snake_case__ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) snake_case__ : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) snake_case__ : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) snake_case__ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) snake_case__ : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) snake_case__ : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_MAPPING snake_case__ : List[Any] = auto_class_update(FlaxAutoModel) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_PRETRAINING_MAPPING snake_case__ : List[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc='''pretraining''') class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING snake_case__ : List[Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc='''causal language modeling''') class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_MASKED_LM_MAPPING snake_case__ : Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='''masked language modeling''') class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING snake_case__ : Any = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='''sequence-to-sequence language modeling''', checkpoint_for_example='''t5-base''' ) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING snake_case__ : Dict = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='''sequence classification''' ) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING snake_case__ : str = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='''question answering''') class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING snake_case__ : int = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='''token classification''' ) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING snake_case__ : Dict = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='''multiple choice''') class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING snake_case__ : Tuple = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='''next sentence prediction''' ) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING snake_case__ : List[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc='''image classification''' ) class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING snake_case__ : Optional[Any] = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='''vision-to-text modeling''') class snake_case_( _BaseAutoModelClass ): __UpperCamelCase = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING snake_case__ : List[str] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='''sequence-to-sequence speech-to-text modeling''' )
60
'''simple docstring''' from math import factorial, radians def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 18 , _SCREAMING_SNAKE_CASE = 10 ): _snake_case = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians _snake_case = radians(_SCREAMING_SNAKE_CASE ) _snake_case = angle_in_radians _snake_case = 3 _snake_case = -1 for _ in range(_SCREAMING_SNAKE_CASE ): result += (b * (angle_in_radians**a)) / factorial(_SCREAMING_SNAKE_CASE ) _snake_case = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __import__('doctest').testmod()
341
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = { 'configuration_roformer': ['ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoFormerConfig', 'RoFormerOnnxConfig'], 'tokenization_roformer': ['RoFormerTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['RoFormerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoFormerForCausalLM', 'RoFormerForMaskedLM', 'RoFormerForMultipleChoice', 'RoFormerForQuestionAnswering', 'RoFormerForSequenceClassification', 'RoFormerForTokenClassification', 'RoFormerLayer', 'RoFormerModel', 'RoFormerPreTrainedModel', 'load_tf_weights_in_roformer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRoFormerForCausalLM', 'TFRoFormerForMaskedLM', 'TFRoFormerForMultipleChoice', 'TFRoFormerForQuestionAnswering', 'TFRoFormerForSequenceClassification', 'TFRoFormerForTokenClassification', 'TFRoFormerLayer', 'TFRoFormerModel', 'TFRoFormerPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxRoFormerForMaskedLM', 'FlaxRoFormerForMultipleChoice', 'FlaxRoFormerForQuestionAnswering', 'FlaxRoFormerForSequenceClassification', 'FlaxRoFormerForTokenClassification', 'FlaxRoFormerModel', 'FlaxRoFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
61
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __lowerCAmelCase = '\\n@inproceedings{popovic-2015-chrf,\n title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation",\n month = sep,\n year = "2015",\n address = "Lisbon, Portugal",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W15-3049",\n doi = "10.18653/v1/W15-3049",\n pages = "392--395",\n}\n@inproceedings{popovic-2017-chrf,\n title = "chr{F}++: words helping character n-grams",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Second Conference on Machine Translation",\n month = sep,\n year = "2017",\n address = "Copenhagen, Denmark",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W17-4770",\n doi = "10.18653/v1/W17-4770",\n pages = "612--618",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' __lowerCAmelCase = '\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n' __lowerCAmelCase = '\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n \'score\' (float): The chrF (chrF++) score,\n \'char_order\' (int): The character n-gram order,\n \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n \'beta\' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Tuple: if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/mjpost/sacreBLEU#chrf--chrf""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#chrf--chrf"""] , reference_urls=[ """https://github.com/m-popovic/chrF""", ] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = CHRF.CHAR_ORDER , UpperCAmelCase = CHRF.WORD_ORDER , UpperCAmelCase = CHRF.BETA , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , ) -> int: _snake_case = len(references[0] ) if any(len(UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) _snake_case = [[refs[i] for refs in references] for i in range(UpperCAmelCase )] _snake_case = CHRF(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _snake_case = sb_chrf.corpus_score(UpperCAmelCase , UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
341
0
import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import OpenLlamaForCausalLM, OpenLlamaForSequenceClassification, OpenLlamaModel class UpperCAmelCase__ : """simple docstring""" def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=False , A_=True , A_=99 , A_=32 , A_=5 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ) -> int: __UpperCamelCase =parent __UpperCamelCase =batch_size __UpperCamelCase =seq_length __UpperCamelCase =is_training __UpperCamelCase =use_input_mask __UpperCamelCase =use_token_type_ids __UpperCamelCase =use_labels __UpperCamelCase =vocab_size __UpperCamelCase =hidden_size __UpperCamelCase =num_hidden_layers __UpperCamelCase =num_attention_heads __UpperCamelCase =intermediate_size __UpperCamelCase =hidden_act __UpperCamelCase =hidden_dropout_prob __UpperCamelCase =attention_probs_dropout_prob __UpperCamelCase =max_position_embeddings __UpperCamelCase =type_vocab_size __UpperCamelCase =type_sequence_label_size __UpperCamelCase =initializer_range __UpperCamelCase =num_labels __UpperCamelCase =num_choices __UpperCamelCase =scope def _a ( self ) -> Dict: __UpperCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCamelCase =None if self.use_input_mask: __UpperCamelCase =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase =None if self.use_token_type_ids: __UpperCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCamelCase =None __UpperCamelCase =None __UpperCamelCase =None if self.use_labels: __UpperCamelCase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCamelCase =ids_tensor([self.batch_size] , self.num_choices ) __UpperCamelCase =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a ( self ) -> int: return OpenLlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=A_ , initializer_range=self.initializer_range , use_stable_embedding=A_ , ) def _a ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Dict: __UpperCamelCase =OpenLlamaModel(config=A_ ) model.to(A_ ) model.eval() __UpperCamelCase =model(A_ , attention_mask=A_ ) __UpperCamelCase =model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ) -> Optional[int]: __UpperCamelCase =True __UpperCamelCase =OpenLlamaModel(A_ ) model.to(A_ ) model.eval() __UpperCamelCase =model( A_ , attention_mask=A_ , encoder_hidden_states=A_ , encoder_attention_mask=A_ , ) __UpperCamelCase =model( A_ , attention_mask=A_ , encoder_hidden_states=A_ , ) __UpperCamelCase =model(A_ , attention_mask=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ) -> Union[str, Any]: __UpperCamelCase =OpenLlamaForCausalLM(config=A_ ) model.to(A_ ) model.eval() __UpperCamelCase =model(A_ , attention_mask=A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , ) -> List[Any]: __UpperCamelCase =True __UpperCamelCase =True __UpperCamelCase =OpenLlamaForCausalLM(config=A_ ) model.to(A_ ) model.eval() # first forward pass __UpperCamelCase =model( A_ , attention_mask=A_ , encoder_hidden_states=A_ , encoder_attention_mask=A_ , use_cache=A_ , ) __UpperCamelCase =outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __UpperCamelCase =ids_tensor((self.batch_size, 3) , config.vocab_size ) __UpperCamelCase =ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __UpperCamelCase =torch.cat([input_ids, next_tokens] , dim=-1 ) __UpperCamelCase =torch.cat([input_mask, next_mask] , dim=-1 ) __UpperCamelCase =model( A_ , attention_mask=A_ , encoder_hidden_states=A_ , encoder_attention_mask=A_ , output_hidden_states=A_ , )['hidden_states'][0] __UpperCamelCase =model( A_ , attention_mask=A_ , encoder_hidden_states=A_ , encoder_attention_mask=A_ , past_key_values=A_ , output_hidden_states=A_ , )['hidden_states'][0] # select random slice __UpperCamelCase =ids_tensor((1,) , output_from_past.shape[-1] ).item() __UpperCamelCase =output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCamelCase =output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(A_ , A_ , atol=1E-3 ) ) def _a ( self ) -> List[str]: __UpperCamelCase =self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) =config_and_inputs __UpperCamelCase ={'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class UpperCAmelCase__ ( A_ , A_ , A_ , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : Optional[int] = ( (OpenLlamaModel, OpenLlamaForCausalLM, OpenLlamaForSequenceClassification) if is_torch_available() else () ) UpperCAmelCase__ : List[Any] = (OpenLlamaForCausalLM,) if is_torch_available() else () UpperCAmelCase__ : str = ( { "feature-extraction": OpenLlamaModel, "text-classification": OpenLlamaForSequenceClassification, "text-generation": OpenLlamaForCausalLM, "zero-shot": OpenLlamaForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[int] = False def _a ( self ) -> List[str]: __UpperCamelCase =OpenLlamaModelTester(self ) __UpperCamelCase =ConfigTester(self , config_class=A_ , hidden_size=37 ) def _a ( self ) -> Tuple: self.config_tester.run_common_tests() def _a ( self ) -> Any: __UpperCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def _a ( self ) -> str: __UpperCamelCase =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __UpperCamelCase =type self.model_tester.create_and_check_model(*A_ ) def _a ( self ) -> Dict: __UpperCamelCase , __UpperCamelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase =3 __UpperCamelCase =input_dict['input_ids'] __UpperCamelCase =input_ids.ne(1 ).to(A_ ) __UpperCamelCase =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __UpperCamelCase =OpenLlamaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __UpperCamelCase =model(A_ , attention_mask=A_ , labels=A_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _a ( self ) -> Tuple: __UpperCamelCase , __UpperCamelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase =3 __UpperCamelCase ='single_label_classification' __UpperCamelCase =input_dict['input_ids'] __UpperCamelCase =input_ids.ne(1 ).to(A_ ) __UpperCamelCase =ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __UpperCamelCase =OpenLlamaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __UpperCamelCase =model(A_ , attention_mask=A_ , labels=A_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _a ( self ) -> Tuple: __UpperCamelCase , __UpperCamelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase =3 __UpperCamelCase ='multi_label_classification' __UpperCamelCase =input_dict['input_ids'] __UpperCamelCase =input_ids.ne(1 ).to(A_ ) __UpperCamelCase =ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __UpperCamelCase =OpenLlamaForSequenceClassification(A_ ) model.to(A_ ) model.eval() __UpperCamelCase =model(A_ , attention_mask=A_ , labels=A_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('Open-Llama buffers include complex numbers, which breaks this test' ) def _a ( self ) -> List[Any]: pass @parameterized.expand([('linear',), ('dynamic',)] ) def _a ( self , A_ ) -> Tuple: __UpperCamelCase , __UpperCamelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase =ids_tensor([1, 10] , config.vocab_size ) __UpperCamelCase =ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __UpperCamelCase =OpenLlamaModel(A_ ) original_model.to(A_ ) original_model.eval() __UpperCamelCase =original_model(A_ ).last_hidden_state __UpperCamelCase =original_model(A_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __UpperCamelCase ={'type': scaling_type, 'factor': 10.0} __UpperCamelCase =OpenLlamaModel(A_ ) scaled_model.to(A_ ) scaled_model.eval() __UpperCamelCase =scaled_model(A_ ).last_hidden_state __UpperCamelCase =scaled_model(A_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(A_ , A_ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(A_ , A_ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(A_ , A_ , atol=1E-5 ) )
62
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCAmelCase = '\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n' __lowerCAmelCase = '\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {\'spearmanr\': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results[\'spearmanr\'])\n -0.7\n >>> print(round(results[\'spearmanr_pvalue\'], 2))\n 0.19\n' __lowerCAmelCase = r'\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Optional[Any]: _snake_case = spearmanr(UpperCAmelCase , UpperCAmelCase ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
341
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _lowerCamelCase ( lowercase : Dict ) -> Any: _a = filter(lambda lowercase : p.requires_grad , model.parameters() ) _a = sum([np.prod(p.size() ) for p in model_parameters] ) return params lowerCAmelCase_ : int = logging.getLogger(__name__) def _lowerCamelCase ( lowercase : List[Any] , lowercase : Any ) -> Any: if metric == "rouge2": _a = "{val_avg_rouge2:.4f}-{step_count}" elif metric == "bleu": _a = "{val_avg_bleu:.4f}-{step_count}" elif metric == "em": _a = "{val_avg_em:.4f}-{step_count}" else: raise NotImplementedError( F'seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this' " function." ) _a = ModelCheckpoint( dirpath=lowercase , filename=lowercase , monitor=F'val_{metric}' , mode="max" , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def _lowerCamelCase ( lowercase : Optional[int] , lowercase : Optional[int] ) -> Union[str, Any]: return EarlyStopping( monitor=F'val_{metric}' , mode="min" if "loss" in metric else "max" , patience=lowercase , verbose=lowercase , ) class __SCREAMING_SNAKE_CASE (pl.Callback ): """simple docstring""" def UpperCamelCase__ ( self : Optional[int] , __a : str , __a : List[Any] ): _a = {f'lr_group_{i}': param["lr"] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(__a ) @rank_zero_only def UpperCamelCase__ ( self : Optional[int] , __a : pl.Trainer , __a : pl.LightningModule , __a : str , __a : Tuple=True ): logger.info(f'***** {type_path} results at step {trainer.global_step:05d} *****' ) _a = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results _a = Path(pl_module.hparams.output_dir ) if type_path == "test": _a = od / "test_results.txt" _a = od / "test_generations.txt" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. _a = od / f'{type_path}_results/{trainer.global_step:05d}.txt' _a = od / f'{type_path}_generations/{trainer.global_step:05d}.txt' results_file.parent.mkdir(exist_ok=__a ) generations_file.parent.mkdir(exist_ok=__a ) with open(__a , "a+" ) as writer: for key in sorted(__a ): if key in ["log", "progress_bar", "preds"]: continue _a = metrics[key] if isinstance(__a , torch.Tensor ): _a = val.item() _a = f'{key}: {val:.6f}\n' writer.write(__a ) if not save_generations: return if "preds" in metrics: _a = "\n".join(metrics["preds"] ) generations_file.open("w+" ).write(__a ) @rank_zero_only def UpperCamelCase__ ( self : int , __a : List[Any] , __a : Union[str, Any] ): try: _a = pl_module.model.model.num_parameters() except AttributeError: _a = pl_module.model.num_parameters() _a = count_trainable_parameters(__a ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1e6, "grad_mp": n_trainable_pars / 1e6} ) @rank_zero_only def UpperCamelCase__ ( self : Union[str, Any] , __a : pl.Trainer , __a : pl.LightningModule ): save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(__a , __a , "test" ) @rank_zero_only def UpperCamelCase__ ( self : Any , __a : pl.Trainer , __a : int ): save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
63
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=4 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[2, 2, 3, 2] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase=37 , UpperCAmelCase="gelu" , UpperCAmelCase=10 , UpperCAmelCase=0.02 , UpperCAmelCase=["stage2", "stage3", "stage4"] , UpperCAmelCase=3 , UpperCAmelCase=None , ) -> List[Any]: _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = num_stages _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = intermediate_size _snake_case = hidden_act _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = out_features _snake_case = num_labels _snake_case = scope _snake_case = num_stages def lowercase (self ) -> List[Any]: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def lowercase (self ) -> Tuple: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def lowercase (self ) -> Any: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCAmelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCAmelCase , loss_ignore_index=255 , num_labels=self.num_labels , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: _snake_case = UperNetForSemanticSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowercase (self ) -> Tuple: _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ), ( _snake_case ), ( _snake_case ), ) = config_and_inputs _snake_case = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCAmelCase_ = {"image-segmentation": UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> Optional[Any]: _snake_case = UperNetModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=37 ) def lowercase (self ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase (self ) -> Union[str, Any]: return def lowercase (self ) -> Union[str, Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def lowercase (self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> str: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> int: pass def lowercase (self ) -> List[str]: def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = _config_zero_init(UpperCAmelCase ) _snake_case = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _snake_case = model_class(config=UpperCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def lowercase (self ) -> Optional[Any]: pass @slow def lowercase (self ) -> Tuple: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = UperNetForSemanticSegmentation.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) _snake_case = Image.open(_SCREAMING_SNAKE_CASE ).convert("""RGB""" ) return image @require_torch @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) ) def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
341
0
"""simple docstring""" import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration A_ = 50_00_00 A_ , A_ = os.path.split(__file__) A_ = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def UpperCAmelCase__ (snake_case__ : datasets.Dataset , **snake_case__ : Optional[int] ): """simple docstring""" _snake_case : Tuple = dataset.map(**snake_case__ ) @get_duration def UpperCAmelCase__ (snake_case__ : datasets.Dataset , **snake_case__ : Any ): """simple docstring""" _snake_case : List[str] = dataset.filter(**snake_case__ ) def UpperCAmelCase__ (): """simple docstring""" _snake_case : Dict = {"""num examples""": SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: _snake_case : Dict = datasets.Features({"""text""": datasets.Value("""string""" ), """numbers""": datasets.Value("""float32""" )} ) _snake_case : List[Any] = generate_example_dataset( os.path.join(snake_case__ , """dataset.arrow""" ) , snake_case__ , num_examples=snake_case__ ) _snake_case : List[Any] = transformers.AutoTokenizer.from_pretrained("""bert-base-cased""" , use_fast=snake_case__ ) def tokenize(snake_case__ : Optional[int] ): return tokenizer(examples["""text"""] ) _snake_case : str = map(snake_case__ ) _snake_case : Optional[int] = map(snake_case__ , batched=snake_case__ ) _snake_case : int = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""numpy""" ): _snake_case : Dict = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""pandas""" ): _snake_case : List[str] = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""torch""" , columns="""numbers""" ): _snake_case : Union[str, Any] = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) with dataset.formatted_as(type="""tensorflow""" , columns="""numbers""" ): _snake_case : List[str] = map(snake_case__ , function=lambda snake_case__ : None , batched=snake_case__ ) _snake_case : Dict = map(snake_case__ , function=snake_case__ , batched=snake_case__ ) _snake_case : List[str] = filter(snake_case__ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(snake_case__ , """wb""" ) as f: f.write(json.dumps(snake_case__ ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
64
'''simple docstring''' import argparse from collections import defaultdict def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = f"""class {class_name}(""" _snake_case = f"""{4 * " "}def {test_name}(""" _snake_case = f"""{8 * " "}{correct_line.split()[0]}""" _snake_case = f"""{16 * " "}{correct_line.split()[0]}""" _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = 0 _snake_case = 0 _snake_case = [] for line in lines: if line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and in_func and (line.startswith(_SCREAMING_SNAKE_CASE ) or line.startswith(_SCREAMING_SNAKE_CASE )): _snake_case = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: _snake_case = True if in_class and in_func and in_line: if ")" not in line: continue else: _snake_case = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) _snake_case = _snake_case = _snake_case = _snake_case = False else: new_lines.append(_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: for line in new_lines: f.write(_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): if fail is not None: with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = {l.strip() for l in f.readlines()} else: _snake_case = None with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = defaultdict(_SCREAMING_SNAKE_CASE ) for line in correct_lines: _snake_case, _snake_case, _snake_case, _snake_case = line.split(""";""" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) __lowerCAmelCase = parser.parse_args() main(args.correct_filename, args.fail_filename)
341
0
from pathlib import Path import numpy as np from PIL import Image def lowerCAmelCase_ ( __A ) -> np.ndarray: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2989 * r + 0.5870 * g + 0.1140 * b def lowerCAmelCase_ ( __A ) -> np.ndarray: '''simple docstring''' return (gray > 127) & (gray <= 255) def lowerCAmelCase_ ( __A, __A ) -> np.ndarray: '''simple docstring''' UpperCAmelCase__ = np.zeros_like(__A ) UpperCAmelCase__ = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image UpperCAmelCase__ = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): UpperCAmelCase__ = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() UpperCAmelCase__ = int(summation > 0 ) return output if __name__ == "__main__": # read original image UpperCamelCase__ = Path(__file__).resolve().parent / 'image_data' / 'lena.jpg' UpperCamelCase__ = np.array(Image.open(lena_path)) # kernel to be applied UpperCamelCase__ = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) UpperCamelCase__ = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image UpperCamelCase__ = Image.fromarray(output).convert('RGB') pil_img.save('result_dilation.png')
65
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A_ ( _lowercase, _lowercase ): '''simple docstring''' snake_case_ :Optional[int] = torch.load(_lowercase, map_location="""cpu""" ) snake_case_ :Any = chkpt["""model"""] # We have the base model one level deeper than the original XLM repository snake_case_ :Dict = {} for k, v in state_dict.items(): if "pred_layer" in k: snake_case_ :Optional[Any] = v else: snake_case_ :List[str] = v snake_case_ :List[Any] = chkpt["""params"""] snake_case_ :str = {n: v for n, v in config.items() if not isinstance(_lowercase, (torch.FloatTensor, numpy.ndarray) )} snake_case_ :List[Any] = chkpt["""dico_word2id"""] snake_case_ :Optional[Any] = {s + """</w>""" if s.find("""@@""" ) == -1 and i > 13 else s.replace("""@@""", """""" ): i for s, i in vocab.items()} # Save pytorch-model snake_case_ :Dict = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME snake_case_ :List[Any] = pytorch_dump_folder_path + """/""" + CONFIG_NAME snake_case_ :Optional[int] = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""vocab_file"""] print(f"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(_lowercase, _lowercase ) print(f"""Save configuration file to {pytorch_config_dump_path}""" ) with open(_lowercase, """w""", encoding="""utf-8""" ) as f: f.write(json.dumps(_lowercase, indent=2 ) + """\n""" ) print(f"""Save vocab file to {pytorch_config_dump_path}""" ) with open(_lowercase, """w""", encoding="""utf-8""" ) as f: f.write(json.dumps(_lowercase, indent=2 ) + """\n""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( "--xlm_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __a = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
66
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = ["pixel_values"] def __init__(self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: super().__init__(**UpperCAmelCase ) _snake_case = size if size is not None else {"""height""": 256, """width""": 256} _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( UpperCAmelCase , size=(size["""height"""], size["""width"""]) , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(UpperCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> List[Any]: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> PIL.Image.Image: _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = size if size is not None else self.size _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=UpperCAmelCase , size=UpperCAmelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] _snake_case = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _snake_case = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
341
0
'''simple docstring''' import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class a__ ( tf.keras.optimizers.schedules.LearningRateSchedule ): def __init__( self : str , a : float , a : Callable , a : int , a : float = 1.0 , a : str = None , ): """simple docstring""" super().__init__() __lowerCamelCase = initial_learning_rate __lowerCamelCase = warmup_steps __lowerCamelCase = power __lowerCamelCase = decay_schedule_fn __lowerCamelCase = name def __call__( self : Dict , a : str ): """simple docstring""" with tf.name_scope(self.name or '''WarmUp''' ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. __lowerCamelCase = tf.cast(a , tf.floataa ) __lowerCamelCase = tf.cast(self.warmup_steps , tf.floataa ) __lowerCamelCase = global_step_float / warmup_steps_float __lowerCamelCase = self.initial_learning_rate * tf.math.pow(a , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=a , ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = 0.0 , UpperCamelCase__ = 0.9 , UpperCamelCase__ = 0.9_9_9 , UpperCamelCase__ = 1E-8 , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__ = 0.0 , UpperCamelCase__ = 1.0 , UpperCamelCase__ = None , ) -> Dict: __lowerCamelCase = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=UpperCamelCase__ , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=UpperCamelCase__ , ) if num_warmup_steps: __lowerCamelCase = WarmUp( initial_learning_rate=UpperCamelCase__ , decay_schedule_fn=UpperCamelCase__ , warmup_steps=UpperCamelCase__ , ) if weight_decay_rate > 0.0: __lowerCamelCase = AdamWeightDecay( learning_rate=UpperCamelCase__ , weight_decay_rate=UpperCamelCase__ , beta_a=UpperCamelCase__ , beta_a=UpperCamelCase__ , epsilon=UpperCamelCase__ , clipnorm=UpperCamelCase__ , global_clipnorm=UpperCamelCase__ , exclude_from_weight_decay=['''LayerNorm''', '''layer_norm''', '''bias'''] , include_in_weight_decay=UpperCamelCase__ , ) else: __lowerCamelCase = tf.keras.optimizers.Adam( learning_rate=UpperCamelCase__ , beta_a=UpperCamelCase__ , beta_a=UpperCamelCase__ , epsilon=UpperCamelCase__ , clipnorm=UpperCamelCase__ , global_clipnorm=UpperCamelCase__ , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class a__ ( UpperCAmelCase__ ): def __init__( self : List[Any] , a : Union[float, tf.keras.optimizers.schedules.LearningRateSchedule] = 0.0_01 , a : float = 0.9 , a : float = 0.9_99 , a : float = 1e-7 , a : bool = False , a : float = 0.0 , a : Optional[List[str]] = None , a : Optional[List[str]] = None , a : str = "AdamWeightDecay" , **a : Union[str, Any] , ): """simple docstring""" super().__init__(a , a , a , a , a , a , **a ) __lowerCamelCase = weight_decay_rate __lowerCamelCase = include_in_weight_decay __lowerCamelCase = exclude_from_weight_decay @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Any , a : int ): """simple docstring""" __lowerCamelCase = {'''WarmUp''': WarmUp} return super(a , cls ).from_config(a , custom_objects=a ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , a : int , a : List[str] , a : Any ): """simple docstring""" super(a , self )._prepare_local(a , a , a ) __lowerCamelCase = tf.constant( self.weight_decay_rate , name='''adam_weight_decay_rate''' ) def SCREAMING_SNAKE_CASE__ ( self : str , a : List[str] , a : List[str] , a : str ): """simple docstring""" __lowerCamelCase = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]['''weight_decay_rate'''] , use_locking=self._use_locking , ) return tf.no_op() def SCREAMING_SNAKE_CASE__ ( self : List[str] , a : List[Any] , a : int=None , **a : Optional[int] ): """simple docstring""" __lowerCamelCase , __lowerCamelCase = list(zip(*a ) ) return super(a , self ).apply_gradients(zip(a , a ) , name=a , **a ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : Any , a : Tuple , a : int ): """simple docstring""" if apply_state is None: return self._decayed_lr_t[var_dtype], {} __lowerCamelCase = apply_state or {} __lowerCamelCase = apply_state.get((var_device, var_dtype) ) if coefficients is None: __lowerCamelCase = self._fallback_apply_state(a , a ) __lowerCamelCase = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def SCREAMING_SNAKE_CASE__ ( self : str , a : Optional[int] , a : Tuple , a : Tuple=None ): """simple docstring""" __lowerCamelCase , __lowerCamelCase = self._get_lr(var.device , var.dtype.base_dtype , a ) __lowerCamelCase = self._decay_weights_op(a , a , a ) with tf.control_dependencies([decay] ): return super(a , self )._resource_apply_dense(a , a , **a ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : int , a : List[str] , a : str , a : Optional[int]=None ): """simple docstring""" __lowerCamelCase , __lowerCamelCase = self._get_lr(var.device , var.dtype.base_dtype , a ) __lowerCamelCase = self._decay_weights_op(a , a , a ) with tf.control_dependencies([decay] ): return super(a , self )._resource_apply_sparse(a , a , a , **a ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" __lowerCamelCase = super().get_config() config.update({'''weight_decay_rate''': self.weight_decay_rate} ) return config def SCREAMING_SNAKE_CASE__ ( self : int , a : List[str] ): """simple docstring""" if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(a , a ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(a , a ) is not None: return False return True class a__ ( UpperCAmelCase__ ): def __init__( self : Tuple ): """simple docstring""" __lowerCamelCase = [] __lowerCamelCase = None @property def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" if self._accum_steps is None: __lowerCamelCase = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=a , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" if not self._gradients: raise ValueError('''The accumulator should be called first to initialize the gradients''' ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self : int , a : Dict ): """simple docstring""" if not self._gradients: __lowerCamelCase = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(a ) , trainable=a , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(a ) != len(self._gradients ): raise ValueError(f"""Expected {len(self._gradients )} gradients, but got {len(a )}""" ) for accum_gradient, gradient in zip(self._gradients , a ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(a ) self._accum_steps.assign_add(1 ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(a ) )
67
'''simple docstring''' __lowerCAmelCase = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure the supplied data is a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_SCREAMING_SNAKE_CASE ) _snake_case = """""".join(bin(_SCREAMING_SNAKE_CASE )[2:].zfill(8 ) for byte in data ) _snake_case = len(_SCREAMING_SNAKE_CASE ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b"""=""" * ((6 - len(_SCREAMING_SNAKE_CASE ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_SCREAMING_SNAKE_CASE ) % 6) else: _snake_case = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 6 ) ).encode() + padding ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = ( """argument should be a bytes-like object or ASCII string, """ f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_SCREAMING_SNAKE_CASE ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): try: _snake_case = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) _snake_case = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_SCREAMING_SNAKE_CASE ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 8 ) ] return bytes(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
from typing import TYPE_CHECKING from ...utils import _LazyModule lowerCAmelCase__ = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
68
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A__ : Optional[Any] ={ '''configuration_whisper''': ['''WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WhisperConfig''', '''WhisperOnnxConfig'''], '''feature_extraction_whisper''': ['''WhisperFeatureExtractor'''], '''processing_whisper''': ['''WhisperProcessor'''], '''tokenization_whisper''': ['''WhisperTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : str =['''WhisperTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Optional[Any] =[ '''WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WhisperForConditionalGeneration''', '''WhisperModel''', '''WhisperPreTrainedModel''', '''WhisperForAudioClassification''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : str =[ '''TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWhisperForConditionalGeneration''', '''TFWhisperModel''', '''TFWhisperPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Dict =[ '''FlaxWhisperForConditionalGeneration''', '''FlaxWhisperModel''', '''FlaxWhisperPreTrainedModel''', '''FlaxWhisperForAudioClassification''', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys A__ : Optional[int] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __lowerCAmelCase = TypeVar('T') __lowerCAmelCase = Union[List[T], Tuple[T, ...]] __lowerCAmelCase = Union[T, List[T], Dict[str, T]] __lowerCAmelCase = Union[str, bytes, os.PathLike]
341
0
from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def A ( ) -> int: import os as original_os from os import path as original_path from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join __UpperCamelCase : Dict ='__test_patch_submodule_mock__' with patch_submodule(_test_patching ,'os.path.join' ,a_ ): # Every way to access os.path.join must be patched, and the rest must stay untouched # check os.path.join assert isinstance(_test_patching.os ,_PatchedModuleObj ) assert isinstance(_test_patching.os.path ,_PatchedModuleObj ) assert _test_patching.os.path.join is mock # check path.join assert isinstance(_test_patching.path ,_PatchedModuleObj ) assert _test_patching.path.join is mock # check join assert _test_patching.join is mock # check that the other attributes are untouched assert _test_patching.os.rename is original_rename assert _test_patching.path.dirname is original_dirname assert _test_patching.os.path.dirname is original_dirname # Even renamed modules or objects must be patched # check renamed_os.path.join assert isinstance(_test_patching.renamed_os ,_PatchedModuleObj ) assert isinstance(_test_patching.renamed_os.path ,_PatchedModuleObj ) assert _test_patching.renamed_os.path.join is mock # check renamed_path.join assert isinstance(_test_patching.renamed_path ,_PatchedModuleObj ) assert _test_patching.renamed_path.join is mock # check renamed_join assert _test_patching.renamed_join is mock # check that the other attributes are untouched assert _test_patching.renamed_os.rename is original_rename assert _test_patching.renamed_path.dirname is original_dirname assert _test_patching.renamed_os.path.dirname is original_dirname # check that everthing is back to normal when the patch is over assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join def A ( ) -> Any: assert _test_patching.open is open __UpperCamelCase : Optional[int] ='__test_patch_submodule_builtin_mock__' # _test_patching has "open" in its globals assert _test_patching.open is open with patch_submodule(_test_patching ,'open' ,a_ ): assert _test_patching.open is mock # check that everthing is back to normal when the patch is over assert _test_patching.open is open def A ( ) -> Union[str, Any]: # pandas.read_csv is not present in _test_patching __UpperCamelCase : List[str] ='__test_patch_submodule_missing_mock__' with patch_submodule(_test_patching ,'pandas.read_csv' ,a_ ): pass def A ( ) -> str: # builtin should always be mocked even if they're not in the globals # in case they're loaded at one point __UpperCamelCase : Any ='__test_patch_submodule_missing_builtin_mock__' # _test_patching doesn't have "len" in its globals assert getattr(_test_patching ,'len' ,a_ ) is None with patch_submodule(_test_patching ,'len' ,a_ ): assert _test_patching.len is mock assert _test_patching.len is len def A ( ) -> Dict: __UpperCamelCase : List[Any] ='__test_patch_submodule_start_and_stop_mock__' __UpperCamelCase : Tuple =patch_submodule(_test_patching ,'open' ,a_ ) assert _test_patching.open is open patch.start() assert _test_patching.open is mock patch.stop() assert _test_patching.open is open def A ( ) -> Optional[int]: from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join __UpperCamelCase : str ='__test_patch_submodule_successive_join__' __UpperCamelCase : Optional[int] ='__test_patch_submodule_successive_dirname__' __UpperCamelCase : Any ='__test_patch_submodule_successive_rename__' assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename with patch_submodule(_test_patching ,'os.path.join' ,a_ ): with patch_submodule(_test_patching ,'os.rename' ,a_ ): with patch_submodule(_test_patching ,'os.path.dirname' ,a_ ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename # try another order with patch_submodule(_test_patching ,'os.rename' ,a_ ): with patch_submodule(_test_patching ,'os.path.join' ,a_ ): with patch_submodule(_test_patching ,'os.path.dirname' ,a_ ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename def A ( ) -> Any: __UpperCamelCase : Optional[Any] ='__test_patch_submodule_doesnt_exist_mock__' with patch_submodule(_test_patching ,'__module_that_doesn_exist__.__attribute_that_doesn_exist__' ,a_ ): pass with patch_submodule(_test_patching ,'os.__attribute_that_doesn_exist__' ,a_ ): pass
71
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: _snake_case = data _snake_case = previous _snake_case = next_node def __str__(self ) -> str: return f"""{self.data}""" def lowercase (self ) -> int: return self.data def lowercase (self ) -> Dict: return self.next def lowercase (self ) -> Union[str, Any]: return self.previous class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase ) -> List[str]: _snake_case = head def __iter__(self ) -> Optional[Any]: return self def lowercase (self ) -> str: if not self.current: raise StopIteration else: _snake_case = self.current.get_data() _snake_case = self.current.get_next() return value class _lowerCAmelCase : '''simple docstring''' def __init__(self ) -> Optional[int]: _snake_case = None # First node in list _snake_case = None # Last node in list def __str__(self ) -> Optional[int]: _snake_case = self.head _snake_case = [] while current is not None: nodes.append(current.get_data() ) _snake_case = current.get_next() return " ".join(str(UpperCAmelCase ) for node in nodes ) def __contains__(self , UpperCAmelCase ) -> int: _snake_case = self.head while current: if current.get_data() == value: return True _snake_case = current.get_next() return False def __iter__(self ) -> Union[str, Any]: return LinkedListIterator(self.head ) def lowercase (self ) -> str: if self.head: return self.head.get_data() return None def lowercase (self ) -> List[Any]: if self.tail: return self.tail.get_data() return None def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: _snake_case = node _snake_case = node else: self.insert_before_node(self.head , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: self.set_head(UpperCAmelCase ) else: self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: _snake_case = Node(UpperCAmelCase ) if self.head is None: self.set_head(UpperCAmelCase ) else: self.set_tail(UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.previous if node.get_previous() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.next if node.get_next() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = 1 _snake_case = Node(UpperCAmelCase ) _snake_case = self.head while node: if current_position == position: self.insert_before_node(UpperCAmelCase , UpperCAmelCase ) return current_position += 1 _snake_case = node.next self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> Node: _snake_case = self.head while node: if node.get_data() == item: return node _snake_case = node.get_next() raise Exception("""Node not found""" ) def lowercase (self , UpperCAmelCase ) -> Optional[int]: if (node := self.get_node(UpperCAmelCase )) is not None: if node == self.head: _snake_case = self.head.get_next() if node == self.tail: _snake_case = self.tail.get_previous() self.remove_node_pointers(UpperCAmelCase ) @staticmethod def lowercase (UpperCAmelCase ) -> None: if node.get_next(): _snake_case = node.previous if node.get_previous(): _snake_case = node.next _snake_case = None _snake_case = None def lowercase (self ) -> Dict: return self.head is None def __SCREAMING_SNAKE_CASE ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" import numpy as np from PIL import Image def snake_case_ ( A_ : np.ndarray, A_ : int, A_ : int ): '''simple docstring''' _lowerCamelCase : List[Any] = np.array(A_ ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) _lowerCamelCase : Any = 0 _lowerCamelCase : int = 0 _lowerCamelCase : Union[str, Any] = 0 _lowerCamelCase : Any = 0 # compute the shape of the output matrix _lowerCamelCase : int = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape _lowerCamelCase : List[Any] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix _lowerCamelCase : int = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 _lowerCamelCase : Dict = 0 _lowerCamelCase : Any = 0 return updated_arr def snake_case_ ( A_ : np.ndarray, A_ : int, A_ : int ): '''simple docstring''' _lowerCamelCase : List[Any] = np.array(A_ ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) _lowerCamelCase : Optional[Any] = 0 _lowerCamelCase : Dict = 0 _lowerCamelCase : str = 0 _lowerCamelCase : str = 0 # compute the shape of the output matrix _lowerCamelCase : List[str] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape _lowerCamelCase : List[Any] = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix _lowerCamelCase : Dict = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 _lowerCamelCase : Any = 0 _lowerCamelCase : Optional[int] = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name='''avgpooling''', verbose=True) # Loading the image lowerCAmelCase__ = Image.open('''path_to_image''') # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
72
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput __lowerCAmelCase = 8 def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x * 255).int().clamp(0 , 255 ) _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c h w -> b c 1 h w""" ) _snake_case = ((x & mask) != 0).float() _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c d h w -> b (c d) h w""" ) _snake_case = bits * 2 - 1 return bits def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x > 0).int() _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE , dtype=torch.intaa ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b (c d) h w -> b c d h w""" , d=8 ) _snake_case = reduce(x * mask , """b c d h w -> b c h w""" , """sum""" ) return (dec / 255).clamp(0.0 , 1.0 ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _snake_case = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _snake_case = self.alphas_cumprod[timestep] _snake_case = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _snake_case = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _snake_case = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _snake_case = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _snake_case = model_output.device if torch.is_tensor(_SCREAMING_SNAKE_CASE ) else """cpu""" _snake_case = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise _snake_case = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="epsilon" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): _snake_case = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _snake_case, _snake_case = torch.split(_SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: _snake_case = None # 1. compute alphas, betas _snake_case = self.alphas_cumprod[t] _snake_case = self.alphas_cumprod[t - 1] if t > 0 else self.one _snake_case = 1 - alpha_prod_t _snake_case = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _snake_case = model_output else: raise ValueError(f"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _snake_case = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _snake_case = 0 if t > 0: _snake_case = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_SCREAMING_SNAKE_CASE ).to(model_output.device ) _snake_case = (self._get_variance(_SCREAMING_SNAKE_CASE , predicted_variance=_SCREAMING_SNAKE_CASE ) ** 0.5) * noise _snake_case = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 1.0 , ) -> Tuple: super().__init__() _snake_case = bit_scale _snake_case = ( ddim_bit_scheduler_step if isinstance(UpperCAmelCase , UpperCAmelCase ) else ddpm_bit_scheduler_step ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) @torch.no_grad() def __call__(self , UpperCAmelCase = 256 , UpperCAmelCase = 256 , UpperCAmelCase = 50 , UpperCAmelCase = None , UpperCAmelCase = 1 , UpperCAmelCase = "pil" , UpperCAmelCase = True , **UpperCAmelCase , ) -> Union[Tuple, ImagePipelineOutput]: _snake_case = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=UpperCAmelCase , ) _snake_case = decimal_to_bits(UpperCAmelCase ) * self.bit_scale _snake_case = latents.to(self.device ) self.scheduler.set_timesteps(UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _snake_case = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _snake_case = self.scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ).prev_sample _snake_case = bits_to_decimal(UpperCAmelCase ) if output_type == "pil": _snake_case = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase )
341
0
from math import ceil def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ = 1_0_0_1 ) -> int: __lowerCamelCase : Tuple = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): __lowerCamelCase : Any = 2 * i + 1 __lowerCamelCase : Tuple = 2 * i __lowerCamelCase : Tuple = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: a =int(sys.argv[1]) print(solution(n)) except ValueError: print("""Invalid entry - please enter a number""")
73
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 10**9 ): _snake_case = 1 _snake_case = 2 _snake_case = 0 _snake_case = 0 _snake_case = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value _snake_case = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f'''{solution() = }''')
341
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: UNetaDModel _lowerCamelCase: KarrasVeScheduler def __init__( self : List[Any] ,A_ : UNetaDModel ,A_ : KarrasVeScheduler ) -> Dict: super().__init__() self.register_modules(unet=A_ ,scheduler=A_ ) @torch.no_grad() def __call__( self : Optional[Any] ,A_ : int = 1 ,A_ : int = 50 ,A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,A_ : Optional[str] = "pil" ,A_ : bool = True ,**A_ : Optional[Any] ,) -> Union[Tuple, ImagePipelineOutput]: A = self.unet.config.sample_size A = (batch_size, 3, img_size, img_size) A = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) A = randn_tensor(A_ ,generator=A_ ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A_ ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper A = self.scheduler.schedule[t] A = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat A , A = self.scheduler.add_noise_to_input(A_ ,A_ ,generator=A_ ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. A = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev A = self.scheduler.step(A_ ,A_ ,A_ ,A_ ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. A = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample A = self.scheduler.step_correct( A_ ,A_ ,A_ ,A_ ,step_output.prev_sample ,step_output['derivative'] ,) A = step_output.prev_sample A = (sample / 2 + 0.5).clamp(0 ,1 ) A = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": A = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
74
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "deberta-v2" def __init__(self , UpperCAmelCase=128100 , UpperCAmelCase=1536 , UpperCAmelCase=24 , UpperCAmelCase=24 , UpperCAmelCase=6144 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-7 , UpperCAmelCase=False , UpperCAmelCase=-1 , UpperCAmelCase=0 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=0 , UpperCAmelCase="gelu" , **UpperCAmelCase , ) -> List[str]: super().__init__(**UpperCAmelCase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = relative_attention _snake_case = max_relative_positions _snake_case = pad_token_id _snake_case = position_biased_input # Backwards compatibility if type(UpperCAmelCase ) == str: _snake_case = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case = pos_att_type _snake_case = vocab_size _snake_case = layer_norm_eps _snake_case = kwargs.get("""pooler_hidden_size""" , UpperCAmelCase ) _snake_case = pooler_dropout _snake_case = pooler_hidden_act class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def lowercase (self ) -> int: return 12 def lowercase (self , UpperCAmelCase , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = 3 , UpperCAmelCase = 40 , UpperCAmelCase = 40 , UpperCAmelCase = None , ) -> Mapping[str, Any]: _snake_case = super().generate_dummy_inputs(preprocessor=UpperCAmelCase , framework=UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
0
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def a_ ( __snake_case : Dict=None ) -> Dict: """simple docstring""" if subparsers is not None: lowerCamelCase_ =subparsers.add_parser('''test''' ) else: lowerCamelCase_ =argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=__snake_case , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=__snake_case ) return parser def a_ ( __snake_case : Optional[Any] ) -> Any: """simple docstring""" lowerCamelCase_ =os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: lowerCamelCase_ =script_name else: lowerCamelCase_ =F'''--config_file={args.config_file} {script_name}''' lowerCamelCase_ =['''accelerate-launch'''] + test_args.split() lowerCamelCase_ =execute_subprocess_async(__snake_case , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def a_ ( ) -> List[Any]: """simple docstring""" lowerCamelCase_ =test_command_parser() lowerCamelCase_ =parser.parse_args() test_command(__snake_case ) if __name__ == "__main__": main()
75
'''simple docstring''' __lowerCAmelCase = [ (1_000, 'M'), (900, 'CM'), (500, 'D'), (400, 'CD'), (100, 'C'), (90, 'XC'), (50, 'L'), (40, 'XL'), (10, 'X'), (9, 'IX'), (5, 'V'), (4, 'IV'), (1, 'I'), ] def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1000} _snake_case = 0 _snake_case = 0 while place < len(_SCREAMING_SNAKE_CASE ): if (place + 1 < len(_SCREAMING_SNAKE_CASE )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = [] for arabic, roman in ROMAN: ((_snake_case), (_snake_case)) = divmod(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) result.append(roman * factor ) if number == 0: break return "".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
import requests def lowerCamelCase__ ( _a , _a): SCREAMING_SNAKE_CASE : Dict = {"Content-Type": "application/json"} SCREAMING_SNAKE_CASE : Optional[int] = requests.post(_a , json={"text": message_body} , headers=_a) if response.status_code != 200: SCREAMING_SNAKE_CASE : Dict = ( "Request to slack returned an error " f"{response.status_code}, the response is:\n{response.text}" ) raise ValueError(_a) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
76
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __lowerCAmelCase = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['PerceiverFeatureExtractor'] __lowerCAmelCase = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __lowerCAmelCase = 'platform' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ): if attention_mask is None: _snake_case = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _snake_case = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _snake_case = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=7 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=99 , UpperCAmelCase=16 , UpperCAmelCase=2 , UpperCAmelCase=4 , UpperCAmelCase=4 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , ) -> Union[str, Any]: _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id _snake_case = initializer_range def lowercase (self ) -> str: _snake_case = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _snake_case = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=UpperCAmelCase , ) _snake_case = prepare_blenderbot_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, inputs_dict def lowercase (self ) -> Dict: _snake_case, _snake_case = self.prepare_config_and_inputs() return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) @require_flax class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = 99 def lowercase (self ) -> Any: _snake_case = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _snake_case = input_ids.shape[0] _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case = self._get_config_and_data() _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = lm_model(input_ids=UpperCAmelCase ) _snake_case = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _snake_case = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _snake_case = lm_model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) _snake_case = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(UpperCAmelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class _lowerCAmelCase ( __snake_case , unittest.TestCase , __snake_case ): '''simple docstring''' lowerCAmelCase_ = True lowerCAmelCase_ = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCAmelCase_ = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def lowercase (self ) -> Any: _snake_case = FlaxBlenderbotModelTester(self ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _snake_case = model_class(UpperCAmelCase ) @jax.jit def encode_jitted(UpperCAmelCase , UpperCAmelCase=None , **UpperCAmelCase ): return model.encode(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase ) with self.subTest("""JIT Enabled""" ): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = model_class(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""] ) _snake_case = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): return model.decode( decoder_input_ids=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , encoder_outputs=UpperCAmelCase , ) with self.subTest("""JIT Enabled""" ): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def lowercase (self ) -> Any: for model_class_name in self.all_model_classes: _snake_case = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _snake_case = np.ones((1, 1) ) * model.config.eos_token_id _snake_case = model(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) @unittest.skipUnless(jax_device != """cpu""" , """3B test too slow on CPU.""" ) @slow def lowercase (self ) -> Dict: _snake_case = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 15, """max_length""": 25} _snake_case = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} _snake_case = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""" , from_pt=UpperCAmelCase ) _snake_case = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) _snake_case = ["""Sam"""] _snake_case = tokenizer(UpperCAmelCase , return_tensors="""jax""" ) _snake_case = model.generate(**UpperCAmelCase , **UpperCAmelCase ) _snake_case = """Sam is a great name. It means \"sun\" in Gaelic.""" _snake_case = tokenizer.batch_decode(UpperCAmelCase , **UpperCAmelCase ) assert generated_txt[0].strip() == tgt_text
341
0
"""simple docstring""" import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class A_ : """simple docstring""" def __init__( self :str , lowercase_ :str , lowercase_ :int=13 , lowercase_ :Optional[Any]=7 , lowercase_ :List[Any]=True , lowercase_ :List[Any]=True , lowercase_ :List[str]=False , lowercase_ :Optional[Any]=True , lowercase_ :List[Any]=99 , lowercase_ :List[str]=64 , lowercase_ :int=5 , lowercase_ :List[str]=4 , lowercase_ :Any=64 , lowercase_ :int="gelu" , lowercase_ :Optional[int]=0.1 , lowercase_ :Union[str, Any]=0.1 , lowercase_ :Union[str, Any]=5_12 , lowercase_ :List[Any]=16 , lowercase_ :Optional[Any]=2 , lowercase_ :str=0.02 , lowercase_ :Any=3 , lowercase_ :Tuple=4 , lowercase_ :Optional[Any]=None , ) -> int: UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope def UpperCAmelCase__ ( self :Union[str, Any] ) -> Union[str, Any]: return MPNetConfig.from_pretrained('microsoft/mpnet-base' ) def UpperCAmelCase__ ( self :Tuple ) -> int: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase__ ( self :List[str] ) -> List[str]: return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self :Optional[int] , lowercase_ :Optional[int] , lowercase_ :List[Any] , lowercase_ :Tuple , lowercase_ :Union[str, Any] , lowercase_ :str , lowercase_ :Any ) -> Any: UpperCAmelCase = MPNetModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase = model(lowercase_ , lowercase_ ) UpperCAmelCase = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCAmelCase__ ( self :Dict , lowercase_ :str , lowercase_ :Any , lowercase_ :Any , lowercase_ :Union[str, Any] , lowercase_ :str , lowercase_ :int ) -> int: UpperCAmelCase = MPNetForQuestionAnswering(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase = model( lowercase_ , attention_mask=lowercase_ , start_positions=lowercase_ , end_positions=lowercase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase__ ( self :Dict , lowercase_ :str , lowercase_ :List[Any] , lowercase_ :Optional[int] , lowercase_ :Dict , lowercase_ :List[Any] , lowercase_ :List[str] ) -> Optional[int]: UpperCAmelCase = self.num_labels UpperCAmelCase = MPNetForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase = model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self :List[str] , lowercase_ :Any , lowercase_ :Union[str, Any] , lowercase_ :Dict , lowercase_ :Optional[Any] , lowercase_ :Optional[int] , lowercase_ :Optional[int] ) -> Any: UpperCAmelCase = self.num_choices UpperCAmelCase = MPNetForMultipleChoice(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase = model( lowercase_ , attention_mask=lowercase_ , labels=lowercase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase__ ( self :List[Any] , lowercase_ :Tuple , lowercase_ :str , lowercase_ :Tuple , lowercase_ :Any , lowercase_ :str , lowercase_ :Union[str, Any] ) -> Dict: UpperCAmelCase = self.num_labels UpperCAmelCase = MPNetForTokenClassification(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase = model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase__ ( self :Any ) -> Dict: UpperCAmelCase = self.prepare_config_and_inputs() ((UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase) , (UpperCAmelCase)) = config_and_inputs UpperCAmelCase = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A_ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" __UpperCamelCase = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __UpperCamelCase = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __UpperCamelCase = False __UpperCamelCase = True def UpperCAmelCase__ ( self :Optional[Any] ) -> int: UpperCAmelCase = MPNetModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=lowercase_ , hidden_size=37 ) def UpperCAmelCase__ ( self :Optional[int] ) -> Dict: self.config_tester.run_common_tests() def UpperCAmelCase__ ( self :int ) -> Optional[int]: UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*lowercase_ ) def UpperCAmelCase__ ( self :Optional[Any] ) -> List[str]: UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*lowercase_ ) def UpperCAmelCase__ ( self :Any ) -> int: UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*lowercase_ ) def UpperCAmelCase__ ( self :List[Any] ) -> List[str]: UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*lowercase_ ) def UpperCAmelCase__ ( self :Dict ) -> List[str]: UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*lowercase_ ) @require_torch class A_ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self :Dict ) -> List[str]: UpperCAmelCase = MPNetModel.from_pretrained('microsoft/mpnet-base' ) UpperCAmelCase = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) UpperCAmelCase = model(lowercase_ )[0] UpperCAmelCase = torch.Size((1, 11, 7_68) ) self.assertEqual(output.shape , lowercase_ ) UpperCAmelCase = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase_ , atol=1E-4 ) )
78
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=2 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=10 , UpperCAmelCase=3 , UpperCAmelCase=32 * 4 , UpperCAmelCase=32 * 6 , UpperCAmelCase=4 , UpperCAmelCase=32 , ) -> Optional[Any]: _snake_case = parent _snake_case = batch_size _snake_case = is_training _snake_case = use_auxiliary_loss _snake_case = num_queries _snake_case = num_channels _snake_case = min_size _snake_case = max_size _snake_case = num_labels _snake_case = mask_feature_size def lowercase (self ) -> str: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( UpperCAmelCase ) _snake_case = torch.ones([self.batch_size, self.min_size, self.max_size] , device=UpperCAmelCase ) _snake_case = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=UpperCAmelCase ) > 0.5 ).float() _snake_case = (torch.rand((self.batch_size, self.num_labels) , device=UpperCAmelCase ) > 0.5).long() _snake_case = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowercase (self ) -> Tuple: return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.prepare_config_and_inputs() _snake_case = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> int: _snake_case = output.encoder_hidden_states _snake_case = output.pixel_decoder_hidden_states _snake_case = output.transformer_decoder_hidden_states self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , config.decoder_config.decoder_layers ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Union[str, Any]: with torch.no_grad(): _snake_case = MaskFormerModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase , output_hidden_states=UpperCAmelCase ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(UpperCAmelCase , UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: _snake_case = MaskFormerForInstanceSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() def comm_check_on_output(UpperCAmelCase ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) _snake_case = model( pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowerCAmelCase_ = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> int: _snake_case = MaskFormerModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def lowercase (self ) -> int: self.config_tester.run_common_tests() def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> Any: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*UpperCAmelCase ) @unittest.skip(reason="""MaskFormer does not use inputs_embeds""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip(reason="""MaskFormer does not have a get_input_embeddings method""" ) def lowercase (self ) -> Optional[int]: pass @unittest.skip(reason="""MaskFormer is not a generative model""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""MaskFormer does not use token embeddings""" ) def lowercase (self ) -> Optional[int]: pass @require_torch_multi_gpu @unittest.skip( reason="""MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> Tuple: pass def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @slow def lowercase (self ) -> int: for model_name in ["facebook/maskformer-swin-small-coco"]: _snake_case = MaskFormerModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = (self.model_tester.min_size,) * 2 _snake_case = { """pixel_values""": torch.randn((2, 3, *size) , device=UpperCAmelCase ), """mask_labels""": torch.randn((2, 10, *size) , device=UpperCAmelCase ), """class_labels""": torch.zeros(2 , 10 , device=UpperCAmelCase ).long(), } _snake_case = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertTrue(outputs.attentions is not None ) def lowercase (self ) -> Tuple: if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ).loss loss.backward() def lowercase (self ) -> List[str]: # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = True _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) _snake_case = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _snake_case = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _snake_case = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _snake_case = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=UpperCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1E-4 def __SCREAMING_SNAKE_CASE ( ): _snake_case = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase (self ) -> Optional[int]: return ( MaskFormerImageProcessor.from_pretrained("""facebook/maskformer-swin-small-coco""" ) if is_vision_available() else None ) def lowercase (self ) -> str: _snake_case = MaskFormerModel.from_pretrained("""facebook/maskformer-swin-small-coco""" ).to(UpperCAmelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[str]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [ [1.6_5_1_2e0_0, -5.2_5_7_2e0_0, -3.3_5_1_9e0_0], [3.6_1_6_9e-0_2, -5.9_0_2_5e0_0, -2.9_3_1_3e0_0], [1.0_7_6_6e-0_4, -7.7_6_3_0e0_0, -5.1_2_6_3e0_0], ] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[Any]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-resnet101-coco-stuff""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> Tuple: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) _snake_case = inputs["""pixel_values"""].to(UpperCAmelCase ) _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""mask_labels"""]] _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""class_labels"""]] with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None )
341
0
'''simple docstring''' import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class _UpperCAmelCase : """simple docstring""" def __init__( self : Optional[Any] , __UpperCAmelCase : str = "cpu" , __UpperCAmelCase : str = "openai/clip-vit-large-patch14" ): '''simple docstring''' _A = device _A = CLIPTokenizerFast.from_pretrained(__UpperCAmelCase ) _A = [0.48145466, 0.4578275, 0.40821073] _A = [0.26862954, 0.26130258, 0.27577711] _A = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _A = torchvision.transforms.Resize(224 ) _A = torchvision.transforms.CenterCrop(224 ) def lowerCAmelCase ( self : Any , __UpperCAmelCase : Union[str, Any] ): '''simple docstring''' _A = self.resize(__UpperCAmelCase ) _A = self.center_crop(__UpperCAmelCase ) _A = self.normalize(__UpperCAmelCase ) return images def __call__( self : Any , __UpperCAmelCase : int=None , __UpperCAmelCase : Union[str, Any]=None , **__UpperCAmelCase : int ): '''simple docstring''' _A = self.tokenizer(text=__UpperCAmelCase , **__UpperCAmelCase ) _A = self.preprocess_img(__UpperCAmelCase ) _A = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class _UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , __UpperCAmelCase : Optional[Any]=10 , __UpperCAmelCase : List[str]=0.01 , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : str=None , __UpperCAmelCase : str=False , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : List[Any]="image" , __UpperCAmelCase : Optional[int]=True , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Union[str, Any]=False , ): '''simple docstring''' super().__init__() _A = None _A = device if device else get_device() if vqgan: _A = vqgan else: _A = load_vqgan(self.device , conf_path=__UpperCAmelCase , ckpt_path=__UpperCAmelCase ) self.vqgan.eval() if clip: _A = clip else: _A = CLIPModel.from_pretrained("openai/clip-vit-base-patch32" ) self.clip.to(self.device ) _A = ProcessorGradientFlow(device=self.device ) _A = iterations _A = lr _A = log _A = make_grid _A = return_val _A = quantize _A = self.vqgan.decoder.z_shape def lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : Optional[Any]=None , __UpperCAmelCase : Optional[Any]=None , __UpperCAmelCase : int=5 , __UpperCAmelCase : Union[str, Any]=True ): '''simple docstring''' _A = [] if output_path is None: _A = "./animation.gif" if input_path is None: _A = self.save_path _A = sorted(glob(input_path + "/*" ) ) if not len(__UpperCAmelCase ): raise ValueError( "No images found in save path, aborting (did you pass save_intermediate=True to the generate" " function?)" ) if len(__UpperCAmelCase ) == 1: print("Only one image found in save path, (did you pass save_intermediate=True to the generate function?)" ) _A = total_duration / len(__UpperCAmelCase ) _A = [frame_duration] * len(__UpperCAmelCase ) if extend_frames: _A = 1.5 _A = 3 for file_name in paths: if file_name.endswith(".png" ): images.append(imageio.imread(__UpperCAmelCase ) ) imageio.mimsave(__UpperCAmelCase , __UpperCAmelCase , duration=__UpperCAmelCase ) print(f'''gif saved to {output_path}''' ) def lowerCAmelCase ( self : int , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : List[str]=None ): '''simple docstring''' if not (path or img): raise ValueError("Input either path or tensor" ) if img is not None: raise NotImplementedError _A = preprocess(Image.open(__UpperCAmelCase ) , target_image_size=256 ).to(self.device ) _A = preprocess_vqgan(__UpperCAmelCase ) _A , *_A = self.vqgan.encode(__UpperCAmelCase ) return z def lowerCAmelCase ( self : str , __UpperCAmelCase : int ): '''simple docstring''' _A = self.latent.detach().requires_grad_() _A = base_latent + transform_vector if self.quantize: _A , *_A = self.vqgan.quantize(__UpperCAmelCase ) else: _A = trans_latent return self.vqgan.decode(__UpperCAmelCase ) def lowerCAmelCase ( self : List[str] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any]=None ): '''simple docstring''' _A = self.clip_preprocessor(text=__UpperCAmelCase , images=__UpperCAmelCase , return_tensors="pt" , padding=__UpperCAmelCase ) _A = self.clip(**__UpperCAmelCase ) _A = clip_outputs.logits_per_image if weights is not None: _A = similarity_logits * weights return similarity_logits.sum() def lowerCAmelCase ( self : List[str] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Any ): '''simple docstring''' _A = self._get_clip_similarity(pos_prompts["prompts"] , __UpperCAmelCase , weights=(1 / pos_prompts["weights"]) ) if neg_prompts: _A = self._get_clip_similarity(neg_prompts["prompts"] , __UpperCAmelCase , weights=neg_prompts["weights"] ) else: _A = torch.tensor([1] , device=self.device ) _A = -torch.log(__UpperCAmelCase ) + torch.log(__UpperCAmelCase ) return loss def lowerCAmelCase ( self : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int ): '''simple docstring''' _A = torch.randn_like(self.latent , requires_grad=__UpperCAmelCase , device=self.device ) _A = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _A = self._add_vector(__UpperCAmelCase ) _A = loop_post_process(__UpperCAmelCase ) _A = self._get_CLIP_loss(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) print("CLIP loss" , __UpperCAmelCase ) if self.log: wandb.log({"CLIP Loss": clip_loss} ) clip_loss.backward(retain_graph=__UpperCAmelCase ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Dict , __UpperCAmelCase : Tuple ): '''simple docstring''' wandb.init(reinit=__UpperCAmelCase , project="face-editor" ) wandb.config.update({"Positive Prompts": positive_prompts} ) wandb.config.update({"Negative Prompts": negative_prompts} ) wandb.config.update({"lr": self.lr, "iterations": self.iterations} ) if image_path: _A = Image.open(__UpperCAmelCase ) _A = image.resize((256, 256) ) wandb.log("Original Image" , wandb.Image(__UpperCAmelCase ) ) def lowerCAmelCase ( self : Any , __UpperCAmelCase : Optional[int] ): '''simple docstring''' if not prompts: return [] _A = [] _A = [] if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _A = [prompt.strip() for prompt in prompts.split("|" )] for prompt in prompts: if isinstance(__UpperCAmelCase , (tuple, list) ): _A = prompt[0] _A = float(prompt[1] ) elif ":" in prompt: _A , _A = prompt.split(":" ) _A = float(__UpperCAmelCase ) else: _A = prompt _A = 1.0 processed_prompts.append(__UpperCAmelCase ) weights.append(__UpperCAmelCase ) return { "prompts": processed_prompts, "weights": torch.tensor(__UpperCAmelCase , device=self.device ), } def lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Tuple=True , __UpperCAmelCase : Optional[Any]=False , __UpperCAmelCase : Dict=True , __UpperCAmelCase : Dict=True , __UpperCAmelCase : Tuple=None , ): '''simple docstring''' if image_path: _A = self._get_latent(__UpperCAmelCase ) else: _A = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) assert pos_prompts, "You must provide at least one positive prompt." _A = self.process_prompts(__UpperCAmelCase ) _A = self.process_prompts(__UpperCAmelCase ) if save_final and save_path is None: _A = os.path.join("./outputs/" , "_".join(pos_prompts["prompts"] ) ) if not os.path.exists(__UpperCAmelCase ): os.makedirs(__UpperCAmelCase ) else: _A = save_path + "_" + get_timestamp() os.makedirs(__UpperCAmelCase ) _A = save_path _A = self.vqgan.decode(self.latent )[0] if show_intermediate: print("Original Image" ) show_pil(custom_to_pil(__UpperCAmelCase ) ) _A = loop_post_process(__UpperCAmelCase ) for iter, transformed_img in enumerate(self._optimize_CLIP(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) ): if show_intermediate: show_pil(__UpperCAmelCase ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}.png''' ) ) if self.log: wandb.log({"Image": wandb.Image(__UpperCAmelCase )} ) if show_final: show_pil(__UpperCAmelCase ) if save_final: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}_final.png''' ) )
79
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self , UpperCAmelCase ) -> Union[str, Any]: for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): _snake_case = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sgugger/tiny-distilbert-classification""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , only_pretrain_model=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , torchscript=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , fpaa=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) # set architectures equal to `None` _snake_case = None _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def lowercase (self ) -> Tuple: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Any: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> int: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> str: _snake_case = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , save_to_csv=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCAmelCase , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(UpperCAmelCase , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(UpperCAmelCase , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(UpperCAmelCase , """train_time.csv""" ) , env_info_csv_file=os.path.join(UpperCAmelCase , """env.csv""" ) , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """env.csv""" ) ).exists() ) def lowercase (self ) -> int: _snake_case = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCAmelCase ): self.assertTrue(hasattr(UpperCAmelCase , """sequential""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """cumulative""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """current""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCAmelCase , """log.txt""" ) , log_print=UpperCAmelCase , trace_memory_line_by_line=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """log.txt""" ) ).exists() )
341
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : Any = logging.get_logger(__name__) a__ : Dict = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowercase_ ( a__ ): __UpperCAmelCase = 'speech_to_text_2' __UpperCAmelCase = ['past_key_values'] __UpperCAmelCase = {'num_attention_heads': 'decoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , a=1_00_00 , a=6 , a=20_48 , a=4 , a=0.0 , a=True , a="relu" , a=2_56 , a=0.1 , a=0.0 , a=0.0 , a=0.02 , a=2 , a=True , a=1 , a=0 , a=2 , a=10_24 , **a , ): UpperCamelCase__ = vocab_size UpperCamelCase__ = d_model UpperCamelCase__ = decoder_ffn_dim UpperCamelCase__ = decoder_layers UpperCamelCase__ = decoder_attention_heads UpperCamelCase__ = dropout UpperCamelCase__ = attention_dropout UpperCamelCase__ = activation_dropout UpperCamelCase__ = activation_function UpperCamelCase__ = init_std UpperCamelCase__ = decoder_layerdrop UpperCamelCase__ = use_cache UpperCamelCase__ = decoder_layers UpperCamelCase__ = scale_embedding # scale factor will be sqrt(d_model) if True UpperCamelCase__ = max_target_positions super().__init__( pad_token_id=a , bos_token_id=a , eos_token_id=a , decoder_start_token_id=a , **a , )
80
'''simple docstring''' from __future__ import annotations def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): if len(_SCREAMING_SNAKE_CASE ) == 0: return [] _snake_case, _snake_case = min(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) _snake_case = int(max_value - min_value ) + 1 _snake_case = [[] for _ in range(_SCREAMING_SNAKE_CASE )] for i in my_list: buckets[int(i - min_value )].append(_SCREAMING_SNAKE_CASE ) return [v for bucket in buckets for v in sorted(_SCREAMING_SNAKE_CASE )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
341
0
"""simple docstring""" def _A ( lowercase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" a =set() # Replace all the whitespace in our sentence a =input_str.replace(''' ''' , '''''' ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(lowercase ) == 26 def _A ( lowercase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" a =[False] * 26 for char in input_str: if char.islower(): a =True elif char.isupper(): a =True return all(lowercase ) def _A ( lowercase = "The quick brown fox jumps over the lazy dog" , ): """simple docstring""" return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def _A ( ): """simple docstring""" from timeit import timeit a ='''from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest''' print(timeit('''is_pangram()''' , setup=lowercase ) ) print(timeit('''is_pangram_faster()''' , setup=lowercase ) ) print(timeit('''is_pangram_fastest()''' , setup=lowercase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
81
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: if not conversation_id: _snake_case = uuid.uuida() if past_user_inputs is None: _snake_case = [] if generated_responses is None: _snake_case = [] _snake_case = conversation_id _snake_case = past_user_inputs _snake_case = generated_responses _snake_case = text def __eq__(self , UpperCAmelCase ) -> Dict: if not isinstance(UpperCAmelCase , UpperCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = False ) -> int: if self.new_user_input: if overwrite: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ f"""with: \"{text}\".""" ) _snake_case = text else: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ f"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: _snake_case = text def lowercase (self ) -> int: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _snake_case = None def lowercase (self , UpperCAmelCase ) -> Any: self.generated_responses.append(UpperCAmelCase ) def lowercase (self ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__(self ) -> Optional[int]: _snake_case = f"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): _snake_case = """user""" if is_user else """bot""" output += f"""{name} >> {text} \n""" return output @add_end_docstrings( __snake_case , r"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , *UpperCAmelCase , **UpperCAmelCase ) -> Union[str, Any]: super().__init__(*UpperCAmelCase , **UpperCAmelCase ) if self.tokenizer.pad_token_id is None: _snake_case = self.tokenizer.eos_token def lowercase (self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase ) -> Dict: _snake_case = {} _snake_case = {} _snake_case = {} if min_length_for_response is not None: _snake_case = min_length_for_response if minimum_tokens is not None: _snake_case = minimum_tokens if "max_length" in generate_kwargs: _snake_case = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _snake_case = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__(self , UpperCAmelCase , UpperCAmelCase=0 , **UpperCAmelCase ) -> Union[str, Any]: _snake_case = super().__call__(UpperCAmelCase , num_workers=UpperCAmelCase , **UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) and len(UpperCAmelCase ) == 1: return outputs[0] return outputs def lowercase (self , UpperCAmelCase , UpperCAmelCase=32 ) -> Dict[str, Any]: if not isinstance(UpperCAmelCase , UpperCAmelCase ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( f"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): _snake_case = self.tokenizer._build_conversation_input_ids(UpperCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _snake_case = self._legacy_parse_and_tokenize(UpperCAmelCase ) if self.framework == "pt": _snake_case = torch.LongTensor([input_ids] ) elif self.framework == "tf": _snake_case = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=10 , **UpperCAmelCase ) -> Optional[int]: _snake_case = generate_kwargs.get("""max_length""" , self.model.config.max_length ) _snake_case = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(f"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) _snake_case = max_length - minimum_tokens _snake_case = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: _snake_case = model_inputs["""attention_mask"""][:, -trim:] _snake_case = model_inputs.pop("""conversation""" ) _snake_case = max_length _snake_case = self.model.generate(**UpperCAmelCase , **UpperCAmelCase ) if self.model.config.is_encoder_decoder: _snake_case = 1 else: _snake_case = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=True ) -> List[str]: _snake_case = model_outputs["""output_ids"""] _snake_case = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCAmelCase , clean_up_tokenization_spaces=UpperCAmelCase , ) _snake_case = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCAmelCase ) return conversation def lowercase (self , UpperCAmelCase ) -> Dict: _snake_case = self.tokenizer.eos_token_id _snake_case = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) ) if len(UpperCAmelCase ) > self.tokenizer.model_max_length: _snake_case = input_ids[-self.tokenizer.model_max_length :] return input_ids
341
0
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __lowerCAmelCase : def __init__( self , _snake_case , _snake_case=12 , _snake_case=7 , _snake_case=True , _snake_case=True , _snake_case=True , _snake_case=99 , _snake_case=32 , _snake_case=32 , _snake_case=2 , _snake_case=4 , _snake_case=37 , _snake_case=0.1 , _snake_case=0.1 , _snake_case=512 , _snake_case=0.02 , _snake_case=0 , _snake_case=None , ): """simple docstring""" _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = seq_length _lowerCAmelCase = is_training _lowerCAmelCase = use_input_mask _lowerCAmelCase = use_labels _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = projection_dim _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = dropout _lowerCAmelCase = attention_dropout _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = initializer_range _lowerCAmelCase = scope _lowerCAmelCase = bos_token_id def snake_case ( self ): """simple docstring""" _lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase = None if self.use_input_mask: _lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: _lowerCAmelCase = input_mask.numpy() _lowerCAmelCase , _lowerCAmelCase = input_mask.shape _lowerCAmelCase = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_snake_case ): _lowerCAmelCase = 1 _lowerCAmelCase = 0 _lowerCAmelCase = self.get_config() return config, input_ids, tf.convert_to_tensor(_snake_case ) def snake_case ( self ): """simple docstring""" return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def snake_case ( self , _snake_case , _snake_case , _snake_case ): """simple docstring""" _lowerCAmelCase = TFBlipTextModel(config=_snake_case ) _lowerCAmelCase = model(_snake_case , attention_mask=_snake_case , training=_snake_case ) _lowerCAmelCase = model(_snake_case , training=_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = config_and_inputs _lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase = (TFBlipTextModel,) if is_tf_available() else () __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def snake_case ( self ): """simple docstring""" _lowerCAmelCase = BlipTextModelTester(self ) _lowerCAmelCase = ConfigTester(self , config_class=_snake_case , hidden_size=37 ) def snake_case ( self ): """simple docstring""" self.config_tester.run_common_tests() def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" pass @unittest.skip(reason="""Blip does not use inputs_embeds""" ) def snake_case ( self ): """simple docstring""" pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def snake_case ( self ): """simple docstring""" pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def snake_case ( self ): """simple docstring""" pass @slow def snake_case ( self ): """simple docstring""" for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase = TFBlipTextModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def snake_case ( self , _snake_case=True ): """simple docstring""" super().test_pt_tf_model_equivalence(allow_missing_keys=_snake_case )
82
'''simple docstring''' from math import factorial, radians def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 18 , _SCREAMING_SNAKE_CASE = 10 ): _snake_case = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians _snake_case = radians(_SCREAMING_SNAKE_CASE ) _snake_case = angle_in_radians _snake_case = 3 _snake_case = -1 for _ in range(_SCREAMING_SNAKE_CASE ): result += (b * (angle_in_radians**a)) / factorial(_SCREAMING_SNAKE_CASE ) _snake_case = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __import__('doctest').testmod()
341
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() snake_case_ : Tuple = logging.get_logger(__name__) snake_case_ : Dict = torch.device('cpu') def A__ ( ): _UpperCamelCase : str = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCamelCase : Dict = Image.open(requests.get(UpperCAmelCase_ , stream=UpperCAmelCase_ ).raw ) return im def A__ ( UpperCAmelCase_ ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1703E00, 2.1107E00, -2.0811E00, 8.8685E-01, 2.4360E-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9636E-01, 2.3478E-01, -1.6963E00, -1.7381E00, -8.6337E-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2768E-01, -4.7429E-01, -1.0897E00, -1.0248E00, 3.5523E-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5330E-01, 2.4211E-01, -6.0185E-01, -8.2789E-01, -6.0446E-02] ) def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Dict = dct.pop(UpperCAmelCase_ ) _UpperCamelCase : str = val def A__ ( UpperCAmelCase_ ): _UpperCamelCase : str = [] for k in state_dict.keys(): _UpperCamelCase : str = k if ".pwconv" in k: _UpperCamelCase : Dict = k_new.replace('.pwconv' , '.point_wise_conv' ) if ".dwconv" in k: _UpperCamelCase : Dict = k_new.replace('.dwconv' , '.depth_wise_conv' ) if ".Proj." in k: _UpperCamelCase : str = k_new.replace('.Proj.' , '.proj.' ) if "patch_embed" in k_new: _UpperCamelCase : List[str] = k_new.replace('patch_embed' , 'swiftformer.patch_embed.patch_embedding' ) if "network" in k_new: _UpperCamelCase : int = k_new.split('.' ) if ls[2].isdigit(): _UpperCamelCase : Tuple = 'swiftformer.encoder.network.' + ls[1] + '.blocks.' + ls[2] + '.' + '.'.join(ls[3:] ) else: _UpperCamelCase : int = k_new.replace('network' , 'swiftformer.encoder.network' ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : str = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _UpperCamelCase : List[Any] = 1_0_0_0 _UpperCamelCase : List[str] = 'huggingface/label-files' _UpperCamelCase : Optional[Any] = 'imagenet-1k-id2label.json' _UpperCamelCase : Union[str, Any] = json.load(open(hf_hub_download(UpperCAmelCase_ , UpperCAmelCase_ , repo_type='dataset' ) , 'r' ) ) _UpperCamelCase : Dict = {int(UpperCAmelCase_ ): v for k, v in idalabel.items()} _UpperCamelCase : Any = idalabel _UpperCamelCase : Any = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _UpperCamelCase : Optional[int] = [3, 3, 6, 4] _UpperCamelCase : Optional[int] = [4_8, 5_6, 1_1_2, 2_2_0] elif swiftformer_name == "swiftformer_s": _UpperCamelCase : Any = [3, 3, 9, 6] _UpperCamelCase : List[Any] = [4_8, 6_4, 1_6_8, 2_2_4] elif swiftformer_name == "swiftformer_l1": _UpperCamelCase : Any = [4, 3, 1_0, 5] _UpperCamelCase : List[Any] = [4_8, 9_6, 1_9_2, 3_8_4] elif swiftformer_name == "swiftformer_l3": _UpperCamelCase : Union[str, Any] = [4, 4, 1_2, 6] _UpperCamelCase : str = [6_4, 1_2_8, 3_2_0, 5_1_2] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith('https' ): _UpperCamelCase : Tuple = torch.hub.load_state_dict_from_url(UpperCAmelCase_ , map_location='cpu' , check_hash=UpperCAmelCase_ ) else: _UpperCamelCase : Optional[Any] = torch.load(UpperCAmelCase_ , map_location='cpu' ) _UpperCamelCase : Optional[int] = checkpoint _UpperCamelCase : Dict = create_rename_keys(UpperCAmelCase_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # load HuggingFace model _UpperCamelCase : Union[str, Any] = SwiftFormerForImageClassification(UpperCAmelCase_ ).eval() hf_model.load_state_dict(UpperCAmelCase_ ) # prepare test inputs _UpperCamelCase : List[str] = prepare_img() _UpperCamelCase : str = ViTImageProcessor.from_pretrained('preprocessor_config' ) _UpperCamelCase : Optional[int] = processor(images=UpperCAmelCase_ , return_tensors='pt' ) # compare outputs from both models _UpperCamelCase : Tuple = get_expected_output(UpperCAmelCase_ ) _UpperCamelCase : str = hf_model(inputs['pixel_values'] ).logits assert hf_logits.shape == torch.Size([1, 1_0_0_0] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCAmelCase_ , atol=1E-3 ) Path(UpperCAmelCase_ ).mkdir(exist_ok=UpperCAmelCase_ ) print(f'Saving model {swiftformer_name} to {pytorch_dump_folder_path}' ) hf_model.save_pretrained(UpperCAmelCase_ ) if __name__ == "__main__": snake_case_ : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') snake_case_ : Union[str, Any] = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
83
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __lowerCAmelCase = '\\n@inproceedings{popovic-2015-chrf,\n title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation",\n month = sep,\n year = "2015",\n address = "Lisbon, Portugal",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W15-3049",\n doi = "10.18653/v1/W15-3049",\n pages = "392--395",\n}\n@inproceedings{popovic-2017-chrf,\n title = "chr{F}++: words helping character n-grams",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Second Conference on Machine Translation",\n month = sep,\n year = "2017",\n address = "Copenhagen, Denmark",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W17-4770",\n doi = "10.18653/v1/W17-4770",\n pages = "612--618",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' __lowerCAmelCase = '\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n' __lowerCAmelCase = '\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n \'score\' (float): The chrF (chrF++) score,\n \'char_order\' (int): The character n-gram order,\n \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n \'beta\' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Tuple: if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/mjpost/sacreBLEU#chrf--chrf""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#chrf--chrf"""] , reference_urls=[ """https://github.com/m-popovic/chrF""", ] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = CHRF.CHAR_ORDER , UpperCAmelCase = CHRF.WORD_ORDER , UpperCAmelCase = CHRF.BETA , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , ) -> int: _snake_case = len(references[0] ) if any(len(UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) _snake_case = [[refs[i] for refs in references] for i in range(UpperCAmelCase )] _snake_case = CHRF(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _snake_case = sb_chrf.corpus_score(UpperCAmelCase , UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
341
0
"""simple docstring""" import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import require_torch if is_torch_available(): from transformers.trainer import DEFAULT_CALLBACKS from .test_trainer import RegressionDataset, RegressionModelConfig, RegressionPreTrainedModel class _SCREAMING_SNAKE_CASE ( A__ ): def __init__( self ) -> Tuple: lowerCAmelCase_ :Dict = [] def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> List[str]: self.events.append("""on_init_end""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Any: self.events.append("""on_train_begin""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Tuple: self.events.append("""on_train_end""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Any: self.events.append("""on_epoch_begin""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Optional[Any]: self.events.append("""on_epoch_end""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Optional[Any]: self.events.append("""on_step_begin""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Any: self.events.append("""on_step_end""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Any: self.events.append("""on_evaluate""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> int: self.events.append("""on_predict""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Union[str, Any]: self.events.append("""on_save""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Dict: self.events.append("""on_log""" ) def __lowerCAmelCase ( self , __A , __A , __A , **__A ) -> Optional[Any]: self.events.append("""on_prediction_step""" ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Tuple: lowerCAmelCase_ :Optional[int] = tempfile.mkdtemp() def __lowerCAmelCase ( self ) -> Union[str, Any]: shutil.rmtree(self.output_dir ) def __lowerCAmelCase ( self , __A=0 , __A=0 , __A=64 , __A=64 , __A=None , __A=False , **__A ) -> Optional[int]: # disable_tqdm in TrainingArguments has a flaky default since it depends on the level of logging. We make sure # its set to False since the tests later on depend on its value. lowerCAmelCase_ :Optional[Any] = RegressionDataset(length=__A ) lowerCAmelCase_ :str = RegressionDataset(length=__A ) lowerCAmelCase_ :Dict = RegressionModelConfig(a=__A , b=__A ) lowerCAmelCase_ :Dict = RegressionPreTrainedModel(__A ) lowerCAmelCase_ :Any = TrainingArguments(self.output_dir , disable_tqdm=__A , report_to=[] , **__A ) return Trainer( __A , __A , train_dataset=__A , eval_dataset=__A , callbacks=__A , ) def __lowerCAmelCase ( self , __A , __A ) -> List[Any]: self.assertEqual(len(__A ) , len(__A ) ) # Order doesn't matter lowerCAmelCase_ :Optional[int] = sorted(__A , key=lambda __A : cb.__name__ if isinstance(__A , __A ) else cb.__class__.__name__ ) lowerCAmelCase_ :List[Any] = sorted(__A , key=lambda __A : cb.__name__ if isinstance(__A , __A ) else cb.__class__.__name__ ) for cba, cba in zip(__A , __A ): if isinstance(__A , __A ) and isinstance(__A , __A ): self.assertEqual(__A , __A ) elif isinstance(__A , __A ) and not isinstance(__A , __A ): self.assertEqual(__A , cba.__class__ ) elif not isinstance(__A , __A ) and isinstance(__A , __A ): self.assertEqual(cba.__class__ , __A ) else: self.assertEqual(__A , __A ) def __lowerCAmelCase ( self , __A ) -> int: lowerCAmelCase_ :Optional[int] = ["""on_init_end""", """on_train_begin"""] lowerCAmelCase_ :Any = 0 lowerCAmelCase_ :Optional[Any] = len(trainer.get_eval_dataloader() ) lowerCAmelCase_ :List[Any] = ["""on_prediction_step"""] * len(trainer.get_eval_dataloader() ) + ["""on_log""", """on_evaluate"""] for _ in range(trainer.state.num_train_epochs ): expected_events.append("""on_epoch_begin""" ) for _ in range(__A ): step += 1 expected_events += ["on_step_begin", "on_step_end"] if step % trainer.args.logging_steps == 0: expected_events.append("""on_log""" ) if trainer.args.evaluation_strategy == IntervalStrategy.STEPS and step % trainer.args.eval_steps == 0: expected_events += evaluation_events.copy() if step % trainer.args.save_steps == 0: expected_events.append("""on_save""" ) expected_events.append("""on_epoch_end""" ) if trainer.args.evaluation_strategy == IntervalStrategy.EPOCH: expected_events += evaluation_events.copy() expected_events += ["on_log", "on_train_end"] return expected_events def __lowerCAmelCase ( self ) -> Tuple: lowerCAmelCase_ :Union[str, Any] = self.get_trainer() lowerCAmelCase_ :Optional[int] = DEFAULT_CALLBACKS.copy() + [ProgressCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) # Callbacks passed at init are added to the default callbacks lowerCAmelCase_ :Optional[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] ) expected_callbacks.append(__A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) # TrainingArguments.disable_tqdm controls if use ProgressCallback or PrinterCallback lowerCAmelCase_ :List[Any] = self.get_trainer(disable_tqdm=__A ) lowerCAmelCase_ :Tuple = DEFAULT_CALLBACKS.copy() + [PrinterCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) def __lowerCAmelCase ( self ) -> Tuple: lowerCAmelCase_ :Dict = DEFAULT_CALLBACKS.copy() + [ProgressCallback] lowerCAmelCase_ :Optional[int] = self.get_trainer() # We can add, pop, or remove by class name trainer.remove_callback(__A ) expected_callbacks.remove(__A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) lowerCAmelCase_ :int = self.get_trainer() lowerCAmelCase_ :Tuple = trainer.pop_callback(__A ) self.assertEqual(cb.__class__ , __A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) trainer.add_callback(__A ) expected_callbacks.insert(0 , __A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) # We can also add, pop, or remove by instance lowerCAmelCase_ :str = self.get_trainer() lowerCAmelCase_ :Optional[int] = trainer.callback_handler.callbacks[0] trainer.remove_callback(__A ) expected_callbacks.remove(__A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) lowerCAmelCase_ :int = self.get_trainer() lowerCAmelCase_ :Tuple = trainer.callback_handler.callbacks[0] lowerCAmelCase_ :Dict = trainer.pop_callback(__A ) self.assertEqual(__A , __A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) trainer.add_callback(__A ) expected_callbacks.insert(0 , __A ) self.check_callbacks_equality(trainer.callback_handler.callbacks , __A ) def __lowerCAmelCase ( self ) -> Optional[Any]: import warnings # XXX: for now ignore scatter_gather warnings in this test since it's not relevant to what's being tested warnings.simplefilter(action="""ignore""" , category=__A ) lowerCAmelCase_ :Tuple = self.get_trainer(callbacks=[MyTestTrainerCallback] ) trainer.train() lowerCAmelCase_ :Tuple = trainer.callback_handler.callbacks[-2].events self.assertEqual(__A , self.get_expected_events(__A ) ) # Independent log/save/eval lowerCAmelCase_ :Union[str, Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , logging_steps=5 ) trainer.train() lowerCAmelCase_ :Optional[int] = trainer.callback_handler.callbacks[-2].events self.assertEqual(__A , self.get_expected_events(__A ) ) lowerCAmelCase_ :Optional[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , save_steps=5 ) trainer.train() lowerCAmelCase_ :Dict = trainer.callback_handler.callbacks[-2].events self.assertEqual(__A , self.get_expected_events(__A ) ) lowerCAmelCase_ :Any = self.get_trainer(callbacks=[MyTestTrainerCallback] , eval_steps=5 , evaluation_strategy="""steps""" ) trainer.train() lowerCAmelCase_ :List[Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(__A , self.get_expected_events(__A ) ) lowerCAmelCase_ :Optional[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , evaluation_strategy="""epoch""" ) trainer.train() lowerCAmelCase_ :Optional[Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(__A , self.get_expected_events(__A ) ) # A bit of everything lowerCAmelCase_ :Any = self.get_trainer( callbacks=[MyTestTrainerCallback] , logging_steps=3 , save_steps=10 , eval_steps=5 , evaluation_strategy="""steps""" , ) trainer.train() lowerCAmelCase_ :Union[str, Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(__A , self.get_expected_events(__A ) ) # warning should be emitted for duplicated callbacks with patch("""transformers.trainer_callback.logger.warning""" ) as warn_mock: lowerCAmelCase_ :Optional[int] = self.get_trainer( callbacks=[MyTestTrainerCallback, MyTestTrainerCallback] , ) assert str(__A ) in warn_mock.call_args[0][0]
84
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCAmelCase = '\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n' __lowerCAmelCase = '\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {\'spearmanr\': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results[\'spearmanr\'])\n -0.7\n >>> print(round(results[\'spearmanr_pvalue\'], 2))\n 0.19\n' __lowerCAmelCase = r'\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Optional[Any]: _snake_case = spearmanr(UpperCAmelCase , UpperCAmelCase ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
341
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Optional[int] = {"vocab_file": "sentencepiece.model"} _SCREAMING_SNAKE_CASE : Dict = { "vocab_file": { "google/rembert": "https://huggingface.co/google/rembert/resolve/main/sentencepiece.model", }, } _SCREAMING_SNAKE_CASE : str = { "google/rembert": 256, } class _snake_case ( lowercase_ ): lowerCAmelCase_ : Union[str, Any] = VOCAB_FILES_NAMES lowerCAmelCase_ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , a__ , a__=False , a__=True , a__=True , a__="[CLS]" , a__="[SEP]" , a__="[UNK]" , a__="[SEP]" , a__="[PAD]" , a__="[CLS]" , a__="[MASK]" , **a__ , ) -> Tuple: '''simple docstring''' super().__init__( do_lower_case=a__ , remove_space=a__ , keep_accents=a__ , bos_token=a__ , eos_token=a__ , unk_token=a__ , sep_token=a__ , pad_token=a__ , cls_token=a__ , mask_token=a__ , **a__ , ) snake_case_ = do_lower_case snake_case_ = remove_space snake_case_ = keep_accents snake_case_ = vocab_file snake_case_ = spm.SentencePieceProcessor() self.sp_model.Load(a__ ) @property def lowerCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' return len(self.sp_model ) def lowerCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ = {self.convert_ids_to_tokens(a__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Optional[int]: '''simple docstring''' snake_case_ = self.__dict__.copy() snake_case_ = None return state def __setstate__( self , a__ ) -> Optional[Any]: '''simple docstring''' snake_case_ = d snake_case_ = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def lowerCAmelCase__ ( self , a__ , a__=False ) -> int: '''simple docstring''' snake_case_ = self.sp_model.EncodeAsPieces(a__ ) return pieces def lowerCAmelCase__ ( self , a__ ) -> Tuple: '''simple docstring''' return self.sp_model.PieceToId(a__ ) def lowerCAmelCase__ ( self , a__ ) -> List[Any]: '''simple docstring''' return self.sp_model.IdToPiece(a__ ) def lowerCAmelCase__ ( self , a__ ) -> List[str]: '''simple docstring''' snake_case_ = self.sp_model.decode_pieces(a__ ) return out_string def lowerCAmelCase__ ( self , a__ , a__ = None ) -> List[int]: '''simple docstring''' snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCAmelCase__ ( self , a__ , a__ = None , a__ = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a__ )) + [1] + ([0] * len(a__ )) + [1] return [1] + ([0] * len(a__ )) + [1] def lowerCAmelCase__ ( self , a__ , a__ = None ) -> List[int]: '''simple docstring''' snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase__ ( self , a__ , a__ = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(a__ ): logger.error("Vocabulary path ({}) should be a directory".format(a__ ) ) return snake_case_ = os.path.join( a__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a__ ): copyfile(self.vocab_file , a__ ) return (out_vocab_file,)
85
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=4 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[2, 2, 3, 2] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase=37 , UpperCAmelCase="gelu" , UpperCAmelCase=10 , UpperCAmelCase=0.02 , UpperCAmelCase=["stage2", "stage3", "stage4"] , UpperCAmelCase=3 , UpperCAmelCase=None , ) -> List[Any]: _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = num_stages _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = intermediate_size _snake_case = hidden_act _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = out_features _snake_case = num_labels _snake_case = scope _snake_case = num_stages def lowercase (self ) -> List[Any]: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def lowercase (self ) -> Tuple: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def lowercase (self ) -> Any: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCAmelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCAmelCase , loss_ignore_index=255 , num_labels=self.num_labels , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: _snake_case = UperNetForSemanticSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowercase (self ) -> Tuple: _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ), ( _snake_case ), ( _snake_case ), ) = config_and_inputs _snake_case = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCAmelCase_ = {"image-segmentation": UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> Optional[Any]: _snake_case = UperNetModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=37 ) def lowercase (self ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase (self ) -> Union[str, Any]: return def lowercase (self ) -> Union[str, Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def lowercase (self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> str: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> int: pass def lowercase (self ) -> List[str]: def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = _config_zero_init(UpperCAmelCase ) _snake_case = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _snake_case = model_class(config=UpperCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def lowercase (self ) -> Optional[Any]: pass @slow def lowercase (self ) -> Tuple: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = UperNetForSemanticSegmentation.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) _snake_case = Image.open(_SCREAMING_SNAKE_CASE ).convert("""RGB""" ) return image @require_torch @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) ) def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
341
0
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase ): return 10 - x * x def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase ): # Bolzano theory in order to find if there is a root between a and b if equation(_UpperCamelCase ) * equation(_UpperCamelCase ) >= 0: raise ValueError('Wrong space!' ) __lowerCAmelCase : List[Any] = a while (b - a) >= 0.01: # Find middle point __lowerCAmelCase : int = (a + b) / 2 # Check if middle point is root if equation(_UpperCamelCase ) == 0.0: break # Decide the side to repeat the steps if equation(_UpperCamelCase ) * equation(_UpperCamelCase ) < 0: __lowerCAmelCase : Dict = c else: __lowerCAmelCase : str = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
86
'''simple docstring''' import argparse from collections import defaultdict def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = f"""class {class_name}(""" _snake_case = f"""{4 * " "}def {test_name}(""" _snake_case = f"""{8 * " "}{correct_line.split()[0]}""" _snake_case = f"""{16 * " "}{correct_line.split()[0]}""" _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = 0 _snake_case = 0 _snake_case = [] for line in lines: if line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and in_func and (line.startswith(_SCREAMING_SNAKE_CASE ) or line.startswith(_SCREAMING_SNAKE_CASE )): _snake_case = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: _snake_case = True if in_class and in_func and in_line: if ")" not in line: continue else: _snake_case = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) _snake_case = _snake_case = _snake_case = _snake_case = False else: new_lines.append(_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: for line in new_lines: f.write(_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): if fail is not None: with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = {l.strip() for l in f.readlines()} else: _snake_case = None with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = defaultdict(_SCREAMING_SNAKE_CASE ) for line in correct_lines: _snake_case, _snake_case, _snake_case, _snake_case = line.split(""";""" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) __lowerCAmelCase = parser.parse_args() main(args.correct_filename, args.fail_filename)
341
0
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def lowercase_ ( _lowerCamelCase : str , _lowerCamelCase : float | Decimal , _lowerCamelCase : float = 10**-10): lowercase__ : Optional[int] = a while True: lowercase__ : int = Decimal(_lowerCamelCase) - ( Decimal(eval(_lowerCamelCase)) / Decimal(eval(str(diff(_lowerCamelCase)))) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(_lowerCamelCase)) < precision: # noqa: S307 return float(_lowerCamelCase) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(f"The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}") # Find root of polynomial print(f"The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}") # Find Square Root of 5 print(f"The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}") # Exponential Roots print(f"The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}")
87
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance __lowerCAmelCase : str = 637_8137.0 __lowerCAmelCase : Optional[Any] = 635_6752.31_4245 __lowerCAmelCase : List[str] = 6378137 def a__ ( A_, A_, A_, A_ ): '''simple docstring''' __magic_name__ = (AXIS_A - AXIS_B) / AXIS_A # Parametric latitudes # https://en.wikipedia.org/wiki/Latitude#Parametric_(or_reduced)_latitude __magic_name__ = atan((1 - flattening) * tan(radians(A_ ) ) ) __magic_name__ = atan((1 - flattening) * tan(radians(A_ ) ) ) # Compute central angle between two points # using haversine theta. sigma = haversine_distance / equatorial radius __magic_name__ = haversine_distance(A_, A_, A_, A_ ) / EQUATORIAL_RADIUS # Intermediate P and Q values __magic_name__ = (b_lata + b_lata) / 2 __magic_name__ = (b_lata - b_lata) / 2 # Intermediate X value # X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2) __magic_name__ = (sin(A_ ) ** 2) * (cos(A_ ) ** 2) __magic_name__ = cos(sigma / 2 ) ** 2 __magic_name__ = (sigma - sin(A_ )) * (x_numerator / x_demonimator) # Intermediate Y value # Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2) __magic_name__ = (cos(A_ ) ** 2) * (sin(A_ ) ** 2) __magic_name__ = sin(sigma / 2 ) ** 2 __magic_name__ = (sigma + sin(A_ )) * (y_numerator / y_denominator) return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value))) if __name__ == "__main__": import doctest doctest.testmod()
88
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = ["pixel_values"] def __init__(self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: super().__init__(**UpperCAmelCase ) _snake_case = size if size is not None else {"""height""": 256, """width""": 256} _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( UpperCAmelCase , size=(size["""height"""], size["""width"""]) , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(UpperCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> List[Any]: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> PIL.Image.Image: _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = size if size is not None else self.size _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=UpperCAmelCase , size=UpperCAmelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] _snake_case = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _snake_case = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
341
0
'''simple docstring''' import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __magic_name__ : def __init__( self : Optional[Any] ,_UpperCAmelCase : Any ,_UpperCAmelCase : Tuple=13 ,_UpperCAmelCase : Optional[int]=30 ,_UpperCAmelCase : int=2 ,_UpperCAmelCase : Union[str, Any]=3 ,_UpperCAmelCase : int=True ,_UpperCAmelCase : Optional[Any]=True ,_UpperCAmelCase : Optional[Any]=32 ,_UpperCAmelCase : int=5 ,_UpperCAmelCase : List[str]=4 ,_UpperCAmelCase : Any=37 ,_UpperCAmelCase : List[str]="gelu" ,_UpperCAmelCase : Any=0.1 ,_UpperCAmelCase : Any=0.1 ,_UpperCAmelCase : Union[str, Any]=10 ,_UpperCAmelCase : Union[str, Any]=0.02 ,_UpperCAmelCase : List[str]=3 ,_UpperCAmelCase : Optional[Any]=0.6 ,_UpperCAmelCase : str=None ,): _a : Union[str, Any] = parent _a : Optional[Any] = batch_size _a : Any = image_size _a : int = patch_size _a : int = num_channels _a : Tuple = is_training _a : Any = use_labels _a : int = hidden_size _a : Tuple = num_hidden_layers _a : Any = num_attention_heads _a : List[Any] = intermediate_size _a : Dict = hidden_act _a : Tuple = hidden_dropout_prob _a : List[str] = attention_probs_dropout_prob _a : List[Any] = type_sequence_label_size _a : List[str] = initializer_range _a : Optional[int] = mask_ratio _a : List[Any] = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _a : int = (image_size // patch_size) ** 2 _a : str = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def __lowercase ( self : Any ): _a : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a : Dict = None if self.use_labels: _a : Tuple = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) _a : Dict = self.get_config() return config, pixel_values, labels def __lowercase ( self : Optional[int] ): return ViTMAEConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=_UpperCAmelCase ,initializer_range=self.initializer_range ,mask_ratio=self.mask_ratio ,) def __lowercase ( self : Union[str, Any] ,_UpperCAmelCase : Tuple ,_UpperCAmelCase : str ,_UpperCAmelCase : Any ): _a : Any = ViTMAEModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _a : Tuple = model(_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def __lowercase ( self : Any ,_UpperCAmelCase : Optional[Any] ,_UpperCAmelCase : int ,_UpperCAmelCase : Tuple ): _a : List[str] = ViTMAEForPreTraining(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _a : Union[str, Any] = model(_UpperCAmelCase ) _a : List[str] = (self.image_size // self.patch_size) ** 2 _a : int = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _a : Optional[int] = 1 _a : Union[str, Any] = ViTMAEForPreTraining(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _a : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _a : Optional[int] = model(_UpperCAmelCase ) _a : int = self.patch_size**2 self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) def __lowercase ( self : int ): _a : List[str] = self.prepare_config_and_inputs() _a , _a , _a : str = config_and_inputs _a : List[str] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __magic_name__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): lowerCAmelCase : List[str] = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () lowerCAmelCase : str = {'feature-extraction': ViTMAEModel} if is_torch_available() else {} lowerCAmelCase : Union[str, Any] = False lowerCAmelCase : str = False lowerCAmelCase : Dict = False lowerCAmelCase : List[str] = False def __lowercase ( self : Dict ): _a : Any = ViTMAEModelTester(self ) _a : Optional[int] = ConfigTester(self ,config_class=_UpperCAmelCase ,has_text_modality=_UpperCAmelCase ,hidden_size=37 ) def __lowercase ( self : Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def __lowercase ( self : List[str] ): pass def __lowercase ( self : List[Any] ): _a , _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : List[str] = model_class(_UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) _a : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_UpperCAmelCase ,nn.Linear ) ) def __lowercase ( self : Optional[int] ): _a , _a : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : List[str] = model_class(_UpperCAmelCase ) _a : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : str = [*signature.parameters.keys()] _a : str = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_UpperCAmelCase ) def __lowercase ( self : Optional[int] ): _a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def __lowercase ( self : Dict ): _a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_UpperCAmelCase ) def __lowercase ( self : List[Any] ,_UpperCAmelCase : Dict ,_UpperCAmelCase : List[str] ,_UpperCAmelCase : Tuple ): # make masks reproducible np.random.seed(2 ) _a : Tuple = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) _a : Union[str, Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _a : Optional[int] = torch.from_numpy(_UpperCAmelCase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _a : Optional[int] = pt_noise super().check_pt_tf_models(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) def __lowercase ( self : List[str] ): _a , _a : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : List[str] = model_class(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): _a : Any = model(**self._prepare_for_class(_UpperCAmelCase ,_UpperCAmelCase ) ) _a : str = outputs[0].cpu().numpy() _a : Any = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_UpperCAmelCase ) _a : Dict = model_class.from_pretrained(_UpperCAmelCase ) model.to(_UpperCAmelCase ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): _a : str = model(**self._prepare_for_class(_UpperCAmelCase ,_UpperCAmelCase ) ) # Make sure we don't have nans _a : Any = after_outputs[0].cpu().numpy() _a : int = 0 _a : Optional[int] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_UpperCAmelCase ,1E-5 ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def __lowercase ( self : str ): pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def __lowercase ( self : Dict ): pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def __lowercase ( self : Optional[Any] ): pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def __lowercase ( self : Union[str, Any] ): pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : Optional[Any] ): pass @slow def __lowercase ( self : Any ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : Union[str, Any] = ViTMAEModel.from_pretrained(_UpperCAmelCase ) self.assertIsNotNone(_UpperCAmelCase ) def __lowerCamelCase ( ) -> Optional[int]: _a : str = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class __magic_name__ ( unittest.TestCase ): @cached_property def __lowercase ( self : Union[str, Any] ): return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def __lowercase ( self : Optional[int] ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _a : int = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ).to(_UpperCAmelCase ) _a : Union[str, Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : List[Any] = image_processor(images=_UpperCAmelCase ,return_tensors='pt' ).to(_UpperCAmelCase ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _a : Dict = ViTMAEConfig() _a : Any = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _a : Any = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): _a : Dict = model(**_UpperCAmelCase ,noise=torch.from_numpy(_UpperCAmelCase ).to(device=_UpperCAmelCase ) ) # verify the logits _a : int = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape ,_UpperCAmelCase ) _a : int = torch.tensor( [[-0.05_48, -1.70_23, -0.93_25], [0.37_21, -0.56_70, -0.22_33], [0.82_35, -1.38_78, -0.35_24]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] ,expected_slice.to(_UpperCAmelCase ) ,atol=1E-4 ) )
89
'''simple docstring''' __lowerCAmelCase = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure the supplied data is a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_SCREAMING_SNAKE_CASE ) _snake_case = """""".join(bin(_SCREAMING_SNAKE_CASE )[2:].zfill(8 ) for byte in data ) _snake_case = len(_SCREAMING_SNAKE_CASE ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b"""=""" * ((6 - len(_SCREAMING_SNAKE_CASE ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_SCREAMING_SNAKE_CASE ) % 6) else: _snake_case = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 6 ) ).encode() + padding ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = ( """argument should be a bytes-like object or ASCII string, """ f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_SCREAMING_SNAKE_CASE ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): try: _snake_case = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) _snake_case = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_SCREAMING_SNAKE_CASE ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 8 ) ] return bytes(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowerCAmelCase : """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=13 , lowerCamelCase__=32 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=16 , lowerCamelCase__=[32, 64, 128] , lowerCamelCase__=[1, 2, 1] , lowerCamelCase__=[2, 2, 4] , lowerCamelCase__=2 , lowerCamelCase__=2.0 , lowerCamelCase__=True , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.1 , lowerCamelCase__="gelu" , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=0.02 , lowerCamelCase__=1e-5 , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=10 , lowerCamelCase__=8 , lowerCamelCase__=["stage1", "stage2"] , lowerCamelCase__=[1, 2] , ) -> int: '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = embed_dim __lowerCamelCase = hidden_sizes __lowerCamelCase = depths __lowerCamelCase = num_heads __lowerCamelCase = window_size __lowerCamelCase = mlp_ratio __lowerCamelCase = qkv_bias __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = drop_path_rate __lowerCamelCase = hidden_act __lowerCamelCase = use_absolute_embeddings __lowerCamelCase = patch_norm __lowerCamelCase = layer_norm_eps __lowerCamelCase = initializer_range __lowerCamelCase = is_training __lowerCamelCase = scope __lowerCamelCase = use_labels __lowerCamelCase = type_sequence_label_size __lowerCamelCase = encoder_stride __lowerCamelCase = out_features __lowerCamelCase = out_indices def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = self.get_config() return config, pixel_values, labels def lowercase_ ( self ) -> List[str]: '''simple docstring''' return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> int: '''simple docstring''' __lowerCamelCase = FocalNetModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) __lowerCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __lowerCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> List[Any]: '''simple docstring''' __lowerCamelCase = FocalNetBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None __lowerCamelCase = None __lowerCamelCase = FocalNetBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> int: '''simple docstring''' __lowerCamelCase = FocalNetForMaskedImageModeling(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowerCamelCase = 1 __lowerCamelCase = FocalNetForMaskedImageModeling(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> str: '''simple docstring''' __lowerCamelCase = self.type_sequence_label_size __lowerCamelCase = FocalNetForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowerCamelCase = 1 __lowerCamelCase = FocalNetForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = config_and_inputs __lowerCamelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( __magic_name__ , __magic_name__ , unittest.TestCase ): """simple docstring""" snake_case_ = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) snake_case_ = ( {'''feature-extraction''': FocalNetModel, '''image-classification''': FocalNetForImageClassification} if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase = FocalNetModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , embed_dim=37 , has_text_modality=lowerCamelCase__ ) def lowercase_ ( self ) -> str: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase_ ( self ) -> str: '''simple docstring''' return def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase__ ) def lowercase_ ( self ) -> Any: '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @unittest.skip(reason='FocalNet does not use inputs_embeds' ) def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason='FocalNet does not use feedforward chunking' ) def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' pass def lowercase_ ( self ) -> Dict: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: __lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: __lowerCamelCase = model_class(lowerCamelCase__ ) __lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase = [*signature.parameters.keys()] __lowerCamelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> int: '''simple docstring''' __lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): __lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) __lowerCamelCase = outputs.hidden_states __lowerCamelCase = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # FocalNet has a different seq_length __lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowerCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) __lowerCamelCase = outputs.reshaped_hidden_states self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = reshaped_hidden_states[0].shape __lowerCamelCase = ( reshaped_hidden_states[0].view(lowerCamelCase__ , lowerCamelCase__ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def lowercase_ ( self ) -> Dict: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: __lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = 3 __lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __lowerCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __lowerCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: __lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) @slow def lowercase_ ( self ) -> str: '''simple docstring''' for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = FocalNetModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' __lowerCamelCase , __lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase = _config_zero_init(lowerCamelCase__ ) for model_class in self.all_model_classes: __lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def lowercase_ ( self ) -> List[str]: '''simple docstring''' # TODO update organization return AutoImageProcessor.from_pretrained('microsoft/focalnet-tiny' ) if is_vision_available() else None @slow def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = FocalNetForImageClassification.from_pretrained('microsoft/focalnet-tiny' ).to(lowerCamelCase__ ) __lowerCamelCase = self.default_image_processor __lowerCamelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) __lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): __lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits __lowerCamelCase = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) __lowerCamelCase = torch.tensor([0.21_66, -0.43_68, 0.21_91] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 281 ) @require_torch class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): """simple docstring""" snake_case_ = (FocalNetBackbone,) if is_torch_available() else () snake_case_ = FocalNetConfig snake_case_ = False def lowercase_ ( self ) -> Any: '''simple docstring''' __lowerCamelCase = FocalNetModelTester(self )
90
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase_ : Any = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] UpperCAmelCase_ : Union[str, Any] = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] UpperCAmelCase_ : Optional[Any] = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): UpperCAmelCase_ : Any = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCAmelCase_ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
91
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract UpperCamelCase__ = logging.get_logger(__name__) def _a ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : int ): return [ int(10_00 * (box[0] / width) ), int(10_00 * (box[1] / height) ), int(10_00 * (box[2] / width) ), int(10_00 * (box[3] / height) ), ] def _a ( SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : Optional[str] , SCREAMING_SNAKE_CASE_ : Optional[str] ): __lowerCAmelCase = to_pil_image(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase , __lowerCAmelCase = pil_image.size __lowerCAmelCase = pytesseract.image_to_data(SCREAMING_SNAKE_CASE_ , lang=SCREAMING_SNAKE_CASE_ , output_type="dict" , config=SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates __lowerCAmelCase = [idx for idx, word in enumerate(SCREAMING_SNAKE_CASE_ ) if not word.strip()] __lowerCAmelCase = [word for idx, word in enumerate(SCREAMING_SNAKE_CASE_ ) if idx not in irrelevant_indices] __lowerCAmelCase = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE_ ) if idx not in irrelevant_indices] __lowerCAmelCase = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE_ ) if idx not in irrelevant_indices] __lowerCAmelCase = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE_ ) if idx not in irrelevant_indices] __lowerCAmelCase = [coord for idx, coord in enumerate(SCREAMING_SNAKE_CASE_ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format __lowerCAmelCase = [] for x, y, w, h in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): __lowerCAmelCase = [x, y, x + w, y + h] actual_boxes.append(SCREAMING_SNAKE_CASE_ ) # finally, normalize the bounding boxes __lowerCAmelCase = [] for box in actual_boxes: normalized_boxes.append(normalize_box(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) assert len(SCREAMING_SNAKE_CASE_ ) == len(SCREAMING_SNAKE_CASE_ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class a__ ( snake_case__ ): _a : Dict = ["""pixel_values"""] def __init__( self , _A = True , _A = None , _A = PILImageResampling.BILINEAR , _A = True , _A = 1 / 2_5_5 , _A = True , _A = None , _A = None , _A = True , _A = None , _A = "" , **_A , ): """simple docstring""" super().__init__(**_A ) __lowerCAmelCase = size if size is not None else {"height": 2_2_4, "width": 2_2_4} __lowerCAmelCase = get_size_dict(_A ) __lowerCAmelCase = do_resize __lowerCAmelCase = size __lowerCAmelCase = resample __lowerCAmelCase = do_rescale __lowerCAmelCase = rescale_value __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowerCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD __lowerCAmelCase = apply_ocr __lowerCAmelCase = ocr_lang __lowerCAmelCase = tesseract_config def __SCREAMING_SNAKE_CASE( self , _A , _A , _A = PILImageResampling.BILINEAR , _A = None , **_A , ): """simple docstring""" __lowerCAmelCase = get_size_dict(_A ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) __lowerCAmelCase = (size["height"], size["width"]) return resize(_A , size=_A , resample=_A , data_format=_A , **_A ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A = None , **_A , ): """simple docstring""" return rescale(_A , scale=_A , data_format=_A , **_A ) def __SCREAMING_SNAKE_CASE( self , _A , _A , _A , _A = None , **_A , ): """simple docstring""" return normalize(_A , mean=_A , std=_A , data_format=_A , **_A ) def __SCREAMING_SNAKE_CASE( self , _A , _A = None , _A = None , _A=None , _A = None , _A = None , _A = None , _A = None , _A = None , _A = None , _A = None , _A = None , _A = None , _A = ChannelDimension.FIRST , **_A , ): """simple docstring""" __lowerCAmelCase = do_resize if do_resize is not None else self.do_resize __lowerCAmelCase = size if size is not None else self.size __lowerCAmelCase = get_size_dict(_A ) __lowerCAmelCase = resample if resample is not None else self.resample __lowerCAmelCase = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize __lowerCAmelCase = image_mean if image_mean is not None else self.image_mean __lowerCAmelCase = image_std if image_std is not None else self.image_std __lowerCAmelCase = apply_ocr if apply_ocr is not None else self.apply_ocr __lowerCAmelCase = ocr_lang if ocr_lang is not None else self.ocr_lang __lowerCAmelCase = tesseract_config if tesseract_config is not None else self.tesseract_config __lowerCAmelCase = make_list_of_images(_A ) if not valid_images(_A ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("If do_normalize is True, image_mean and image_std must be specified." ) # All transformations expect numpy arrays. __lowerCAmelCase = [to_numpy_array(_A ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , "pytesseract" ) __lowerCAmelCase = [] __lowerCAmelCase = [] for image in images: __lowerCAmelCase , __lowerCAmelCase = apply_tesseract(_A , _A , _A ) words_batch.append(_A ) boxes_batch.append(_A ) if do_resize: __lowerCAmelCase = [self.resize(image=_A , size=_A , resample=_A ) for image in images] if do_rescale: __lowerCAmelCase = [self.rescale(image=_A , scale=_A ) for image in images] if do_normalize: __lowerCAmelCase = [self.normalize(image=_A , mean=_A , std=_A ) for image in images] __lowerCAmelCase = [to_channel_dimension_format(_A , _A ) for image in images] __lowerCAmelCase = BatchFeature(data={"pixel_values": images} , tensor_type=_A ) if apply_ocr: __lowerCAmelCase = words_batch __lowerCAmelCase = boxes_batch return data
92
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __lowerCAmelCase = TypeVar('T') __lowerCAmelCase = Union[List[T], Tuple[T, ...]] __lowerCAmelCase = Union[T, List[T], Dict[str, T]] __lowerCAmelCase = Union[str, bytes, os.PathLike]
341
0
'''simple docstring''' import inspect import unittest class lowerCAmelCase__ ( unittest.TestCase ): def _snake_case ( self ): """simple docstring""" try: import diffusers # noqa: F401 except ImportError: assert False def _snake_case ( self ): """simple docstring""" import diffusers from diffusers.dependency_versions_table import deps lowercase_ : List[Any] = inspect.getmembers(__SCREAMING_SNAKE_CASE , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": lowercase_ : Any = '''k-diffusion''' elif backend == "invisible_watermark": lowercase_ : Dict = '''invisible-watermark''' assert backend in deps, F'''{backend} is not in the deps table!'''
93
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: _snake_case = data _snake_case = previous _snake_case = next_node def __str__(self ) -> str: return f"""{self.data}""" def lowercase (self ) -> int: return self.data def lowercase (self ) -> Dict: return self.next def lowercase (self ) -> Union[str, Any]: return self.previous class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase ) -> List[str]: _snake_case = head def __iter__(self ) -> Optional[Any]: return self def lowercase (self ) -> str: if not self.current: raise StopIteration else: _snake_case = self.current.get_data() _snake_case = self.current.get_next() return value class _lowerCAmelCase : '''simple docstring''' def __init__(self ) -> Optional[int]: _snake_case = None # First node in list _snake_case = None # Last node in list def __str__(self ) -> Optional[int]: _snake_case = self.head _snake_case = [] while current is not None: nodes.append(current.get_data() ) _snake_case = current.get_next() return " ".join(str(UpperCAmelCase ) for node in nodes ) def __contains__(self , UpperCAmelCase ) -> int: _snake_case = self.head while current: if current.get_data() == value: return True _snake_case = current.get_next() return False def __iter__(self ) -> Union[str, Any]: return LinkedListIterator(self.head ) def lowercase (self ) -> str: if self.head: return self.head.get_data() return None def lowercase (self ) -> List[Any]: if self.tail: return self.tail.get_data() return None def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: _snake_case = node _snake_case = node else: self.insert_before_node(self.head , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: self.set_head(UpperCAmelCase ) else: self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: _snake_case = Node(UpperCAmelCase ) if self.head is None: self.set_head(UpperCAmelCase ) else: self.set_tail(UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.previous if node.get_previous() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.next if node.get_next() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = 1 _snake_case = Node(UpperCAmelCase ) _snake_case = self.head while node: if current_position == position: self.insert_before_node(UpperCAmelCase , UpperCAmelCase ) return current_position += 1 _snake_case = node.next self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> Node: _snake_case = self.head while node: if node.get_data() == item: return node _snake_case = node.get_next() raise Exception("""Node not found""" ) def lowercase (self , UpperCAmelCase ) -> Optional[int]: if (node := self.get_node(UpperCAmelCase )) is not None: if node == self.head: _snake_case = self.head.get_next() if node == self.tail: _snake_case = self.tail.get_previous() self.remove_node_pointers(UpperCAmelCase ) @staticmethod def lowercase (UpperCAmelCase ) -> None: if node.get_next(): _snake_case = node.previous if node.get_previous(): _snake_case = node.next _snake_case = None _snake_case = None def lowercase (self ) -> Dict: return self.head is None def __SCREAMING_SNAKE_CASE ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
341
0
import math from collections.abc import Iterator from itertools import takewhile def __lowerCamelCase ( UpperCAmelCase_ : int ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(UpperCAmelCase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __lowerCamelCase ( ): """simple docstring""" a :Optional[int] = 2 while True: if is_prime(UpperCAmelCase_ ): yield num num += 1 def __lowerCamelCase ( UpperCAmelCase_ : int = 200_0000 ): """simple docstring""" return sum(takewhile(lambda UpperCAmelCase_ : x < n , prime_generator() ) ) if __name__ == "__main__": print(F"""{solution() = }""")
94
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput __lowerCAmelCase = 8 def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x * 255).int().clamp(0 , 255 ) _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c h w -> b c 1 h w""" ) _snake_case = ((x & mask) != 0).float() _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c d h w -> b (c d) h w""" ) _snake_case = bits * 2 - 1 return bits def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x > 0).int() _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE , dtype=torch.intaa ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b (c d) h w -> b c d h w""" , d=8 ) _snake_case = reduce(x * mask , """b c d h w -> b c h w""" , """sum""" ) return (dec / 255).clamp(0.0 , 1.0 ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _snake_case = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _snake_case = self.alphas_cumprod[timestep] _snake_case = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _snake_case = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _snake_case = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _snake_case = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _snake_case = model_output.device if torch.is_tensor(_SCREAMING_SNAKE_CASE ) else """cpu""" _snake_case = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise _snake_case = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="epsilon" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): _snake_case = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _snake_case, _snake_case = torch.split(_SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: _snake_case = None # 1. compute alphas, betas _snake_case = self.alphas_cumprod[t] _snake_case = self.alphas_cumprod[t - 1] if t > 0 else self.one _snake_case = 1 - alpha_prod_t _snake_case = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _snake_case = model_output else: raise ValueError(f"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _snake_case = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _snake_case = 0 if t > 0: _snake_case = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_SCREAMING_SNAKE_CASE ).to(model_output.device ) _snake_case = (self._get_variance(_SCREAMING_SNAKE_CASE , predicted_variance=_SCREAMING_SNAKE_CASE ) ** 0.5) * noise _snake_case = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 1.0 , ) -> Tuple: super().__init__() _snake_case = bit_scale _snake_case = ( ddim_bit_scheduler_step if isinstance(UpperCAmelCase , UpperCAmelCase ) else ddpm_bit_scheduler_step ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) @torch.no_grad() def __call__(self , UpperCAmelCase = 256 , UpperCAmelCase = 256 , UpperCAmelCase = 50 , UpperCAmelCase = None , UpperCAmelCase = 1 , UpperCAmelCase = "pil" , UpperCAmelCase = True , **UpperCAmelCase , ) -> Union[Tuple, ImagePipelineOutput]: _snake_case = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=UpperCAmelCase , ) _snake_case = decimal_to_bits(UpperCAmelCase ) * self.bit_scale _snake_case = latents.to(self.device ) self.scheduler.set_timesteps(UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _snake_case = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _snake_case = self.scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ).prev_sample _snake_case = bits_to_decimal(UpperCAmelCase ) if output_type == "pil": _snake_case = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase )
341
0
def _A ( SCREAMING_SNAKE_CASE : str ): """simple docstring""" a__ : Optional[int] =[int(SCREAMING_SNAKE_CASE ) for i in ip_va_address.split("." ) if i.isdigit()] return len(SCREAMING_SNAKE_CASE ) == 4 and all(0 <= int(SCREAMING_SNAKE_CASE ) <= 254 for octet in octets ) if __name__ == "__main__": UpperCAmelCase : Optional[int] = input().strip() UpperCAmelCase : Optional[int] = """valid""" if is_ip_va_address_valid(ip) else """invalid""" print(F"""{ip} is a {valid_or_invalid} IP v4 address.""")
95
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 10**9 ): _snake_case = 1 _snake_case = 2 _snake_case = 0 _snake_case = 0 _snake_case = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value _snake_case = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f'''{solution() = }''')
341
0
"""simple docstring""" import os def _snake_case ( ): _lowerCamelCase : Dict = os.path.dirname(os.path.realpath(lowercase__ ) ) _lowerCamelCase : int = os.path.join(lowercase__ , 'triangle.txt' ) with open(lowercase__ ) as f: _lowerCamelCase : str = f.readlines() _lowerCamelCase : Optional[Any] = [] for line in triangle: _lowerCamelCase : Optional[int] = [] for number in line.strip().split(' ' ): numbers_from_line.append(int(lowercase__ ) ) a.append(lowercase__ ) for i in range(1 , len(lowercase__ ) ): for j in range(len(a[i] ) ): _lowerCamelCase : str = a[i - 1][j] if j != len(a[i - 1] ) else 0 _lowerCamelCase : Any = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(lowercase__ , lowercase__ ) return max(a[-1] ) if __name__ == "__main__": print(solution())
96
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "deberta-v2" def __init__(self , UpperCAmelCase=128100 , UpperCAmelCase=1536 , UpperCAmelCase=24 , UpperCAmelCase=24 , UpperCAmelCase=6144 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-7 , UpperCAmelCase=False , UpperCAmelCase=-1 , UpperCAmelCase=0 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=0 , UpperCAmelCase="gelu" , **UpperCAmelCase , ) -> List[str]: super().__init__(**UpperCAmelCase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = relative_attention _snake_case = max_relative_positions _snake_case = pad_token_id _snake_case = position_biased_input # Backwards compatibility if type(UpperCAmelCase ) == str: _snake_case = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case = pos_att_type _snake_case = vocab_size _snake_case = layer_norm_eps _snake_case = kwargs.get("""pooler_hidden_size""" , UpperCAmelCase ) _snake_case = pooler_dropout _snake_case = pooler_hidden_act class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def lowercase (self ) -> int: return 12 def lowercase (self , UpperCAmelCase , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = 3 , UpperCAmelCase = 40 , UpperCAmelCase = 40 , UpperCAmelCase = None , ) -> Mapping[str, Any]: _snake_case = super().generate_dummy_inputs(preprocessor=UpperCAmelCase , framework=UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): """simple docstring""" @slow def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = TFAutoModelForSeqaSeqLM.from_pretrained('''google/mt5-small''' ) UpperCamelCase__ :Union[str, Any] = AutoTokenizer.from_pretrained('''google/mt5-small''' ) UpperCamelCase__ :List[Any] = tokenizer('''Hello there''' , return_tensors='''tf''' ).input_ids UpperCamelCase__ :Union[str, Any] = tokenizer('''Hi I am''' , return_tensors='''tf''' ).input_ids UpperCamelCase__ :Union[str, Any] = model(UpperCamelCase_ , labels=UpperCamelCase_ ).loss UpperCamelCase__ :Union[str, Any] = -tf.math.reduce_mean(UpperCamelCase_ ).numpy() UpperCamelCase__ :Tuple = -21.228168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
97
'''simple docstring''' __lowerCAmelCase = [ (1_000, 'M'), (900, 'CM'), (500, 'D'), (400, 'CD'), (100, 'C'), (90, 'XC'), (50, 'L'), (40, 'XL'), (10, 'X'), (9, 'IX'), (5, 'V'), (4, 'IV'), (1, 'I'), ] def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1000} _snake_case = 0 _snake_case = 0 while place < len(_SCREAMING_SNAKE_CASE ): if (place + 1 < len(_SCREAMING_SNAKE_CASE )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = [] for arabic, roman in ROMAN: ((_snake_case), (_snake_case)) = divmod(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) result.append(roman * factor ) if number == 0: break return "".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( *lowerCamelCase__ : List[Any] ,**lowerCamelCase__ : Union[str, Any] ): pass def a_ ( lowerCamelCase ): UpperCAmelCase__ = hashlib.mda(image.tobytes() ) return m.hexdigest()[:1_0] def a_ ( lowerCamelCase ): UpperCAmelCase__ = np.array(lowerCamelCase ) UpperCAmelCase__ = npimg.shape return {"hash": hashimage(lowerCamelCase ), "shape": shape} @is_pipeline_test @require_vision @require_torch class snake_case ( unittest.TestCase ): """simple docstring""" snake_case__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) snake_case__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : str ): UpperCAmelCase__ = MaskGenerationPipeline(model=lowerCamelCase__ ,image_processor=lowerCamelCase__ ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : str ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def __lowerCAmelCase ( self : Optional[Any] ): pass @slow @require_torch def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = pipeline('mask-generation' ,model='facebook/sam-vit-huge' ) UpperCAmelCase__ = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(lowerCamelCase__ ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(lowerCamelCase__ ,decimals=4 ) ,[ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] ,) # fmt: on @require_torch @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'facebook/sam-vit-huge' UpperCAmelCase__ = pipeline('mask-generation' ,model=lowerCamelCase__ ) UpperCAmelCase__ = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' ,pred_iou_thresh=1 ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(lowerCamelCase__ ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(lowerCamelCase__ ,decimals=4 ) ,[ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] ,)
98
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __lowerCAmelCase = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['PerceiverFeatureExtractor'] __lowerCAmelCase = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
import os import sys lowercase : List[str] = os.path.join(os.path.dirname(__file__), """src""") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) lowercase : Optional[int] = [ """torch""", """numpy""", """tokenizers""", """filelock""", """requests""", """tqdm""", """regex""", """sentencepiece""", """sacremoses""", """importlib_metadata""", """huggingface_hub""", ] @add_start_docstrings(AutoConfig.__doc__ ) def A_ ( *A__ , **A__ ) -> str: return AutoConfig.from_pretrained(*A__ , **A__ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def A_ ( *A__ , **A__ ) -> Tuple: return AutoTokenizer.from_pretrained(*A__ , **A__ ) @add_start_docstrings(AutoModel.__doc__ ) def A_ ( *A__ , **A__ ) -> Any: return AutoModel.from_pretrained(*A__ , **A__ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def A_ ( *A__ , **A__ ) -> Tuple: return AutoModelForCausalLM.from_pretrained(*A__ , **A__ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def A_ ( *A__ , **A__ ) -> Optional[Any]: return AutoModelForMaskedLM.from_pretrained(*A__ , **A__ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def A_ ( *A__ , **A__ ) -> Dict: return AutoModelForSequenceClassification.from_pretrained(*A__ , **A__ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def A_ ( *A__ , **A__ ) -> Dict: return AutoModelForQuestionAnswering.from_pretrained(*A__ , **A__ )
99
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __lowerCAmelCase = 'platform' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ): if attention_mask is None: _snake_case = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _snake_case = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _snake_case = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=7 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=99 , UpperCAmelCase=16 , UpperCAmelCase=2 , UpperCAmelCase=4 , UpperCAmelCase=4 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , ) -> Union[str, Any]: _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id _snake_case = initializer_range def lowercase (self ) -> str: _snake_case = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _snake_case = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=UpperCAmelCase , ) _snake_case = prepare_blenderbot_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, inputs_dict def lowercase (self ) -> Dict: _snake_case, _snake_case = self.prepare_config_and_inputs() return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) @require_flax class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = 99 def lowercase (self ) -> Any: _snake_case = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _snake_case = input_ids.shape[0] _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case = self._get_config_and_data() _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = lm_model(input_ids=UpperCAmelCase ) _snake_case = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _snake_case = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _snake_case = lm_model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) _snake_case = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(UpperCAmelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class _lowerCAmelCase ( __snake_case , unittest.TestCase , __snake_case ): '''simple docstring''' lowerCAmelCase_ = True lowerCAmelCase_ = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCAmelCase_ = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def lowercase (self ) -> Any: _snake_case = FlaxBlenderbotModelTester(self ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _snake_case = model_class(UpperCAmelCase ) @jax.jit def encode_jitted(UpperCAmelCase , UpperCAmelCase=None , **UpperCAmelCase ): return model.encode(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase ) with self.subTest("""JIT Enabled""" ): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = model_class(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""] ) _snake_case = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): return model.decode( decoder_input_ids=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , encoder_outputs=UpperCAmelCase , ) with self.subTest("""JIT Enabled""" ): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def lowercase (self ) -> Any: for model_class_name in self.all_model_classes: _snake_case = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _snake_case = np.ones((1, 1) ) * model.config.eos_token_id _snake_case = model(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) @unittest.skipUnless(jax_device != """cpu""" , """3B test too slow on CPU.""" ) @slow def lowercase (self ) -> Dict: _snake_case = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 15, """max_length""": 25} _snake_case = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} _snake_case = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""" , from_pt=UpperCAmelCase ) _snake_case = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) _snake_case = ["""Sam"""] _snake_case = tokenizer(UpperCAmelCase , return_tensors="""jax""" ) _snake_case = model.generate(**UpperCAmelCase , **UpperCAmelCase ) _snake_case = """Sam is a great name. It means \"sun\" in Gaelic.""" _snake_case = tokenizer.batch_decode(UpperCAmelCase , **UpperCAmelCase ) assert generated_txt[0].strip() == tgt_text
341
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __magic_name__ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_=False ): __SCREAMING_SNAKE_CASE = [] # fmt: off # stem: rename_keys.append(("""cls_token""", """vit.embeddings.cls_token""") ) rename_keys.append(("""pos_embed""", """vit.embeddings.position_embeddings""") ) rename_keys.append(("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight""") ) rename_keys.append(("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias""") ) # backbone rename_keys.append(("""patch_embed.backbone.stem.conv.weight""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight""") ) rename_keys.append(("""patch_embed.backbone.stem.norm.weight""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight""") ) rename_keys.append(("""patch_embed.backbone.stem.norm.bias""", """vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias""") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight") ) rename_keys.append((f"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias", f"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"blocks.{i}.norm1.weight", f"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((f"blocks.{i}.norm1.bias", f"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((f"blocks.{i}.attn.proj.weight", f"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((f"blocks.{i}.attn.proj.bias", f"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((f"blocks.{i}.norm2.weight", f"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((f"blocks.{i}.norm2.bias", f"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((f"blocks.{i}.mlp.fc1.weight", f"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((f"blocks.{i}.mlp.fc1.bias", f"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((f"blocks.{i}.mlp.fc2.weight", f"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((f"blocks.{i}.mlp.fc2.bias", f"vit.encoder.layer.{i}.output.dense.bias") ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __SCREAMING_SNAKE_CASE = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) # fmt: on return rename_keys def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=False ): for i in range(config.num_hidden_layers ): if base_model: __SCREAMING_SNAKE_CASE = """""" else: __SCREAMING_SNAKE_CASE = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __SCREAMING_SNAKE_CASE = state_dict.pop(f"blocks.{i}.attn.qkv.weight" ) __SCREAMING_SNAKE_CASE = state_dict.pop(f"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __SCREAMING_SNAKE_CASE = in_proj_weight[ : config.hidden_size, : ] __SCREAMING_SNAKE_CASE = in_proj_bias[: config.hidden_size] __SCREAMING_SNAKE_CASE = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __SCREAMING_SNAKE_CASE = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __SCREAMING_SNAKE_CASE = in_proj_weight[ -config.hidden_size :, : ] __SCREAMING_SNAKE_CASE = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(UpperCamelCase_ , UpperCamelCase_ ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = dct.pop(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = val def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = """http://images.cocodataset.org/val2017/000000039769.jpg""" __SCREAMING_SNAKE_CASE = Image.open(requests.get(UpperCamelCase_ , stream=UpperCamelCase_ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=False ): __SCREAMING_SNAKE_CASE = BitConfig( global_padding="""same""" , layer_type="""bottleneck""" , depths=(3, 4, 9) , out_features=["""stage3"""] , embedding_dynamic_padding=UpperCamelCase_ , ) __SCREAMING_SNAKE_CASE = ViTHybridConfig(backbone_config=UpperCamelCase_ , image_size=384 , num_labels=1000 ) __SCREAMING_SNAKE_CASE = False # load original model from timm __SCREAMING_SNAKE_CASE = timm.create_model(UpperCamelCase_ , pretrained=UpperCamelCase_ ) timm_model.eval() # load state_dict of original model, remove and rename some keys __SCREAMING_SNAKE_CASE = timm_model.state_dict() if base_model: remove_classification_head_(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = create_rename_keys(UpperCamelCase_ , UpperCamelCase_ ) for src, dest in rename_keys: rename_key(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) read_in_q_k_v(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = """huggingface/label-files""" __SCREAMING_SNAKE_CASE = """imagenet-1k-id2label.json""" __SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(UpperCamelCase_ , UpperCamelCase_ , repo_type="""dataset""" ) , """r""" ) ) __SCREAMING_SNAKE_CASE = {int(UpperCamelCase_ ): v for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = idalabel __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": __SCREAMING_SNAKE_CASE = ViTHybridModel(UpperCamelCase_ ).eval() else: __SCREAMING_SNAKE_CASE = ViTHybridForImageClassification(UpperCamelCase_ ).eval() model.load_state_dict(UpperCamelCase_ ) # create image processor __SCREAMING_SNAKE_CASE = create_transform(**resolve_data_config({} , model=UpperCamelCase_ ) ) __SCREAMING_SNAKE_CASE = transform.transforms __SCREAMING_SNAKE_CASE = { """bilinear""": PILImageResampling.BILINEAR, """bicubic""": PILImageResampling.BICUBIC, """nearest""": PILImageResampling.NEAREST, } __SCREAMING_SNAKE_CASE = ViTHybridImageProcessor( do_resize=UpperCamelCase_ , size={"""shortest_edge""": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=UpperCamelCase_ , crop_size={"""height""": timm_transforms[1].size[0], """width""": timm_transforms[1].size[1]} , do_normalize=UpperCamelCase_ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) __SCREAMING_SNAKE_CASE = prepare_img() __SCREAMING_SNAKE_CASE = transform(UpperCamelCase_ ).unsqueeze(0 ) __SCREAMING_SNAKE_CASE = processor(UpperCamelCase_ , return_tensors="""pt""" ).pixel_values # verify pixel values assert torch.allclose(UpperCamelCase_ , UpperCamelCase_ ) # verify logits with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = outputs.logits print("""Predicted class:""" , logits.argmax(-1 ).item() ) if base_model: __SCREAMING_SNAKE_CASE = timm_model.forward_features(UpperCamelCase_ ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(UpperCamelCase_ , outputs.pooler_output , atol=1e-3 ) else: __SCREAMING_SNAKE_CASE = timm_model(UpperCamelCase_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCamelCase_ , outputs.logits , atol=1e-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(UpperCamelCase_ ).mkdir(exist_ok=UpperCamelCase_ ) print(f"Saving model {vit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(UpperCamelCase_ ) print(f"Saving processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(UpperCamelCase_ ) if push_to_hub: print(f"Pushing model and processor to the hub {vit_name}" ) model.push_to_hub(f"ybelkada/{vit_name}" ) processor.push_to_hub(f"ybelkada/{vit_name}" ) if __name__ == "__main__": __magic_name__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) __magic_name__ = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
100
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=2 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=10 , UpperCAmelCase=3 , UpperCAmelCase=32 * 4 , UpperCAmelCase=32 * 6 , UpperCAmelCase=4 , UpperCAmelCase=32 , ) -> Optional[Any]: _snake_case = parent _snake_case = batch_size _snake_case = is_training _snake_case = use_auxiliary_loss _snake_case = num_queries _snake_case = num_channels _snake_case = min_size _snake_case = max_size _snake_case = num_labels _snake_case = mask_feature_size def lowercase (self ) -> str: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( UpperCAmelCase ) _snake_case = torch.ones([self.batch_size, self.min_size, self.max_size] , device=UpperCAmelCase ) _snake_case = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=UpperCAmelCase ) > 0.5 ).float() _snake_case = (torch.rand((self.batch_size, self.num_labels) , device=UpperCAmelCase ) > 0.5).long() _snake_case = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowercase (self ) -> Tuple: return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.prepare_config_and_inputs() _snake_case = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> int: _snake_case = output.encoder_hidden_states _snake_case = output.pixel_decoder_hidden_states _snake_case = output.transformer_decoder_hidden_states self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , config.decoder_config.decoder_layers ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Union[str, Any]: with torch.no_grad(): _snake_case = MaskFormerModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase , output_hidden_states=UpperCAmelCase ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(UpperCAmelCase , UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: _snake_case = MaskFormerForInstanceSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() def comm_check_on_output(UpperCAmelCase ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) _snake_case = model( pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowerCAmelCase_ = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> int: _snake_case = MaskFormerModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def lowercase (self ) -> int: self.config_tester.run_common_tests() def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> Any: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*UpperCAmelCase ) @unittest.skip(reason="""MaskFormer does not use inputs_embeds""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip(reason="""MaskFormer does not have a get_input_embeddings method""" ) def lowercase (self ) -> Optional[int]: pass @unittest.skip(reason="""MaskFormer is not a generative model""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""MaskFormer does not use token embeddings""" ) def lowercase (self ) -> Optional[int]: pass @require_torch_multi_gpu @unittest.skip( reason="""MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> Tuple: pass def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @slow def lowercase (self ) -> int: for model_name in ["facebook/maskformer-swin-small-coco"]: _snake_case = MaskFormerModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = (self.model_tester.min_size,) * 2 _snake_case = { """pixel_values""": torch.randn((2, 3, *size) , device=UpperCAmelCase ), """mask_labels""": torch.randn((2, 10, *size) , device=UpperCAmelCase ), """class_labels""": torch.zeros(2 , 10 , device=UpperCAmelCase ).long(), } _snake_case = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertTrue(outputs.attentions is not None ) def lowercase (self ) -> Tuple: if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ).loss loss.backward() def lowercase (self ) -> List[str]: # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = True _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) _snake_case = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _snake_case = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _snake_case = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _snake_case = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=UpperCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1E-4 def __SCREAMING_SNAKE_CASE ( ): _snake_case = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase (self ) -> Optional[int]: return ( MaskFormerImageProcessor.from_pretrained("""facebook/maskformer-swin-small-coco""" ) if is_vision_available() else None ) def lowercase (self ) -> str: _snake_case = MaskFormerModel.from_pretrained("""facebook/maskformer-swin-small-coco""" ).to(UpperCAmelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[str]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [ [1.6_5_1_2e0_0, -5.2_5_7_2e0_0, -3.3_5_1_9e0_0], [3.6_1_6_9e-0_2, -5.9_0_2_5e0_0, -2.9_3_1_3e0_0], [1.0_7_6_6e-0_4, -7.7_6_3_0e0_0, -5.1_2_6_3e0_0], ] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[Any]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-resnet101-coco-stuff""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> Tuple: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) _snake_case = inputs["""pixel_values"""].to(UpperCAmelCase ) _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""mask_labels"""]] _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""class_labels"""]] with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None )
341
0
import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if isinstance(lowerCAmelCase__ , collections.abc.Iterable ): return x return (x, x) @require_flax class lowercase : def A__ ( self ,A__ ,A__): pass def A__ ( self): pass def A__ ( self): pass def A__ ( self ,A__ ,A__ ,A__): lowercase = np.abs((a - b)).max() self.assertLessEqual(A__ ,A__ ,f'Difference between torch and flax is {diff} (>= {tol}).') def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__=None ,**A__): lowercase = VisionTextDualEncoderConfig.from_vision_text_configs(A__ ,A__) lowercase = FlaxVisionTextDualEncoderModel(A__) lowercase = model(input_ids=A__ ,pixel_values=A__ ,attention_mask=A__) self.assertEqual(output['''text_embeds'''].shape ,(input_ids.shape[0], config.projection_dim)) self.assertEqual(output['''image_embeds'''].shape ,(pixel_values.shape[0], config.projection_dim)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__=None ,**A__): lowercase , lowercase = self.get_vision_text_model(A__ ,A__) lowercase = {'''vision_model''': vision_model, '''text_model''': text_model} lowercase = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**A__) lowercase = model(input_ids=A__ ,pixel_values=A__ ,attention_mask=A__) self.assertEqual(output['''text_embeds'''].shape ,(input_ids.shape[0], model.config.projection_dim)) self.assertEqual(output['''image_embeds'''].shape ,(pixel_values.shape[0], model.config.projection_dim)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__=None ,**A__): lowercase , lowercase = self.get_vision_text_model(A__ ,A__) lowercase = {'''vision_model''': vision_model, '''text_model''': text_model} lowercase = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**A__) lowercase = model(input_ids=A__ ,pixel_values=A__ ,attention_mask=A__) lowercase = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A__) lowercase = FlaxVisionTextDualEncoderModel.from_pretrained(A__) lowercase = model(input_ids=A__ ,pixel_values=A__ ,attention_mask=A__) lowercase = after_output[0] lowercase = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(A__ ,1E-3) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__=None ,**A__): lowercase , lowercase = self.get_vision_text_model(A__ ,A__) lowercase = {'''vision_model''': vision_model, '''text_model''': text_model} lowercase = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**A__) lowercase = model( input_ids=A__ ,pixel_values=A__ ,attention_mask=A__ ,output_attentions=A__) lowercase = output.vision_model_output.attentions self.assertEqual(len(A__) ,vision_config.num_hidden_layers) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) lowercase = to_atuple(vision_model.config.image_size) lowercase = to_atuple(vision_model.config.patch_size) lowercase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) lowercase = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] ,(vision_config.num_attention_heads, seq_len, seq_len)) lowercase = output.text_model_output.attentions self.assertEqual(len(A__) ,text_config.num_hidden_layers) self.assertEqual( text_attentions[0].shape[-3:] ,(text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) ,) def A__ ( self ,A__ ,A__ ,A__): pt_model.to(A__) pt_model.eval() # prepare inputs lowercase = inputs_dict lowercase = {k: torch.tensor(v.tolist()) for k, v in flax_inputs.items()} with torch.no_grad(): lowercase = pt_model(**A__).to_tuple() lowercase = fx_model(**A__).to_tuple() self.assertEqual(len(A__) ,len(A__) ,'''Output lengths differ between Flax and PyTorch''') for fx_output, pt_output in zip(fx_outputs[:4] ,pt_outputs[:4]): self.assert_almost_equals(A__ ,pt_output.numpy() ,4E-2) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(A__) lowercase = FlaxVisionTextDualEncoderModel.from_pretrained(A__ ,from_pt=A__) lowercase = fx_model_loaded(**A__).to_tuple() self.assertEqual(len(A__) ,len(A__) ,'''Output lengths differ between Flax and PyTorch''') for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] ,pt_outputs[:4]): self.assert_almost_equals(A__ ,pt_output.numpy() ,4E-2) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(A__) lowercase = VisionTextDualEncoderModel.from_pretrained(A__ ,from_flax=A__) pt_model_loaded.to(A__) pt_model_loaded.eval() with torch.no_grad(): lowercase = pt_model_loaded(**A__).to_tuple() self.assertEqual(len(A__) ,len(A__) ,'''Output lengths differ between Flax and PyTorch''') for fx_output, pt_output_loaded in zip(fx_outputs[:4] ,pt_outputs_loaded[:4]): self.assert_almost_equals(A__ ,pt_output_loaded.numpy() ,4E-2) def A__ ( self ,A__ ,A__ ,A__): lowercase = VisionTextDualEncoderConfig.from_vision_text_configs(A__ ,A__) lowercase = VisionTextDualEncoderModel(A__) lowercase = FlaxVisionTextDualEncoderModel(A__) lowercase = convert_pytorch_state_dict_to_flax(pt_model.state_dict() ,A__) lowercase = fx_state self.check_pt_flax_equivalence(A__ ,A__ ,A__) def A__ ( self ,A__ ,A__ ,A__): lowercase = VisionTextDualEncoderConfig.from_vision_text_configs(A__ ,A__) lowercase = VisionTextDualEncoderModel(A__) lowercase = FlaxVisionTextDualEncoderModel(A__) lowercase = load_flax_weights_in_pytorch_model(A__ ,fx_model.params) self.check_pt_flax_equivalence(A__ ,A__ ,A__) def A__ ( self): lowercase = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**A__) def A__ ( self): lowercase = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**A__) def A__ ( self): lowercase = self.prepare_config_and_inputs() self.check_save_load(**A__) def A__ ( self): lowercase = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**A__) @is_pt_flax_cross_test def A__ ( self): lowercase = self.prepare_config_and_inputs() lowercase = config_inputs_dict.pop('''vision_config''') lowercase = config_inputs_dict.pop('''text_config''') lowercase = config_inputs_dict self.check_equivalence_pt_to_flax(A__ ,A__ ,A__) self.check_equivalence_flax_to_pt(A__ ,A__ ,A__) @slow def A__ ( self): lowercase , lowercase = self.get_pretrained_model_and_inputs() lowercase = model_a(**A__) lowercase = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(A__) lowercase = FlaxVisionTextDualEncoderModel.from_pretrained(A__) lowercase = model_a(**A__) lowercase = after_outputs[0] lowercase = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(A__ ,1E-5) @require_flax class lowercase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): def A__ ( self): lowercase = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-vit''' ,'''hf-internal-testing/tiny-bert''' ,vision_from_pt=A__ ,text_from_pt=A__ ,) lowercase = 1_3 lowercase = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) lowercase = ids_tensor([batch_size, 4] ,model.config.text_config.vocab_size) lowercase = random_attention_mask([batch_size, 4]) lowercase = {'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def A__ ( self ,A__ ,A__): lowercase = FlaxViTModel(A__) lowercase = FlaxBertModel(A__) return vision_model, text_model def A__ ( self): lowercase = FlaxViTModelTester(self) lowercase = FlaxBertModelTester(self) lowercase = vit_model_tester.prepare_config_and_inputs() lowercase = bert_model_tester.prepare_config_and_inputs() lowercase , lowercase = vision_config_and_inputs lowercase , lowercase , lowercase , lowercase = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class lowercase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): def A__ ( self): lowercase = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( '''hf-internal-testing/tiny-random-clip''' ,'''hf-internal-testing/tiny-bert''' ,vision_from_pt=A__ ,text_from_pt=A__ ,) lowercase = 1_3 lowercase = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ]) lowercase = ids_tensor([batch_size, 4] ,model.config.text_config.vocab_size) lowercase = random_attention_mask([batch_size, 4]) lowercase = {'''pixel_values''': pixel_values, '''input_ids''': input_ids, '''attention_mask''': attention_mask} return model, inputs def A__ ( self ,A__ ,A__): lowercase = FlaxCLIPVisionModel(A__) lowercase = FlaxBertModel(A__) return vision_model, text_model def A__ ( self): lowercase = FlaxCLIPVisionModelTester(self) lowercase = FlaxBertModelTester(self) lowercase = clip_model_tester.prepare_config_and_inputs() lowercase = bert_model_tester.prepare_config_and_inputs() lowercase , lowercase = vision_config_and_inputs lowercase , lowercase , lowercase , lowercase = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class lowercase ( unittest.TestCase ): @slow def A__ ( self): lowercase = FlaxVisionTextDualEncoderModel.from_pretrained('''clip-italian/clip-italian''' ,logit_scale_init_value=1.0) lowercase = VisionTextDualEncoderProcessor.from_pretrained('''clip-italian/clip-italian''') lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') lowercase = processor( text=['''una foto di un gatto''', '''una foto di un cane'''] ,images=A__ ,padding=A__ ,return_tensors='''np''') lowercase = model(**A__) # verify the logits self.assertEqual(outputs.logits_per_image.shape ,(inputs.pixel_values.shape[0], inputs.input_ids.shape[0])) self.assertEqual( outputs.logits_per_text.shape ,(inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) ,) lowercase = np.array([[1.2284727, 0.3104122]]) self.assertTrue(np.allclose(outputs.logits_per_image ,A__ ,atol=1E-3))
101
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self , UpperCAmelCase ) -> Union[str, Any]: for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): _snake_case = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sgugger/tiny-distilbert-classification""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , only_pretrain_model=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , torchscript=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , fpaa=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) # set architectures equal to `None` _snake_case = None _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def lowercase (self ) -> Tuple: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Any: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> int: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> str: _snake_case = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , save_to_csv=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCAmelCase , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(UpperCAmelCase , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(UpperCAmelCase , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(UpperCAmelCase , """train_time.csv""" ) , env_info_csv_file=os.path.join(UpperCAmelCase , """env.csv""" ) , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """env.csv""" ) ).exists() ) def lowercase (self ) -> int: _snake_case = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCAmelCase ): self.assertTrue(hasattr(UpperCAmelCase , """sequential""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """cumulative""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """current""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCAmelCase , """log.txt""" ) , log_print=UpperCAmelCase , trace_memory_line_by_line=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """log.txt""" ) ).exists() )
341
0
"""simple docstring""" import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE : str = """▁""" SCREAMING_SNAKE_CASE : List[str] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =BigBirdTokenizer lowerCamelCase__ =BigBirdTokenizerFast lowerCamelCase__ =True lowerCamelCase__ =True def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().setUp() __snake_case : List[Any] = self.tokenizer_class(a_ , keep_accents=a_ ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = '''<s>''' __snake_case : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''[MASK]''' ) self.assertEqual(len(a_ ) , 10_04 ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' if not self.test_rust_tokenizer: return __snake_case : str = self.get_tokenizer() __snake_case : Dict = self.get_rust_tokenizer() __snake_case : Dict = '''I was born in 92000, and this is falsé.''' __snake_case : int = tokenizer.tokenize(a_ ) __snake_case : str = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_ , a_ ) __snake_case : Tuple = tokenizer.encode(a_ , add_special_tokens=a_ ) __snake_case : Tuple = rust_tokenizer.encode(a_ , add_special_tokens=a_ ) self.assertListEqual(a_ , a_ ) __snake_case : Optional[Any] = self.get_rust_tokenizer() __snake_case : Optional[int] = tokenizer.encode(a_ ) __snake_case : Dict = rust_tokenizer.encode(a_ ) self.assertListEqual(a_ , a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = BigBirdTokenizer(a_ , keep_accents=a_ ) __snake_case : Optional[int] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(a_ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a_ ) , [2_85, 46, 10, 1_70, 3_82] , ) __snake_case : Union[str, Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( a_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __snake_case : Tuple = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __snake_case : Optional[Any] = tokenizer.convert_ids_to_tokens(a_ ) self.assertListEqual( a_ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = '''Hello World!''' __snake_case : List[Any] = [65, 1_85_36, 22_60, 1_01, 66] self.assertListEqual(a_ , self.big_tokenizer.encode(a_ ) ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) # fmt: off __snake_case : Optional[int] = [65, 8_71, 4_19, 3_58, 9_46, 9_91, 25_21, 4_52, 3_58, 13_57, 3_87, 77_51, 35_36, 1_12, 9_85, 4_56, 1_26, 8_65, 9_38, 54_00, 57_34, 4_58, 13_68, 4_67, 7_86, 24_62, 52_46, 11_59, 6_33, 8_65, 45_19, 4_57, 5_82, 8_52, 25_57, 4_27, 9_16, 5_08, 4_05, 3_43_24, 4_97, 3_91, 4_08, 1_13_42, 12_44, 3_85, 1_00, 9_38, 9_85, 4_56, 5_74, 3_62, 1_25_97, 32_00, 31_29, 11_72, 66] # noqa: E231 # fmt: on self.assertListEqual(a_ , self.big_tokenizer.encode(a_ ) ) @require_torch @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __snake_case : str = list(self.big_tokenizer.get_vocab().keys() )[:10] __snake_case : Tuple = ''' '''.join(a_ ) __snake_case : Tuple = self.big_tokenizer.encode_plus(a_ , return_tensors='''pt''' , return_token_type_ids=a_ ) __snake_case : List[Any] = self.big_tokenizer.batch_encode_plus( [sequence + ''' ''' + sequence] , return_tensors='''pt''' , return_token_type_ids=a_ ) __snake_case : Optional[int] = BigBirdConfig(attention_type='''original_full''' ) __snake_case : str = BigBirdModel(a_ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**a_ ) model(**a_ ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) __snake_case : Any = tokenizer.decode(tokenizer('''Paris is the [MASK].''' ).input_ids ) self.assertTrue(decoded_text == '''[CLS] Paris is the[MASK].[SEP]''' ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Tuple = {'''input_ids''': [[65, 3_92_86, 4_58, 3_63_35, 20_01, 4_56, 1_30_73, 1_32_66, 4_55, 1_13, 77_46, 17_41, 1_11_57, 3_91, 1_30_73, 1_32_66, 4_55, 1_13, 39_67, 3_54_12, 1_13, 49_36, 1_09, 38_70, 23_77, 1_13, 3_00_84, 4_57_20, 4_58, 1_34, 1_74_96, 1_12, 5_03, 1_16_72, 1_13, 1_18, 1_12, 56_65, 1_33_47, 3_86_87, 1_12, 14_96, 3_13_89, 1_12, 32_68, 4_72_64, 1_34, 9_62, 1_12, 1_63_77, 80_35, 2_31_30, 4_30, 1_21_69, 1_55_18, 2_85_92, 4_58, 1_46, 4_16_97, 1_09, 3_91, 1_21_69, 1_55_18, 1_66_89, 4_58, 1_46, 4_13_58, 1_09, 4_52, 7_26, 40_34, 1_11, 7_63, 3_54_12, 50_82, 3_88, 19_03, 1_11, 90_51, 3_91, 28_70, 4_89_18, 19_00, 11_23, 5_50, 9_98, 1_12, 95_86, 1_59_85, 4_55, 3_91, 4_10, 2_29_55, 3_76_36, 1_14, 66], [65, 4_48, 1_74_96, 4_19, 36_63, 3_85, 7_63, 1_13, 2_75_33, 28_70, 32_83, 1_30_43, 16_39, 2_47_13, 5_23, 6_56, 2_40_13, 1_85_50, 25_21, 5_17, 2_70_14, 2_12_44, 4_20, 12_12, 14_65, 3_91, 9_27, 48_33, 3_88, 5_78, 1_17_86, 1_14, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 4_84, 21_69, 76_87, 2_19_32, 1_81_46, 7_26, 3_63, 1_70_32, 33_91, 1_14, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name='''google/bigbird-roberta-base''' , revision='''215c99f1600e06f83acce68422f2035b2b5c3510''' , )
102
'''simple docstring''' from __future__ import annotations def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): if len(_SCREAMING_SNAKE_CASE ) == 0: return [] _snake_case, _snake_case = min(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) _snake_case = int(max_value - min_value ) + 1 _snake_case = [[] for _ in range(_SCREAMING_SNAKE_CASE )] for i in my_list: buckets[int(i - min_value )].append(_SCREAMING_SNAKE_CASE ) return [v for bucket in buckets for v in sorted(_SCREAMING_SNAKE_CASE )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
341
0
import math import random from typing import Any from .hill_climbing import SearchProblem def UpperCamelCase( __UpperCamelCase : Optional[Any] ,__UpperCamelCase : bool = True ,__UpperCamelCase : float = math.inf ,__UpperCamelCase : float = -math.inf ,__UpperCamelCase : float = math.inf ,__UpperCamelCase : float = -math.inf ,__UpperCamelCase : bool = False ,__UpperCamelCase : float = 100 ,__UpperCamelCase : float = 0.0_1 ,__UpperCamelCase : float = 1 ,): lowerCAmelCase_ : str = False lowerCAmelCase_ : Optional[Any] = search_prob lowerCAmelCase_ : int = start_temperate lowerCAmelCase_ : Any = [] lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Union[str, Any] = None while not search_end: lowerCAmelCase_ : int = current_state.score() if best_state is None or current_score > best_state.score(): lowerCAmelCase_ : Optional[int] = current_state scores.append(__UpperCamelCase ) iterations += 1 lowerCAmelCase_ : Tuple = None lowerCAmelCase_ : Union[str, Any] = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to lowerCAmelCase_ : Any = random.randint(0 ,len(__UpperCamelCase ) - 1 ) # picking a random neighbor lowerCAmelCase_ : Optional[int] = neighbors.pop(__UpperCamelCase ) lowerCAmelCase_ : Any = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: lowerCAmelCase_ : Optional[Any] = change * -1 # in case we are finding minimum if change > 0: # improves the solution lowerCAmelCase_ : Union[str, Any] = picked_neighbor else: lowerCAmelCase_ : Optional[int] = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability lowerCAmelCase_ : Optional[int] = picked_neighbor lowerCAmelCase_ : int = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor lowerCAmelCase_ : Optional[int] = True else: lowerCAmelCase_ : Optional[Any] = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(__UpperCamelCase ) ,__UpperCamelCase ) plt.xlabel('''Iterations''' ) plt.ylabel('''Function values''' ) plt.show() return best_state if __name__ == "__main__": def UpperCamelCase( __UpperCamelCase : Dict ,__UpperCamelCase : str ): return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) A__ : str = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) A__ : Tuple = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) # starting the problem with initial coordinates (12, 47) A__ : Union[str, Any] = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) A__ : Tuple = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) def UpperCamelCase( __UpperCamelCase : Optional[Any] ,__UpperCamelCase : Dict ): return (3 * x**2) - (6 * y) A__ : Optional[int] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) A__ : Optional[Any] = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F'''{local_min.score()}''' ) A__ : Union[str, Any] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) A__ : Any = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F'''{local_min.score()}''' )
103
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: if not conversation_id: _snake_case = uuid.uuida() if past_user_inputs is None: _snake_case = [] if generated_responses is None: _snake_case = [] _snake_case = conversation_id _snake_case = past_user_inputs _snake_case = generated_responses _snake_case = text def __eq__(self , UpperCAmelCase ) -> Dict: if not isinstance(UpperCAmelCase , UpperCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = False ) -> int: if self.new_user_input: if overwrite: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ f"""with: \"{text}\".""" ) _snake_case = text else: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ f"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: _snake_case = text def lowercase (self ) -> int: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _snake_case = None def lowercase (self , UpperCAmelCase ) -> Any: self.generated_responses.append(UpperCAmelCase ) def lowercase (self ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__(self ) -> Optional[int]: _snake_case = f"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): _snake_case = """user""" if is_user else """bot""" output += f"""{name} >> {text} \n""" return output @add_end_docstrings( __snake_case , r"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , *UpperCAmelCase , **UpperCAmelCase ) -> Union[str, Any]: super().__init__(*UpperCAmelCase , **UpperCAmelCase ) if self.tokenizer.pad_token_id is None: _snake_case = self.tokenizer.eos_token def lowercase (self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase ) -> Dict: _snake_case = {} _snake_case = {} _snake_case = {} if min_length_for_response is not None: _snake_case = min_length_for_response if minimum_tokens is not None: _snake_case = minimum_tokens if "max_length" in generate_kwargs: _snake_case = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _snake_case = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__(self , UpperCAmelCase , UpperCAmelCase=0 , **UpperCAmelCase ) -> Union[str, Any]: _snake_case = super().__call__(UpperCAmelCase , num_workers=UpperCAmelCase , **UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) and len(UpperCAmelCase ) == 1: return outputs[0] return outputs def lowercase (self , UpperCAmelCase , UpperCAmelCase=32 ) -> Dict[str, Any]: if not isinstance(UpperCAmelCase , UpperCAmelCase ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( f"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): _snake_case = self.tokenizer._build_conversation_input_ids(UpperCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _snake_case = self._legacy_parse_and_tokenize(UpperCAmelCase ) if self.framework == "pt": _snake_case = torch.LongTensor([input_ids] ) elif self.framework == "tf": _snake_case = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=10 , **UpperCAmelCase ) -> Optional[int]: _snake_case = generate_kwargs.get("""max_length""" , self.model.config.max_length ) _snake_case = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(f"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) _snake_case = max_length - minimum_tokens _snake_case = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: _snake_case = model_inputs["""attention_mask"""][:, -trim:] _snake_case = model_inputs.pop("""conversation""" ) _snake_case = max_length _snake_case = self.model.generate(**UpperCAmelCase , **UpperCAmelCase ) if self.model.config.is_encoder_decoder: _snake_case = 1 else: _snake_case = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=True ) -> List[str]: _snake_case = model_outputs["""output_ids"""] _snake_case = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCAmelCase , clean_up_tokenization_spaces=UpperCAmelCase , ) _snake_case = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCAmelCase ) return conversation def lowercase (self , UpperCAmelCase ) -> Dict: _snake_case = self.tokenizer.eos_token_id _snake_case = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) ) if len(UpperCAmelCase ) > self.tokenizer.model_max_length: _snake_case = input_ids[-self.tokenizer.model_max_length :] return input_ids
341
0
'''simple docstring''' from __future__ import annotations def _A ( A__ ): """simple docstring""" __lowercase = len(A__ ) # We need to create solution object to save path. __lowercase = [[0 for _ in range(A__ )] for _ in range(A__ )] __lowercase = run_maze(A__ , 0 , 0 , A__ ) if solved: print('''\n'''.join(str(A__ ) for row in solutions ) ) else: print('''No solution exists!''' ) return solved def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = len(A__ ) # Final check point. if i == j == (size - 1): __lowercase = 1 return True __lowercase = (not i < 0) and (not j < 0) # Check lower bounds __lowercase = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. __lowercase = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited __lowercase = 1 # check for directions if ( run_maze(A__ , i + 1 , A__ , A__ ) or run_maze(A__ , A__ , j + 1 , A__ ) or run_maze(A__ , i - 1 , A__ , A__ ) or run_maze(A__ , A__ , j - 1 , A__ ) ): return True __lowercase = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
104
'''simple docstring''' from math import factorial, radians def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 18 , _SCREAMING_SNAKE_CASE = 10 ): _snake_case = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians _snake_case = radians(_SCREAMING_SNAKE_CASE ) _snake_case = angle_in_radians _snake_case = 3 _snake_case = -1 for _ in range(_SCREAMING_SNAKE_CASE ): result += (b * (angle_in_radians**a)) / factorial(_SCREAMING_SNAKE_CASE ) _snake_case = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __import__('doctest').testmod()
341
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) a : Tuple = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys a : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
105
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __lowerCAmelCase = '\\n@inproceedings{popovic-2015-chrf,\n title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation",\n month = sep,\n year = "2015",\n address = "Lisbon, Portugal",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W15-3049",\n doi = "10.18653/v1/W15-3049",\n pages = "392--395",\n}\n@inproceedings{popovic-2017-chrf,\n title = "chr{F}++: words helping character n-grams",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Second Conference on Machine Translation",\n month = sep,\n year = "2017",\n address = "Copenhagen, Denmark",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W17-4770",\n doi = "10.18653/v1/W17-4770",\n pages = "612--618",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' __lowerCAmelCase = '\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n' __lowerCAmelCase = '\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n \'score\' (float): The chrF (chrF++) score,\n \'char_order\' (int): The character n-gram order,\n \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n \'beta\' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Tuple: if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/mjpost/sacreBLEU#chrf--chrf""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#chrf--chrf"""] , reference_urls=[ """https://github.com/m-popovic/chrF""", ] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = CHRF.CHAR_ORDER , UpperCAmelCase = CHRF.WORD_ORDER , UpperCAmelCase = CHRF.BETA , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , ) -> int: _snake_case = len(references[0] ) if any(len(UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) _snake_case = [[refs[i] for refs in references] for i in range(UpperCAmelCase )] _snake_case = CHRF(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _snake_case = sb_chrf.corpus_score(UpperCAmelCase , UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
341
0
"""simple docstring""" import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets __UpperCamelCase : Any = datasets.logging.get_logger(__name__) __UpperCamelCase : Tuple = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' __UpperCamelCase : Dict = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' __UpperCamelCase : Tuple = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_=False , A_=False , A_=True , A_=False , A_="dummy_doc" ): lowerCAmelCase__ : Dict = {doc: key_lines} lowerCAmelCase__ : List[str] = {doc: sys_lines} lowerCAmelCase__ : Optional[int] = {} lowerCAmelCase__ : Union[str, Any] = 0 lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Dict = 0 lowerCAmelCase__ : Any = 0 lowerCAmelCase__ : Union[str, Any] = 0 lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ ,lowerCAmelCase__ : str = reader.get_doc_mentions(A_ , key_doc_lines[doc] , A_ ) key_singletons_num += singletons_num if NP_only or min_span: lowerCAmelCase__ : int = reader.set_annotated_parse_trees(A_ , key_doc_lines[doc] , A_ , A_ ) lowerCAmelCase__ ,lowerCAmelCase__ : Optional[Any] = reader.get_doc_mentions(A_ , sys_doc_lines[doc] , A_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowerCAmelCase__ : List[str] = reader.set_annotated_parse_trees(A_ , key_doc_lines[doc] , A_ , A_ ) if remove_nested: lowerCAmelCase__ ,lowerCAmelCase__ : int = reader.remove_nested_coref_mentions(A_ , A_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowerCAmelCase__ ,lowerCAmelCase__ : str = reader.remove_nested_coref_mentions(A_ , A_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowerCAmelCase__ : Optional[Any] = reader.get_mention_assignments(A_ , A_ ) lowerCAmelCase__ : Optional[Any] = reader.get_mention_assignments(A_ , A_ ) lowerCAmelCase__ : Optional[int] = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ , A_ , A_ , A_ ): lowerCAmelCase__ : List[str] = get_coref_infos(A_ , A_ , A_ , A_ , A_ , A_ ) lowerCAmelCase__ : int = {} lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : Optional[Any] = 0 for name, metric in metrics: lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ : Any = evaluator.evaluate_documents(A_ , A_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(10 ) , f'Recall: {recall * 1_00:.2f}' , f' Precision: {precision * 1_00:.2f}' , f' F1: {fa * 1_00:.2f}' , ) if conll_subparts_num == 3: lowerCAmelCase__ : str = (conll / 3) * 1_00 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Optional[int] = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowerCAmelCase__ : List[str] = line.split()[5] if not parse_col == "-": lowerCAmelCase__ : int = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE ( datasets.Metric ): """simple docstring""" def __lowerCAmelCase ( self : int ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) ,codebase_urls=['''https://github.com/ns-moosavi/coval'''] ,reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] ,) def __lowerCAmelCase ( self : int ,lowercase_ : Any ,lowercase_ : List[str] ,lowercase_ : List[str]=True ,lowercase_ : Union[str, Any]=False ,lowercase_ : List[str]=False ,lowercase_ : Optional[int]=False ): lowerCAmelCase__ : List[str] = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowerCAmelCase__ : int = util.check_gold_parse_annotation(lowercase_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowerCAmelCase__ : Dict = evaluate( key_lines=lowercase_ ,sys_lines=lowercase_ ,metrics=lowercase_ ,NP_only=lowercase_ ,remove_nested=lowercase_ ,keep_singletons=lowercase_ ,min_span=lowercase_ ,) return score
106
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCAmelCase = '\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n' __lowerCAmelCase = '\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {\'spearmanr\': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results[\'spearmanr\'])\n -0.7\n >>> print(round(results[\'spearmanr_pvalue\'], 2))\n 0.19\n' __lowerCAmelCase = r'\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Optional[Any]: _snake_case = spearmanr(UpperCAmelCase , UpperCAmelCase ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
341
0
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __lowerCAmelCase : Any = logging.get_logger(__name__) class snake_case__ (enum.Enum ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = 0 SCREAMING_SNAKE_CASE_ : Optional[Any] = 1 @add_end_docstrings(_UpperCamelCase ) class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = """generated""" def __init__( self : Dict , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> List[Any]: super().__init__(*__lowerCamelCase , **__lowerCamelCase ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : List[str]=None , __lowerCamelCase : List[str]=None , __lowerCamelCase : Any=None , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : str=None , **__lowerCamelCase : str , ) -> Tuple: a = {} if truncation is not None: a = truncation a = generate_kwargs a = {} if return_tensors is not None and return_type is None: a = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: a = return_type if clean_up_tokenization_spaces is not None: a = clean_up_tokenization_spaces if stop_sequence is not None: a = self.tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) if len(__lowerCamelCase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) a = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int ) -> Union[str, Any]: return True def __UpperCAmelCase ( self : Optional[int] , *__lowerCamelCase : str , __lowerCamelCase : Optional[int] ) -> Union[str, Any]: a = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , __lowerCamelCase ): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input" ) a = ([prefix + arg for arg in args[0]],) a = True elif isinstance(args[0] , __lowerCamelCase ): a = (prefix + args[0],) a = False else: raise ValueError( f""" `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`""" ) a = self.tokenizer(*__lowerCamelCase , padding=__lowerCamelCase , truncation=__lowerCamelCase , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Any , *__lowerCamelCase : List[Any] , **__lowerCamelCase : int ) -> Optional[int]: a = super().__call__(*__lowerCamelCase , **__lowerCamelCase ) if ( isinstance(args[0] , __lowerCamelCase ) and all(isinstance(__lowerCamelCase , __lowerCamelCase ) for el in args[0] ) and all(len(__lowerCamelCase ) == 1 for res in result ) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : str , __lowerCamelCase : Any , __lowerCamelCase : List[Any]=TruncationStrategy.DO_NOT_TRUNCATE , **__lowerCamelCase : str ) -> Optional[int]: a = self._parse_and_tokenize(__lowerCamelCase , truncation=__lowerCamelCase , **__lowerCamelCase ) return inputs def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Optional[int] , **__lowerCamelCase : Dict ) -> Optional[int]: if self.framework == "pt": a , a = model_inputs["input_ids"].shape elif self.framework == "tf": a , a = tf.shape(model_inputs["input_ids"] ).numpy() a = generate_kwargs.get("min_length" , self.model.config.min_length ) a = generate_kwargs.get("max_length" , self.model.config.max_length ) self.check_inputs(__lowerCamelCase , generate_kwargs["min_length"] , generate_kwargs["max_length"] ) a = self.model.generate(**__lowerCamelCase , **__lowerCamelCase ) a = output_ids.shape[0] if self.framework == "pt": a = output_ids.reshape(__lowerCamelCase , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": a = tf.reshape(__lowerCamelCase , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any]=ReturnType.TEXT , __lowerCamelCase : List[Any]=False ) -> int: a = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: a = {f"""{self.return_name}_token_ids""": output_ids} elif return_type == ReturnType.TEXT: a = { f"""{self.return_name}_text""": self.tokenizer.decode( __lowerCamelCase , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase , ) } records.append(__lowerCamelCase ) return records @add_end_docstrings(_UpperCamelCase ) class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = """summary""" def __call__( self : Dict , *__lowerCamelCase : List[Any] , **__lowerCamelCase : str ) -> Optional[Any]: return super().__call__(*__lowerCamelCase , **__lowerCamelCase ) def __UpperCAmelCase ( self : Tuple , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int ) -> bool: if max_length < min_length: logger.warning(f"""Your min_length={min_length} must be inferior than your max_length={max_length}.""" ) if input_length < max_length: logger.warning( f"""Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is """ "a summarization task, where outputs shorter than the input are typically wanted, you might " f"""consider decreasing max_length manually, e.g. summarizer('...', max_length={input_length//2})""" ) @add_end_docstrings(_UpperCamelCase ) class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = """translation""" def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int ) -> Dict: if input_length > 0.9 * max_length: logger.warning( f"""Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider """ "increasing your max_length manually, e.g. translator('...', max_length=400)" ) return True def __UpperCAmelCase ( self : Union[str, Any] , *__lowerCamelCase : Optional[int] , __lowerCamelCase : Any=TruncationStrategy.DO_NOT_TRUNCATE , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : str=None ) -> Optional[Any]: if getattr(self.tokenizer , "_build_translation_inputs" , __lowerCamelCase ): return self.tokenizer._build_translation_inputs( *__lowerCamelCase , return_tensors=self.framework , truncation=__lowerCamelCase , src_lang=__lowerCamelCase , tgt_lang=__lowerCamelCase ) else: return super()._parse_and_tokenize(*__lowerCamelCase , truncation=__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : Any=None , __lowerCamelCase : str=None , **__lowerCamelCase : Union[str, Any] ) -> List[Any]: a , a , a = super()._sanitize_parameters(**__lowerCamelCase ) if src_lang is not None: a = src_lang if tgt_lang is not None: a = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. a = kwargs.get("task" , self.task ) a = task.split("_" ) if task and len(__lowerCamelCase ) == 4: # translation, XX, to YY a = items[1] a = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , *__lowerCamelCase : int , **__lowerCamelCase : List[str] ) -> Dict: return super().__call__(*__lowerCamelCase , **__lowerCamelCase )
107
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=4 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[2, 2, 3, 2] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase=37 , UpperCAmelCase="gelu" , UpperCAmelCase=10 , UpperCAmelCase=0.02 , UpperCAmelCase=["stage2", "stage3", "stage4"] , UpperCAmelCase=3 , UpperCAmelCase=None , ) -> List[Any]: _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = num_stages _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = intermediate_size _snake_case = hidden_act _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = out_features _snake_case = num_labels _snake_case = scope _snake_case = num_stages def lowercase (self ) -> List[Any]: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def lowercase (self ) -> Tuple: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def lowercase (self ) -> Any: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCAmelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCAmelCase , loss_ignore_index=255 , num_labels=self.num_labels , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: _snake_case = UperNetForSemanticSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowercase (self ) -> Tuple: _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ), ( _snake_case ), ( _snake_case ), ) = config_and_inputs _snake_case = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCAmelCase_ = {"image-segmentation": UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> Optional[Any]: _snake_case = UperNetModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=37 ) def lowercase (self ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase (self ) -> Union[str, Any]: return def lowercase (self ) -> Union[str, Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def lowercase (self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> str: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> int: pass def lowercase (self ) -> List[str]: def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = _config_zero_init(UpperCAmelCase ) _snake_case = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _snake_case = model_class(config=UpperCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def lowercase (self ) -> Optional[Any]: pass @slow def lowercase (self ) -> Tuple: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = UperNetForSemanticSegmentation.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) _snake_case = Image.open(_SCREAMING_SNAKE_CASE ).convert("""RGB""" ) return image @require_torch @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) ) def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
341
0
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' def count_of_possible_combinations(SCREAMING_SNAKE_CASE : int ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' def count_of_possible_combinations_with_dp_array( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[int] ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase : Optional[Any] = sum( count_of_possible_combinations_with_dp_array(target - item , SCREAMING_SNAKE_CASE ) for item in array ) lowerCAmelCase : Any = answer return answer lowerCAmelCase : List[Any] = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' lowerCAmelCase : Dict = [0] * (target + 1) lowerCAmelCase : Dict = 1 for i in range(1 , target + 1 ): for j in range(SCREAMING_SNAKE_CASE ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = 3 lowerCAmelCase__ = 5 lowerCAmelCase__ = [1, 2, 5] print(combination_sum_iv(n, array, target))
108
'''simple docstring''' import argparse from collections import defaultdict def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = f"""class {class_name}(""" _snake_case = f"""{4 * " "}def {test_name}(""" _snake_case = f"""{8 * " "}{correct_line.split()[0]}""" _snake_case = f"""{16 * " "}{correct_line.split()[0]}""" _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = 0 _snake_case = 0 _snake_case = [] for line in lines: if line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and in_func and (line.startswith(_SCREAMING_SNAKE_CASE ) or line.startswith(_SCREAMING_SNAKE_CASE )): _snake_case = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: _snake_case = True if in_class and in_func and in_line: if ")" not in line: continue else: _snake_case = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) _snake_case = _snake_case = _snake_case = _snake_case = False else: new_lines.append(_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: for line in new_lines: f.write(_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): if fail is not None: with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = {l.strip() for l in f.readlines()} else: _snake_case = None with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = defaultdict(_SCREAMING_SNAKE_CASE ) for line in correct_lines: _snake_case, _snake_case, _snake_case, _snake_case = line.split(""";""" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) __lowerCAmelCase = parser.parse_args() main(args.correct_filename, args.fail_filename)
341
0
"""simple docstring""" import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib A: Dict = { "debug": logging.DEBUG, "info": logging.INFO, "warning": logging.WARNING, "error": logging.ERROR, "critical": logging.CRITICAL, } A: str = logging.WARNING def _snake_case ( ): UpperCAmelCase : Union[str, Any] = os.getenv("""DATASETS_VERBOSITY""" , UpperCamelCase ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"Unknown option DATASETS_VERBOSITY={env_level_str}, " F"has to be one of: { ', '.join(log_levels.keys() ) }" ) return _default_log_level def _snake_case ( ): return __name__.split(""".""" )[0] def _snake_case ( ): return logging.getLogger(_get_library_name() ) def _snake_case ( ): # Apply our default configuration to the library root logger. UpperCAmelCase : List[Any] = _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level() ) def _snake_case ( ): UpperCAmelCase : Optional[Any] = _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET ) def _snake_case ( UpperCamelCase : Optional[str] = None ): if name is None: UpperCAmelCase : int = _get_library_name() return logging.getLogger(UpperCamelCase ) def _snake_case ( ): return _get_library_root_logger().getEffectiveLevel() def _snake_case ( UpperCamelCase : int ): _get_library_root_logger().setLevel(UpperCamelCase ) def _snake_case ( ): return set_verbosity(UpperCamelCase ) def _snake_case ( ): return set_verbosity(UpperCamelCase ) def _snake_case ( ): return set_verbosity(UpperCamelCase ) def _snake_case ( ): return set_verbosity(UpperCamelCase ) def _snake_case ( ): UpperCAmelCase : Optional[Any] = False def _snake_case ( ): UpperCAmelCase : List[Any] = True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class SCREAMING_SNAKE_CASE__ : def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> int: # pylint: disable=unused-argument '''simple docstring''' UpperCAmelCase : str = args[0] if args else None def __iter__( self ) -> List[Any]: '''simple docstring''' return iter(self._iterator ) def __getattr__( self , _SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' def empty_fn(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): # pylint: disable=unused-argument return return empty_fn def __enter__( self ) -> int: '''simple docstring''' return self def __exit__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' return A: List[Any] = True class SCREAMING_SNAKE_CASE__ : def __call__( self , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , **_SCREAMING_SNAKE_CASE ) -> Optional[int]: '''simple docstring''' if _tqdm_active and not disable: return tqdm_lib.tqdm(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) else: return EmptyTqdm(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' UpperCAmelCase : str = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm.get_lock() A: List[Any] = _tqdm_cls() def _snake_case ( ): global _tqdm_active return bool(_tqdm_active ) def _snake_case ( ): global _tqdm_active UpperCAmelCase : Optional[Any] = True def _snake_case ( ): global _tqdm_active UpperCAmelCase : Tuple = False
109
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import requests from bsa import BeautifulSoup def lowercase ( A_ = "https://www.worldometers.info/coronavirus" )-> int: '''simple docstring''' a : Any = BeautifulSoup(requests.get(_SCREAMING_SNAKE_CASE ).text , "html.parser" ) a : Tuple = soup.findAll("h1" ) a : int = soup.findAll("div" , {"class": "maincounter-number"} ) keys += soup.findAll("span" , {"class": "panel-title"} ) values += soup.findAll("div" , {"class": "number-table-main"} ) return {key.text.strip(): value.text.strip() for key, value in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )} if __name__ == "__main__": print("""\033[1m""" + """COVID-19 Status of the World""" + """\033[0m\n""") for key, value in world_covidaa_stats().items(): print(f'''{key}\n{value}\n''')
40
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = ["pixel_values"] def __init__(self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: super().__init__(**UpperCAmelCase ) _snake_case = size if size is not None else {"""height""": 256, """width""": 256} _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( UpperCAmelCase , size=(size["""height"""], size["""width"""]) , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(UpperCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> List[Any]: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> PIL.Image.Image: _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = size if size is not None else self.size _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=UpperCAmelCase , size=UpperCAmelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] _snake_case = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _snake_case = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
341
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available lowerCAmelCase_ = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
308
'''simple docstring''' __lowerCAmelCase = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure the supplied data is a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_SCREAMING_SNAKE_CASE ) _snake_case = """""".join(bin(_SCREAMING_SNAKE_CASE )[2:].zfill(8 ) for byte in data ) _snake_case = len(_SCREAMING_SNAKE_CASE ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b"""=""" * ((6 - len(_SCREAMING_SNAKE_CASE ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_SCREAMING_SNAKE_CASE ) % 6) else: _snake_case = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 6 ) ).encode() + padding ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = ( """argument should be a bytes-like object or ASCII string, """ f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_SCREAMING_SNAKE_CASE ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): try: _snake_case = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) _snake_case = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_SCREAMING_SNAKE_CASE ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 8 ) ] return bytes(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" def A ( snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = 1 for i in range(1 , num + 1 ): fact *= i return fact def A ( snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = 0 while number > 0: SCREAMING_SNAKE_CASE__ = number % 10 sum_of_digits += last_digit SCREAMING_SNAKE_CASE__ = number // 10 # Removing the last_digit from the given number return sum_of_digits def A ( snake_case__ = 1_00 ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = factorial(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ = split_and_add(_SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
165
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING lowerCAmelCase__ = logging.get_logger(__name__) @add_end_docstrings(__snake_case ) class a__ ( __snake_case ): """simple docstring""" def __init__( self , **lowercase ) -> int: '''simple docstring''' super().__init__(**lowercase ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , "vision" ) self.check_model_type(lowercase ) def __call__( self , lowercase , lowercase = None , **lowercase , ) -> int: '''simple docstring''' if "text_queries" in kwargs: A__ = kwargs.pop("text_queries" ) if isinstance(lowercase , (str, Image.Image) ): A__ = {"image": image, "candidate_labels": candidate_labels} else: A__ = image A__ = super().__call__(lowercase , **lowercase ) return results def UpperCamelCase ( self , **lowercase ) -> int: '''simple docstring''' A__ = {} if "threshold" in kwargs: A__ = kwargs["threshold"] if "top_k" in kwargs: A__ = kwargs["top_k"] return {}, {}, postprocess_params def UpperCamelCase ( self , lowercase ) -> int: '''simple docstring''' A__ = load_image(inputs["image"] ) A__ = inputs["candidate_labels"] if isinstance(lowercase , lowercase ): A__ = candidate_labels.split("," ) A__ = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(lowercase ): A__ = self.tokenizer(lowercase , return_tensors=self.framework ) A__ = self.image_processor(lowercase , return_tensors=self.framework ) yield { "is_last": i == len(lowercase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def UpperCamelCase ( self , lowercase ) -> Dict: '''simple docstring''' A__ = model_inputs.pop("target_size" ) A__ = model_inputs.pop("candidate_label" ) A__ = model_inputs.pop("is_last" ) A__ = self.model(**lowercase ) A__ = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def UpperCamelCase ( self , lowercase , lowercase=0.1 , lowercase=None ) -> int: '''simple docstring''' A__ = [] for model_output in model_outputs: A__ = model_output["candidate_label"] A__ = BaseModelOutput(lowercase ) A__ = self.image_processor.post_process_object_detection( outputs=lowercase , threshold=lowercase , target_sizes=model_output["target_size"] )[0] for index in outputs["scores"].nonzero(): A__ = outputs["scores"][index].item() A__ = self._get_bounding_box(outputs["boxes"][index][0] ) A__ = {"score": score, "label": label, "box": box} results.append(lowercase ) A__ = sorted(lowercase , key=lambda lowercase : x["score"] , reverse=lowercase ) if top_k: A__ = results[:top_k] return results def UpperCamelCase ( self , lowercase ) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch." ) A__ , A__ , A__ , A__ = box.int().tolist() A__ = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
68
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> Dict: if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError('Cash flows list cannot be empty' ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
42
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __lowerCAmelCase = TypeVar('T') __lowerCAmelCase = Union[List[T], Tuple[T, ...]] __lowerCAmelCase = Union[T, List[T], Dict[str, T]] __lowerCAmelCase = Union[str, bytes, os.PathLike]
341
0
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy a__ = logging.getLogger(__name__) a__ = '''pytorch_model.bin''' @dataclasses.dataclass class UpperCAmelCase_ : """simple docstring""" UpperCAmelCase__ : Any = dataclasses.field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models."} ) UpperCAmelCase__ : str = dataclasses.field( default=__snake_case , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co."} , ) @dataclasses.dataclass class UpperCAmelCase_ : """simple docstring""" UpperCAmelCase__ : Tuple = dataclasses.field(metadata={"help": "A csv or a json file containing the training data."} ) UpperCAmelCase__ : int = dataclasses.field(metadata={"help": "A csv or a json file containing the data to predict on."} ) UpperCAmelCase__ : Union[str, Any] = dataclasses.field( default=__snake_case , metadata={"help": "A csv or a json file containing the validation data."} ) UpperCAmelCase__ : List[Any] = dataclasses.field( default=__snake_case , metadata={"help": "The name of the task to train on."} , ) UpperCAmelCase__ : str = dataclasses.field( default=__snake_case , metadata={"help": "The list of labels for the task."} ) @dataclasses.dataclass class UpperCAmelCase_ : """simple docstring""" UpperCAmelCase__ : Optional[int] = dataclasses.field( metadata={"help": "The output directory where the model predictions and checkpoints will be written."} ) UpperCAmelCase__ : List[str] = dataclasses.field( default="accuracy" , metadata={"help": "The evaluation metric used for the task."} ) UpperCAmelCase__ : Optional[Any] = dataclasses.field( default="no" , metadata={ "help": "The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]" } , ) UpperCAmelCase__ : int = dataclasses.field( default=10 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Dict = dataclasses.field( default=0.0 , metadata={ "help": "How much the specified evaluation metric must improve to satisfy early stopping conditions." } , ) UpperCAmelCase__ : List[str] = dataclasses.field( default=__snake_case , metadata={"help": "Whether to filter the pseudo-labeled data based on the confidence score."} , ) UpperCAmelCase__ : Union[str, Any] = dataclasses.field( default=__snake_case , metadata={"help": "Whether to filter the pseudo-labeled data based on the validation performance."} , ) UpperCAmelCase__ : str = dataclasses.field( default=__snake_case , metadata={"help": "Whether to fine-tune on labeled data after pseudo training."} , ) UpperCAmelCase__ : Optional[Any] = dataclasses.field( default=0.0 , metadata={"help": "Confidence threshold for pseudo-labeled data filtering."} , ) UpperCAmelCase__ : Tuple = dataclasses.field( default=100 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Union[str, Any] = dataclasses.field( default=__snake_case , metadata={"help": "Random seed for initialization."} , ) def __UpperCAmelCase ( __a : Dict ,__a : Dict ,__a : Optional[Any] ,__a : str ,__a : Dict ,__a : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" _a : int = datasets.concatenate_datasets([infer_input, infer_output] ,axis=1 ) if args.do_filter_by_confidence: _a : List[Any] = dataset.filter(lambda __a : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 _a : int = int(eval_result * len(_SCREAMING_SNAKE_CASE ) ) print(_SCREAMING_SNAKE_CASE ) _a : List[Any] = dataset.sort('''probability''' ,reverse=_SCREAMING_SNAKE_CASE ) _a : Optional[int] = dataset.select(range(_SCREAMING_SNAKE_CASE ) ) _a : Dict = dataset.remove_columns(['''label''', '''probability'''] ) _a : Dict = dataset.rename_column('''prediction''' ,'''label''' ) _a : Union[str, Any] = dataset.map(lambda __a : {"label": idalabel[example["label"]]} ) _a : str = dataset.shuffle(seed=args.seed ) _a : Dict = os.path.join(_SCREAMING_SNAKE_CASE ,F"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(_SCREAMING_SNAKE_CASE ,index=_SCREAMING_SNAKE_CASE ) else: dataset.to_json(_SCREAMING_SNAKE_CASE ) def __UpperCAmelCase ( __a : List[str] ,__a : str ,__a : Union[str, Any] ,__a : Union[str, Any] ,**__a : Dict ) -> Optional[int]: """simple docstring""" _a : Any = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' ,datefmt='''%m/%d/%Y %H:%M:%S''' ,level=logging.INFO ,) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() _a : Union[str, Any] = STModelArguments(model_name_or_path=_SCREAMING_SNAKE_CASE ) _a : Dict = STDataArguments(train_file=_SCREAMING_SNAKE_CASE ,infer_file=_SCREAMING_SNAKE_CASE ) _a : Any = STTrainingArguments(output_dir=_SCREAMING_SNAKE_CASE ) _a : Any = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(_SCREAMING_SNAKE_CASE ).items(): setattr(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) for key, value in kwargs.items(): if hasattr(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): setattr(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # Sanity checks _a : Optional[int] = {} _a : str = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None _a : int = args.train_file _a : Any = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None _a : List[Any] = args.eval_file for key in data_files: _a : Any = data_files[key].split('''.''' )[-1] assert extension in ["csv", "json"], F"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: _a : str = extension else: assert extension == args.data_file_extension, F"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), F"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info('''Creating the initial data directory for self-training...''' ) _a : List[str] = F"""{args.output_dir}/self-train_iter-{{}}""".format _a : List[Any] = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir ,exist_ok=_SCREAMING_SNAKE_CASE ) os.makedirs(_SCREAMING_SNAKE_CASE ,exist_ok=_SCREAMING_SNAKE_CASE ) accelerator.wait_for_everyone() _a : Any = None _a : str = None _a : str = 0 _a : Dict = False # Show the progress bar _a : Dict = tqdm(range(args.max_selftrain_iterations ) ,disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 ,int(args.max_selftrain_iterations ) ): _a : int = data_dir_format(_SCREAMING_SNAKE_CASE ) assert os.path.exists(_SCREAMING_SNAKE_CASE ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 _a : List[Any] = os.path.join(_SCREAMING_SNAKE_CASE ,'''stage-1''' ) _a : int = { '''accelerator''': accelerator, '''model_name_or_path''': args.model_name_or_path, '''cache_dir''': args.cache_dir, '''do_train''': True, '''train_file''': data_files['''train'''] if iteration == 0 else data_files['''train_pseudo'''], '''do_eval''': True if args.eval_file is not None else False, '''eval_file''': data_files['''eval'''], '''do_predict''': True, '''infer_file''': data_files['''infer'''], '''task_name''': args.task_name, '''label_list''': args.label_list, '''output_dir''': current_output_dir, '''eval_metric''': args.eval_metric, '''evaluation_strategy''': args.evaluation_strategy, '''early_stopping_patience''': args.early_stopping_patience, '''early_stopping_threshold''': args.early_stopping_threshold, '''seed''': args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): arguments_dict.update({key: value} ) _a : str = os.path.join(_SCREAMING_SNAKE_CASE ,'''best-checkpoint''' ,_SCREAMING_SNAKE_CASE ) if os.path.exists(_SCREAMING_SNAKE_CASE ): logger.info( '''Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.''' ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,) else: logger.info('''***** Running self-training: iteration: %d, stage: 1 *****''' ,_SCREAMING_SNAKE_CASE ) finetune(**_SCREAMING_SNAKE_CASE ) accelerator.wait_for_everyone() assert os.path.exists(_SCREAMING_SNAKE_CASE ) logger.info('''Self-training job completed: iteration: %d, stage: 1.''' ,_SCREAMING_SNAKE_CASE ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data _a : Dict = os.path.join(_SCREAMING_SNAKE_CASE ,'''best-checkpoint''' ) _a : int = os.path.join(_SCREAMING_SNAKE_CASE ,'''stage-2''' ) # Update arguments_dict _a : Any = model_path _a : Optional[int] = data_files['''train'''] _a : int = current_output_dir _a : List[Any] = os.path.join(_SCREAMING_SNAKE_CASE ,'''best-checkpoint''' ,_SCREAMING_SNAKE_CASE ) if os.path.exists(_SCREAMING_SNAKE_CASE ): logger.info( '''Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.''' ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,) else: logger.info('''***** Running self-training: iteration: %d, stage: 2 *****''' ,_SCREAMING_SNAKE_CASE ) finetune(**_SCREAMING_SNAKE_CASE ) accelerator.wait_for_everyone() assert os.path.exists(_SCREAMING_SNAKE_CASE ) logger.info('''Self-training job completed: iteration: %d, stage: 2.''' ,_SCREAMING_SNAKE_CASE ) _a : List[str] = iteration _a : Union[str, Any] = data_dir_format(iteration + 1 ) _a : List[str] = AutoConfig.from_pretrained(os.path.join(_SCREAMING_SNAKE_CASE ,'''best-checkpoint''' ) ) _a : List[str] = config.idalabel _a : Union[str, Any] = os.path.join(_SCREAMING_SNAKE_CASE ,'''eval_results_best-checkpoint.json''' ) _a : Tuple = os.path.join(_SCREAMING_SNAKE_CASE ,'''test_results_best-checkpoint.json''' ) assert os.path.exists(_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE ,'''r''' ) as f: _a : Tuple = float(json.load(_SCREAMING_SNAKE_CASE )[args.eval_metric] ) _a : Tuple = os.path.join(_SCREAMING_SNAKE_CASE ,'''infer_output_best-checkpoint.csv''' ) assert os.path.exists(_SCREAMING_SNAKE_CASE ) # Loading the dataset from local csv or json files. _a : Any = load_dataset(args.data_file_extension ,data_files={'''data''': data_files['''infer''']} )['''data'''] _a : List[Any] = load_dataset('''csv''' ,data_files={'''data''': infer_output_file} )['''data'''] if accelerator.is_main_process: os.makedirs(_SCREAMING_SNAKE_CASE ,exist_ok=_SCREAMING_SNAKE_CASE ) shutil.copy(_SCREAMING_SNAKE_CASE ,os.path.join(_SCREAMING_SNAKE_CASE ,F"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(_SCREAMING_SNAKE_CASE ): shutil.copy(_SCREAMING_SNAKE_CASE ,os.path.join(_SCREAMING_SNAKE_CASE ,F"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) accelerator.wait_for_everyone() _a : Optional[int] = os.path.join(_SCREAMING_SNAKE_CASE ,F"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: _a : Optional[Any] = eval_result if best_iteration is None: _a : int = new_iteration _a : str = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: _a : Optional[Any] = new_iteration _a : Tuple = new_eval_result _a : Any = 0 else: if new_eval_result == best_eval_result: _a : Optional[int] = new_iteration _a : Dict = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: _a : int = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info('''Best iteration: %d''' ,_SCREAMING_SNAKE_CASE ) logger.info('''Best evaluation result: %s = %f''' ,args.eval_metric ,_SCREAMING_SNAKE_CASE ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_SCREAMING_SNAKE_CASE ,F"""eval_results_iter-{iteration}.json""" ) ,os.path.join(_SCREAMING_SNAKE_CASE ,'''eval_results_best-iteration.json''' ) ,) else: # Assume that the last iteration is the best logger.info('''Best iteration: %d''' ,args.max_selftrain_iterations - 1 ) logger.info('''Best evaluation result: %s = %f''' ,args.eval_metric ,_SCREAMING_SNAKE_CASE ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(_SCREAMING_SNAKE_CASE ,F"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) ,os.path.join(_SCREAMING_SNAKE_CASE ,'''eval_results_best-iteration.json''' ) ,)
235
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: _snake_case = data _snake_case = previous _snake_case = next_node def __str__(self ) -> str: return f"""{self.data}""" def lowercase (self ) -> int: return self.data def lowercase (self ) -> Dict: return self.next def lowercase (self ) -> Union[str, Any]: return self.previous class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase ) -> List[str]: _snake_case = head def __iter__(self ) -> Optional[Any]: return self def lowercase (self ) -> str: if not self.current: raise StopIteration else: _snake_case = self.current.get_data() _snake_case = self.current.get_next() return value class _lowerCAmelCase : '''simple docstring''' def __init__(self ) -> Optional[int]: _snake_case = None # First node in list _snake_case = None # Last node in list def __str__(self ) -> Optional[int]: _snake_case = self.head _snake_case = [] while current is not None: nodes.append(current.get_data() ) _snake_case = current.get_next() return " ".join(str(UpperCAmelCase ) for node in nodes ) def __contains__(self , UpperCAmelCase ) -> int: _snake_case = self.head while current: if current.get_data() == value: return True _snake_case = current.get_next() return False def __iter__(self ) -> Union[str, Any]: return LinkedListIterator(self.head ) def lowercase (self ) -> str: if self.head: return self.head.get_data() return None def lowercase (self ) -> List[Any]: if self.tail: return self.tail.get_data() return None def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: _snake_case = node _snake_case = node else: self.insert_before_node(self.head , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: self.set_head(UpperCAmelCase ) else: self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: _snake_case = Node(UpperCAmelCase ) if self.head is None: self.set_head(UpperCAmelCase ) else: self.set_tail(UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.previous if node.get_previous() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.next if node.get_next() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = 1 _snake_case = Node(UpperCAmelCase ) _snake_case = self.head while node: if current_position == position: self.insert_before_node(UpperCAmelCase , UpperCAmelCase ) return current_position += 1 _snake_case = node.next self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> Node: _snake_case = self.head while node: if node.get_data() == item: return node _snake_case = node.get_next() raise Exception("""Node not found""" ) def lowercase (self , UpperCAmelCase ) -> Optional[int]: if (node := self.get_node(UpperCAmelCase )) is not None: if node == self.head: _snake_case = self.head.get_next() if node == self.tail: _snake_case = self.tail.get_previous() self.remove_node_pointers(UpperCAmelCase ) @staticmethod def lowercase (UpperCAmelCase ) -> None: if node.get_next(): _snake_case = node.previous if node.get_previous(): _snake_case = node.next _snake_case = None _snake_case = None def lowercase (self ) -> Dict: return self.head is None def __SCREAMING_SNAKE_CASE ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
341
0
import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: _lowercase : Dict = os.path.join(args.tf_model_dir , 'parameters.json' ) _lowercase : Optional[Any] = json.loads(open(_SCREAMING_SNAKE_CASE ).read() ) if not params: raise ValueError( F'''It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.''' ) if not args.output.endswith('.pt' ): _lowercase : int = args.output + '.pt' _lowercase : Any = OrderedDict() with tf.device('/CPU:0' ): _lowercase : Union[str, Any] = tf.train.load_checkpoint(args.tf_model_dir ) _lowercase : List[str] = reader.get_variable_to_shape_map() for key_name in shapes.keys(): _lowercase : List[str] = reader.get_tensor(_SCREAMING_SNAKE_CASE ).astype(np.floataa ) if key_name.endswith('/adam_m' ) or key_name.endswith('/adam_v' ): continue if key_name.startswith('pasts/' ): if key_name.startswith('pasts/mlp' ): _lowercase : int = int(key_name[9] ) elif key_name.startswith('pasts/out' ): _lowercase : Optional[Any] = 8 _lowercase : Optional[Any] = 'model.sqout.%d.weight' % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time _lowercase : List[str] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _lowercase : List[str] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.startswith('model/moe' ): _lowercase : List[Any] = int(key_name[9:].split('/' )[0] ) if key_name.endswith('/switch_gating/kernel' ): _lowercase : Union[str, Any] = 'model.blocks.%d.feed_forward.mlp.router.classifier.weight' % player _lowercase : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _lowercase : Dict = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/softmlp/kernel' ): _lowercase : str = 'model.blocks.%d.feed_forward.soft_bypass_mlp.weight' % player _lowercase : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _lowercase : Union[str, Any] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/wo/kernel' ) or key_name.endswith('/wi/kernel' ): _lowercase : str = key_name[-9:-7] for i in range(16 ): _lowercase : Optional[int] = 'model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight' % (player, i, nlayer) _lowercase : Optional[int] = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided _lowercase : Any = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.startswith('model/mlp' ): _lowercase : int = int(key_name[9:].split('/' )[0] ) if key_name.endswith('/p1/kernel' ): _lowercase : Optional[Any] = 'model.blocks.%d.feed_forward.mlp.wi.weight' % player _lowercase : int = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _lowercase : List[str] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/p1/bias' ): _lowercase : Optional[Any] = 'model.blocks.%d.feed_forward.mlp.wi.bias' % player _lowercase : Any = vnp.copy() # same because it is one dimensional _lowercase : Dict = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/p2/kernel' ): _lowercase : Optional[int] = 'model.blocks.%d.feed_forward.mlp.wo.weight' % player _lowercase : Optional[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _lowercase : List[Any] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/p2/bias' ): _lowercase : Optional[Any] = 'model.blocks.%d.feed_forward.mlp.wo.bias' % player _lowercase : List[str] = vnp.copy() # same because it is one dimensional _lowercase : Tuple = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.startswith('model/ln' ): _lowercase : List[Any] = int(key_name[8:].split('/' )[0] ) if key_name.endswith('/b' ): _lowercase : Union[str, Any] = 'model.blocks.%d.feed_forward.norm.bias' % player _lowercase : List[str] = vnp.copy() # same because it is one dimensional _lowercase : List[str] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/g' ): _lowercase : List[Any] = 'model.blocks.%d.feed_forward.norm.weight' % player _lowercase : Tuple = vnp.copy() # same because it is one dimensional _lowercase : Dict = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.startswith('model/att' ): _lowercase : Dict = int(key_name[9:].split('/' )[0] ) if key_name.endswith('/qkv/kernel' ): _lowercase : List[Any] = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum _lowercase : Tuple = state[:, 0, :, :] _lowercase : Any = state[:, 1, :, :] _lowercase : Optional[Any] = state[:, 2, :, :] _lowercase : Optional[int] = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _lowercase : List[Any] = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _lowercase : str = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _lowercase : Optional[Any] = 'model.blocks.%d.self_attn.self_attn.q_proj.weight' % player _lowercase : List[Any] = torch.tensor(_SCREAMING_SNAKE_CASE ) _lowercase : int = 'model.blocks.%d.self_attn.self_attn.k_proj.weight' % player _lowercase : Optional[int] = torch.tensor(_SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = 'model.blocks.%d.self_attn.self_attn.v_proj.weight' % player _lowercase : Union[str, Any] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/o/kernel' ): _lowercase : Dict = 'model.blocks.%d.self_attn.self_attn.out_proj.weight' % player _lowercase : List[str] = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix _lowercase : Any = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.startswith('model/an' ): _lowercase : int = int(key_name[8:].split('/' )[0] ) if key_name.endswith('/b' ): _lowercase : Optional[int] = 'model.blocks.%d.self_attn.norm.bias' % player _lowercase : Optional[Any] = vnp.copy() # same because it is one dimensional _lowercase : List[str] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.endswith('/g' ): _lowercase : Tuple = 'model.blocks.%d.self_attn.norm.weight' % player _lowercase : Union[str, Any] = vnp.copy() # same because it is one dimensional _lowercase : Tuple = torch.tensor(_SCREAMING_SNAKE_CASE ) elif ( key_name.startswith('model/wte' ) or key_name.startswith('model/wpe' ) or key_name.startswith('model/ete' ) ): _lowercase : Optional[int] = {'wte': 'embed_tokens', 'wpe': 'position_embeddings', 'ete': 'extra_position_embeddings'}[ key_name[-3:] ] _lowercase : str = 'model.%s.weight' % nlayer _lowercase : Optional[int] = vnp.copy() # same in embedded _lowercase : List[Any] = torch.tensor(_SCREAMING_SNAKE_CASE ) if key_name.startswith('model/wte' ): _lowercase : Tuple = 'lm_head.weight' _lowercase : Any = vnp.copy() # same in embedded _lowercase : Union[str, Any] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name.startswith('model/wob' ): _lowercase : Tuple = 'final_logits_bias' _lowercase : List[str] = vnp.copy() # same in embedded _lowercase : int = state.reshape((1, -1) ) _lowercase : Any = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name == "model/dense/kernel": _lowercase : str = 'model.last_project.weight' _lowercase : Tuple = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _lowercase : Optional[int] = torch.tensor(_SCREAMING_SNAKE_CASE ) elif key_name == "model/dense_1/bias": _lowercase : Tuple = 'model.last_project.bias' _lowercase : int = vnp.copy() # same because it is one dimensional _lowercase : Optional[int] = torch.tensor(_SCREAMING_SNAKE_CASE ) torch.save(_SCREAMING_SNAKE_CASE , args.output ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[Any] = argparse.ArgumentParser( description="model converter.", formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument("--tf_model_dir", metavar="PATH", type=str, required=True, help="import model") parser.add_argument("--output", metavar="PATH", type=str, required=True, help="output model") SCREAMING_SNAKE_CASE : List[Any] = parser.parse_args() convert_tf_gptsan_to_pt(args)
21
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput __lowerCAmelCase = 8 def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x * 255).int().clamp(0 , 255 ) _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c h w -> b c 1 h w""" ) _snake_case = ((x & mask) != 0).float() _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c d h w -> b (c d) h w""" ) _snake_case = bits * 2 - 1 return bits def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x > 0).int() _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE , dtype=torch.intaa ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b (c d) h w -> b c d h w""" , d=8 ) _snake_case = reduce(x * mask , """b c d h w -> b c h w""" , """sum""" ) return (dec / 255).clamp(0.0 , 1.0 ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _snake_case = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _snake_case = self.alphas_cumprod[timestep] _snake_case = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _snake_case = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _snake_case = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _snake_case = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _snake_case = model_output.device if torch.is_tensor(_SCREAMING_SNAKE_CASE ) else """cpu""" _snake_case = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise _snake_case = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="epsilon" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): _snake_case = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _snake_case, _snake_case = torch.split(_SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: _snake_case = None # 1. compute alphas, betas _snake_case = self.alphas_cumprod[t] _snake_case = self.alphas_cumprod[t - 1] if t > 0 else self.one _snake_case = 1 - alpha_prod_t _snake_case = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _snake_case = model_output else: raise ValueError(f"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _snake_case = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _snake_case = 0 if t > 0: _snake_case = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_SCREAMING_SNAKE_CASE ).to(model_output.device ) _snake_case = (self._get_variance(_SCREAMING_SNAKE_CASE , predicted_variance=_SCREAMING_SNAKE_CASE ) ** 0.5) * noise _snake_case = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 1.0 , ) -> Tuple: super().__init__() _snake_case = bit_scale _snake_case = ( ddim_bit_scheduler_step if isinstance(UpperCAmelCase , UpperCAmelCase ) else ddpm_bit_scheduler_step ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) @torch.no_grad() def __call__(self , UpperCAmelCase = 256 , UpperCAmelCase = 256 , UpperCAmelCase = 50 , UpperCAmelCase = None , UpperCAmelCase = 1 , UpperCAmelCase = "pil" , UpperCAmelCase = True , **UpperCAmelCase , ) -> Union[Tuple, ImagePipelineOutput]: _snake_case = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=UpperCAmelCase , ) _snake_case = decimal_to_bits(UpperCAmelCase ) * self.bit_scale _snake_case = latents.to(self.device ) self.scheduler.set_timesteps(UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _snake_case = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _snake_case = self.scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ).prev_sample _snake_case = bits_to_decimal(UpperCAmelCase ) if output_type == "pil": _snake_case = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase )
341
0
'''simple docstring''' from __future__ import annotations import os from typing import Any import requests __snake_case = '''https://api.github.com''' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user __snake_case = BASE_URL + '''/user''' # https://github.com/settings/tokens __snake_case = os.environ.get('''USER_TOKEN''', '''''') def a ( __a ) -> str: '''simple docstring''' UpperCamelCase__ :Optional[Any] = { '''Authorization''': f'''token {auth_token}''', '''Accept''': '''application/vnd.github.v3+json''', } return requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F"""{key}: {value}""") else: raise ValueError('''\'USER_TOKEN\' field cannot be empty.''')
97
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 10**9 ): _snake_case = 1 _snake_case = 2 _snake_case = 0 _snake_case = 0 _snake_case = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value _snake_case = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f'''{solution() = }''')
341
0
def lowercase_ ( _A : Tuple , _A : Optional[int] ): """simple docstring""" lowerCamelCase__ : Any = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): lowerCamelCase__ : Union[str, Any] = n - k # Calculate C(n,k) for i in range(_SCREAMING_SNAKE_CASE ): result *= n - i result //= i + 1 return result def lowercase_ ( _A : List[Any] ): """simple docstring""" return binomial_coefficient(2 * node_count , _SCREAMING_SNAKE_CASE ) // (node_count + 1) def lowercase_ ( _A : Dict ): """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) lowerCamelCase__ : Tuple = 1 for i in range(1 , n + 1 ): result *= i return result def lowercase_ ( _A : Tuple ): """simple docstring""" return catalan_number(_SCREAMING_SNAKE_CASE ) * factorial(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": A : Dict = int(input("Enter the number of nodes: ").strip() or 0) if node_count <= 0: raise ValueError("We need some nodes to work with.") print( f'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' f'binary trees and {catalan_number(node_count)} binary search trees.' )
184
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "deberta-v2" def __init__(self , UpperCAmelCase=128100 , UpperCAmelCase=1536 , UpperCAmelCase=24 , UpperCAmelCase=24 , UpperCAmelCase=6144 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-7 , UpperCAmelCase=False , UpperCAmelCase=-1 , UpperCAmelCase=0 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=0 , UpperCAmelCase="gelu" , **UpperCAmelCase , ) -> List[str]: super().__init__(**UpperCAmelCase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = relative_attention _snake_case = max_relative_positions _snake_case = pad_token_id _snake_case = position_biased_input # Backwards compatibility if type(UpperCAmelCase ) == str: _snake_case = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case = pos_att_type _snake_case = vocab_size _snake_case = layer_norm_eps _snake_case = kwargs.get("""pooler_hidden_size""" , UpperCAmelCase ) _snake_case = pooler_dropout _snake_case = pooler_hidden_act class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def lowercase (self ) -> int: return 12 def lowercase (self , UpperCAmelCase , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = 3 , UpperCAmelCase = 40 , UpperCAmelCase = 40 , UpperCAmelCase = None , ) -> Mapping[str, Any]: _snake_case = super().generate_dummy_inputs(preprocessor=UpperCAmelCase , framework=UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
0
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def A_ ( snake_case : Optional[Any] ) -> Optional[Any]: # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def A_ ( ) -> Dict: '''simple docstring''' with parallel_backend('''spark''' ): assert ParallelBackendConfig.backend_name == "spark" __UpperCamelCase = [1, 2, 3] with pytest.raises(_SCREAMING_SNAKE_CASE ): with parallel_backend('''unsupported backend''' ): map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=2 ) with pytest.raises(_SCREAMING_SNAKE_CASE ): with parallel_backend('''unsupported backend''' ): map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize('''num_proc''' , [2, -1] ) def A_ ( snake_case : int ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase = [1, 2] __UpperCamelCase = {'''a''': 1, '''b''': 2} __UpperCamelCase = {'''a''': [1, 2], '''b''': [3, 4]} __UpperCamelCase = {'''a''': {'''1''': 1}, '''b''': 2} __UpperCamelCase = {'''a''': 1, '''b''': 2, '''c''': 3, '''d''': 4} __UpperCamelCase = [2, 3] __UpperCamelCase = {'''a''': 2, '''b''': 3} __UpperCamelCase = {'''a''': [2, 3], '''b''': [4, 5]} __UpperCamelCase = {'''a''': {'''1''': 2}, '''b''': 3} __UpperCamelCase = {'''a''': 2, '''b''': 3, '''c''': 4, '''d''': 5} with parallel_backend('''spark''' ): assert map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE ) == expected_map_nested_sa assert map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE ) == expected_map_nested_sa assert map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE ) == expected_map_nested_sa assert map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE ) == expected_map_nested_sa assert map_nested(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE ) == expected_map_nested_sa
328
'''simple docstring''' __lowerCAmelCase = [ (1_000, 'M'), (900, 'CM'), (500, 'D'), (400, 'CD'), (100, 'C'), (90, 'XC'), (50, 'L'), (40, 'XL'), (10, 'X'), (9, 'IX'), (5, 'V'), (4, 'IV'), (1, 'I'), ] def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = {"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1000} _snake_case = 0 _snake_case = 0 while place < len(_SCREAMING_SNAKE_CASE ): if (place + 1 < len(_SCREAMING_SNAKE_CASE )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = [] for arabic, roman in ROMAN: ((_snake_case), (_snake_case)) = divmod(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) result.append(roman * factor ) if number == 0: break return "".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( __snake_case ): lowercase__ = (EulerDiscreteScheduler,) lowercase__ = 10 def _UpperCAmelCase ( self : Optional[int] , **lowerCAmelCase_ : Any): """simple docstring""" lowercase_ = { """num_train_timesteps""": 1_1_0_0, """beta_start""": 0.0_001, """beta_end""": 0.02, """beta_schedule""": """linear""", } config.update(**lowerCAmelCase_) return config def _UpperCAmelCase ( self : List[Any]): """simple docstring""" for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=lowerCAmelCase_) def _UpperCAmelCase ( self : List[str]): """simple docstring""" for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02]): self.check_over_configs(beta_start=lowerCAmelCase_ , beta_end=lowerCAmelCase_) def _UpperCAmelCase ( self : Union[str, Any]): """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=lowerCAmelCase_) def _UpperCAmelCase ( self : int): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase_) def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) scheduler.set_timesteps(self.num_inference_steps) lowercase_ = torch.manual_seed(0) lowercase_ = self.dummy_model() lowercase_ = self.dummy_sample_deter * scheduler.init_noise_sigma lowercase_ = sample.to(lowerCAmelCase_) for i, t in enumerate(scheduler.timesteps): lowercase_ = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_) lowercase_ = output.prev_sample lowercase_ = torch.sum(torch.abs(lowerCAmelCase_)) lowercase_ = torch.mean(torch.abs(lowerCAmelCase_)) assert abs(result_sum.item() - 10.0_807) < 1E-2 assert abs(result_mean.item() - 0.0_131) < 1E-3 def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config(prediction_type="""v_prediction""") lowercase_ = scheduler_class(**lowerCAmelCase_) scheduler.set_timesteps(self.num_inference_steps) lowercase_ = torch.manual_seed(0) lowercase_ = self.dummy_model() lowercase_ = self.dummy_sample_deter * scheduler.init_noise_sigma lowercase_ = sample.to(lowerCAmelCase_) for i, t in enumerate(scheduler.timesteps): lowercase_ = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_) lowercase_ = output.prev_sample lowercase_ = torch.sum(torch.abs(lowerCAmelCase_)) lowercase_ = torch.mean(torch.abs(lowerCAmelCase_)) assert abs(result_sum.item() - 0.0_002) < 1E-2 assert abs(result_mean.item() - 2.2676E-06) < 1E-3 def _UpperCAmelCase ( self : Union[str, Any]): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) scheduler.set_timesteps(self.num_inference_steps , device=lowerCAmelCase_) lowercase_ = torch.manual_seed(0) lowercase_ = self.dummy_model() lowercase_ = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() lowercase_ = sample.to(lowerCAmelCase_) for t in scheduler.timesteps: lowercase_ = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_) lowercase_ = output.prev_sample lowercase_ = torch.sum(torch.abs(lowerCAmelCase_)) lowercase_ = torch.mean(torch.abs(lowerCAmelCase_)) assert abs(result_sum.item() - 10.0_807) < 1E-2 assert abs(result_mean.item() - 0.0_131) < 1E-3 def _UpperCAmelCase ( self : Union[str, Any]): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_ , use_karras_sigmas=lowerCAmelCase_) scheduler.set_timesteps(self.num_inference_steps , device=lowerCAmelCase_) lowercase_ = torch.manual_seed(0) lowercase_ = self.dummy_model() lowercase_ = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() lowercase_ = sample.to(lowerCAmelCase_) for t in scheduler.timesteps: lowercase_ = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_) lowercase_ = output.prev_sample lowercase_ = torch.sum(torch.abs(lowerCAmelCase_)) lowercase_ = torch.mean(torch.abs(lowerCAmelCase_)) assert abs(result_sum.item() - 124.52_299_499_511_719) < 1E-2 assert abs(result_mean.item() - 0.16_213_932_633_399_963) < 1E-3
136
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __lowerCAmelCase = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['PerceiverFeatureExtractor'] __lowerCAmelCase = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = DownBlockaD # noqa F405 UpperCAmelCase : List[str] = """down""" def __snake_case ( self : List[str]): a : List[Any] = [-0.0_232, -0.9_869, 0.8_054, -0.0_637, -0.1_688, -1.4_264, 0.4_470, -1.3_394, 0.0_904] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : List[Any] = ResnetDownsampleBlockaD # noqa F405 UpperCAmelCase : Any = """down""" def __snake_case ( self : Tuple): a : Dict = [0.0_710, 0.2_410, -0.7_320, -1.0_757, -1.1_343, 0.3_540, -0.0_133, -0.2_576, 0.0_948] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = AttnDownBlockaD # noqa F405 UpperCAmelCase : Optional[Any] = """down""" def __snake_case ( self : Optional[Any]): a : int = [0.0_636, 0.8_964, -0.6_234, -1.0_131, 0.0_844, 0.4_935, 0.3_437, 0.0_911, -0.2_957] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Optional[Any] = CrossAttnDownBlockaD # noqa F405 UpperCAmelCase : List[str] = """down""" def __snake_case ( self : Union[str, Any]): a , a : Optional[Any] = super().prepare_init_args_and_inputs_for_common() a : Optional[Any] = 32 return init_dict, inputs_dict def __snake_case ( self : Union[str, Any]): a : Optional[Any] = [0.2_238, -0.7_396, -0.2_255, -0.3_829, 0.1_925, 1.1_665, 0.0_603, -0.7_295, 0.1_983] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : List[Any] = SimpleCrossAttnDownBlockaD # noqa F405 UpperCAmelCase : int = """down""" @property def __snake_case ( self : Dict): return super().get_dummy_input(include_encoder_hidden_states=__UpperCAmelCase) def __snake_case ( self : Union[str, Any]): a , a : Union[str, Any] = super().prepare_init_args_and_inputs_for_common() a : Optional[int] = 32 return init_dict, inputs_dict @unittest.skipIf(torch_device == "mps" , "MPS result is not consistent") def __snake_case ( self : str): a : Dict = [0.7_921, -0.0_992, -0.1_962, -0.7_695, -0.4_242, 0.7_804, 0.4_737, 0.2_765, 0.3_338] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = SkipDownBlockaD # noqa F405 UpperCAmelCase : Tuple = """down""" @property def __snake_case ( self : Tuple): return super().get_dummy_input(include_skip_sample=__UpperCAmelCase) def __snake_case ( self : Union[str, Any]): a : List[str] = [-0.0_845, -0.2_087, -0.2_465, 0.0_971, 0.1_900, -0.0_484, 0.2_664, 0.4_179, 0.5_069] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : List[str] = AttnSkipDownBlockaD # noqa F405 UpperCAmelCase : int = """down""" @property def __snake_case ( self : List[str]): return super().get_dummy_input(include_skip_sample=__UpperCAmelCase) def __snake_case ( self : Optional[int]): a : Dict = [0.5_539, 0.1_609, 0.4_924, 0.0_537, -0.1_995, 0.4_050, 0.0_979, -0.2_721, -0.0_642] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = DownEncoderBlockaD # noqa F405 UpperCAmelCase : Tuple = """down""" @property def __snake_case ( self : List[Any]): return super().get_dummy_input(include_temb=__UpperCAmelCase) def __snake_case ( self : List[Any]): a : int = { "in_channels": 32, "out_channels": 32, } a : List[Any] = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : List[str]): a : Dict = [1.1_102, 0.5_302, 0.4_872, -0.0_023, -0.8_042, 0.0_483, -0.3_489, -0.5_632, 0.7_626] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = AttnDownEncoderBlockaD # noqa F405 UpperCAmelCase : List[str] = """down""" @property def __snake_case ( self : int): return super().get_dummy_input(include_temb=__UpperCAmelCase) def __snake_case ( self : Optional[Any]): a : int = { "in_channels": 32, "out_channels": 32, } a : Optional[Any] = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : List[Any]): a : List[Any] = [0.8_966, -0.1_486, 0.8_568, 0.8_141, -0.9_046, -0.1_342, -0.0_972, -0.7_417, 0.1_538] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : List[str] = UNetMidBlockaD # noqa F405 UpperCAmelCase : Tuple = """mid""" def __snake_case ( self : Dict): a : List[str] = { "in_channels": 32, "temb_channels": 128, } a : List[Any] = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str): a : Dict = [-0.1_062, 1.7_248, 0.3_494, 1.4_569, -0.0_910, -1.2_421, -0.9_984, 0.6_736, 1.0_028] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : int = UNetMidBlockaDCrossAttn # noqa F405 UpperCAmelCase : Tuple = """mid""" def __snake_case ( self : Union[str, Any]): a , a : str = super().prepare_init_args_and_inputs_for_common() a : Optional[Any] = 32 return init_dict, inputs_dict def __snake_case ( self : List[Any]): a : str = [0.0_187, 2.4_220, 0.4_484, 1.1_203, -0.6_121, -1.5_122, -0.8_270, 0.7_851, 1.8_335] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Any = UNetMidBlockaDSimpleCrossAttn # noqa F405 UpperCAmelCase : str = """mid""" @property def __snake_case ( self : Dict): return super().get_dummy_input(include_encoder_hidden_states=__UpperCAmelCase) def __snake_case ( self : Tuple): a , a : List[str] = super().prepare_init_args_and_inputs_for_common() a : Union[str, Any] = 32 return init_dict, inputs_dict def __snake_case ( self : List[str]): a : Union[str, Any] = [0.7_143, 1.9_974, 0.5_448, 1.3_977, 0.1_282, -1.1_237, -1.4_238, 0.5_530, 0.8_880] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = UpBlockaD # noqa F405 UpperCAmelCase : Any = """up""" @property def __snake_case ( self : Optional[Any]): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase) def __snake_case ( self : List[str]): a : str = [-0.2_041, -0.4_165, -0.3_022, 0.0_041, -0.6_628, -0.7_053, 0.1_928, -0.0_325, 0.0_523] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : int = ResnetUpsampleBlockaD # noqa F405 UpperCAmelCase : Optional[Any] = """up""" @property def __snake_case ( self : int): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase) def __snake_case ( self : str): a : Optional[int] = [0.2_287, 0.3_549, -0.1_346, 0.4_797, -0.1_715, -0.9_649, 0.7_305, -0.5_864, -0.6_244] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Tuple = CrossAttnUpBlockaD # noqa F405 UpperCAmelCase : Dict = """up""" @property def __snake_case ( self : str): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase) def __snake_case ( self : str): a , a : Optional[Any] = super().prepare_init_args_and_inputs_for_common() a : Dict = 32 return init_dict, inputs_dict def __snake_case ( self : Optional[int]): a : Optional[Any] = [-0.1_403, -0.3_515, -0.0_420, -0.1_425, 0.3_167, 0.5_094, -0.2_181, 0.5_931, 0.5_582] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Union[str, Any] = SimpleCrossAttnUpBlockaD # noqa F405 UpperCAmelCase : Any = """up""" @property def __snake_case ( self : int): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase , include_encoder_hidden_states=__UpperCAmelCase) def __snake_case ( self : int): a , a : Union[str, Any] = super().prepare_init_args_and_inputs_for_common() a : Dict = 32 return init_dict, inputs_dict def __snake_case ( self : str): a : int = [0.2_645, 0.1_480, 0.0_909, 0.8_044, -0.9_758, -0.9_083, 0.0_994, -1.1_453, -0.7_402] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Any = AttnUpBlockaD # noqa F405 UpperCAmelCase : str = """up""" @property def __snake_case ( self : List[Any]): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase) @unittest.skipIf(torch_device == "mps" , "MPS result is not consistent") def __snake_case ( self : int): a : Optional[Any] = [0.0_979, 0.1_326, 0.0_021, 0.0_659, 0.2_249, 0.0_059, 0.1_132, 0.5_952, 0.1_033] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Any = SkipUpBlockaD # noqa F405 UpperCAmelCase : Optional[int] = """up""" @property def __snake_case ( self : List[Any]): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase) def __snake_case ( self : int): a : Any = [-0.0_893, -0.1_234, -0.1_506, -0.0_332, 0.0_123, -0.0_211, 0.0_566, 0.0_143, 0.0_362] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : str = AttnSkipUpBlockaD # noqa F405 UpperCAmelCase : Dict = """up""" @property def __snake_case ( self : Optional[int]): return super().get_dummy_input(include_res_hidden_states_tuple=__UpperCAmelCase) def __snake_case ( self : Optional[Any]): a : Dict = [0.0_361, 0.0_617, 0.2_787, -0.0_350, 0.0_342, 0.3_421, -0.0_843, 0.0_913, 0.3_015] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Any = UpDecoderBlockaD # noqa F405 UpperCAmelCase : List[Any] = """up""" @property def __snake_case ( self : Optional[Any]): return super().get_dummy_input(include_temb=__UpperCAmelCase) def __snake_case ( self : Tuple): a : str = {"in_channels": 32, "out_channels": 32} a : int = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : List[str]): a : str = [0.4_404, 0.1_998, -0.9_886, -0.3_320, -0.3_128, -0.7_034, -0.6_955, -0.2_338, -0.3_137] super().test_output(__UpperCAmelCase) class _A ( __snake_case ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : int = AttnUpDecoderBlockaD # noqa F405 UpperCAmelCase : Optional[int] = """up""" @property def __snake_case ( self : Optional[int]): return super().get_dummy_input(include_temb=__UpperCAmelCase) def __snake_case ( self : List[Any]): a : Dict = {"in_channels": 32, "out_channels": 32} a : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : List[str]): a : Dict = [0.6_738, 0.4_491, 0.1_055, 1.0_710, 0.7_316, 0.3_339, 0.3_352, 0.1_023, 0.3_568] super().test_output(__UpperCAmelCase)
40
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __lowerCAmelCase = 'platform' import jax import jax.numpy as jnp from transformers import BlenderbotTokenizer from transformers.models.blenderbot.modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, shift_tokens_right, ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ): if attention_mask is None: _snake_case = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _snake_case = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _snake_case = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _snake_case = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=7 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=99 , UpperCAmelCase=16 , UpperCAmelCase=2 , UpperCAmelCase=4 , UpperCAmelCase=4 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=32 , UpperCAmelCase=2 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , ) -> Union[str, Any]: _snake_case = parent _snake_case = batch_size _snake_case = seq_length _snake_case = is_training _snake_case = use_labels _snake_case = vocab_size _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = eos_token_id _snake_case = pad_token_id _snake_case = bos_token_id _snake_case = initializer_range def lowercase (self ) -> str: _snake_case = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _snake_case = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=UpperCAmelCase , ) _snake_case = prepare_blenderbot_inputs_dict(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return config, inputs_dict def lowercase (self ) -> Dict: _snake_case, _snake_case = self.prepare_config_and_inputs() return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> List[Any]: _snake_case = 20 _snake_case = model_class_name(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] ) _snake_case, _snake_case = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) _snake_case = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _snake_case = model.init_cache(decoder_input_ids.shape[0] , UpperCAmelCase , UpperCAmelCase ) _snake_case = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _snake_case = model.decode( decoder_input_ids[:, :-1] , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , past_key_values=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""" ) _snake_case = model.decode( decoder_input_ids[:, -1:] , UpperCAmelCase , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=UpperCAmelCase , decoder_position_ids=UpperCAmelCase , ) _snake_case = model.decode(UpperCAmelCase , UpperCAmelCase , decoder_attention_mask=UpperCAmelCase ) _snake_case = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) @require_flax class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = 99 def lowercase (self ) -> Any: _snake_case = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _snake_case = input_ids.shape[0] _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case = self._get_config_and_data() _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = lm_model(input_ids=UpperCAmelCase ) _snake_case = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = BlenderbotConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _snake_case = FlaxBlenderbotForConditionalGeneration(UpperCAmelCase ) _snake_case = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _snake_case = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _snake_case = lm_model(input_ids=UpperCAmelCase , decoder_input_ids=UpperCAmelCase ) _snake_case = (*summary.shape, config.vocab_size) self.assertEqual(outputs["""logits"""].shape , UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _snake_case = shift_tokens_right(UpperCAmelCase , 1 , 2 ) _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() _snake_case = np.equal(UpperCAmelCase , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(UpperCAmelCase , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class _lowerCAmelCase ( __snake_case , unittest.TestCase , __snake_case ): '''simple docstring''' lowerCAmelCase_ = True lowerCAmelCase_ = ( ( FlaxBlenderbotModel, FlaxBlenderbotForConditionalGeneration, ) if is_flax_available() else () ) lowerCAmelCase_ = (FlaxBlenderbotForConditionalGeneration,) if is_flax_available() else () def lowercase (self ) -> Any: _snake_case = FlaxBlenderbotModelTester(self ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) _snake_case = model_class(UpperCAmelCase ) @jax.jit def encode_jitted(UpperCAmelCase , UpperCAmelCase=None , **UpperCAmelCase ): return model.encode(input_ids=UpperCAmelCase , attention_mask=UpperCAmelCase ) with self.subTest("""JIT Enabled""" ): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = encode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) def lowercase (self ) -> str: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _snake_case = model_class(UpperCAmelCase ) _snake_case = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""] ) _snake_case = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): return model.decode( decoder_input_ids=UpperCAmelCase , decoder_attention_mask=UpperCAmelCase , encoder_outputs=UpperCAmelCase , ) with self.subTest("""JIT Enabled""" ): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): _snake_case = decode_jitted(**UpperCAmelCase ).to_tuple() self.assertEqual(len(UpperCAmelCase ) , len(UpperCAmelCase ) ) for jitted_output, output in zip(UpperCAmelCase , UpperCAmelCase ): self.assertEqual(jitted_output.shape , output.shape ) @slow def lowercase (self ) -> Any: for model_class_name in self.all_model_classes: _snake_case = model_class_name.from_pretrained("""facebook/blenderbot-400M-distill""" ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _snake_case = np.ones((1, 1) ) * model.config.eos_token_id _snake_case = model(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) @unittest.skipUnless(jax_device != """cpu""" , """3B test too slow on CPU.""" ) @slow def lowercase (self ) -> Dict: _snake_case = {"""num_beams""": 1, """early_stopping""": True, """min_length""": 15, """max_length""": 25} _snake_case = {"""skip_special_tokens""": True, """clean_up_tokenization_spaces""": True} _snake_case = FlaxBlenderbotForConditionalGeneration.from_pretrained("""facebook/blenderbot-3B""" , from_pt=UpperCAmelCase ) _snake_case = BlenderbotTokenizer.from_pretrained("""facebook/blenderbot-3B""" ) _snake_case = ["""Sam"""] _snake_case = tokenizer(UpperCAmelCase , return_tensors="""jax""" ) _snake_case = model.generate(**UpperCAmelCase , **UpperCAmelCase ) _snake_case = """Sam is a great name. It means \"sun\" in Gaelic.""" _snake_case = tokenizer.batch_decode(UpperCAmelCase , **UpperCAmelCase ) assert generated_txt[0].strip() == tgt_text
341
0
import socket def snake_case( ) -> Tuple: '''simple docstring''' lowercase : Dict = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) lowercase : List[str] = socket.gethostname() lowercase : int = 1_23_12 sock.connect((host, port) ) sock.send(b'''Hello server!''' ) with open('''Received_file''' , '''wb''' ) as out_file: print('''File opened''' ) print('''Receiving data...''' ) while True: lowercase : int = sock.recv(10_24 ) if not data: break out_file.write(_SCREAMING_SNAKE_CASE ) print('''Successfully received the file''' ) sock.close() print('''Connection closed''' ) if __name__ == "__main__": main()
308
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=2 , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=10 , UpperCAmelCase=3 , UpperCAmelCase=32 * 4 , UpperCAmelCase=32 * 6 , UpperCAmelCase=4 , UpperCAmelCase=32 , ) -> Optional[Any]: _snake_case = parent _snake_case = batch_size _snake_case = is_training _snake_case = use_auxiliary_loss _snake_case = num_queries _snake_case = num_channels _snake_case = min_size _snake_case = max_size _snake_case = num_labels _snake_case = mask_feature_size def lowercase (self ) -> str: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( UpperCAmelCase ) _snake_case = torch.ones([self.batch_size, self.min_size, self.max_size] , device=UpperCAmelCase ) _snake_case = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=UpperCAmelCase ) > 0.5 ).float() _snake_case = (torch.rand((self.batch_size, self.num_labels) , device=UpperCAmelCase ) > 0.5).long() _snake_case = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def lowercase (self ) -> Tuple: return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def lowercase (self ) -> Optional[Any]: _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.prepare_config_and_inputs() _snake_case = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> int: _snake_case = output.encoder_hidden_states _snake_case = output.pixel_decoder_hidden_states _snake_case = output.transformer_decoder_hidden_states self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(UpperCAmelCase ) , config.decoder_config.decoder_layers ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Union[str, Any]: with torch.no_grad(): _snake_case = MaskFormerModel(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase , output_hidden_states=UpperCAmelCase ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(UpperCAmelCase , UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: _snake_case = MaskFormerForInstanceSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() def comm_check_on_output(UpperCAmelCase ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _snake_case = model(pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase ) _snake_case = model(UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) _snake_case = model( pixel_values=UpperCAmelCase , pixel_mask=UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) comm_check_on_output(UpperCAmelCase ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowerCAmelCase_ = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> int: _snake_case = MaskFormerModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase ) def lowercase (self ) -> int: self.config_tester.run_common_tests() def lowercase (self ) -> List[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> Any: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*UpperCAmelCase ) @unittest.skip(reason="""MaskFormer does not use inputs_embeds""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip(reason="""MaskFormer does not have a get_input_embeddings method""" ) def lowercase (self ) -> Optional[int]: pass @unittest.skip(reason="""MaskFormer is not a generative model""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""MaskFormer does not use token embeddings""" ) def lowercase (self ) -> Optional[int]: pass @require_torch_multi_gpu @unittest.skip( reason="""MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> Optional[Any]: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> Tuple: pass def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) @slow def lowercase (self ) -> int: for model_name in ["facebook/maskformer-swin-small-coco"]: _snake_case = MaskFormerModel.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Tuple: _snake_case = (self.model_tester.min_size,) * 2 _snake_case = { """pixel_values""": torch.randn((2, 3, *size) , device=UpperCAmelCase ), """mask_labels""": torch.randn((2, 10, *size) , device=UpperCAmelCase ), """class_labels""": torch.zeros(2 , 10 , device=UpperCAmelCase ).long(), } _snake_case = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None ) def lowercase (self ) -> Dict: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCAmelCase , **UpperCAmelCase , output_hidden_states=UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ).to(UpperCAmelCase ) _snake_case = model(**UpperCAmelCase , output_attentions=UpperCAmelCase ) self.assertTrue(outputs.attentions is not None ) def lowercase (self ) -> Tuple: if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ).loss loss.backward() def lowercase (self ) -> List[str]: # only MaskFormerForInstanceSegmentation has the loss _snake_case = self.all_model_classes[1] _snake_case, _snake_case, _snake_case, _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs() _snake_case = True _snake_case = True _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.train() _snake_case = model(UpperCAmelCase , mask_labels=UpperCAmelCase , class_labels=UpperCAmelCase ) _snake_case = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _snake_case = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _snake_case = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _snake_case = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=UpperCAmelCase ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1E-4 def __SCREAMING_SNAKE_CASE ( ): _snake_case = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def lowercase (self ) -> Optional[int]: return ( MaskFormerImageProcessor.from_pretrained("""facebook/maskformer-swin-small-coco""" ) if is_vision_available() else None ) def lowercase (self ) -> str: _snake_case = MaskFormerModel.from_pretrained("""facebook/maskformer-swin-small-coco""" ).to(UpperCAmelCase ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) _snake_case = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(UpperCAmelCase ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[str]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [ [1.6_5_1_2e0_0, -5.2_5_7_2e0_0, -3.3_5_1_9e0_0], [3.6_1_6_9e-0_2, -5.9_0_2_5e0_0, -2.9_3_1_3e0_0], [1.0_7_6_6e-0_4, -7.7_6_3_0e0_0, -5.1_2_6_3e0_0], ] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> List[Any]: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-resnet101-coco-stuff""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) _snake_case = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(UpperCAmelCase , (1, 3, 800, 1088) ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) # masks_queries_logits _snake_case = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _snake_case = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _snake_case = torch.tensor(UpperCAmelCase ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) # class_queries_logits _snake_case = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _snake_case = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCAmelCase , atol=UpperCAmelCase ) ) def lowercase (self ) -> Tuple: _snake_case = ( MaskFormerForInstanceSegmentation.from_pretrained("""facebook/maskformer-swin-small-coco""" ) .to(UpperCAmelCase ) .eval() ) _snake_case = self.default_image_processor _snake_case = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) _snake_case = inputs["""pixel_values"""].to(UpperCAmelCase ) _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""mask_labels"""]] _snake_case = [el.to(UpperCAmelCase ) for el in inputs["""class_labels"""]] with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) self.assertTrue(outputs.loss is not None )
341
0
"""simple docstring""" import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = StableDiffusionPipeline.from_pretrained(_SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors SCREAMING_SNAKE_CASE__ = load_file(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: SCREAMING_SNAKE_CASE__ = key.split(""".""" )[0].split(LORA_PREFIX_TEXT_ENCODER + """_""" )[-1].split("""_""" ) SCREAMING_SNAKE_CASE__ = pipeline.text_encoder else: SCREAMING_SNAKE_CASE__ = key.split(""".""" )[0].split(LORA_PREFIX_UNET + """_""" )[-1].split("""_""" ) SCREAMING_SNAKE_CASE__ = pipeline.unet # find the target layer SCREAMING_SNAKE_CASE__ = layer_infos.pop(0 ) while len(_SCREAMING_SNAKE_CASE ) > -1: try: SCREAMING_SNAKE_CASE__ = curr_layer.__getattr__(_SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: SCREAMING_SNAKE_CASE__ = layer_infos.pop(0 ) elif len(_SCREAMING_SNAKE_CASE ) == 0: break except Exception: if len(_SCREAMING_SNAKE_CASE ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: SCREAMING_SNAKE_CASE__ = layer_infos.pop(0 ) SCREAMING_SNAKE_CASE__ = [] if "lora_down" in key: pair_keys.append(key.replace("""lora_down""" , """lora_up""" ) ) pair_keys.append(_SCREAMING_SNAKE_CASE ) else: pair_keys.append(_SCREAMING_SNAKE_CASE ) pair_keys.append(key.replace("""lora_up""" , """lora_down""" ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: SCREAMING_SNAKE_CASE__ = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) SCREAMING_SNAKE_CASE__ = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).unsqueeze(2 ).unsqueeze(3 ) else: SCREAMING_SNAKE_CASE__ = state_dict[pair_keys[0]].to(torch.floataa ) SCREAMING_SNAKE_CASE__ = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # update visited list for item in pair_keys: visited.append(_SCREAMING_SNAKE_CASE ) return pipeline if __name__ == "__main__": A_ : Tuple = argparse.ArgumentParser() parser.add_argument( "--base_model_path", default=None, type=str, required=True, help="Path to the base model in diffusers format." ) parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument( "--lora_prefix_unet", default="lora_unet", type=str, help="The prefix of UNet weight in safetensors" ) parser.add_argument( "--lora_prefix_text_encoder", default="lora_te", type=str, help="The prefix of text encoder weight in safetensors", ) parser.add_argument("--alpha", default=0.75, type=float, help="The merging ratio in W = W0 + alpha * deltaW") parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not." ) parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") A_ : Dict = parser.parse_args() A_ : List[str] = args.base_model_path A_ : List[str] = args.checkpoint_path A_ : Tuple = args.dump_path A_ : Union[str, Any] = args.lora_prefix_unet A_ : Optional[int] = args.lora_prefix_text_encoder A_ : Any = args.alpha A_ : int = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) A_ : int = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
165
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self , UpperCAmelCase ) -> Union[str, Any]: for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] , model_result["""ss"""] ): _snake_case = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCAmelCase ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sgugger/tiny-distilbert-classification""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , only_pretrain_model=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , torchscript=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , fpaa=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) # set architectures equal to `None` _snake_case = None _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Optional[int]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == """cpu""" , """Can't do half precision""" ) def lowercase (self ) -> Tuple: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> Union[str, Any]: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Dict: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowercase (self ) -> Any: _snake_case = """sshleifer/tiny-gpt2""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> int: _snake_case = """sshleifer/tinier_bart""" _snake_case = AutoConfig.from_pretrained(UpperCAmelCase ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowercase (self ) -> str: _snake_case = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , save_to_csv=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCAmelCase , """inf_time.csv""" ) , train_memory_csv_file=os.path.join(UpperCAmelCase , """train_mem.csv""" ) , inference_memory_csv_file=os.path.join(UpperCAmelCase , """inf_mem.csv""" ) , train_time_csv_file=os.path.join(UpperCAmelCase , """train_time.csv""" ) , env_info_csv_file=os.path.join(UpperCAmelCase , """env.csv""" ) , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """train_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """env.csv""" ) ).exists() ) def lowercase (self ) -> int: _snake_case = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCAmelCase ): self.assertTrue(hasattr(UpperCAmelCase , """sequential""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """cumulative""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """current""" ) ) self.assertTrue(hasattr(UpperCAmelCase , """total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCAmelCase , inference=UpperCAmelCase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCAmelCase , """log.txt""" ) , log_print=UpperCAmelCase , trace_memory_line_by_line=UpperCAmelCase , multi_process=UpperCAmelCase , ) _snake_case = PyTorchBenchmark(UpperCAmelCase ) _snake_case = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(UpperCAmelCase , """log.txt""" ) ).exists() )
341
0
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class a__ ( unittest.TestCase ): """simple docstring""" def UpperCamelCase ( self ) -> Dict: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' A__ = 1 A__ = 3 A__ = (32, 32) A__ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowercase ) return image @property def UpperCamelCase ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) A__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def UpperCamelCase ( self ) -> int: '''simple docstring''' torch.manual_seed(0 ) A__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def UpperCamelCase ( self ) -> int: '''simple docstring''' torch.manual_seed(0 ) A__ = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(lowercase ) @property def UpperCamelCase ( self ) -> int: '''simple docstring''' def extract(*lowercase , **lowercase ): class a__ : """simple docstring""" def __init__( self ) -> List[Any]: '''simple docstring''' A__ = torch.ones([0] ) def UpperCamelCase ( self , lowercase ) -> Tuple: '''simple docstring''' self.pixel_values.to(lowercase ) return self return Out() return extract def UpperCamelCase ( self ) -> Tuple: '''simple docstring''' A__ = "cpu" # ensure determinism for the device-dependent torch.Generator A__ = self.dummy_cond_unet A__ = PNDMScheduler(skip_prk_steps=lowercase ) A__ = self.dummy_vae A__ = self.dummy_text_encoder A__ = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) A__ = 77 A__ = self.dummy_image.to(lowercase ) A__ = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk A__ = AltDiffusionImgaImgPipeline( unet=lowercase , scheduler=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , safety_checker=lowercase , feature_extractor=self.dummy_extractor , ) A__ = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowercase ) A__ = alt_pipe.to(lowercase ) alt_pipe.set_progress_bar_config(disable=lowercase ) A__ = "A painting of a squirrel eating a burger" A__ = torch.Generator(device=lowercase ).manual_seed(0 ) A__ = alt_pipe( [prompt] , generator=lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowercase , ) A__ = output.images A__ = torch.Generator(device=lowercase ).manual_seed(0 ) A__ = alt_pipe( [prompt] , generator=lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=lowercase , return_dict=lowercase , )[0] A__ = image[0, -3:, -3:, -1] A__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) A__ = np.array([0.4427, 0.3731, 0.4249, 0.4941, 0.4546, 0.4148, 0.4193, 0.4666, 0.4499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5e-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' A__ = self.dummy_cond_unet A__ = PNDMScheduler(skip_prk_steps=lowercase ) A__ = self.dummy_vae A__ = self.dummy_text_encoder A__ = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) A__ = 77 A__ = self.dummy_image.to(lowercase ) # put models in fp16 A__ = unet.half() A__ = vae.half() A__ = bert.half() # make sure here that pndm scheduler skips prk A__ = AltDiffusionImgaImgPipeline( unet=lowercase , scheduler=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , safety_checker=lowercase , feature_extractor=self.dummy_extractor , ) A__ = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=lowercase ) A__ = alt_pipe.to(lowercase ) alt_pipe.set_progress_bar_config(disable=lowercase ) A__ = "A painting of a squirrel eating a burger" A__ = torch.manual_seed(0 ) A__ = alt_pipe( [prompt] , generator=lowercase , num_inference_steps=2 , output_type="np" , image=lowercase , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def UpperCamelCase ( self ) -> Any: '''simple docstring''' A__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 A__ = init_image.resize((760, 504) ) A__ = "BAAI/AltDiffusion" A__ = AltDiffusionImgaImgPipeline.from_pretrained( lowercase , safety_checker=lowercase , ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() A__ = "A fantasy landscape, trending on artstation" A__ = torch.manual_seed(0 ) A__ = pipe( prompt=lowercase , image=lowercase , strength=0.75 , guidance_scale=7.5 , generator=lowercase , output_type="np" , ) A__ = output.images[0] A__ = image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) A__ = np.array([0.9358, 0.9397, 0.9599, 0.9901, 1.0000, 1.0000, 0.9882, 1.0000, 1.0000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class a__ ( unittest.TestCase ): """simple docstring""" def UpperCamelCase ( self ) -> Dict: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self ) -> Dict: '''simple docstring''' A__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) A__ = init_image.resize((768, 512) ) A__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) A__ = "BAAI/AltDiffusion" A__ = AltDiffusionImgaImgPipeline.from_pretrained( lowercase , safety_checker=lowercase , ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() A__ = "A fantasy landscape, trending on artstation" A__ = torch.manual_seed(0 ) A__ = pipe( prompt=lowercase , image=lowercase , strength=0.75 , guidance_scale=7.5 , generator=lowercase , output_type="np" , ) A__ = output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1e-2
68
'''simple docstring''' from __future__ import annotations def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): if len(_SCREAMING_SNAKE_CASE ) == 0: return [] _snake_case, _snake_case = min(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) _snake_case = int(max_value - min_value ) + 1 _snake_case = [[] for _ in range(_SCREAMING_SNAKE_CASE )] for i in my_list: buckets[int(i - min_value )].append(_SCREAMING_SNAKE_CASE ) return [v for bucket in buckets for v in sorted(_SCREAMING_SNAKE_CASE )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
341
0
'''simple docstring''' import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class __UpperCAmelCase ( __snake_case ): __lowercase = """microsoft/speecht5_tts""" __lowercase = ( """This is a tool that reads an English text out loud. It takes an input named `text` which should contain the """ """text to read (in English) and returns a waveform object containing the sound.""" ) __lowercase = """text_reader""" __lowercase = SpeechTaProcessor __lowercase = SpeechTaForTextToSpeech __lowercase = SpeechTaHifiGan __lowercase = ["""text"""] __lowercase = ["""audio"""] def lowerCamelCase ( self ): """simple docstring""" if self.post_processor is None: _snake_case = 'microsoft/speecht5_hifigan' super().setup() def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = self.pre_processor(text=lowerCAmelCase_ , return_tensors='pt' , truncation=lowerCAmelCase_ ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError('Datasets needs to be installed if not passing speaker embeddings.' ) _snake_case = load_dataset('Matthijs/cmu-arctic-xvectors' , split='validation' ) _snake_case = torch.tensor(embeddings_dataset[73_05]['xvector'] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" with torch.no_grad(): return self.model.generate_speech(**lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" with torch.no_grad(): return self.post_processor(lowerCAmelCase_ ).cpu().detach()
42
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: if not conversation_id: _snake_case = uuid.uuida() if past_user_inputs is None: _snake_case = [] if generated_responses is None: _snake_case = [] _snake_case = conversation_id _snake_case = past_user_inputs _snake_case = generated_responses _snake_case = text def __eq__(self , UpperCAmelCase ) -> Dict: if not isinstance(UpperCAmelCase , UpperCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = False ) -> int: if self.new_user_input: if overwrite: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ f"""with: \"{text}\".""" ) _snake_case = text else: logger.warning( f"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ f"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: _snake_case = text def lowercase (self ) -> int: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _snake_case = None def lowercase (self , UpperCAmelCase ) -> Any: self.generated_responses.append(UpperCAmelCase ) def lowercase (self ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__(self ) -> Optional[int]: _snake_case = f"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): _snake_case = """user""" if is_user else """bot""" output += f"""{name} >> {text} \n""" return output @add_end_docstrings( __snake_case , r"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , *UpperCAmelCase , **UpperCAmelCase ) -> Union[str, Any]: super().__init__(*UpperCAmelCase , **UpperCAmelCase ) if self.tokenizer.pad_token_id is None: _snake_case = self.tokenizer.eos_token def lowercase (self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase ) -> Dict: _snake_case = {} _snake_case = {} _snake_case = {} if min_length_for_response is not None: _snake_case = min_length_for_response if minimum_tokens is not None: _snake_case = minimum_tokens if "max_length" in generate_kwargs: _snake_case = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _snake_case = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(UpperCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__(self , UpperCAmelCase , UpperCAmelCase=0 , **UpperCAmelCase ) -> Union[str, Any]: _snake_case = super().__call__(UpperCAmelCase , num_workers=UpperCAmelCase , **UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) and len(UpperCAmelCase ) == 1: return outputs[0] return outputs def lowercase (self , UpperCAmelCase , UpperCAmelCase=32 ) -> Dict[str, Any]: if not isinstance(UpperCAmelCase , UpperCAmelCase ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( f"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): _snake_case = self.tokenizer._build_conversation_input_ids(UpperCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version _snake_case = self._legacy_parse_and_tokenize(UpperCAmelCase ) if self.framework == "pt": _snake_case = torch.LongTensor([input_ids] ) elif self.framework == "tf": _snake_case = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=10 , **UpperCAmelCase ) -> Optional[int]: _snake_case = generate_kwargs.get("""max_length""" , self.model.config.max_length ) _snake_case = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(f"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) _snake_case = max_length - minimum_tokens _snake_case = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: _snake_case = model_inputs["""attention_mask"""][:, -trim:] _snake_case = model_inputs.pop("""conversation""" ) _snake_case = max_length _snake_case = self.model.generate(**UpperCAmelCase , **UpperCAmelCase ) if self.model.config.is_encoder_decoder: _snake_case = 1 else: _snake_case = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def lowercase (self , UpperCAmelCase , UpperCAmelCase=True ) -> List[str]: _snake_case = model_outputs["""output_ids"""] _snake_case = self.tokenizer.decode( output_ids[0] , skip_special_tokens=UpperCAmelCase , clean_up_tokenization_spaces=UpperCAmelCase , ) _snake_case = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(UpperCAmelCase ) return conversation def lowercase (self , UpperCAmelCase ) -> Dict: _snake_case = self.tokenizer.eos_token_id _snake_case = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) ) if len(UpperCAmelCase ) > self.tokenizer.model_max_length: _snake_case = input_ids[-self.tokenizer.model_max_length :] return input_ids
341
0
def __UpperCAmelCase ( __a : Any ) -> Union[str, Any]: """simple docstring""" if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) _a : Union[str, Any] = sorted(string.lower() ) return len(_SCREAMING_SNAKE_CASE ) == len(set(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": a__ = input('''Enter a string ''').strip() a__ = is_isogram(input_str) print(f'''{input_str} is {"an" if isogram else "not an"} isogram.''')
235
'''simple docstring''' from math import factorial, radians def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 18 , _SCREAMING_SNAKE_CASE = 10 ): _snake_case = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians _snake_case = radians(_SCREAMING_SNAKE_CASE ) _snake_case = angle_in_radians _snake_case = 3 _snake_case = -1 for _ in range(_SCREAMING_SNAKE_CASE ): result += (b * (angle_in_radians**a)) / factorial(_SCREAMING_SNAKE_CASE ) _snake_case = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __import__('doctest').testmod()
341
0
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class _lowerCamelCase( __snake_case, unittest.TestCase ): lowercase_ : int = RoFormerTokenizer lowercase_ : List[Any] = RoFormerTokenizerFast lowercase_ : Optional[Any] = True lowercase_ : str = True def UpperCamelCase ( self) -> Tuple: """simple docstring""" super().setUp() def UpperCamelCase ( self, **lowerCamelCase) -> Optional[Any]: """simple docstring""" return self.tokenizer_class.from_pretrained('junnyu/roformer_chinese_base', **lowerCamelCase) def UpperCamelCase ( self, **lowerCamelCase) -> Tuple: """simple docstring""" return self.rust_tokenizer_class.from_pretrained('junnyu/roformer_chinese_base', **lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = '永和服装饰品有限公司,今天天气非常好' _lowercase : Any = '永和 服装 饰品 有限公司 , 今 天 天 气 非常 好' return input_text, output_text def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Optional[Any] = self.get_tokenizer() _lowercase , _lowercase : Any = self.get_chinese_input_output_texts() _lowercase : Tuple = tokenizer.tokenize(lowerCamelCase) self.assertListEqual(lowerCamelCase, output_text.split()) _lowercase : str = tokens + [tokenizer.unk_token] _lowercase : List[Any] = [2_29_43, 2_13_32, 3_44_31, 4_59_04, 1_17, 3_06, 12_31, 12_31, 26_53, 3_39_94, 12_66, 1_00] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase), lowerCamelCase) def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = self.get_rust_tokenizer() _lowercase , _lowercase : Dict = self.get_chinese_input_output_texts() _lowercase : int = tokenizer.tokenize(lowerCamelCase) self.assertListEqual(lowerCamelCase, output_text.split()) _lowercase : Optional[Any] = tokens + [tokenizer.unk_token] _lowercase : Optional[Any] = [2_29_43, 2_13_32, 3_44_31, 4_59_04, 1_17, 3_06, 12_31, 12_31, 26_53, 3_39_94, 12_66, 1_00] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase), lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" pass def UpperCamelCase ( self) -> List[Any]: """simple docstring""" pass def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" pass
21
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __lowerCAmelCase = '\\n@inproceedings{popovic-2015-chrf,\n title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation",\n month = sep,\n year = "2015",\n address = "Lisbon, Portugal",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W15-3049",\n doi = "10.18653/v1/W15-3049",\n pages = "392--395",\n}\n@inproceedings{popovic-2017-chrf,\n title = "chr{F}++: words helping character n-grams",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Second Conference on Machine Translation",\n month = sep,\n year = "2017",\n address = "Copenhagen, Denmark",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W17-4770",\n doi = "10.18653/v1/W17-4770",\n pages = "612--618",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' __lowerCAmelCase = '\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n' __lowerCAmelCase = '\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n \'score\' (float): The chrF (chrF++) score,\n \'char_order\' (int): The character n-gram order,\n \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n \'beta\' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Tuple: if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/mjpost/sacreBLEU#chrf--chrf""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#chrf--chrf"""] , reference_urls=[ """https://github.com/m-popovic/chrF""", ] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = CHRF.CHAR_ORDER , UpperCAmelCase = CHRF.WORD_ORDER , UpperCAmelCase = CHRF.BETA , UpperCAmelCase = False , UpperCAmelCase = False , UpperCAmelCase = False , ) -> int: _snake_case = len(references[0] ) if any(len(UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) _snake_case = [[refs[i] for refs in references] for i in range(UpperCAmelCase )] _snake_case = CHRF(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _snake_case = sb_chrf.corpus_score(UpperCAmelCase , UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
341
0
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class lowercase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = '''ZinengTang/tvlt-base''' UpperCamelCase__ :Any = tempfile.mkdtemp() def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **UpperCamelCase_ ) def lowerCAmelCase__ ( self , **UpperCamelCase_ ): '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.get_image_processor() UpperCamelCase__ :Any = self.get_feature_extractor() UpperCamelCase__ :int = TvltProcessor(image_processor=UpperCamelCase_ , feature_extractor=UpperCamelCase_ ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase__ :Union[str, Any] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , UpperCamelCase_ ) self.assertIsInstance(processor.image_processor , UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[int] = self.get_image_processor() UpperCamelCase__ :List[str] = self.get_feature_extractor() UpperCamelCase__ :List[str] = TvltProcessor(image_processor=UpperCamelCase_ , feature_extractor=UpperCamelCase_ ) UpperCamelCase__ :Optional[Any] = np.ones([12000] ) UpperCamelCase__ :Optional[Any] = feature_extractor(UpperCamelCase_ , return_tensors='''np''' ) UpperCamelCase__ :Any = processor(audio=UpperCamelCase_ , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :int = self.get_image_processor() UpperCamelCase__ :Optional[int] = self.get_feature_extractor() UpperCamelCase__ :Optional[Any] = TvltProcessor(image_processor=UpperCamelCase_ , feature_extractor=UpperCamelCase_ ) UpperCamelCase__ :int = np.ones([3, 224, 224] ) UpperCamelCase__ :Dict = image_processor(UpperCamelCase_ , return_tensors='''np''' ) UpperCamelCase__ :int = processor(images=UpperCamelCase_ , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.get_image_processor() UpperCamelCase__ :Optional[Any] = self.get_feature_extractor() UpperCamelCase__ :Union[str, Any] = TvltProcessor(image_processor=UpperCamelCase_ , feature_extractor=UpperCamelCase_ ) UpperCamelCase__ :Optional[Any] = np.ones([12000] ) UpperCamelCase__ :Dict = np.ones([3, 224, 224] ) UpperCamelCase__ :List[Any] = processor(audio=UpperCamelCase_ , images=UpperCamelCase_ ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_ ): processor() def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = self.get_image_processor() UpperCamelCase__ :Optional[Any] = self.get_feature_extractor() UpperCamelCase__ :int = TvltProcessor(image_processor=UpperCamelCase_ , feature_extractor=UpperCamelCase_ ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
97
'''simple docstring''' from scipy.stats import spearmanr import datasets __lowerCAmelCase = '\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n' __lowerCAmelCase = '\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {\'spearmanr\': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric("spearmanr")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results[\'spearmanr\'])\n -0.7\n >>> print(round(results[\'spearmanr_pvalue\'], 2))\n 0.19\n' __lowerCAmelCase = r'\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): '''simple docstring''' def lowercase (self ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) -> Optional[Any]: _snake_case = spearmanr(UpperCAmelCase , UpperCAmelCase ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
341
0
def lowercase_ ( _A : Dict , _A : int ): """simple docstring""" lowerCamelCase__ : Dict = len(_SCREAMING_SNAKE_CASE ) lowerCamelCase__ : int = [] for i in range(len(_SCREAMING_SNAKE_CASE ) - pat_len + 1 ): lowerCamelCase__ : int = True for j in range(_SCREAMING_SNAKE_CASE ): if s[i + j] != pattern[j]: lowerCamelCase__ : List[Any] = False break if match_found: position.append(_SCREAMING_SNAKE_CASE ) return position if __name__ == "__main__": assert naive_pattern_search("ABCDEFG", "DE") == [3] print(naive_pattern_search("ABAAABCDBBABCDDEBCABC", "ABC"))
184
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=32 , UpperCAmelCase=3 , UpperCAmelCase=4 , UpperCAmelCase=[10, 20, 30, 40] , UpperCAmelCase=[2, 2, 3, 2] , UpperCAmelCase=True , UpperCAmelCase=True , UpperCAmelCase=37 , UpperCAmelCase="gelu" , UpperCAmelCase=10 , UpperCAmelCase=0.02 , UpperCAmelCase=["stage2", "stage3", "stage4"] , UpperCAmelCase=3 , UpperCAmelCase=None , ) -> List[Any]: _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = num_channels _snake_case = num_stages _snake_case = hidden_sizes _snake_case = depths _snake_case = is_training _snake_case = use_labels _snake_case = intermediate_size _snake_case = hidden_act _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = out_features _snake_case = num_labels _snake_case = scope _snake_case = num_stages def lowercase (self ) -> List[Any]: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case = None if self.use_labels: _snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case = self.get_config() return config, pixel_values, labels def lowercase (self ) -> Tuple: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def lowercase (self ) -> Any: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCAmelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCAmelCase , loss_ignore_index=255 , num_labels=self.num_labels , ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> str: _snake_case = UperNetForSemanticSegmentation(config=UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() _snake_case = model(UpperCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowercase (self ) -> Tuple: _snake_case = self.prepare_config_and_inputs() ( ( _snake_case ), ( _snake_case ), ( _snake_case ), ) = config_and_inputs _snake_case = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowerCAmelCase_ = {"image-segmentation": UperNetForSemanticSegmentation} if is_torch_available() else {} lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def lowercase (self ) -> Optional[Any]: _snake_case = UperNetModelTester(self ) _snake_case = ConfigTester(self , config_class=UpperCAmelCase , has_text_modality=UpperCAmelCase , hidden_size=37 ) def lowercase (self ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase (self ) -> Union[str, Any]: return def lowercase (self ) -> Union[str, Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(UpperCAmelCase ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase ) def lowercase (self ) -> int: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def lowercase (self ) -> int: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def lowercase (self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def lowercase (self ) -> Union[str, Any]: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def lowercase (self ) -> str: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase (self ) -> int: pass def lowercase (self ) -> List[str]: def check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): _snake_case = model_class(UpperCAmelCase ) model.to(UpperCAmelCase ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(UpperCAmelCase , UpperCAmelCase ) ) _snake_case = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case = self.model_tester.num_stages self.assertEqual(len(UpperCAmelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) def lowercase (self ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = _config_zero_init(UpperCAmelCase ) _snake_case = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _snake_case = model_class(config=UpperCAmelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def lowercase (self ) -> Optional[Any]: pass @slow def lowercase (self ) -> Tuple: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = UperNetForSemanticSegmentation.from_pretrained(UpperCAmelCase ) self.assertIsNotNone(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) _snake_case = Image.open(_SCREAMING_SNAKE_CASE ).convert("""RGB""" ) return image @require_torch @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) ) def lowercase (self ) -> Any: _snake_case = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) _snake_case = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(UpperCAmelCase ) _snake_case = prepare_img() _snake_case = processor(images=UpperCAmelCase , return_tensors="""pt""" ).to(UpperCAmelCase ) with torch.no_grad(): _snake_case = model(**UpperCAmelCase ) _snake_case = torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase ) _snake_case = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ).to(UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCAmelCase , atol=1e-4 ) )
341
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( __snake_case , unittest.TestCase ): """simple docstring""" _snake_case = KandinskyInpaintPipeline _snake_case = ['prompt', 'image_embeds', 'negative_image_embeds', 'image', 'mask_image'] _snake_case = [ 'prompt', 'negative_prompt', 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] _snake_case = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'negative_prompt', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] _snake_case = False @property def A__ ( self )-> Tuple: '''simple docstring''' return 32 @property def A__ ( self )-> str: '''simple docstring''' return 32 @property def A__ ( self )-> Union[str, Any]: '''simple docstring''' return self.time_input_dim @property def A__ ( self )-> Tuple: '''simple docstring''' return self.time_input_dim * 4 @property def A__ ( self )-> Any: '''simple docstring''' return 100 @property def A__ ( self )-> str: '''simple docstring''' __UpperCamelCase = XLMRobertaTokenizerFast.from_pretrained('''YiYiXu/tiny-random-mclip-base''' ) return tokenizer @property def A__ ( self )-> Tuple: '''simple docstring''' torch.manual_seed(0 ) __UpperCamelCase = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1005 , ) __UpperCamelCase = MultilingualCLIP(SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = text_encoder.eval() return text_encoder @property def A__ ( self )-> Dict: '''simple docstring''' torch.manual_seed(0 ) __UpperCamelCase = { '''in_channels''': 9, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''text_image''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''text_image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } __UpperCamelCase = UNetaDConditionModel(**SCREAMING_SNAKE_CASE_ ) return model @property def A__ ( self )-> Dict: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def A__ ( self )-> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) __UpperCamelCase = VQModel(**self.dummy_movq_kwargs ) return model def A__ ( self )-> Tuple: '''simple docstring''' __UpperCamelCase = self.dummy_text_encoder __UpperCamelCase = self.dummy_tokenizer __UpperCamelCase = self.dummy_unet __UpperCamelCase = self.dummy_movq __UpperCamelCase = DDIMScheduler( num_train_timesteps=1000 , beta_schedule='''linear''' , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=SCREAMING_SNAKE_CASE_ , set_alpha_to_one=SCREAMING_SNAKE_CASE_ , steps_offset=1 , prediction_type='''epsilon''' , thresholding=SCREAMING_SNAKE_CASE_ , ) __UpperCamelCase = { '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def A__ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=0 )-> List[str]: '''simple docstring''' __UpperCamelCase = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(SCREAMING_SNAKE_CASE_ ) ).to(SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(SCREAMING_SNAKE_CASE_ ) # create init_image __UpperCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(SCREAMING_SNAKE_CASE_ ) ).to(SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase = Image.fromarray(np.uinta(SCREAMING_SNAKE_CASE_ ) ).convert('''RGB''' ).resize((256, 256) ) # create mask __UpperCamelCase = np.ones((64, 64) , dtype=np.floataa ) __UpperCamelCase = 0 if str(SCREAMING_SNAKE_CASE_ ).startswith('''mps''' ): __UpperCamelCase = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) else: __UpperCamelCase = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = { '''prompt''': '''horse''', '''image''': init_image, '''mask_image''': mask, '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''generator''': generator, '''height''': 64, '''width''': 64, '''num_inference_steps''': 2, '''guidance_scale''': 4.0, '''output_type''': '''np''', } return inputs def A__ ( self )-> List[str]: '''simple docstring''' __UpperCamelCase = '''cpu''' __UpperCamelCase = self.get_dummy_components() __UpperCamelCase = self.pipeline_class(**SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = pipe(**self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) ) __UpperCamelCase = output.images __UpperCamelCase = pipe( **self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) , return_dict=SCREAMING_SNAKE_CASE_ , )[0] __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] print(F"image.shape {image.shape}" ) assert image.shape == (1, 64, 64, 3) __UpperCamelCase = np.array( [0.8_3_2_6_9_1_9, 0.7_3_7_9_0_4_6_7, 0.2_0_9_1_8_5_8_1, 0.9_3_0_9_6_1_2, 0.5_5_1_1_7_9_1, 0.4_3_7_1_3_3_2_8, 0.5_5_1_3_3_2_1, 0.4_9_9_2_2_9_3_4, 0.5_9_4_9_7_7_8_6] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" def A__ ( self )-> Optional[int]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def A__ ( self )-> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self )-> Tuple: '''simple docstring''' __UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy''' ) __UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinsky/cat.png''' ) __UpperCamelCase = np.ones((768, 768) , dtype=np.floataa ) __UpperCamelCase = 0 __UpperCamelCase = '''a hat''' __UpperCamelCase = KandinskyPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-1-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = KandinskyInpaintPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-1-inpaint''' , torch_dtype=torch.floataa ) __UpperCamelCase = pipeline.to(SCREAMING_SNAKE_CASE_ ) pipeline.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = torch.Generator(device='''cpu''' ).manual_seed(0 ) __UpperCamelCase , __UpperCamelCase = pipe_prior( SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , num_inference_steps=5 , negative_prompt='''''' , ).to_tuple() __UpperCamelCase = pipeline( SCREAMING_SNAKE_CASE_ , image=SCREAMING_SNAKE_CASE_ , mask_image=SCREAMING_SNAKE_CASE_ , image_embeds=SCREAMING_SNAKE_CASE_ , negative_image_embeds=SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , num_inference_steps=100 , height=768 , width=768 , output_type='''np''' , ) __UpperCamelCase = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )
328
'''simple docstring''' import argparse from collections import defaultdict def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = f"""class {class_name}(""" _snake_case = f"""{4 * " "}def {test_name}(""" _snake_case = f"""{8 * " "}{correct_line.split()[0]}""" _snake_case = f"""{16 * " "}{correct_line.split()[0]}""" _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = 0 _snake_case = 0 _snake_case = [] for line in lines: if line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and line.startswith(_SCREAMING_SNAKE_CASE ): _snake_case = True elif in_class and in_func and (line.startswith(_SCREAMING_SNAKE_CASE ) or line.startswith(_SCREAMING_SNAKE_CASE )): _snake_case = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: _snake_case = True if in_class and in_func and in_line: if ")" not in line: continue else: _snake_case = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) _snake_case = _snake_case = _snake_case = _snake_case = False else: new_lines.append(_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: for line in new_lines: f.write(_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): if fail is not None: with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = {l.strip() for l in f.readlines()} else: _snake_case = None with open(_SCREAMING_SNAKE_CASE , """r""" ) as f: _snake_case = f.readlines() _snake_case = defaultdict(_SCREAMING_SNAKE_CASE ) for line in correct_lines: _snake_case, _snake_case, _snake_case, _snake_case = line.split(""";""" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) __lowerCAmelCase = parser.parse_args() main(args.correct_filename, args.fail_filename)
341
0
"""simple docstring""" def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> Any: '''simple docstring''' return int((input_a, input_a).count(1 ) != 0 ) def _SCREAMING_SNAKE_CASE () -> Any: '''simple docstring''' assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
136
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) __lowerCAmelCase = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder __lowercase = """__DUMMY_TRANSFORMERS_USER__""" __lowercase = """Dummy User""" __lowercase = """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""" __lowercase = """https://hub-ci.huggingface.co""" __lowercase = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" __lowercase = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" __lowercase = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def lowercase ( A_ )-> Tuple: '''simple docstring''' monkeypatch.setattr( "huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE" , _SCREAMING_SNAKE_CASE ) @pytest.fixture def lowercase ( A_ )-> List[Any]: '''simple docstring''' monkeypatch.setattr("datasets.config.HF_ENDPOINT" , _SCREAMING_SNAKE_CASE ) monkeypatch.setattr("datasets.config.HUB_DATASETS_URL" , _SCREAMING_SNAKE_CASE ) @pytest.fixture def lowercase ( A_ )-> List[str]: '''simple docstring''' monkeypatch.setattr("huggingface_hub.hf_api.HfFolder.path_token" , _SCREAMING_SNAKE_CASE ) @pytest.fixture def lowercase ( A_ , A_ )-> Union[str, Any]: '''simple docstring''' HfFolder.save_token(_SCREAMING_SNAKE_CASE ) yield HfFolder.delete_token() @pytest.fixture(scope="session" ) def lowercase ( )-> Any: '''simple docstring''' return HfApi(endpoint=_SCREAMING_SNAKE_CASE ) @pytest.fixture(scope="session" ) def lowercase ( A_ )-> Dict: '''simple docstring''' a : Optional[Any] = HfFolder.get_token() HfFolder.save_token(_SCREAMING_SNAKE_CASE ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(_SCREAMING_SNAKE_CASE ) @pytest.fixture def lowercase ( A_ )-> Union[str, Any]: '''simple docstring''' def _cleanup_repo(A_ ): hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" ) return _cleanup_repo @pytest.fixture def lowercase ( A_ )-> int: '''simple docstring''' @contextmanager def _temporary_repo(A_ ): try: yield repo_id finally: cleanup_repo(_SCREAMING_SNAKE_CASE ) return _temporary_repo @pytest.fixture(scope="session" ) def lowercase ( A_ , A_ , A_ )-> List[Any]: '''simple docstring''' a : int = F'''repo_txt_data-{int(time.time() * 10e3 )}''' a : Tuple = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" , private=_SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=_SCREAMING_SNAKE_CASE , path_or_fileobj=str(_SCREAMING_SNAKE_CASE ) , path_in_repo="data/text_data.txt" , repo_id=_SCREAMING_SNAKE_CASE , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowercase ( A_ , A_ , A_ )-> Optional[int]: '''simple docstring''' return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="session" ) def lowercase ( A_ , A_ , A_ )-> str: '''simple docstring''' a : List[Any] = F'''repo_zipped_txt_data-{int(time.time() * 10e3 )}''' a : Any = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" , private=_SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=_SCREAMING_SNAKE_CASE , path_or_fileobj=str(_SCREAMING_SNAKE_CASE ) , path_in_repo="data.zip" , repo_id=_SCREAMING_SNAKE_CASE , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowercase ( A_ , A_ , A_ )-> Tuple: '''simple docstring''' return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="session" ) def lowercase ( A_ , A_ , A_ )-> Optional[Any]: '''simple docstring''' a : Optional[Any] = F'''repo_zipped_img_data-{int(time.time() * 10e3 )}''' a : List[Any] = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" , private=_SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=_SCREAMING_SNAKE_CASE , path_or_fileobj=str(_SCREAMING_SNAKE_CASE ) , path_in_repo="data.zip" , repo_id=_SCREAMING_SNAKE_CASE , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(_SCREAMING_SNAKE_CASE , token=_SCREAMING_SNAKE_CASE , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowercase ( A_ , A_ , A_ )-> Union[str, Any]: '''simple docstring''' return hf_private_dataset_repo_zipped_img_data_
40
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __lowerCAmelCase = logging.get_logger(__name__) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = ["pixel_values"] def __init__(self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = None , **UpperCAmelCase , ) -> None: super().__init__(**UpperCAmelCase ) _snake_case = size if size is not None else {"""height""": 256, """width""": 256} _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = do_resize _snake_case = size _snake_case = resample _snake_case = do_center_crop _snake_case = crop_size _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_normalize _snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PIL.Image.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( UpperCAmelCase , size=(size["""height"""], size["""width"""]) , resample=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: _snake_case = get_size_dict(UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(UpperCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> List[Any]: return rescale(UpperCAmelCase , scale=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: return normalize(UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase , data_format=UpperCAmelCase , **UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase=None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> PIL.Image.Image: _snake_case = do_resize if do_resize is not None else self.do_resize _snake_case = resample if resample is not None else self.resample _snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case = do_rescale if do_rescale is not None else self.do_rescale _snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case = do_normalize if do_normalize is not None else self.do_normalize _snake_case = image_mean if image_mean is not None else self.image_mean _snake_case = image_std if image_std is not None else self.image_std _snake_case = size if size is not None else self.size _snake_case = get_size_dict(UpperCAmelCase ) _snake_case = crop_size if crop_size is not None else self.crop_size _snake_case = get_size_dict(UpperCAmelCase , param_name="""crop_size""" ) _snake_case = make_list_of_images(UpperCAmelCase ) if not valid_images(UpperCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. _snake_case = [to_numpy_array(UpperCAmelCase ) for image in images] if do_resize: _snake_case = [self.resize(image=UpperCAmelCase , size=UpperCAmelCase , resample=UpperCAmelCase ) for image in images] if do_center_crop: _snake_case = [self.center_crop(image=UpperCAmelCase , size=UpperCAmelCase ) for image in images] if do_rescale: _snake_case = [self.rescale(image=UpperCAmelCase , scale=UpperCAmelCase ) for image in images] if do_normalize: _snake_case = [self.normalize(image=UpperCAmelCase , mean=UpperCAmelCase , std=UpperCAmelCase ) for image in images] _snake_case = [to_channel_dimension_format(UpperCAmelCase , UpperCAmelCase ) for image in images] _snake_case = {"""pixel_values""": images} return BatchFeature(data=UpperCAmelCase , tensor_type=UpperCAmelCase )
341
0
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class _A ( unittest.TestCase ): def __a ( self : List[str] ) -> Any: """simple docstring""" lowercase : str = 10 def __a ( self : str ) -> List[Any]: """simple docstring""" lowercase : Optional[Any] = [1, 2, 3, 4] lowercase : List[str] = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(_A , self.block_size , 0 ) , _A ) def __a ( self : Union[str, Any] ) -> List[str]: """simple docstring""" lowercase : Optional[int] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] lowercase : Optional[int] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_A , self.block_size , 0 ) , _A ) def __a ( self : Tuple ) -> Optional[int]: """simple docstring""" lowercase : Optional[int] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] lowercase : Union[str, Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(_A , self.block_size , 0 ) , _A ) def __a ( self : Union[str, Any] ) -> int: """simple docstring""" lowercase : int = '''It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.''' lowercase , lowercase : Optional[Any] = process_story(_A ) self.assertEqual(_A , [] ) def __a ( self : Optional[Any] ) -> List[Any]: """simple docstring""" lowercase : Union[str, Any] = '''''' lowercase , lowercase : int = process_story(_A ) self.assertEqual(_A , [] ) self.assertEqual(_A , [] ) def __a ( self : Union[str, Any] ) -> Tuple: """simple docstring""" lowercase : Union[str, Any] = ( '''It was the year of Our Lord one thousand seven hundred and ''' '''seventy-five\n\nSpiritual revelations were conceded to England ''' '''at that favoured period, as at this.\n@highlight\n\nIt was the best of times''' ) lowercase , lowercase : List[Any] = process_story(_A ) lowercase : List[str] = [ '''It was the year of Our Lord one thousand seven hundred and seventy-five.''', '''Spiritual revelations were conceded to England at that favoured period, as at this.''', ] self.assertEqual(_A , _A ) lowercase : Union[str, Any] = ['''It was the best of times.'''] self.assertEqual(_A , _A ) def __a ( self : List[str] ) -> int: """simple docstring""" lowercase : str = torch.tensor([1, 2, 3, 4] ) lowercase : Any = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(_A , 0 ).numpy() , expected.numpy() ) def __a ( self : int ) -> Union[str, Any]: """simple docstring""" lowercase : List[str] = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) lowercase : Tuple = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(_A , 23 ).numpy() , expected.numpy() ) def __a ( self : Union[str, Any] ) -> str: """simple docstring""" lowercase : Dict = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) lowercase : Optional[Any] = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(_A , 1 ).numpy() , expected.numpy() ) def __a ( self : Any ) -> Optional[int]: """simple docstring""" lowercase : Union[str, Any] = 101 lowercase : List[str] = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]] ) lowercase : List[Any] = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) lowercase : str = compute_token_type_ids(_A , _A ) np.testing.assert_array_equal(_A , _A )
308
'''simple docstring''' __lowerCAmelCase = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure the supplied data is a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_SCREAMING_SNAKE_CASE ) _snake_case = """""".join(bin(_SCREAMING_SNAKE_CASE )[2:].zfill(8 ) for byte in data ) _snake_case = len(_SCREAMING_SNAKE_CASE ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b"""=""" * ((6 - len(_SCREAMING_SNAKE_CASE ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_SCREAMING_SNAKE_CASE ) % 6) else: _snake_case = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 6 ) ).encode() + padding ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _snake_case = ( """argument should be a bytes-like object or ASCII string, """ f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_SCREAMING_SNAKE_CASE ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): try: _snake_case = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) _snake_case = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_SCREAMING_SNAKE_CASE ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = """""".join( bin(B64_CHARSET.index(_SCREAMING_SNAKE_CASE ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_SCREAMING_SNAKE_CASE ) , 8 ) ] return bytes(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ : Optional[Any] = 16 A_ : int = 32 def A ( snake_case__ , snake_case__ = 16 ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained("""bert-base-cased""" ) SCREAMING_SNAKE_CASE__ = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(snake_case__ ): # max_length=None => use the model max length (it's actually the default) SCREAMING_SNAKE_CASE__ = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): SCREAMING_SNAKE_CASE__ = datasets.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library SCREAMING_SNAKE_CASE__ = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case__ ): # On TPU it's best to pad everything to the same length or training will be very slow. SCREAMING_SNAKE_CASE__ = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": SCREAMING_SNAKE_CASE__ = 16 elif accelerator.mixed_precision != "no": SCREAMING_SNAKE_CASE__ = 8 else: SCREAMING_SNAKE_CASE__ = None return tokenizer.pad( _SCREAMING_SNAKE_CASE , padding="""longest""" , max_length=_SCREAMING_SNAKE_CASE , pad_to_multiple_of=_SCREAMING_SNAKE_CASE , return_tensors="""pt""" , ) # Instantiate dataloaders. SCREAMING_SNAKE_CASE__ = DataLoader( tokenized_datasets["""train"""] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ = DataLoader( tokenized_datasets["""validation"""] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", None) == "1": from accelerate.test_utils.training import mocked_dataloaders A_ : List[Any] = mocked_dataloaders # noqa: F811 def A ( snake_case__ , snake_case__ ): '''simple docstring''' if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , _SCREAMING_SNAKE_CASE ) == "1": SCREAMING_SNAKE_CASE__ = 2 # New Code # SCREAMING_SNAKE_CASE__ = int(args.gradient_accumulation_steps ) # Initialize accelerator SCREAMING_SNAKE_CASE__ = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_SCREAMING_SNAKE_CASE ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( """Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`""" ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs SCREAMING_SNAKE_CASE__ = config["""lr"""] SCREAMING_SNAKE_CASE__ = int(config["""num_epochs"""] ) SCREAMING_SNAKE_CASE__ = int(config["""seed"""] ) SCREAMING_SNAKE_CASE__ = int(config["""batch_size"""] ) SCREAMING_SNAKE_CASE__ = evaluate.load("""glue""" , """mrpc""" ) set_seed(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_dataloaders(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) SCREAMING_SNAKE_CASE__ = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=_SCREAMING_SNAKE_CASE ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). SCREAMING_SNAKE_CASE__ = model.to(accelerator.device ) # Instantiate optimizer SCREAMING_SNAKE_CASE__ = AdamW(params=model.parameters() , lr=_SCREAMING_SNAKE_CASE ) # Instantiate scheduler SCREAMING_SNAKE_CASE__ = get_linear_schedule_with_warmup( optimizer=_SCREAMING_SNAKE_CASE , num_warmup_steps=1_00 , num_training_steps=(len(_SCREAMING_SNAKE_CASE ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = accelerator.prepare( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Now we train the model for epoch in range(_SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_SCREAMING_SNAKE_CASE ): SCREAMING_SNAKE_CASE__ = model(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ = output.loss accelerator.backward(_SCREAMING_SNAKE_CASE ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ = outputs.logits.argmax(dim=-1 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=_SCREAMING_SNAKE_CASE , references=_SCREAMING_SNAKE_CASE , ) SCREAMING_SNAKE_CASE__ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , _SCREAMING_SNAKE_CASE ) def A ( ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) # New Code # parser.add_argument( """--gradient_accumulation_steps""" , type=_SCREAMING_SNAKE_CASE , default=1 , help="""The number of minibatches to be ran before gradients are accumulated.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args() SCREAMING_SNAKE_CASE__ = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
165
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
341
0
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Optional[int] ) -> Any: '''simple docstring''' if len(_SCREAMING_SNAKE_CASE ) == 0: return [] A__ , A__ = min(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) A__ = int(max_value - min_value ) + 1 A__ = [[] for _ in range(_SCREAMING_SNAKE_CASE )] for i in my_list: buckets[int(i - min_value )].append(_SCREAMING_SNAKE_CASE ) return [v for bucket in buckets for v in sorted(_SCREAMING_SNAKE_CASE )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -1_0, 1_5, 2, -2]) == [-1_0, -2, 0, 1, 2, 1_5]
68
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
'''simple docstring''' import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] , model_result['ss'] ): _snake_case = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sgugger/tiny-distilbert-classification' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , only_pretrain_model=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , torchscript=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , fpaa=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = AutoConfig.from_pretrained(lowerCAmelCase_ ) # set architectures equal to `None` _snake_case = None _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == 'cpu' , 'Can\'t do half precision' ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , fpaa=lowerCAmelCase_ , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = AutoConfig.from_pretrained(lowerCAmelCase_ ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tinier_bart' _snake_case = AutoConfig.from_pretrained(lowerCAmelCase_ ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' _snake_case = AutoConfig.from_pretrained(lowerCAmelCase_ ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tinier_bart' _snake_case = AutoConfig.from_pretrained(lowerCAmelCase_ ) _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ , configs=[config] ) _snake_case = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , save_to_csv=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(lowerCAmelCase_ , 'inf_time.csv' ) , train_memory_csv_file=os.path.join(lowerCAmelCase_ , 'train_mem.csv' ) , inference_memory_csv_file=os.path.join(lowerCAmelCase_ , 'inf_mem.csv' ) , train_time_csv_file=os.path.join(lowerCAmelCase_ , 'train_time.csv' ) , env_info_csv_file=os.path.join(lowerCAmelCase_ , 'env.csv' ) , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) benchmark.run() self.assertTrue(Path(os.path.join(lowerCAmelCase_ , 'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowerCAmelCase_ , 'train_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowerCAmelCase_ , 'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowerCAmelCase_ , 'train_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(lowerCAmelCase_ , 'env.csv' ) ).exists() ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(lowerCAmelCase_ ): self.assertTrue(hasattr(lowerCAmelCase_ , 'sequential' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'cumulative' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'current' ) ) self.assertTrue(hasattr(lowerCAmelCase_ , 'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: _snake_case = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=lowerCAmelCase_ , inference=lowerCAmelCase_ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(lowerCAmelCase_ , 'log.txt' ) , log_print=lowerCAmelCase_ , trace_memory_line_by_line=lowerCAmelCase_ , multi_process=lowerCAmelCase_ , ) _snake_case = PyTorchBenchmark(lowerCAmelCase_ ) _snake_case = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(lowerCAmelCase_ , 'log.txt' ) ).exists() )
42
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __lowerCAmelCase = TypeVar('T') __lowerCAmelCase = Union[List[T], Tuple[T, ...]] __lowerCAmelCase = Union[T, List[T], Dict[str, T]] __lowerCAmelCase = Union[str, bytes, os.PathLike]
341
0
import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments a__ = logging.getLogger(__name__) @dataclass class UpperCAmelCase_ ( __snake_case ): """simple docstring""" UpperCAmelCase__ : List[str] = field( default=0.0 , metadata={"help": "The label smoothing epsilon to apply (if not zero)."} ) UpperCAmelCase__ : Tuple = field(default=__snake_case , metadata={"help": "Whether to SortishSamler or not."} ) UpperCAmelCase__ : List[Any] = field( default=__snake_case , metadata={"help": "Whether to use generate to calculate generative metrics (ROUGE, BLEU)."} ) UpperCAmelCase__ : int = field(default=__snake_case , metadata={"help": "whether to use adafactor"} ) UpperCAmelCase__ : Union[str, Any] = field( default=__snake_case , metadata={"help": "Encoder layer dropout probability. Goes into model.config."} ) UpperCAmelCase__ : Union[str, Any] = field( default=__snake_case , metadata={"help": "Decoder layer dropout probability. Goes into model.config."} ) UpperCAmelCase__ : str = field(default=__snake_case , metadata={"help": "Dropout probability. Goes into model.config."} ) UpperCAmelCase__ : Optional[Any] = field( default=__snake_case , metadata={"help": "Attention dropout probability. Goes into model.config."} ) UpperCAmelCase__ : Dict = field( default="linear" , metadata={"help": F'''Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}'''} , )
235
'''simple docstring''' class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=None ) -> int: _snake_case = data _snake_case = previous _snake_case = next_node def __str__(self ) -> str: return f"""{self.data}""" def lowercase (self ) -> int: return self.data def lowercase (self ) -> Dict: return self.next def lowercase (self ) -> Union[str, Any]: return self.previous class _lowerCAmelCase : '''simple docstring''' def __init__(self , UpperCAmelCase ) -> List[str]: _snake_case = head def __iter__(self ) -> Optional[Any]: return self def lowercase (self ) -> str: if not self.current: raise StopIteration else: _snake_case = self.current.get_data() _snake_case = self.current.get_next() return value class _lowerCAmelCase : '''simple docstring''' def __init__(self ) -> Optional[int]: _snake_case = None # First node in list _snake_case = None # Last node in list def __str__(self ) -> Optional[int]: _snake_case = self.head _snake_case = [] while current is not None: nodes.append(current.get_data() ) _snake_case = current.get_next() return " ".join(str(UpperCAmelCase ) for node in nodes ) def __contains__(self , UpperCAmelCase ) -> int: _snake_case = self.head while current: if current.get_data() == value: return True _snake_case = current.get_next() return False def __iter__(self ) -> Union[str, Any]: return LinkedListIterator(self.head ) def lowercase (self ) -> str: if self.head: return self.head.get_data() return None def lowercase (self ) -> List[Any]: if self.tail: return self.tail.get_data() return None def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: _snake_case = node _snake_case = node else: self.insert_before_node(self.head , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: if self.head is None: self.set_head(UpperCAmelCase ) else: self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> None: _snake_case = Node(UpperCAmelCase ) if self.head is None: self.set_head(UpperCAmelCase ) else: self.set_tail(UpperCAmelCase ) def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.previous if node.get_previous() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = node _snake_case = node.next if node.get_next() is None: _snake_case = node_to_insert else: _snake_case = node_to_insert _snake_case = node_to_insert def lowercase (self , UpperCAmelCase , UpperCAmelCase ) -> None: _snake_case = 1 _snake_case = Node(UpperCAmelCase ) _snake_case = self.head while node: if current_position == position: self.insert_before_node(UpperCAmelCase , UpperCAmelCase ) return current_position += 1 _snake_case = node.next self.insert_after_node(self.tail , UpperCAmelCase ) def lowercase (self , UpperCAmelCase ) -> Node: _snake_case = self.head while node: if node.get_data() == item: return node _snake_case = node.get_next() raise Exception("""Node not found""" ) def lowercase (self , UpperCAmelCase ) -> Optional[int]: if (node := self.get_node(UpperCAmelCase )) is not None: if node == self.head: _snake_case = self.head.get_next() if node == self.tail: _snake_case = self.tail.get_previous() self.remove_node_pointers(UpperCAmelCase ) @staticmethod def lowercase (UpperCAmelCase ) -> None: if node.get_next(): _snake_case = node.previous if node.get_previous(): _snake_case = node.next _snake_case = None _snake_case = None def lowercase (self ) -> Dict: return self.head is None def __SCREAMING_SNAKE_CASE ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
341
0
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> Union[str, Any]: try: _lowercase : int = os.environ[key] except KeyError: # KEY isn't set, default to `default`. _lowercase : List[str] = default else: # KEY is set, convert it to True or False. try: _lowercase : List[Any] = strtobool(_SCREAMING_SNAKE_CASE ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F'''If set, {key} must be yes or no.''' ) return _value SCREAMING_SNAKE_CASE : List[str] = parse_flag_from_env("RUN_SLOW", default=False) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: return unittest.skip('Test was skipped' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: return unittest.skipUnless(_run_slow_tests , 'test is slow' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> str: return unittest.skipUnless(not torch.cuda.is_available() , 'test requires only a CPU' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: return unittest.skipUnless(torch.cuda.is_available() , 'test requires a GPU' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: return unittest.skipUnless(is_xpu_available() , 'test requires a XPU' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: return unittest.skipUnless(is_mps_available() , 'test requires a `mps` backend support in `torch`' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: return unittest.skipUnless( is_transformers_available() and is_datasets_available() , 'test requires the Hugging Face suite' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: return unittest.skipUnless(is_bnb_available() , 'test requires the bitsandbytes library' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: return unittest.skipUnless(is_tpu_available() , 'test requires TPU' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> str: return unittest.skipUnless(torch.cuda.device_count() == 1 , 'test requires a GPU' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: return unittest.skipUnless(torch.xpu.device_count() == 1 , 'test requires a XPU' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> int: return unittest.skipUnless(torch.cuda.device_count() > 1 , 'test requires multiple GPUs' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> int: return unittest.skipUnless(torch.xpu.device_count() > 1 , 'test requires multiple XPUs' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: return unittest.skipUnless(is_safetensors_available() , 'test requires safetensors' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: return unittest.skipUnless(is_deepspeed_available() , 'test requires DeepSpeed' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: return unittest.skipUnless(is_torch_version('>=' , '1.12.0' ) , 'test requires torch version >= 1.12.0' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_=None , lowerCamelCase_=None ) -> Optional[Any]: if test_case is None: return partial(_SCREAMING_SNAKE_CASE , version=_SCREAMING_SNAKE_CASE ) return unittest.skipUnless(is_torch_version('>=' , _SCREAMING_SNAKE_CASE ) , F'''test requires torch version >= {version}''' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: return unittest.skipUnless(is_tensorboard_available() , 'test requires Tensorboard' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: return unittest.skipUnless(is_wandb_available() , 'test requires wandb' )(_SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: return unittest.skipUnless(is_comet_ml_available() , 'test requires comet_ml' )(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Tuple = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: return unittest.skipUnless( _atleast_one_tracker_available , 'test requires at least one tracker to be available and for `comet_ml` to not be installed' , )(_SCREAMING_SNAKE_CASE ) class _lowerCamelCase( unittest.TestCase ): lowercase_ : Optional[int] = True @classmethod def UpperCamelCase ( cls) -> Optional[Any]: """simple docstring""" _lowercase : str = tempfile.mkdtemp() @classmethod def UpperCamelCase ( cls) -> Optional[int]: """simple docstring""" if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" if self.clear_on_setup: for path in Path(self.tmpdir).glob('**/*'): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(lowerCamelCase) class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Dict: """simple docstring""" super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self, lowerCamelCase) -> Tuple: """simple docstring""" _lowercase : Union[str, Any] = mocks if isinstance(lowerCamelCase, (tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: _lowercase : Optional[int] = AcceleratorState() _lowercase : List[Any] = tensor[None].clone().to(state.device ) _lowercase : int = gather(_SCREAMING_SNAKE_CASE ).cpu() _lowercase : Union[str, Any] = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , _SCREAMING_SNAKE_CASE ): return False return True class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : List[Any] = returncode _lowercase : int = stdout _lowercase : Union[str, Any] = stderr async def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> int: while True: _lowercase : Optional[int] = await stream.readline() if line: callback(_SCREAMING_SNAKE_CASE ) else: break async def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=False , lowerCamelCase_=False ) -> int: if echo: print('\nRunning: ' , ' '.join(_SCREAMING_SNAKE_CASE ) ) _lowercase : Optional[Any] = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_SCREAMING_SNAKE_CASE , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_SCREAMING_SNAKE_CASE , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) _lowercase : Any = [] _lowercase : List[str] = [] def tee(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_="" ): _lowercase : List[Any] = line.decode('utf-8' ).rstrip() sink.append(_SCREAMING_SNAKE_CASE ) if not quiet: print(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , file=_SCREAMING_SNAKE_CASE ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowerCamelCase_ : tee(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , sys.stdout , label='stdout:' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowerCamelCase_ : tee(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , sys.stderr , label='stderr:' ) ) ), ] , timeout=_SCREAMING_SNAKE_CASE , ) return _RunOutput(await p.wait() , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=180 , lowerCamelCase_=False , lowerCamelCase_=True ) -> Dict: _lowercase : Optional[Any] = asyncio.get_event_loop() _lowercase : int = loop.run_until_complete( _stream_subprocess(_SCREAMING_SNAKE_CASE , env=_SCREAMING_SNAKE_CASE , stdin=_SCREAMING_SNAKE_CASE , timeout=_SCREAMING_SNAKE_CASE , quiet=_SCREAMING_SNAKE_CASE , echo=_SCREAMING_SNAKE_CASE ) ) _lowercase : List[str] = ' '.join(_SCREAMING_SNAKE_CASE ) if result.returncode > 0: _lowercase : int = '\n'.join(result.stderr ) raise RuntimeError( F'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' F'''The combined stderr from workers follows:\n{stderr}''' ) return result class _lowerCamelCase( __snake_case ): pass def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> Dict: try: _lowercase : Dict = subprocess.check_output(_SCREAMING_SNAKE_CASE , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(_SCREAMING_SNAKE_CASE , 'decode' ): _lowercase : Dict = output.decode('utf-8' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F'''Command `{" ".join(_SCREAMING_SNAKE_CASE )}` failed with the following error:\n\n{e.output.decode()}''' ) from e
21
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput __lowerCAmelCase = 8 def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x * 255).int().clamp(0 , 255 ) _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c h w -> b c 1 h w""" ) _snake_case = ((x & mask) != 0).float() _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b c d h w -> b (c d) h w""" ) _snake_case = bits * 2 - 1 return bits def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ): _snake_case = x.device _snake_case = (x > 0).int() _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE , dtype=torch.intaa ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """d -> d 1 1""" ) _snake_case = rearrange(_SCREAMING_SNAKE_CASE , """b (c d) h w -> b c d h w""" , d=8 ) _snake_case = reduce(x * mask , """b c d h w -> b c h w""" , """sum""" ) return (dec / 255).clamp(0.0 , 1.0 ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _snake_case = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _snake_case = self.alphas_cumprod[timestep] _snake_case = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _snake_case = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _snake_case = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _snake_case = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _snake_case = model_output.device if torch.is_tensor(_SCREAMING_SNAKE_CASE ) else """cpu""" _snake_case = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) _snake_case = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise _snake_case = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="epsilon" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ): _snake_case = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _snake_case, _snake_case = torch.split(_SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: _snake_case = None # 1. compute alphas, betas _snake_case = self.alphas_cumprod[t] _snake_case = self.alphas_cumprod[t - 1] if t > 0 else self.one _snake_case = 1 - alpha_prod_t _snake_case = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _snake_case = model_output else: raise ValueError(f"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _snake_case = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _snake_case = 0 if t > 0: _snake_case = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_SCREAMING_SNAKE_CASE ).to(model_output.device ) _snake_case = (self._get_variance(_SCREAMING_SNAKE_CASE , predicted_variance=_SCREAMING_SNAKE_CASE ) ** 0.5) * noise _snake_case = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) class _lowerCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = 1.0 , ) -> Tuple: super().__init__() _snake_case = bit_scale _snake_case = ( ddim_bit_scheduler_step if isinstance(UpperCAmelCase , UpperCAmelCase ) else ddpm_bit_scheduler_step ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) @torch.no_grad() def __call__(self , UpperCAmelCase = 256 , UpperCAmelCase = 256 , UpperCAmelCase = 50 , UpperCAmelCase = None , UpperCAmelCase = 1 , UpperCAmelCase = "pil" , UpperCAmelCase = True , **UpperCAmelCase , ) -> Union[Tuple, ImagePipelineOutput]: _snake_case = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=UpperCAmelCase , ) _snake_case = decimal_to_bits(UpperCAmelCase ) * self.bit_scale _snake_case = latents.to(self.device ) self.scheduler.set_timesteps(UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _snake_case = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _snake_case = self.scheduler.step(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ).prev_sample _snake_case = bits_to_decimal(UpperCAmelCase ) if output_type == "pil": _snake_case = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase )
341
0
'''simple docstring''' def a ( __a , __a , __a , __a ) -> int: '''simple docstring''' if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def a ( __a , __a ) -> Tuple: '''simple docstring''' print('''moving disk from''' , _SCREAMING_SNAKE_CASE , '''to''' , _SCREAMING_SNAKE_CASE ) def a ( ) -> Any: '''simple docstring''' UpperCamelCase__ :Dict = int(input('''Height of hanoi: ''' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , '''A''' , '''B''' , '''C''' ) if __name__ == "__main__": main()
97
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 10**9 ): _snake_case = 1 _snake_case = 2 _snake_case = 0 _snake_case = 0 _snake_case = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value _snake_case = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f'''{solution() = }''')
341
0
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) A : str = "\\n Text data.\n Second line of data." A : Dict = "file" @pytest.fixture(scope="session" ) def lowercase_ ( _A : List[str] ): """simple docstring""" lowerCamelCase__ : Union[str, Any] = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") lowerCamelCase__ : List[Any] = bytes(_SCREAMING_SNAKE_CASE , "utf-8" ) with zstd.open(_SCREAMING_SNAKE_CASE , "wb" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return path @pytest.fixture def lowercase_ ( _A : List[Any] ): """simple docstring""" with open(os.path.join(tmpfs.local_root_dir , _SCREAMING_SNAKE_CASE ) , "w" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowercase_ ( _A : str , _A : Union[str, Any] , _A : Tuple , _A : Union[str, Any] , _A : str , _A : int ): """simple docstring""" lowerCamelCase__ : Union[str, Any] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} lowerCamelCase__ : Any = input_paths[compression_format] lowerCamelCase__ : Optional[Any] = tmp_path / "cache" lowerCamelCase__ : Dict = DownloadConfig(cache_dir=_SCREAMING_SNAKE_CASE , extract_compressed_file=_SCREAMING_SNAKE_CASE ) lowerCamelCase__ : Optional[Any] = cached_path(_SCREAMING_SNAKE_CASE , download_config=_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE ) as f: lowerCamelCase__ : Union[str, Any] = f.read() with open(_SCREAMING_SNAKE_CASE ) as f: lowerCamelCase__ : str = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowercase_ ( _A : Union[str, Any] , _A : Any , _A : List[str] , _A : Any , _A : Union[str, Any] ): """simple docstring""" lowerCamelCase__ : Any = "custom_cache" lowerCamelCase__ : Optional[Any] = "custom_extracted_dir" lowerCamelCase__ : Tuple = tmp_path / "custom_extracted_path" if default_extracted: lowerCamelCase__ : str = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , _SCREAMING_SNAKE_CASE ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(_SCREAMING_SNAKE_CASE ) ) lowerCamelCase__ : str = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) lowerCamelCase__ : Optional[int] = xz_file lowerCamelCase__ : Optional[Any] = ( DownloadConfig(extract_compressed_file=_SCREAMING_SNAKE_CASE ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=_SCREAMING_SNAKE_CASE ) ) lowerCamelCase__ : str = cached_path(_SCREAMING_SNAKE_CASE , download_config=_SCREAMING_SNAKE_CASE ) assert Path(_SCREAMING_SNAKE_CASE ).parent.parts[-2:] == expected def lowercase_ ( _A : List[Any] ): """simple docstring""" lowerCamelCase__ : Tuple = str(Path(_SCREAMING_SNAKE_CASE ).resolve() ) assert cached_path(_SCREAMING_SNAKE_CASE ) == text_file # relative path lowerCamelCase__ : Tuple = str(Path(_SCREAMING_SNAKE_CASE ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(_SCREAMING_SNAKE_CASE ) == text_file def lowercase_ ( _A : Any ): """simple docstring""" lowerCamelCase__ : Tuple = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(_SCREAMING_SNAKE_CASE ): cached_path(_SCREAMING_SNAKE_CASE ) # relative path lowerCamelCase__ : Any = "./__missing_file__.txt" with pytest.raises(_SCREAMING_SNAKE_CASE ): cached_path(_SCREAMING_SNAKE_CASE ) def lowercase_ ( _A : int ): """simple docstring""" lowerCamelCase__ : Optional[int] = get_from_cache(F"tmp://{tmpfs_file}" ) with open(_SCREAMING_SNAKE_CASE ) as f: lowerCamelCase__ : List[Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , _SCREAMING_SNAKE_CASE ) def lowercase_ ( ): """simple docstring""" with pytest.raises(_SCREAMING_SNAKE_CASE ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , _SCREAMING_SNAKE_CASE ) def lowercase_ ( _A : Optional[int] ): """simple docstring""" lowerCamelCase__ : int = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(_SCREAMING_SNAKE_CASE ): http_get("https://huggingface.co" , temp_file=_SCREAMING_SNAKE_CASE ) with pytest.raises(_SCREAMING_SNAKE_CASE ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , _SCREAMING_SNAKE_CASE ) def lowercase_ ( _A : Union[str, Any] ): """simple docstring""" lowerCamelCase__ : str = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(_SCREAMING_SNAKE_CASE ): ftp_get("ftp://huggingface.co" , temp_file=_SCREAMING_SNAKE_CASE ) with pytest.raises(_SCREAMING_SNAKE_CASE ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , _SCREAMING_SNAKE_CASE ) def lowercase_ ( _A : int ): """simple docstring""" lowerCamelCase__ : Optional[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(_SCREAMING_SNAKE_CASE ): fsspec_get("s3://huggingface.co" , temp_file=_SCREAMING_SNAKE_CASE ) with pytest.raises(_SCREAMING_SNAKE_CASE ): fsspec_head("s3://huggingface.co" )
184
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class _lowerCAmelCase ( __snake_case ): '''simple docstring''' lowerCAmelCase_ = "deberta-v2" def __init__(self , UpperCAmelCase=128100 , UpperCAmelCase=1536 , UpperCAmelCase=24 , UpperCAmelCase=24 , UpperCAmelCase=6144 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=0 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-7 , UpperCAmelCase=False , UpperCAmelCase=-1 , UpperCAmelCase=0 , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=0 , UpperCAmelCase="gelu" , **UpperCAmelCase , ) -> List[str]: super().__init__(**UpperCAmelCase ) _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = max_position_embeddings _snake_case = type_vocab_size _snake_case = initializer_range _snake_case = relative_attention _snake_case = max_relative_positions _snake_case = pad_token_id _snake_case = position_biased_input # Backwards compatibility if type(UpperCAmelCase ) == str: _snake_case = [x.strip() for x in pos_att_type.lower().split("""|""" )] _snake_case = pos_att_type _snake_case = vocab_size _snake_case = layer_norm_eps _snake_case = kwargs.get("""pooler_hidden_size""" , UpperCAmelCase ) _snake_case = pooler_dropout _snake_case = pooler_hidden_act class _lowerCAmelCase ( __snake_case ): '''simple docstring''' @property def lowercase (self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _snake_case = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _snake_case = {0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def lowercase (self ) -> int: return 12 def lowercase (self , UpperCAmelCase , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = -1 , UpperCAmelCase = False , UpperCAmelCase = None , UpperCAmelCase = 3 , UpperCAmelCase = 40 , UpperCAmelCase = 40 , UpperCAmelCase = None , ) -> Mapping[str, Any]: _snake_case = super().generate_dummy_inputs(preprocessor=UpperCAmelCase , framework=UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
0