sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
sequencelengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
sequencelengths 0
201
| languages
sequencelengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
sequencelengths 0
722
| processed_texts
sequencelengths 1
723
| tokens_length
sequencelengths 1
723
| input_texts
sequencelengths 1
61
| embeddings
sequencelengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null | transformers | This model is a model that has successfully merged without warnings from Mergekit. I hope quantization works.
Merging and quantization were successful, but of course the results were not good. | {} | text-generation | rhplus0831/maid-yuzu-v6 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T15:18:09+00:00 | [] | [] | TAGS
#transformers #safetensors #mixtral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| This model is a model that has successfully merged without warnings from Mergekit. I hope quantization works.
Merging and quantization were successful, but of course the results were not good. | [] | [
"TAGS\n#transformers #safetensors #mixtral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
51
] | [
"passage: TAGS\n#transformers #safetensors #mixtral #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.005800837650895119,
-0.0037230453453958035,
-0.006158282980322838,
-0.030133750289678574,
0.14476196467876434,
-0.0036664605140686035,
0.16229654848575592,
0.09592274576425552,
-0.009181147441267967,
-0.006152432411909103,
0.13547320663928986,
0.15334069728851318,
-0.027436882257461548,
0.11783536523580551,
-0.15286561846733093,
-0.19672618806362152,
0.1047123447060585,
-0.007288352120667696,
0.023495066910982132,
0.0933275893330574,
0.10272379219532013,
-0.06828339397907257,
0.07095541059970856,
-0.05726435035467148,
-0.1245780736207962,
0.04088548198342323,
0.05649738386273384,
-0.12458145618438721,
0.11081969738006592,
0.08184432238340378,
0.12858566641807556,
0.04913971945643425,
-0.07726526260375977,
-0.21997801959514618,
0.04154593124985695,
0.009938947856426239,
-0.06427451223134995,
0.02193874679505825,
0.04838162660598755,
-0.11921676248311996,
0.038488876074552536,
0.041708290576934814,
0.00025608809664845467,
0.086815744638443,
-0.1643691211938858,
0.014961781911551952,
-0.01654316671192646,
-0.04934395104646683,
0.11313924938440323,
0.07268321514129639,
-0.04371236637234688,
0.13648377358913422,
-0.04871951416134834,
0.1330050528049469,
0.05804912745952606,
-0.3166044354438782,
-0.009717185981571674,
0.09097786247730255,
0.0653049498796463,
0.0853312686085701,
-0.054129622876644135,
0.08406248688697815,
0.0695362463593483,
-0.01498272456228733,
0.007709707133471966,
-0.06929977238178253,
-0.09682584553956985,
0.024727175012230873,
-0.07560402154922485,
-0.0025499321054667234,
0.2124347984790802,
-0.03343690186738968,
0.052581146359443665,
-0.10072332620620728,
-0.11837567389011383,
-0.00758837116882205,
-0.03918151929974556,
-0.019034985452890396,
-0.06400299072265625,
0.09653863310813904,
0.02677595056593418,
-0.033258724957704544,
-0.1271834671497345,
0.006537920795381069,
-0.17772991955280304,
0.16864904761314392,
-0.0058672791346907616,
0.023905333131551743,
-0.1930292695760727,
0.026287289336323738,
0.04555031657218933,
-0.10524080693721771,
0.049706876277923584,
-0.09550052881240845,
-0.008824318647384644,
-0.030920200049877167,
-0.0563303679227829,
-0.17647072672843933,
0.13722306489944458,
0.09190934151411057,
-0.016288211569190025,
0.056490734219551086,
-0.10322391986846924,
0.06043196842074394,
0.048489172011613846,
0.03301643207669258,
0.04097973182797432,
-0.08371305465698242,
0.04018532112240791,
-0.09741508215665817,
0.038392484188079834,
-0.06434754282236099,
-0.1361418515443802,
0.02236449345946312,
0.04444461315870285,
0.11199680715799332,
0.000050260667194379494,
0.12048058956861496,
-0.05138117074966431,
0.03758550062775612,
0.020058194175362587,
-0.061249226331710815,
-0.004423743579536676,
0.02875768579542637,
0.06378965079784393,
0.06584100425243378,
-0.018303319811820984,
0.05813997611403465,
-0.06447984278202057,
0.015193567611277103,
-0.05934791639447212,
-0.032304611057043076,
-0.029288399964571,
-0.09225255250930786,
0.02197396382689476,
-0.012736152857542038,
0.0006801231065765023,
-0.1881139874458313,
-0.15279826521873474,
-0.014857254922389984,
-0.00008858469664119184,
-0.030219294130802155,
0.01590142771601677,
-0.09135869890451431,
-0.008905895054340363,
0.058585863560438156,
-0.05914311483502388,
-0.08751774579286575,
-0.0761396661400795,
0.0745251253247261,
-0.018931875005364418,
0.06884735822677612,
-0.1266237199306488,
0.05243972688913345,
-0.11282125115394592,
0.013297900557518005,
-0.08208373934030533,
0.09385782480239868,
-0.014601340517401695,
0.1587718278169632,
-0.018055090680718422,
0.04486407712101936,
-0.08716477453708649,
0.08336792886257172,
-0.03145917132496834,
0.24287419021129608,
-0.13750557601451874,
-0.06613865494728088,
0.24425853788852692,
-0.10628853738307953,
-0.19283704459667206,
0.13660746812820435,
-0.016133761033415794,
0.0820322036743164,
0.12620896100997925,
0.18216060101985931,
0.015275371260941029,
-0.007581495214253664,
0.08013312518596649,
0.09365848451852798,
-0.09429137408733368,
-0.002616623882204294,
-0.024356674402952194,
-0.020394043996930122,
-0.1506359726190567,
0.033235687762498856,
0.1735067069530487,
0.07057067006826401,
-0.03652415797114372,
-0.0380864180624485,
-0.026559511199593544,
-0.014898261055350304,
0.09631852805614471,
-0.035881876945495605,
0.0787406638264656,
-0.11060360819101334,
0.0014385797549039125,
-0.024054834619164467,
-0.028224820271134377,
-0.04337170720100403,
0.007721672300249338,
-0.06971678137779236,
0.06263602524995804,
-0.02314375527203083,
0.06682904809713364,
-0.15083909034729004,
-0.13577696681022644,
-0.015824107453227043,
0.1500907838344574,
-0.016257205978035927,
0.07506807148456573,
0.06249907612800598,
-0.00950713362544775,
-0.011881954036653042,
-0.0027462902944535017,
0.2396756112575531,
0.02215125411748886,
-0.08097494393587112,
-0.07281089574098587,
0.11380738019943237,
-0.09682168066501617,
0.06593124568462372,
-0.13627256453037262,
0.021712208166718483,
0.04705327749252319,
0.12700648605823517,
0.047356147319078445,
0.06022628769278526,
0.0029065250419080257,
0.012886370532214642,
-0.0908179059624672,
0.0108116390183568,
0.09249769896268845,
-0.0001868757390184328,
-0.1274716705083847,
0.1960521936416626,
-0.25754809379577637,
0.2289544939994812,
0.1768645942211151,
-0.2303866595029831,
-0.0024039761628955603,
-0.14364126324653625,
0.004479940980672836,
0.0064788032323122025,
0.052027225494384766,
-0.04927888885140419,
0.047664374113082886,
-0.002529296325519681,
0.17434504628181458,
-0.04198553413152695,
-0.016601571813225746,
-0.01738758385181427,
-0.06157572939991951,
-0.06045174226164818,
0.05925477668642998,
-0.016496429219841957,
-0.22637061774730682,
0.18596147000789642,
0.23488567769527435,
0.047555241733789444,
0.18865571916103363,
-0.01411859318614006,
0.020403746515512466,
0.06918303668498993,
0.05946829169988632,
-0.02376592345535755,
-0.09700726717710495,
-0.1647307276725769,
-0.025128021836280823,
0.04753253608942032,
0.04167541116476059,
0.08389326184988022,
-0.08078666031360626,
-0.024501236155629158,
0.007118229288607836,
-0.004164670594036579,
0.06038019433617592,
0.07477652281522751,
0.048284679651260376,
0.1230786070227623,
-0.028676601126790047,
-0.03540462255477905,
0.11009242385625839,
-0.02514282800257206,
-0.11774394661188126,
0.19354572892189026,
-0.17214974761009216,
-0.3268393874168396,
-0.15640904009342194,
-0.18694175779819489,
-0.05782800167798996,
0.06909701228141785,
0.12020381540060043,
-0.10809806734323502,
-0.05169645696878433,
-0.0523100383579731,
0.08934404700994492,
-0.011228695511817932,
0.014810319058597088,
-0.0436498261988163,
0.05171462893486023,
-0.06366898119449615,
-0.0923234224319458,
-0.03323884308338165,
0.009805281646549702,
-0.02587759681046009,
0.13806657493114471,
-0.11138506978750229,
0.07480072230100632,
0.17058347165584564,
0.029143504798412323,
0.005570023320615292,
-0.05995183438062668,
0.1742922067642212,
-0.09340579807758331,
-0.007713443599641323,
0.1618456095457077,
-0.09976048022508621,
0.05691710487008095,
0.20509302616119385,
-0.017758388072252274,
-0.11166847497224808,
0.07821988314390182,
-0.031112246215343475,
-0.05743974447250366,
-0.21581292152404785,
-0.13406024873256683,
-0.1045726016163826,
0.09053955227136612,
-0.0033815414644777775,
0.04976074770092964,
0.14119893312454224,
0.06487458944320679,
-0.042238153517246246,
-0.024518225342035294,
0.08810403943061829,
0.08134426921606064,
0.21335655450820923,
-0.03973308205604553,
0.14395874738693237,
-0.05841333046555519,
-0.15669794380664825,
0.0653231218457222,
0.053122397512197495,
0.07189339399337769,
0.12496715784072876,
0.1082230657339096,
0.014865574426949024,
0.03435389697551727,
0.14011941850185394,
0.12414326518774033,
0.04193396493792534,
-0.055453069508075714,
-0.019216865301132202,
-0.0369708277285099,
-0.054365236312150955,
0.04849531874060631,
0.00713329529389739,
-0.10524705797433853,
-0.04220562055706978,
-0.04228932783007622,
0.11544731259346008,
0.10515124350786209,
0.07164645940065384,
-0.2062009572982788,
0.0020008315332233906,
0.1429404765367508,
-0.023853879421949387,
-0.12644119560718536,
0.10490543395280838,
0.07054047286510468,
-0.07684750109910965,
0.0757846087217331,
-0.04987265542149544,
0.1078341156244278,
-0.036024224013090134,
0.09670703858137131,
-0.1276305764913559,
-0.08111613243818283,
-0.014178982935845852,
0.08636923134326935,
-0.30306750535964966,
0.20376411080360413,
0.022630970925092697,
-0.027451155707240105,
-0.08384303003549576,
-0.0032569679897278547,
0.010781325399875641,
0.1436411589384079,
0.09831896424293518,
-0.042275406420230865,
-0.12852683663368225,
-0.07863204181194305,
0.006580037530511618,
0.023119447752833366,
0.14931689202785492,
0.0032731089740991592,
0.020565969869494438,
-0.06414368003606796,
-0.004848301410675049,
-0.00010057714825961739,
-0.03357640281319618,
-0.06617601215839386,
-0.21025557816028595,
0.029674839228391647,
0.11005092412233353,
0.1500563770532608,
-0.006108025088906288,
0.028506504371762276,
-0.09801730513572693,
0.2116306871175766,
-0.08718197047710419,
-0.04434308037161827,
-0.12397494167089462,
-0.08614499121904373,
0.010069925338029861,
-0.027370614930987358,
0.034889668226242065,
-0.06975515186786652,
0.0817720964550972,
-0.09952014684677124,
-0.19159959256649017,
0.1445152312517166,
-0.10322751849889755,
-0.028608916327357292,
-0.06512434035539627,
0.16118115186691284,
-0.06313972175121307,
-0.031207241117954254,
0.05518730357289314,
0.02993057295680046,
-0.05678690969944,
-0.06914331018924713,
0.0002050183102255687,
0.0019557292107492685,
0.04565997049212456,
0.07569476962089539,
-0.08764532953500748,
-0.16703404486179352,
-0.03396124020218849,
-0.034103721380233765,
0.3256298005580902,
0.23418977856636047,
-0.035550251603126526,
0.14565888047218323,
0.19174079596996307,
-0.07420279085636139,
-0.3419078588485718,
-0.07357829064130783,
-0.15586134791374207,
-0.03133442997932434,
-0.00636285450309515,
-0.08771949261426926,
0.07166914641857147,
-0.0009618918993510306,
-0.012713017873466015,
0.060830168426036835,
-0.20816311240196228,
-0.10144190490245819,
0.19024433195590973,
0.02269846387207508,
0.3847474455833435,
-0.16804587841033936,
-0.1132955551147461,
-0.12221329659223557,
-0.09737200289964676,
0.13342687487602234,
-0.12765957415103912,
0.0686027929186821,
0.04224312677979469,
0.048710763454437256,
0.05475765839219093,
-0.018267326056957245,
0.10558945685625076,
-0.05442364513874054,
0.05338206887245178,
-0.11045587807893753,
-0.017244374379515648,
0.017584633082151413,
-0.01374255120754242,
0.020332399755716324,
-0.10458814352750778,
0.015908852219581604,
-0.0379900299012661,
-0.04132906720042229,
-0.0134358499199152,
0.05766068398952484,
0.030492205172777176,
-0.05003593489527702,
0.0013545445399358869,
-0.08518907427787781,
0.0173235684633255,
-0.0008922514389269054,
0.2880125939846039,
-0.0932076945900917,
0.24015918374061584,
0.15964823961257935,
0.17706266045570374,
-0.12076102197170258,
0.10646545141935349,
-0.022093217819929123,
-0.07511031627655029,
0.07725369930267334,
-0.06907165050506592,
0.09023798257112503,
0.07358505576848984,
-0.07687368243932724,
0.11972593516111374,
0.0906536653637886,
0.03415843844413757,
0.0155512485653162,
0.1353500783443451,
-0.23077096045017242,
-0.09743180125951767,
-0.047230880707502365,
0.04695291072130203,
0.04546894133090973,
0.09246910363435745,
0.20222127437591553,
0.022257240489125252,
0.00416206568479538,
-0.015512548387050629,
0.03796568512916565,
-0.030062340199947357,
0.050254032015800476,
-0.021222732961177826,
0.03387105464935303,
-0.1266549676656723,
0.10109065473079681,
0.019064918160438538,
-0.1907520294189453,
0.03220897167921066,
0.13092628121376038,
-0.12741588056087494,
-0.13303011655807495,
-0.016583334654569626,
0.18969379365444183,
-0.0476936511695385,
-0.056202542036771774,
-0.035120781511068344,
-0.18684342503547668,
0.026535192504525185,
0.23096288740634918,
0.04576188325881958,
0.10620924085378647,
-0.021466216072440147,
-0.042051732540130615,
-0.025384046137332916,
0.034116048365831375,
-0.002467074664309621,
0.009755841456353664,
-0.08802871406078339,
0.016671592369675636,
-0.06509646773338318,
0.06557222455739975,
-0.10681498795747757,
-0.07220232486724854,
-0.17690061032772064,
0.017222195863723755,
-0.19288209080696106,
-0.041437383741140366,
-0.10342452675104141,
-0.03142404556274414,
0.01771537959575653,
-0.006088911555707455,
-0.03821616992354393,
-0.08029630035161972,
-0.0920756533741951,
0.04198124259710312,
-0.02468748949468136,
0.05142378434538841,
-0.09125258773565292,
-0.014776421710848808,
0.0596148855984211,
-0.038273829966783524,
0.12445179373025894,
0.10307559370994568,
-0.11014258861541748,
0.09309541434049606,
-0.24811165034770966,
-0.08018241077661514,
0.13974665105342865,
-0.012274859473109245,
0.04137380048632622,
0.07514360547065735,
-0.0223940871655941,
0.10060727596282959,
0.028620736673474312,
0.05461081489920616,
0.03275372087955475,
-0.08476798236370087,
0.05422923341393471,
-0.05384410172700882,
-0.12911638617515564,
-0.03693132847547531,
-0.08276031166315079,
0.04934628680348396,
-0.03412659466266632,
0.16354911029338837,
-0.10385499149560928,
0.08129590004682541,
-0.02668697200715542,
0.021812649443745613,
0.021874170750379562,
-0.19207602739334106,
-0.09233317524194717,
-0.08088872581720352,
0.03394326940178871,
-0.002184416400268674,
0.22881752252578735,
-0.0063036116771399975,
0.006833946797996759,
0.06641337275505066,
0.009306887164711952,
0.025052908807992935,
0.05418117344379425,
0.21890892088413239,
0.10531165450811386,
-0.06278764456510544,
-0.14196747541427612,
0.04822472482919693,
0.0433984249830246,
-0.01768500730395317,
0.08523741364479065,
0.07174747437238693,
-0.0875975638628006,
0.12999862432479858,
-0.03001311980187893,
0.01789293996989727,
0.0010054619051516056,
-0.08869358152151108,
-0.06007562205195427,
0.04227394610643387,
-0.03391311690211296,
0.07754029333591461,
0.17639881372451782,
-0.008532049134373665,
-0.0021799886599183083,
-0.06557418406009674,
-0.053584493696689606,
-0.17924116551876068,
-0.09899032860994339,
-0.11062261462211609,
-0.13509704172611237,
-0.0031246659345924854,
-0.11335091292858124,
0.03045600838959217,
0.02505817823112011,
0.06634732335805893,
-0.03677665442228317,
0.14219939708709717,
0.005272229202091694,
-0.07059774547815323,
0.060609254986047745,
-0.039187293499708176,
0.08489051461219788,
-0.0005837218486703932,
-0.05879097804427147,
-0.07350610941648483,
-0.019591964781284332,
-0.032896608114242554,
0.08127778023481369,
-0.017102094367146492,
0.04861724004149437,
-0.1601472645998001,
-0.11311696469783783,
-0.02544534206390381,
0.09283430129289627,
-0.07150934636592865,
0.11804944276809692,
0.040014397352933884,
-0.03640314191579819,
0.05701145529747009,
0.23363706469535828,
-0.06640453636646271,
-0.05833832174539566,
-0.06949034333229065,
0.1772133857011795,
0.011515109799802303,
0.17556817829608917,
-0.06164545565843582,
-0.03653505817055702,
-0.06739061325788498,
0.2944355010986328,
0.27835631370544434,
-0.08374209702014923,
0.03394955024123192,
-0.041623324155807495,
0.042466435581445694,
0.07614653557538986,
0.11959473788738251,
0.07460001856088638,
0.2395261526107788,
-0.04222880303859711,
0.006043989211320877,
0.008341225795447826,
-0.03936750814318657,
-0.1136622205376625,
0.10810278356075287,
0.009884144179522991,
-0.023359671235084534,
-0.026450419798493385,
0.12967772781848907,
-0.19534580409526825,
0.07861483097076416,
-0.11984775215387344,
-0.14653997123241425,
-0.030937962234020233,
-0.013667594641447067,
0.15530915558338165,
-0.014980345033109188,
0.06503602117300034,
0.0032963177654892206,
-0.11207564920186996,
0.02590802311897278,
0.0011753392172977328,
-0.15004241466522217,
0.0028493020217865705,
0.01628642901778221,
-0.07600481808185577,
-0.0014932762132957578,
-0.005610094405710697,
0.007041973527520895,
0.0839076116681099,
0.03361568599939346,
-0.01694958098232746,
0.10008566081523895,
0.015474583953619003,
-0.04245442897081375,
0.056303855031728745,
0.03336367756128311,
0.00592902023345232,
0.01795760542154312,
0.08802462369203568,
-0.1988658905029297,
0.07018439471721649,
-0.04100437089800835,
-0.10450468212366104,
0.00023289999808184803,
0.03334207460284233,
-0.04662621393799782,
0.0687335729598999,
0.06066638603806496,
-0.013131367973983288,
0.03755770996212959,
-0.002024680143222213,
0.0007067753467708826,
-0.033159565180540085,
-0.058379124850034714,
-0.06549149006605148,
-0.18190479278564453,
-0.08929894119501114,
0.15773873031139374,
0.014779342338442802,
-0.26594996452331543,
0.003054126864299178,
-0.10869979858398438,
0.09269300848245621,
-0.17001447081565857,
0.08308188617229462,
0.16453157365322113,
0.010288641788065434,
-0.029686687514185905,
-0.16980479657649994,
0.06382575631141663,
0.10825681686401367,
-0.03237207978963852,
-0.09691159427165985
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-squad-model2
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 74
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-squad-model2", "results": []}]} | question-answering | varun-v-rao/roberta-base-squad-model2 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-base",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:22:51+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us
|
# roberta-base-squad-model2
This model is a fine-tuned version of roberta-base on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 74
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-base-squad-model2\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 74\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us \n",
"# roberta-base-squad-model2\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 74\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
69,
32,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us \n# roberta-base-squad-model2\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 74\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.059664420783519745,
0.04428571090102196,
-0.0017254542326554656,
0.07213782519102097,
0.18725374341011047,
0.01591227389872074,
0.14863140881061554,
0.07643721997737885,
-0.1482870876789093,
0.05218767374753952,
0.06543336808681488,
0.08545482158660889,
0.01222477201372385,
0.09706069529056549,
-0.039590783417224884,
-0.23071953654289246,
0.005059666465967894,
-0.00716845178976655,
-0.12758409976959229,
0.09683595597743988,
0.09943837672472,
-0.11844311654567719,
0.0601666197180748,
-0.006738197524100542,
-0.21974386274814606,
0.055135928094387054,
0.00858341995626688,
-0.036098502576351166,
0.10669466853141785,
0.020172372460365295,
0.13926173746585846,
0.0020462588872760534,
0.1353551745414734,
-0.20986290276050568,
0.01786944270133972,
0.09066452831029892,
0.02063249982893467,
0.0659184455871582,
0.022782819345593452,
-0.009873321279883385,
0.12566611170768738,
-0.12704713642597198,
0.09355414658784866,
0.019800685346126556,
-0.09184911102056503,
-0.16975881159305573,
-0.09164682775735855,
0.02814938873052597,
0.08522934466600418,
0.10319589078426361,
-0.01312338374555111,
0.17847198247909546,
-0.11749207973480225,
0.07762353867292404,
0.19278931617736816,
-0.2794138491153717,
-0.1049085333943367,
0.10703188180923462,
0.04908840358257294,
0.09233410656452179,
-0.11355265974998474,
-0.0020252817776054144,
0.050466276705265045,
0.031105857342481613,
0.07962843775749207,
-0.02617393061518669,
-0.08318610489368439,
0.007138756103813648,
-0.14600074291229248,
0.03422883152961731,
0.12652605772018433,
0.05486996844410896,
-0.03284251317381859,
-0.021867314353585243,
-0.04313160851597786,
-0.027599552646279335,
-0.03984348848462105,
-0.060237180441617966,
0.04897695779800415,
-0.04806268587708473,
-0.13465064764022827,
-0.03354519233107567,
-0.08242969959974289,
-0.07120497524738312,
-0.025001192465424538,
0.13423079252243042,
0.04346088692545891,
0.012373726814985275,
-0.06595239043235779,
0.08151126652956009,
-0.014952708035707474,
-0.10451830923557281,
0.015350711531937122,
0.0070060500875115395,
-0.07922053337097168,
-0.07536089420318604,
-0.056614167988300323,
-0.05888816714286804,
0.03979514166712761,
0.18202288448810577,
-0.05166971683502197,
0.05633854866027832,
0.004496157635003328,
0.010247522965073586,
-0.05049508064985275,
0.12429701536893845,
-0.06543362140655518,
-0.06716544181108475,
0.015470529906451702,
0.062034331262111664,
0.016277359798550606,
0.008805613033473492,
-0.08080709725618362,
0.011416982859373093,
0.07962561398744583,
0.034921035170555115,
-0.060962677001953125,
0.03894874081015587,
0.0005810571019537747,
-0.007514636963605881,
-0.034262221306562424,
-0.09979644417762756,
0.037225667387247086,
-0.010482225567102432,
-0.06106770411133766,
-0.015384435653686523,
0.013004222884774208,
0.02537190541625023,
0.018355851992964745,
0.11082178354263306,
-0.09876788407564163,
0.014938139356672764,
-0.11229690164327621,
-0.10056951642036438,
-0.0024650765117257833,
-0.06488625705242157,
0.005584089085459709,
-0.08438358455896378,
-0.1414748877286911,
-0.04404919967055321,
0.04585006460547447,
-0.04782074689865112,
-0.00876203365623951,
-0.050900403410196304,
-0.0863136351108551,
-0.0036385587882250547,
0.006367754191160202,
0.16354666650295258,
-0.03937222436070442,
0.06917324662208557,
0.032577820122241974,
0.03662503883242607,
-0.028497682884335518,
0.021212583407759666,
-0.08797552436590195,
0.013805904425680637,
-0.18320104479789734,
0.03423330560326576,
-0.09001751244068146,
0.06757152080535889,
-0.10190799087285995,
-0.08816550672054291,
-0.01772148907184601,
0.007273541763424873,
0.07824338227510452,
0.07255325466394424,
-0.16597069799900055,
-0.03707639500498772,
0.16463041305541992,
-0.09073179960250854,
-0.0880347490310669,
0.09153424948453903,
-0.06748533248901367,
0.06823394447565079,
0.07320016622543335,
0.1588895171880722,
0.0542590469121933,
-0.11017162352800369,
-0.004115677438676357,
-0.038698628544807434,
0.048774197697639465,
-0.0121405478566885,
0.04007052630186081,
0.008027222938835621,
0.028547441586852074,
0.0058181509375572205,
-0.06577887386083603,
0.019674159586429596,
-0.1099616214632988,
-0.07404953986406326,
-0.04319842532277107,
-0.09666077047586441,
-0.03227798640727997,
0.05307953804731369,
0.060190677642822266,
-0.09066574275493622,
-0.09282651543617249,
0.15397599339485168,
0.10762368887662888,
-0.06092867627739906,
0.017325405031442642,
-0.08367946743965149,
0.04803231731057167,
-0.0702606663107872,
-0.02651563659310341,
-0.1910892128944397,
-0.11262010037899017,
-0.007533304393291473,
0.019326750189065933,
0.05827132984995842,
0.04182712361216545,
0.07130014896392822,
0.0572846345603466,
-0.053016774356365204,
0.013516183942556381,
-0.06755074858665466,
0.009059431031346321,
-0.09638532996177673,
-0.239759162068367,
-0.022065196186304092,
-0.028184670954942703,
0.18048836290836334,
-0.2563445568084717,
0.020437873899936676,
-0.05224008113145828,
0.13390932977199554,
0.03303512558341026,
-0.03854971379041672,
-0.04680614545941353,
0.06781961023807526,
-0.0035683063324540854,
-0.07259196043014526,
0.04321175813674927,
-0.003477161517366767,
-0.06279189884662628,
-0.08265507221221924,
-0.13801053166389465,
0.08420101553201675,
0.1012999638915062,
-0.03819631412625313,
-0.10021098703145981,
0.027450483292341232,
-0.06637950241565704,
-0.02862739935517311,
-0.06789440661668777,
0.02079557254910469,
0.12764355540275574,
-0.01173704955726862,
0.1226421371102333,
-0.05362313240766525,
-0.04614917188882828,
0.002214579377323389,
-0.03445117175579071,
0.03839738667011261,
0.0858813226222992,
0.12771494686603546,
-0.12539243698120117,
0.09119583666324615,
0.08415783196687698,
-0.10322752594947815,
0.13373206555843353,
-0.04206828773021698,
-0.06743567436933517,
-0.029934445396065712,
-0.01443968340754509,
0.0075302887707948685,
0.1399291306734085,
-0.08133801817893982,
0.002609278541058302,
-0.0016450464027002454,
0.01526065357029438,
0.041844822466373444,
-0.18387889862060547,
-0.029590101912617683,
0.003050463739782572,
-0.017583459615707397,
0.004172299988567829,
-0.007634324952960014,
0.018552443012595177,
0.10136262327432632,
0.012387480586767197,
-0.03220542520284653,
-0.005118667613714933,
-0.007387581281363964,
-0.08052156865596771,
0.20192433893680573,
-0.079648956656456,
-0.049278732389211655,
-0.08533919602632523,
-0.01379543449729681,
-0.03901219740509987,
-0.030055809766054153,
0.028025101870298386,
-0.11784788221120834,
-0.046944376081228256,
-0.08284234255552292,
0.015665292739868164,
-0.0007983447285369039,
0.0012474415125325322,
0.03428152576088905,
0.02786444127559662,
0.09054359048604965,
-0.1319086253643036,
0.012568861246109009,
-0.08216496556997299,
-0.1352459341287613,
0.007256956771016121,
0.05817374214529991,
0.12441984564065933,
0.11805839091539383,
-0.018697470426559448,
0.011984794400632381,
-0.024807874113321304,
0.23761914670467377,
-0.06191156059503555,
-0.011917372234165668,
0.10255730152130127,
0.008432256057858467,
0.029640594497323036,
0.12523674964904785,
0.05703452602028847,
-0.116801418364048,
0.043357860296964645,
0.10603026300668716,
-0.03246995806694031,
-0.23265911638736725,
-0.02986725978553295,
-0.037019625306129456,
-0.11251623183488846,
0.04645774886012077,
0.04030999913811684,
0.002367333509027958,
0.055854108184576035,
0.02639150433242321,
0.05678655207157135,
-0.024090047925710678,
0.07988601177930832,
0.08564157038927078,
0.04954517260193825,
0.12877526879310608,
-0.05820796266198158,
-0.06360773742198944,
0.03717869892716408,
-0.02420739270746708,
0.30455997586250305,
-0.0021869356278330088,
0.019425248727202415,
0.0978606790304184,
0.11531105637550354,
-0.018117353320121765,
0.01972854509949684,
0.0024381137918680906,
-0.04891646280884743,
0.00023758388124406338,
-0.055349599570035934,
0.009742487221956253,
0.006776202004402876,
-0.0677834153175354,
0.060623664408922195,
-0.06541066616773605,
0.047896649688482285,
0.05469970405101776,
0.21572212874889374,
0.012515202164649963,
-0.2702254056930542,
-0.07944705337285995,
0.01094884891062975,
-0.029185233637690544,
-0.00922022107988596,
0.0058233062736690044,
0.11035653948783875,
-0.09092560410499573,
0.02441471815109253,
-0.05984705314040184,
0.09206435829401016,
0.004050926771014929,
0.009430556558072567,
0.03985491022467613,
0.17245976626873016,
-0.01514029037207365,
0.060150787234306335,
-0.22745539247989655,
0.240970179438591,
0.009832806885242462,
0.14107497036457062,
-0.020807860419154167,
0.0007892405847087502,
0.021214768290519714,
0.07387152314186096,
0.053486280143260956,
-0.005598831456154585,
-0.054036110639572144,
-0.1627846509218216,
-0.012847605161368847,
0.06871451437473297,
0.12375703454017639,
-0.008024864830076694,
0.10959017276763916,
-0.04578874632716179,
0.011585439555346966,
0.06925391405820847,
-0.05698687955737114,
-0.18418438732624054,
-0.09676886349916458,
-0.03643425181508064,
-0.007999933324754238,
-0.09146743267774582,
-0.08333504945039749,
-0.0932026281952858,
-0.06818131357431412,
0.15633606910705566,
0.024159202352166176,
0.012220949865877628,
-0.11891283094882965,
0.1432744860649109,
0.08911500126123428,
-0.059251539409160614,
0.010882392525672913,
0.013359463773667812,
0.0759795755147934,
0.041319895535707474,
-0.06838559359312057,
0.06917917728424072,
-0.07575641572475433,
-0.13912418484687805,
-0.05982252582907677,
0.08191204071044922,
0.07244943827390671,
0.03863053023815155,
0.0026200376451015472,
0.032183531671762466,
0.01794472336769104,
-0.09161783754825592,
0.00020755815785378218,
0.04953621327877045,
0.04900100827217102,
0.04307134822010994,
-0.07987582683563232,
-0.01879020780324936,
-0.04420892894268036,
-0.022663529962301254,
0.11235921829938889,
0.2324729859828949,
-0.07684897631406784,
0.015305932611227036,
0.05332334339618683,
-0.08262304216623306,
-0.17662207782268524,
0.12825343012809753,
0.08088324218988419,
-0.0020223017781972885,
0.06044234707951546,
-0.1449175924062729,
0.1883086860179901,
0.11546465009450912,
-0.01361241564154625,
0.032180801033973694,
-0.2855280935764313,
-0.1353624016046524,
0.07007980346679688,
0.16061067581176758,
0.09939610958099365,
-0.15913355350494385,
-0.016109326854348183,
-0.03376620635390282,
-0.1398797184228897,
0.10928439348936081,
-0.18303938210010529,
0.10089688748121262,
0.011687139980494976,
0.07379239797592163,
-0.0005025044665671885,
-0.04630931839346886,
0.11945383250713348,
0.02188456431031227,
0.11607371270656586,
-0.05557377636432648,
-0.003239624435082078,
0.12653225660324097,
-0.03649722412228584,
-0.012600146234035492,
-0.021279416978359222,
0.052060965448617935,
-0.07567743211984634,
-0.015800634399056435,
-0.07576718926429749,
0.043821658939123154,
-0.05545171722769737,
-0.059408579021692276,
-0.05620303750038147,
0.0626157894730568,
0.03716526925563812,
-0.024067595601081848,
0.07855309545993805,
-0.01584031991660595,
0.1589498072862625,
0.012209806591272354,
0.09134581685066223,
-0.007985721342265606,
-0.019898705184459686,
0.01718929596245289,
-0.019672172144055367,
0.057255033403635025,
-0.14434762299060822,
0.016810264438390732,
0.11052499711513519,
0.035916853696107864,
0.1311032772064209,
0.05994928628206253,
-0.05690544471144676,
0.025458309799432755,
0.07095228135585785,
-0.09780118614435196,
-0.15027429163455963,
0.038454484194517136,
-0.0681183710694313,
-0.11607249826192856,
0.06206018105149269,
0.11473886668682098,
-0.060662854462862015,
-0.006748709362000227,
-0.01458683516830206,
-0.0024765576235949993,
-0.06324968487024307,
0.18199172616004944,
0.06801123917102814,
0.043309248983860016,
-0.08222750574350357,
0.08381610363721848,
0.04686495289206505,
-0.03731769323348999,
0.025128325447440147,
0.07788154482841492,
-0.08194386959075928,
-0.016721423715353012,
0.07527054846286774,
0.2405451387166977,
-0.06870292872190475,
-0.02931744046509266,
-0.13843826949596405,
-0.0909866988658905,
0.030190816149115562,
0.17574800550937653,
0.06620336323976517,
-0.04301454499363899,
-0.03150087222456932,
0.060758817940950394,
-0.16264742612838745,
0.07750499248504639,
0.04202565923333168,
0.07921178638935089,
-0.12126943469047546,
0.10664953291416168,
0.020807914435863495,
-0.0053566619753837585,
-0.02187563292682171,
0.03301527351140976,
-0.13923725485801697,
-0.008970635943114758,
-0.1337497979402542,
-0.037503231316804886,
-0.02230733446776867,
0.000010585313248157036,
0.00004776095011038706,
-0.06494073569774628,
-0.0879226103425026,
0.03782108426094055,
-0.07694458216428757,
-0.03499111160635948,
0.03501763939857483,
0.042023662477731705,
-0.1459743231534958,
0.0067153265699744225,
0.014564364217221737,
-0.07136787474155426,
0.05377053841948509,
0.048472411930561066,
0.043446436524391174,
0.06117366626858711,
-0.20749829709529877,
-0.010168123058974743,
0.026027927175164223,
-0.00043146152165718377,
0.08380263298749924,
-0.0675455629825592,
-0.006991001311689615,
-0.014144963584840298,
0.107964888215065,
0.02397282049059868,
0.0576334074139595,
-0.10461220890283585,
0.022303404286503792,
-0.08823501318693161,
-0.06711059808731079,
-0.05020255967974663,
0.02551879733800888,
0.08499067276716232,
0.04691549763083458,
0.1770429164171219,
-0.10535600036382675,
0.025839904323220253,
-0.21700039505958557,
-0.030431412160396576,
-0.0017823155503720045,
-0.044215381145477295,
-0.08925441652536392,
-0.04650653898715973,
0.06422962993383408,
-0.06722671538591385,
0.11864446103572845,
0.006377561017870903,
0.07904548943042755,
0.05369354784488678,
-0.04462973400950432,
-0.007985088042914867,
0.021827038377523422,
0.18972884118556976,
0.06115582212805748,
-0.0049934531562030315,
0.050530847162008286,
0.014291191473603249,
0.06511300802230835,
0.026707423850893974,
0.19569917023181915,
0.1938318908214569,
-0.04320703446865082,
0.05593947321176529,
0.08021806180477142,
-0.06893293559551239,
-0.077132947742939,
0.15305911004543304,
-0.04342478886246681,
0.056035954505205154,
-0.04887649789452553,
0.14267730712890625,
0.15997932851314545,
-0.1520688235759735,
0.01768592931330204,
-0.07586288452148438,
-0.08787070214748383,
-0.11707994341850281,
0.0006808238103985786,
-0.09615495800971985,
-0.1382574737071991,
0.04803629219532013,
-0.12595640122890472,
0.007375627756118774,
0.11441978812217712,
0.01863645762205124,
0.01923898234963417,
0.16886776685714722,
0.0043761711567640305,
0.04268820211291313,
0.02216808684170246,
-0.010343064554035664,
-0.026980554684996605,
-0.04881894215941429,
-0.043207425624132156,
0.028483398258686066,
-0.011364023201167583,
0.05758032575249672,
-0.04907629266381264,
-0.03443102538585663,
0.04253658652305603,
-0.02406046912074089,
-0.07348835468292236,
0.03600800037384033,
0.039355214685201645,
0.03546546772122383,
0.04287615045905113,
0.03960184380412102,
-0.010268392972648144,
-0.031208712607622147,
0.2820903956890106,
-0.06142258644104004,
-0.12816016376018524,
-0.1383419632911682,
0.22164380550384521,
0.031437866389751434,
0.0005222095642238855,
0.019094666466116905,
-0.1064777597784996,
0.017525769770145416,
0.1814805269241333,
0.1974075436592102,
-0.07849706709384918,
-0.007805196568369865,
0.0023280668538063765,
-0.025113077834248543,
-0.0785435140132904,
0.12744903564453125,
0.1043463945388794,
0.03817174211144447,
-0.05559603497385979,
-0.04786010831594467,
-0.034221064299345016,
-0.02320837415754795,
-0.04600124433636665,
0.030381806194782257,
0.05946917086839676,
0.011455214582383633,
-0.028040636330842972,
0.0939340814948082,
-0.014129430055618286,
-0.20096737146377563,
0.0672835111618042,
-0.15494167804718018,
-0.17500528693199158,
-0.031619708985090256,
0.08992902934551239,
-0.03163544833660126,
0.059516552835702896,
-0.036747824400663376,
-0.015362848527729511,
0.0763024389743805,
-0.018366655334830284,
-0.03157386556267738,
-0.1120508462190628,
0.09065074473619461,
-0.07874812185764313,
0.20296770334243774,
-0.038800548762083054,
0.06023625656962395,
0.13955366611480713,
0.022181037813425064,
-0.09076954424381256,
0.06822694092988968,
0.04663728177547455,
-0.09658943116664886,
0.015117957256734371,
0.12827955186367035,
-0.053924594074487686,
0.1103045791387558,
0.05863950029015541,
-0.14496271312236786,
0.0281552542001009,
-0.0704335868358612,
-0.047028105705976486,
-0.08208759874105453,
0.0020953258499503136,
-0.09120215475559235,
0.1422988623380661,
0.201589435338974,
-0.020855272188782692,
0.04154177010059357,
-0.08963640034198761,
0.03421112149953842,
0.06383796781301498,
0.10236602276563644,
-0.05412612482905388,
-0.24455136060714722,
0.03310072794556618,
0.029699234291911125,
-0.011700665578246117,
-0.2888869345188141,
-0.07593109458684921,
0.033284857869148254,
-0.04908917471766472,
-0.06254459917545319,
0.10642394423484802,
0.12452984601259232,
0.06290861964225769,
-0.050923969596624374,
-0.18409286439418793,
-0.057462554425001144,
0.16721667349338531,
-0.14470909535884857,
-0.06887026131153107
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# image_classification
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 2.0856
- Accuracy: 0.0813
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 40 | 2.0855 | 0.0813 |
| No log | 2.0 | 80 | 2.0880 | 0.0813 |
| No log | 3.0 | 120 | 2.0852 | 0.0813 |
| No log | 4.0 | 160 | 2.0861 | 0.0813 |
| No log | 5.0 | 200 | 2.0858 | 0.0813 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "image_classification", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.08125, "name": "Accuracy"}]}]}]} | image-classification | miifta-hs/image_classification | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:25:32+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| image\_classification
=====================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 2.0856
* Accuracy: 0.0813
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.17.0
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
86,
97,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1"
] | [
-0.12278702855110168,
0.1298970729112625,
-0.0024047885090112686,
0.12514805793762207,
0.1450514793395996,
0.008282636292278767,
0.1435735672712326,
0.13263212144374847,
-0.06685581803321838,
0.07742425799369812,
0.15067428350448608,
0.11514359712600708,
0.02871137298643589,
0.16969464719295502,
-0.046234942972660065,
-0.215292289853096,
0.0270839910954237,
0.04920222610235214,
-0.04663344472646713,
0.12106435745954514,
0.0881650447845459,
-0.1408090740442276,
0.12301945686340332,
0.024633917957544327,
-0.19913575053215027,
-0.009694295935332775,
0.03534740209579468,
-0.04712343588471413,
0.12152542173862457,
0.038227956742048264,
0.10130022466182709,
0.02752482332289219,
0.0629022940993309,
-0.1485932171344757,
0.010196754708886147,
0.0779941976070404,
-0.004749365150928497,
0.09544815868139267,
0.05504314601421356,
0.01414886862039566,
-0.0009633562876842916,
-0.09601643681526184,
0.03970000147819519,
0.025529272854328156,
-0.11624885350465775,
-0.22157680988311768,
-0.07984486222267151,
0.05274517089128494,
0.07736316323280334,
0.0762762501835823,
-0.00807569082826376,
0.14233914017677307,
-0.017629358917474747,
0.09647083282470703,
0.21976514160633087,
-0.2719685435295105,
-0.07805890589952469,
0.04069223627448082,
0.02116214856505394,
0.08504410833120346,
-0.1042770966887474,
0.014638297259807587,
0.060018543154001236,
0.013425169512629509,
0.14729177951812744,
-0.001908890320919454,
-0.010164731182157993,
-0.01658707484602928,
-0.12311939895153046,
-0.05603865534067154,
0.18518967926502228,
0.08851416409015656,
-0.050077132880687714,
-0.0751604288816452,
-0.07748420536518097,
-0.1449243575334549,
-0.039528001099824905,
-0.007429749239236116,
0.051430802792310715,
-0.035003237426280975,
-0.07514654099941254,
-0.03319652006030083,
-0.10504405200481415,
-0.0712091401219368,
-0.015122251585125923,
0.092501699924469,
0.047960080206394196,
0.011791802011430264,
-0.02056511864066124,
0.08843624591827393,
-0.042943328619003296,
-0.14520469307899475,
-0.0041960421949625015,
0.016460560262203217,
0.012189757078886032,
-0.03606167435646057,
-0.030661357566714287,
-0.12165825068950653,
0.01248230691999197,
0.10069967061281204,
-0.049177784472703934,
0.0542658194899559,
-0.013018994592130184,
0.04852667078375816,
-0.1098388135433197,
0.20941124856472015,
-0.06129668280482292,
0.013476784341037273,
0.034230392426252365,
0.10551372170448303,
0.053168609738349915,
-0.0034677584189921618,
-0.1133560985326767,
0.009692314080893993,
0.12108329683542252,
0.0022498853504657745,
-0.04232059419155121,
0.07995876669883728,
-0.05788865312933922,
-0.02956036664545536,
0.06446095556020737,
-0.07734235376119614,
0.02992115542292595,
-0.0057167611084878445,
-0.04857804253697395,
-0.047099769115448,
0.04732353240251541,
-0.007235965225845575,
-0.010025065392255783,
0.05537322163581848,
-0.10828889161348343,
0.010498909279704094,
-0.06961704790592194,
-0.1145118996500969,
0.010502098128199577,
-0.11668343096971512,
0.01739802211523056,
-0.1261342167854309,
-0.14291128516197205,
-0.007355415727943182,
0.05643676593899727,
-0.0292383823543787,
-0.045254264026880264,
-0.041611213237047195,
-0.07668641954660416,
0.02661288157105446,
0.007721186615526676,
0.050254758447408676,
-0.05593841150403023,
0.093110091984272,
0.04787793010473251,
0.07259851694107056,
-0.022515885531902313,
0.04261719807982445,
-0.08696943521499634,
0.0591672845184803,
-0.1946343332529068,
0.034670036286115646,
-0.0493880994617939,
0.08194053173065186,
-0.11545547842979431,
-0.08567306399345398,
0.0009542961488477886,
-0.022086573764681816,
0.07048574090003967,
0.11154641211032867,
-0.14683859050273895,
-0.05573644861578941,
0.1646764576435089,
-0.09767665714025497,
-0.15554340183734894,
0.1257738620042801,
-0.031461309641599655,
0.02674054726958275,
0.061390288174152374,
0.18820807337760925,
0.07831265777349472,
-0.11645180732011795,
-0.010831893421709538,
-0.03503677621483803,
0.04851926863193512,
-0.04387917369604111,
0.08073408156633377,
0.009314869530498981,
-0.016257666051387787,
0.016787922009825706,
-0.10050695389509201,
0.0727115347981453,
-0.07717195898294449,
-0.08353296667337418,
-0.06410567462444305,
-0.09562079608440399,
0.05637821927666664,
0.058858271688222885,
0.06090501323342323,
-0.09790018200874329,
-0.09806380420923233,
0.025940701365470886,
0.07789240777492523,
-0.09112179279327393,
0.01364232785999775,
-0.08467476814985275,
0.11106357723474503,
-0.11947692930698395,
-0.003017863491550088,
-0.12663346529006958,
-0.03757397085428238,
0.044036056846380234,
-0.054171327501535416,
-0.007004484534263611,
-0.03383548557758331,
0.07764986157417297,
0.06884638965129852,
-0.07047022879123688,
-0.07456835359334946,
-0.04602385312318802,
0.0007810284150764346,
-0.10645652562379837,
-0.18466773629188538,
-0.031047185882925987,
-0.024293169379234314,
0.11921913921833038,
-0.22600339353084564,
0.03848474472761154,
0.04202654957771301,
0.10014577955007553,
0.05864056572318077,
-0.03292582929134369,
-0.004781442228704691,
0.019956979900598526,
-0.04271212965250015,
-0.09050516039133072,
0.05571562051773071,
0.01369395386427641,
-0.07655178755521774,
-0.0012536834692582488,
-0.10624085366725922,
0.1708177626132965,
0.1316153109073639,
-0.041373398154973984,
-0.06605841964483261,
0.007162267342209816,
-0.03669450432062149,
-0.037723395973443985,
-0.04681520164012909,
-0.006906796246767044,
0.08322817087173462,
-0.0012927670031785965,
0.15757954120635986,
-0.10064589232206345,
-0.021309079602360725,
0.05752146244049072,
-0.025982147082686424,
-0.03471081331372261,
0.09174586832523346,
0.07108836621046066,
-0.13804754614830017,
0.14487656950950623,
0.16923463344573975,
-0.08024239540100098,
0.12036754190921783,
-0.03966303914785385,
-0.0682150200009346,
-0.027333959937095642,
0.034149978309869766,
0.033533353358507156,
0.13980898261070251,
-0.12383134663105011,
-0.011450767517089844,
0.018845660611987114,
0.0009070493979379535,
-0.012343713082373142,
-0.2037554383277893,
-0.016545044258236885,
0.04466019198298454,
-0.05178726464509964,
0.01782902143895626,
-0.009911661967635155,
-0.019561244174838066,
0.08468637615442276,
0.011974319815635681,
-0.03773633390665054,
0.04385779798030853,
0.008557172492146492,
-0.07158598303794861,
0.19730322062969208,
-0.07538449764251709,
-0.20902825891971588,
-0.13552913069725037,
-0.028884101659059525,
-0.07864686101675034,
0.028941867873072624,
0.053596381098032,
-0.10133517533540726,
-0.0603555329144001,
-0.10342912375926971,
-0.0006873764796182513,
0.02751828171312809,
0.04013544321060181,
0.030305245891213417,
-0.009877590462565422,
0.12673573195934296,
-0.09385901689529419,
-0.003070043632760644,
-0.003665897296741605,
-0.023409847170114517,
0.04532363638281822,
0.017149697989225388,
0.11961537599563599,
0.09114203602075577,
-0.028187131509184837,
0.03065520152449608,
-0.01838121935725212,
0.2496030032634735,
-0.0760965570807457,
-0.004895728547126055,
0.15202100574970245,
0.013650228269398212,
0.06162453442811966,
0.12898463010787964,
0.0394262932240963,
-0.09754053503274918,
0.008194906637072563,
0.024486074224114418,
-0.025900207459926605,
-0.1783175766468048,
-0.015204687602818012,
-0.03968002647161484,
-0.0010861764894798398,
0.15029940009117126,
0.04863089323043823,
0.054395511746406555,
0.09873702377080917,
0.003454151563346386,
0.09695963561534882,
-0.008006764575839043,
0.08686225116252899,
0.10980274528265,
0.04506891593337059,
0.10933882743120193,
-0.039622221142053604,
-0.038867659866809845,
0.029592176899313927,
0.016164256259799004,
0.2162783145904541,
0.0015302281826734543,
0.17438258230686188,
0.036625150591135025,
0.18539024889469147,
0.014208718203008175,
0.061480771750211716,
-0.019521422684192657,
-0.033373527228832245,
-0.005783088505268097,
-0.05393345281481743,
-0.020345861092209816,
0.033796921372413635,
-0.04369072616100311,
0.07025846093893051,
-0.09315548092126846,
0.04101021587848663,
0.059206001460552216,
0.25417521595954895,
0.05563522130250931,
-0.38415294885635376,
-0.10257049649953842,
0.0001892942818813026,
-0.007167657371610403,
-0.06717251986265182,
0.0007052058936096728,
0.1536627560853958,
-0.057463135570287704,
0.04924957454204559,
-0.10494671016931534,
0.07919846475124359,
-0.05068845674395561,
0.02552342787384987,
0.08122701197862625,
0.07958606630563736,
0.004335882607847452,
0.057901717722415924,
-0.24085554480552673,
0.25251907110214233,
0.01698443852365017,
0.0632348358631134,
-0.044268254190683365,
0.0040327757596969604,
0.033095672726631165,
0.11126340180635452,
0.10809005051851273,
-0.003519708989188075,
0.012546989135444164,
-0.18914510309696198,
-0.0909893736243248,
0.008086220361292362,
0.07068915665149689,
-0.051836468279361725,
0.08950339257717133,
-0.034838952124118805,
-0.0224619060754776,
0.05659545585513115,
0.01559155061841011,
-0.08842818439006805,
-0.0928245559334755,
-0.00762437516823411,
0.04637875407934189,
0.018333477899432182,
-0.09305795282125473,
-0.10110682994127274,
-0.10929746180772781,
0.13332559168338776,
-0.017742913216352463,
-0.03654127195477486,
-0.11971423029899597,
0.09327254444360733,
0.061886053532361984,
-0.08904752135276794,
0.07220963388681412,
-0.02403651364147663,
0.14185625314712524,
0.028867149725556374,
-0.06046466901898384,
0.10847845673561096,
-0.054605577141046524,
-0.17366637289524078,
-0.055067725479602814,
0.10245431214570999,
-0.015650760382413864,
0.02598298154771328,
-0.00045515975216403604,
0.03073887713253498,
-0.018526585772633553,
-0.05716301128268242,
0.05464750900864601,
0.00505144102498889,
0.047835350036621094,
-0.0028974032029509544,
-0.014065862633287907,
0.015040641650557518,
-0.062150247395038605,
-0.029265210032463074,
0.13232165575027466,
0.24091880023479462,
-0.09681961685419083,
-0.012576700188219547,
0.015965484082698822,
-0.045739538967609406,
-0.19589705765247345,
0.05146520957350731,
0.061254896223545074,
0.012523573823273182,
0.03504384309053421,
-0.1466376781463623,
0.08310041576623917,
0.07613673806190491,
-0.026607809588313103,
0.09859517216682434,
-0.2614401578903198,
-0.13502521812915802,
0.08145444095134735,
0.18697060644626617,
0.06861665099859238,
-0.1417100429534912,
-0.05066847428679466,
-0.019209621474146843,
-0.10277109593153,
0.09705248475074768,
-0.06557640433311462,
0.10571310669183731,
-0.023529211059212685,
0.021211566403508186,
0.010752649046480656,
-0.05740690231323242,
0.13592416048049927,
-0.03624329715967178,
0.10247301310300827,
-0.05443224310874939,
-0.018735108897089958,
0.07642009854316711,
-0.07744927704334259,
0.05592646077275276,
-0.09111884981393814,
0.059978265315294266,
-0.0558868907392025,
-0.01657266728579998,
-0.07043074071407318,
0.02667560800909996,
-0.02256162464618683,
-0.03026893176138401,
-0.049578890204429626,
0.02891109138727188,
0.04675905033946037,
-0.0051739951595664024,
0.19764302670955658,
0.044853325933218,
0.0896855890750885,
0.12363465875387192,
0.049787599593400955,
-0.07119978219270706,
-0.09608714282512665,
-0.02365751378238201,
-0.02882142923772335,
0.08613590151071548,
-0.17688047885894775,
0.05149596557021141,
0.09792756289243698,
0.013895310461521149,
0.14824652671813965,
0.050987608730793,
-0.031086940318346024,
0.01936529390513897,
0.06785588711500168,
-0.15867570042610168,
-0.16319674253463745,
-0.03470935299992561,
-0.01576690375804901,
-0.12121539562940598,
0.057762738317251205,
0.11973410099744797,
-0.08759941905736923,
0.004827198106795549,
-0.007844008505344391,
0.013003088533878326,
-0.00606570253148675,
0.1636328548192978,
0.07720037549734116,
0.042931657284498215,
-0.08830242604017258,
0.0911063477396965,
0.05849052593111992,
-0.10815457254648209,
0.014330006204545498,
0.01887066289782524,
-0.10005439072847366,
-0.03695818781852722,
0.05656116455793381,
0.15085114538669586,
-0.009405829012393951,
-0.0423177108168602,
-0.14540989696979523,
-0.10245504230260849,
0.05060161277651787,
0.12661011517047882,
0.09059479832649231,
0.021504150703549385,
-0.0181596539914608,
-0.0005341914365999401,
-0.10807254910469055,
0.12481489032506943,
0.03806748241186142,
0.10147184878587723,
-0.22307638823986053,
0.06419514864683151,
0.011358436197042465,
0.031215839087963104,
-0.019420325756072998,
0.0307367704808712,
-0.09792373329401016,
-0.0116913802921772,
-0.08193053305149078,
0.04103265330195427,
-0.02757757157087326,
0.00624421052634716,
-0.008270635269582272,
-0.0739615187048912,
-0.061846647411584854,
0.03316595032811165,
-0.09604858607053757,
-0.044522274285554886,
0.04052363708615303,
0.0645059123635292,
-0.09813778847455978,
-0.02953462488949299,
0.029144762083888054,
-0.08361127227544785,
0.08063619583845139,
0.010631970129907131,
0.005353987216949463,
0.02557472698390484,
-0.10803678631782532,
0.01407899521291256,
0.07511188834905624,
0.004432829562574625,
0.03333457559347153,
-0.10430612415075302,
0.007174383383244276,
-0.009502493776381016,
0.0051411837339401245,
-0.004510160535573959,
0.11051212251186371,
-0.12664100527763367,
-0.023936113342642784,
-0.03157968819141388,
-0.029592912644147873,
-0.06070547550916672,
0.06300317496061325,
0.08060941845178604,
0.004271997604519129,
0.19781270623207092,
-0.0865250825881958,
0.002997100353240967,
-0.22336992621421814,
0.006325787398964167,
-0.0049057528376579285,
-0.12941473722457886,
-0.1276504397392273,
-0.03263981267809868,
0.05531114339828491,
-0.0746862143278122,
0.08764737099409103,
0.008752613328397274,
0.02262062579393387,
0.03725124150514603,
0.011718075722455978,
0.0031094292644411325,
0.02847927249968052,
0.19167347252368927,
-0.009918062947690487,
-0.025762364268302917,
0.0629638209939003,
0.017694173380732536,
0.11817898601293564,
0.10049540549516678,
0.10410145670175552,
0.15787631273269653,
-0.05292106047272682,
0.10583890229463577,
0.04472494125366211,
-0.021789472550153732,
-0.1715007871389389,
0.08797641098499298,
-0.08133774250745773,
0.13711142539978027,
-0.01640157215297222,
0.17644692957401276,
0.11232228577136993,
-0.15540780127048492,
0.020292319357395172,
-0.04004890099167824,
-0.07662557810544968,
-0.06550176441669464,
-0.14246293902397156,
-0.11489930003881454,
-0.17524918913841248,
0.015209940262138844,
-0.09729600697755814,
0.004677677992731333,
0.07055871933698654,
-0.002939539961516857,
-0.029206126928329468,
0.21137772500514984,
0.04825558885931969,
-0.0009786889422684908,
0.06858247518539429,
0.004552985541522503,
-0.05954104661941528,
-0.050764668732881546,
-0.08189984411001205,
0.0387030728161335,
-0.006322448141872883,
0.03204234689474106,
-0.03325853869318962,
-0.007471595425158739,
0.04869880899786949,
0.006990345660597086,
-0.1147676333785057,
0.015129521489143372,
0.012714114040136337,
0.012597896158695221,
0.006560739129781723,
-0.0022024193312972784,
0.0166153721511364,
-0.014992176555097103,
0.18470041453838348,
-0.06015435606241226,
-0.0034500849433243275,
-0.11627183854579926,
0.12443462014198303,
0.016687816008925438,
-0.015556525439023972,
0.026713306084275246,
-0.08519620448350906,
0.02885940670967102,
0.2292078137397766,
0.13895688951015472,
-0.036076609045267105,
-0.005012980662286282,
-0.005183031316846609,
-0.019403737038373947,
-0.03225622698664665,
0.0867382362484932,
0.08340371400117874,
-0.04555077850818634,
-0.06274514645338058,
-0.016789263114333153,
-0.04779066890478134,
-0.01678798720240593,
-0.029041195288300514,
0.04016515240073204,
0.016807809472084045,
0.020991351455450058,
-0.07664187997579575,
0.045506253838539124,
0.01725965552031994,
-0.07140489667654037,
0.0889132022857666,
-0.2014935165643692,
-0.1402713507413864,
-0.034390538930892944,
0.10448431223630905,
-0.013708268292248249,
0.03265117481350899,
-0.027929725125432014,
0.017523126676678658,
0.06875300407409668,
-0.02331601269543171,
-0.07561469078063965,
-0.09508862346410751,
0.05189156159758568,
-0.11681663244962692,
0.24745137989521027,
-0.033571965992450714,
0.0005925163277424872,
0.11301527917385101,
0.023798834532499313,
-0.11888739466667175,
0.052595000714063644,
0.027569809928536415,
-0.043499164283275604,
0.025901509448885918,
0.1163618341088295,
-0.02455432526767254,
0.11552682518959045,
0.03032822348177433,
-0.09560415893793106,
-0.012705275788903236,
-0.04645555838942528,
-0.04193851724267006,
-0.05560345575213432,
-0.02852756343781948,
-0.0662735104560852,
0.13015250861644745,
0.17604254186153412,
-0.04377282038331032,
-0.03008260764181614,
-0.061905451118946075,
0.031984828412532806,
0.09311630576848984,
0.024463176727294922,
-0.011336607858538628,
-0.2271403968334198,
0.01766611449420452,
0.021512774750590324,
0.0010952819138765335,
-0.22889180481433868,
-0.10560157150030136,
-0.014617710374295712,
-0.05362720042467117,
-0.08014020323753357,
0.09380429983139038,
0.12119107693433762,
0.05318662151694298,
-0.06136804446578026,
-0.050950270146131516,
-0.06596160680055618,
0.15859675407409668,
-0.12128744274377823,
-0.09623292088508606
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | shahzebnaveed/t5-large_PREFIX_TUNING_SEQ2SEQ | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:29:44+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | MasakiK/stack-llama-2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T15:29:59+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# emotion_classification
This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 1.1249
- Accuracy: 0.6188
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 15
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 40 | 1.8344 | 0.3 |
| No log | 2.0 | 80 | 1.5609 | 0.4375 |
| No log | 3.0 | 120 | 1.4819 | 0.4562 |
| No log | 4.0 | 160 | 1.3477 | 0.5188 |
| No log | 5.0 | 200 | 1.2618 | 0.5813 |
| No log | 6.0 | 240 | 1.1946 | 0.5813 |
| No log | 7.0 | 280 | 1.1800 | 0.5875 |
| No log | 8.0 | 320 | 1.1921 | 0.5625 |
| No log | 9.0 | 360 | 1.1274 | 0.6 |
| No log | 10.0 | 400 | 1.0886 | 0.65 |
| No log | 11.0 | 440 | 1.0750 | 0.6125 |
| No log | 12.0 | 480 | 1.1349 | 0.575 |
| 1.0832 | 13.0 | 520 | 1.0841 | 0.5875 |
| 1.0832 | 14.0 | 560 | 1.1195 | 0.5813 |
| 1.0832 | 15.0 | 600 | 1.0865 | 0.6188 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "google/vit-base-patch16-224-in21k", "model-index": [{"name": "emotion_classification", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "train", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.61875, "name": "Accuracy"}]}]}]} | image-classification | malifiahm/emotion_classification | [
"transformers",
"tensorboard",
"safetensors",
"vit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:google/vit-base-patch16-224-in21k",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:31:51+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| emotion\_classification
=======================
This model is a fine-tuned version of google/vit-base-patch16-224-in21k on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 1.1249
* Accuracy: 0.6188
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 15
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 15",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 15",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
86,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #vit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-google/vit-base-patch16-224-in21k #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 15### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.11811337620019913,
0.13483555614948273,
-0.002608612412586808,
0.11961443722248077,
0.14219510555267334,
0.0050515602342784405,
0.13914495706558228,
0.13795039057731628,
-0.07118016481399536,
0.08092769235372543,
0.14925982058048248,
0.12972450256347656,
0.02959899976849556,
0.18728971481323242,
-0.049368128180503845,
-0.222564697265625,
0.02601581998169422,
0.04734741523861885,
-0.0522691011428833,
0.12076617032289505,
0.0886019691824913,
-0.1376405656337738,
0.1166982427239418,
0.02592886984348297,
-0.20304442942142487,
-0.007615117356181145,
0.029098207131028175,
-0.056919001042842865,
0.11676168441772461,
0.037896107882261276,
0.09137658774852753,
0.02816101908683777,
0.05190208926796913,
-0.14934077858924866,
0.010603067465126514,
0.07517293095588684,
-0.008613502606749535,
0.097054123878479,
0.055827803909778595,
0.012219731695950031,
0.011749614961445332,
-0.09264149516820908,
0.03983303904533386,
0.02654171548783779,
-0.11214680224657059,
-0.2338249832391739,
-0.08387917280197144,
0.05802435800433159,
0.07740037143230438,
0.07054296880960464,
-0.0028684416320174932,
0.14147241413593292,
-0.008090575225651264,
0.09724876284599304,
0.2288333922624588,
-0.2725082337856293,
-0.07707613706588745,
0.0382843017578125,
0.015430927276611328,
0.0770019143819809,
-0.10192235559225082,
0.011624174192547798,
0.05947103351354599,
0.014623235911130905,
0.1543622612953186,
-0.004538698587566614,
-0.013555784709751606,
-0.02619186043739319,
-0.12524428963661194,
-0.06403586268424988,
0.1917194277048111,
0.08952893316745758,
-0.04724816977977753,
-0.08089526742696762,
-0.08033386617898941,
-0.13759849965572357,
-0.04689684137701988,
-0.011270390823483467,
0.057964008301496506,
-0.03416437655687332,
-0.06219732016324997,
-0.03524642437696457,
-0.09843727201223373,
-0.07024387270212173,
-0.014886156655848026,
0.09846138954162598,
0.05437090992927551,
0.013064629398286343,
-0.020326262339949608,
0.08063661307096481,
-0.041325896978378296,
-0.14483605325222015,
-0.007839716970920563,
0.01815846748650074,
0.02448871172964573,
-0.03133099898695946,
-0.02552160993218422,
-0.1136103942990303,
0.02110660821199417,
0.10973213613033295,
-0.0706816241145134,
0.056373026221990585,
-0.018844274803996086,
0.0508960597217083,
-0.11135609447956085,
0.18870708346366882,
-0.04998411610722542,
0.018273813650012016,
0.04010777920484543,
0.10290872305631638,
0.04924130439758301,
-0.00162196543533355,
-0.10518983751535416,
0.018268786370754242,
0.11962330341339111,
0.00517577538266778,
-0.03565511107444763,
0.08145362883806229,
-0.06206951290369034,
-0.029533717781305313,
0.07810026407241821,
-0.0864962711930275,
0.025279588997364044,
-0.006731956731528044,
-0.05419694632291794,
-0.05826117843389511,
0.04743794724345207,
-0.011842501349747181,
-0.013043711893260479,
0.04374115914106369,
-0.10329610109329224,
0.01217141468077898,
-0.06757714599370956,
-0.1077800989151001,
0.013161982409656048,
-0.11606291681528091,
0.016227610409259796,
-0.12267278134822845,
-0.13887597620487213,
-0.012597721070051193,
0.06051623821258545,
-0.03007390722632408,
-0.05283919721841812,
-0.04174118861556053,
-0.08089839667081833,
0.026779508218169212,
0.0049300664104521275,
0.048026394098997116,
-0.05676327645778656,
0.0882662981748581,
0.04600731283426285,
0.07642398029565811,
-0.018046606332063675,
0.04749411344528198,
-0.08802305161952972,
0.05782405287027359,
-0.20476210117340088,
0.037461649626493454,
-0.05966132506728172,
0.08739495277404785,
-0.12054326385259628,
-0.08681707829236984,
8.245754656854842e-8,
-0.020297737792134285,
0.06496932357549667,
0.10898001492023468,
-0.13716930150985718,
-0.058494143187999725,
0.17197832465171814,
-0.10235631465911865,
-0.1559467613697052,
0.11321426928043365,
-0.03041159175336361,
0.029482407495379448,
0.05470358952879906,
0.19649885594844818,
0.08132990449666977,
-0.11062505096197128,
-0.007079981733113527,
-0.03057224303483963,
0.03357299789786339,
-0.05531175062060356,
0.07652188092470169,
-0.0003474817203823477,
-0.011214455589652061,
0.02246684953570366,
-0.09512591361999512,
0.0643419474363327,
-0.07409214973449707,
-0.08478579670190811,
-0.06557157635688782,
-0.08932899683713913,
0.04289957135915756,
0.05970412492752075,
0.06635873019695282,
-0.09980020672082901,
-0.09068282693624496,
0.03020690195262432,
0.08147106319665909,
-0.09339673817157745,
0.01664392277598381,
-0.08065690100193024,
0.11216481029987335,
-0.11094366014003754,
0.0014294554712250829,
-0.13545817136764526,
-0.03065856173634529,
0.048524606972932816,
-0.06409776210784912,
-0.008567430078983307,
-0.03686706721782684,
0.0737147331237793,
0.06236539036035538,
-0.06374558061361313,
-0.07231000065803528,
-0.039871957153081894,
-0.003146301954984665,
-0.0989849716424942,
-0.19341829419136047,
-0.023916881531476974,
-0.02699757181107998,
0.1048165038228035,
-0.21783189475536346,
0.04208202660083771,
0.05092693492770195,
0.10128840059041977,
0.05917399749159813,
-0.03185688704252243,
0.004048972390592098,
0.016817180439829826,
-0.03980059549212456,
-0.08895774185657501,
0.06217832490801811,
0.01427388284355402,
-0.06635143607854843,
0.0068201846443116665,
-0.09787274897098541,
0.1759055107831955,
0.13049612939357758,
-0.03351129591464996,
-0.06344152987003326,
-0.004213555715978146,
-0.043545614928007126,
-0.03528248518705368,
-0.03702044486999512,
0.00897136703133583,
0.08078207075595856,
-0.00947840977460146,
0.16266374289989471,
-0.10526886582374573,
-0.026199668645858765,
0.0590047612786293,
-0.029783276841044426,
-0.037960272282361984,
0.08977246284484863,
0.06990905851125717,
-0.13621707260608673,
0.14820963144302368,
0.16512958705425262,
-0.06864918768405914,
0.12576285004615784,
-0.04820854216814041,
-0.06320492178201675,
-0.023548301309347153,
0.04079396277666092,
0.03244993835687637,
0.12839201092720032,
-0.12003456056118011,
-0.012698296457529068,
0.023860221728682518,
0.002722275909036398,
-0.007653374224901199,
-0.2003244161605835,
-0.008881805464625359,
0.03923380747437477,
-0.060138143599033356,
0.02479555271565914,
-0.005791828967630863,
-0.021750187501311302,
0.08476974815130234,
0.007745973765850067,
-0.0419430285692215,
0.04653361067175865,
0.011344957165420055,
-0.06956381350755692,
0.19539684057235718,
-0.08346343785524368,
-0.21489690244197845,
-0.13166634738445282,
-0.0214513149112463,
-0.07969480007886887,
0.020822223275899887,
0.05844384431838989,
-0.09474411606788635,
-0.057422202080488205,
-0.1027030199766159,
-0.013067323714494705,
0.030706671997904778,
0.04023214057087898,
0.04244999215006828,
-0.0021424260921776295,
0.1310640424489975,
-0.09935938566923141,
-0.006750526838004589,
-0.010671090334653854,
-0.022405968979001045,
0.047964178025722504,
0.020353814586997032,
0.12129854410886765,
0.08637033402919769,
-0.028014538809657097,
0.03335375338792801,
-0.020929161459207535,
0.24402356147766113,
-0.07317174226045609,
-0.0031845003832131624,
0.15085476636886597,
0.020264659076929092,
0.06783358752727509,
0.13053905963897705,
0.03894012048840523,
-0.10269611328840256,
0.007967564277350903,
0.023408832028508186,
-0.025517692789435387,
-0.18748973309993744,
-0.017082426697015762,
-0.03992816433310509,
-0.005310575012117624,
0.1524132639169693,
0.054956067353487015,
0.059837352484464645,
0.09335610270500183,
0.0006898478604853153,
0.09150495380163193,
-0.005583284888416529,
0.08689933270215988,
0.1108987033367157,
0.04706966504454613,
0.1094338446855545,
-0.044298429042100906,
-0.028626078739762306,
0.032389797270298004,
0.016002558171749115,
0.22706560790538788,
0.00020310455875005573,
0.17483092844486237,
0.04707358777523041,
0.18886423110961914,
0.018033649772405624,
0.057034797966480255,
-0.022385064512491226,
-0.027543995529413223,
-0.009018163196742535,
-0.05375729873776436,
-0.020335029810667038,
0.03607749566435814,
-0.050183698534965515,
0.06412369012832642,
-0.09278424084186554,
0.04310345649719238,
0.06286797672510147,
0.2650319039821625,
0.03749196603894234,
-0.3765833079814911,
-0.09599293023347855,
-0.006457073148339987,
-0.015000157989561558,
-0.06306160241365433,
0.003157390048727393,
0.14556193351745605,
-0.06002454087138176,
0.059487201273441315,
-0.10448174923658371,
0.08238985389471054,
-0.04907115548849106,
0.021390307694673538,
0.07714858651161194,
0.09014001488685608,
0.007975099608302116,
0.056237027049064636,
-0.249930277466774,
0.25884684920310974,
0.015901101753115654,
0.06355836987495422,
-0.046657998114824295,
0.01285374816507101,
0.0358203649520874,
0.10374182462692261,
0.1107485443353653,
-0.005275308154523373,
-0.013981667347252369,
-0.17540819942951202,
-0.08827585726976395,
0.005992108955979347,
0.07163839787244797,
-0.045461300760507584,
0.08091539889574051,
-0.030246488749980927,
-0.023902583867311478,
0.05147729441523552,
-0.004220571834594011,
-0.08984536677598953,
-0.09435426443815231,
-0.008079295046627522,
0.04183362051844597,
0.013285839930176735,
-0.09489520639181137,
-0.09797141700983047,
-0.1038513109087944,
0.13047079741954803,
-0.017963053658604622,
-0.04099510610103607,
-0.12043735384941101,
0.08851757645606995,
0.058595508337020874,
-0.09324067831039429,
0.08146335929632187,
-0.029300404712557793,
0.13662868738174438,
0.029252292588353157,
-0.062240853905677795,
0.10969259589910507,
-0.06036292016506195,
-0.17226964235305786,
-0.04662119969725609,
0.10395712405443192,
-0.019283311441540718,
0.02545684389770031,
-0.0006019543507136405,
0.027186378836631775,
-0.011029436253011227,
-0.0596676804125309,
0.05846496298909187,
0.010062429122626781,
0.0574432834982872,
-0.0137277040630579,
-0.020587317645549774,
0.005209364928305149,
-0.06414493918418884,
-0.030352618545293808,
0.1347113996744156,
0.2454831451177597,
-0.09741272777318954,
0.004012614022940397,
0.021386010572314262,
-0.051253773272037506,
-0.19672775268554688,
0.04747342690825462,
0.06734459847211838,
0.0017792806029319763,
0.030634094029664993,
-0.1560230404138565,
0.07357994467020035,
0.08122958242893219,
-0.030620619654655457,
0.09420695155858994,
-0.26387789845466614,
-0.1334042251110077,
0.08049169182777405,
0.18394631147384644,
0.06955462694168091,
-0.144282728433609,
-0.05479405075311661,
-0.01088814064860344,
-0.09360235929489136,
0.09409633278846741,
-0.057727087289094925,
0.10505452752113342,
-0.028861433267593384,
0.002270476194098592,
0.0059490022249519825,
-0.05787969008088112,
0.12839028239250183,
-0.03276979550719261,
0.10822311788797379,
-0.05622243136167526,
-0.010700024664402008,
0.07683528959751129,
-0.07610113173723221,
0.06443601101636887,
-0.09060078859329224,
0.06328599900007248,
-0.06150584667921066,
-0.01632372848689556,
-0.07098719477653503,
0.03246122971177101,
-0.018721945583820343,
-0.026020457968115807,
-0.050546322017908096,
0.02465064451098442,
0.051601458340883255,
-0.001398571184836328,
0.19979743659496307,
0.04996337741613388,
0.08822537958621979,
0.1389024704694748,
0.04264577850699425,
-0.07611879706382751,
-0.0999656692147255,
-0.027445113286376,
-0.027116796001791954,
0.08754102140665054,
-0.18767961859703064,
0.05024249106645584,
0.09690632671117783,
0.011310224421322346,
0.14395952224731445,
0.04596976563334465,
-0.03312673792243004,
0.018624791875481606,
0.07134110480546951,
-0.15431475639343262,
-0.16505153477191925,
-0.03247111290693283,
-0.02027037926018238,
-0.11648004502058029,
0.06160777062177658,
0.11112837493419647,
-0.08592600375413895,
0.003426074981689453,
-0.008567458018660545,
0.015196014195680618,
-0.0017052257899194956,
0.1637272834777832,
0.08126139640808105,
0.04554775357246399,
-0.09228046983480453,
0.09902362525463104,
0.05416230857372284,
-0.1047559306025505,
0.02179264836013317,
0.025547826662659645,
-0.10391340404748917,
-0.038040343672037125,
0.06732680648565292,
0.14574941992759705,
0.0008763946243561804,
-0.05033043026924133,
-0.14442609250545502,
-0.09389335662126541,
0.058804791420698166,
0.1254853755235672,
0.09292245656251907,
0.014865768142044544,
-0.011163334362208843,
0.0007053414010442793,
-0.10467668622732162,
0.12055911868810654,
0.032729413360357285,
0.09729781746864319,
-0.2187214344739914,
0.05768940970301628,
0.018582215532660484,
0.03169427812099457,
-0.01972782239317894,
0.029827957972884178,
-0.09803614765405655,
-0.0168850626796484,
-0.061290498822927475,
0.041980184614658356,
-0.03650015592575073,
0.0057180942967534065,
-0.006341718137264252,
-0.06969422101974487,
-0.061450861394405365,
0.04156522452831268,
-0.1008782684803009,
-0.045561570674180984,
0.036558620631694794,
0.07008048892021179,
-0.10234192758798599,
-0.030030617490410805,
0.025391951203346252,
-0.08016756176948547,
0.07987046986818314,
0.0136539526283741,
0.0006410190835595131,
0.024142952635884285,
-0.10038962960243225,
0.011165685020387173,
0.08455300331115723,
0.0028227069415152073,
0.030509598553180695,
-0.10393964499235153,
0.0069387974217534065,
-0.002060604514554143,
0.0023920342791825533,
-0.009181283414363861,
0.10368724167346954,
-0.1335068941116333,
-0.024663101881742477,
-0.03798602893948555,
-0.03248472884297371,
-0.05953774228692055,
0.06146005913615227,
0.0861043781042099,
-0.0032658593263477087,
0.20133450627326965,
-0.08557508885860443,
0.0009489953517913818,
-0.2241573929786682,
0.00473367003723979,
-0.005016423296183348,
-0.13707740604877472,
-0.126900315284729,
-0.02808256819844246,
0.052527688443660736,
-0.07319869846105576,
0.09482517838478088,
0.016018196940422058,
0.005833158735185862,
0.03620041534304619,
0.004364409018307924,
-0.00327532016672194,
0.026778578758239746,
0.18415822088718414,
-0.006745863240212202,
-0.02089574746787548,
0.07106412947177887,
0.01989048160612583,
0.11759690195322037,
0.08453572541475296,
0.10050121694803238,
0.16296188533306122,
-0.04347632825374603,
0.10470350086688995,
0.04961124807596207,
-0.022944103926420212,
-0.17300568521022797,
0.10302316397428513,
-0.07557917386293411,
0.14314275979995728,
-0.012961996719241142,
0.1632063090801239,
0.12230784446001053,
-0.15819421410560608,
0.02951657958328724,
-0.028552694246172905,
-0.07274667173624039,
-0.07123821973800659,
-0.14293955266475677,
-0.1163991317152977,
-0.18582849204540253,
0.014975468628108501,
-0.09832064062356949,
0.006791998632252216,
0.0742390975356102,
-0.009317392483353615,
-0.022757386788725853,
0.20721586048603058,
0.050449710339307785,
-0.0017955147195607424,
0.07085927575826645,
-0.0001608180027687922,
-0.06834685802459717,
-0.0601092129945755,
-0.08387629687786102,
0.03922131285071373,
-0.009413119405508041,
0.03239017724990845,
-0.03048616647720337,
-0.00822769571095705,
0.04943614453077316,
-0.000030611128750024363,
-0.11058830469846725,
0.017509929835796356,
0.015752140432596207,
0.00987994484603405,
0.003946311306208372,
0.0035220300778746605,
0.008244358003139496,
-0.009524581953883171,
0.18381452560424805,
-0.05413465201854706,
-0.006990767549723387,
-0.11899719387292862,
0.12412447482347488,
0.028377491980791092,
-0.014865638688206673,
0.029340052977204323,
-0.07932072877883911,
0.026510966941714287,
0.21984656155109406,
0.14379233121871948,
-0.018433962017297745,
-0.0014454583870247006,
-0.0070984624326229095,
-0.020136944949626923,
-0.028865426778793335,
0.09244752675294876,
0.09355153888463974,
-0.04361030086874962,
-0.05534041300415993,
-0.023654403164982796,
-0.047336239367723465,
-0.0166631992906332,
-0.03569130226969719,
0.038284845650196075,
0.018391557037830353,
0.014795871451497078,
-0.06359363347291946,
0.04177533835172653,
0.019448641687631607,
-0.07260351628065109,
0.08854573220014572,
-0.19557501375675201,
-0.1386980414390564,
-0.034064747393131256,
0.09884845465421677,
-0.0047919051721692085,
0.02931911125779152,
-0.02098333276808262,
0.015507815405726433,
0.06938814371824265,
-0.02387051098048687,
-0.08344113826751709,
-0.10004539042711258,
0.05112903192639351,
-0.13180024921894073,
0.2446517050266266,
-0.03228965401649475,
0.006066294386982918,
0.11191846430301666,
0.018580928444862366,
-0.11916393786668777,
0.05102118104696274,
0.025515109300613403,
-0.03693299740552902,
0.0256801825016737,
0.11176413297653198,
-0.0208415687084198,
0.10723520815372467,
0.03339647874236107,
-0.08970727026462555,
-0.018056262284517288,
-0.05356735363602638,
-0.04208860546350479,
-0.056582849472761154,
-0.02354576624929905,
-0.06779967248439789,
0.12323543429374695,
0.17353424429893494,
-0.042273472994565964,
-0.02678775228559971,
-0.06389665603637695,
0.03773527592420578,
0.09117208421230316,
0.018846124410629272,
-0.013722807168960571,
-0.2285662740468979,
0.01779686100780964,
0.02249043434858322,
-0.003642615396529436,
-0.21387918293476105,
-0.10973727703094482,
-0.017915982753038406,
-0.053049616515636444,
-0.08842597901821136,
0.08633087575435638,
0.11947934329509735,
0.05113840848207474,
-0.060327135026454926,
-0.04741908237338066,
-0.06931841373443604,
0.15980486571788788,
-0.12278150022029877,
-0.08498451858758926
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0208_19
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0073
- Wer: 0.1871
- Cer: 0.1524
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 13
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 3.5447 | 1.0 | 120 | 3.4564 | 0.9849 | 0.9954 |
| 1.4868 | 2.0 | 240 | 1.3054 | 0.9996 | 0.7825 |
| 0.7829 | 3.0 | 360 | 0.6057 | 0.8097 | 0.5026 |
| 0.4487 | 4.0 | 480 | 0.2791 | 0.4736 | 0.2457 |
| 0.325 | 5.0 | 600 | 0.1375 | 0.3020 | 0.1677 |
| 0.2221 | 6.0 | 720 | 0.0720 | 0.2321 | 0.1591 |
| 0.2002 | 7.0 | 840 | 0.0372 | 0.2123 | 0.1295 |
| 0.1661 | 8.0 | 960 | 0.0276 | 0.2050 | 0.1247 |
| 0.129 | 9.0 | 1080 | 0.0153 | 0.1955 | 0.1571 |
| 0.0707 | 10.0 | 1200 | 0.0124 | 0.1925 | 0.1378 |
| 0.0737 | 11.0 | 1320 | 0.0107 | 0.1897 | 0.1601 |
| 0.0818 | 12.0 | 1440 | 0.0084 | 0.1879 | 0.1535 |
| 0.0672 | 13.0 | 1560 | 0.0073 | 0.1871 | 0.1524 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0208_19", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0208_19 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:33:16+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0208\_19
========================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0073
* Wer: 0.1871
* Cer: 0.1524
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 6e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 13
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 13",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 13",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 13### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.14091284573078156,
0.15205152332782745,
-0.0005105260061100125,
0.10033951699733734,
0.1181183010339737,
0.008221986703574657,
0.17481480538845062,
0.1506439745426178,
-0.041153956204652786,
0.11094454675912857,
0.11317409574985504,
0.0614452101290226,
0.05511699616909027,
0.19514599442481995,
-0.08128761500120163,
-0.22106203436851501,
0.07670484483242035,
-0.004085349850356579,
0.009533348493278027,
0.11209411919116974,
0.07080043107271194,
-0.11798843741416931,
0.09004402160644531,
-0.007307722698897123,
-0.14268426597118378,
-0.042056888341903687,
0.01633509248495102,
-0.10968092083930969,
0.10900094360113144,
0.009458199143409729,
0.0664912536740303,
0.03489179164171219,
0.08977682143449783,
-0.18663492798805237,
0.0022737146355211735,
0.01670593023300171,
0.014722064137458801,
0.07421526312828064,
0.04303494468331337,
-0.0001994612393900752,
0.003381873946636915,
-0.11351623386144638,
0.03615633025765419,
0.015368473716080189,
-0.1170836091041565,
-0.19776363670825958,
-0.07875938713550568,
0.016152849420905113,
0.09867508709430695,
0.08499042689800262,
-0.020623158663511276,
0.12381237000226974,
0.0014898965600878,
0.07942700386047363,
0.19691592454910278,
-0.31140416860580444,
-0.05503339692950249,
-0.015968721359968185,
0.038639381527900696,
0.08199921250343323,
-0.1019219160079956,
-0.01867164485156536,
0.04946177080273628,
0.022153420373797417,
0.09001468122005463,
-0.03098279982805252,
-0.03578285127878189,
-0.010924735106527805,
-0.12019715458154678,
-0.03889642655849457,
0.18758343160152435,
0.07228991389274597,
-0.06318531185388565,
-0.0794241726398468,
-0.06494627147912979,
-0.12038347125053406,
-0.05479174852371216,
-0.007069990038871765,
0.02657814882695675,
-0.039321690797805786,
-0.09981369227170944,
-0.005235292017459869,
-0.07976865768432617,
-0.09163050353527069,
-0.01816716231405735,
0.1763553023338318,
0.010768347419798374,
0.014017139561474323,
-0.011575044132769108,
0.05442178621888161,
-0.022576941177248955,
-0.18500590324401855,
-0.022519683465361595,
0.028132937848567963,
-0.032301802188158035,
-0.013906015083193779,
-0.043227724730968475,
-0.03310549259185791,
0.04410584270954132,
0.11580466479063034,
-0.022221477702260017,
0.06561435014009476,
-0.024849310517311096,
0.0016406449722126126,
-0.08626354485750198,
0.1823546290397644,
-0.06440720707178116,
-0.0659199133515358,
0.01956027001142502,
0.12664580345153809,
0.06286849826574326,
-0.023285599425435066,
-0.09889284521341324,
-0.009051132015883923,
0.14586183428764343,
0.03575187921524048,
-0.0417223758995533,
0.05045789107680321,
-0.038734037429094315,
-0.01419241726398468,
0.05896828696131706,
-0.12167445570230484,
0.025085192173719406,
0.021283850073814392,
-0.06307906657457352,
-0.023088160902261734,
-0.011692028492689133,
0.012597832828760147,
0.013168255798518658,
0.052073437720537186,
-0.08249229192733765,
0.003705754177644849,
-0.02361825667321682,
-0.09154974669218063,
0.027510233223438263,
-0.0658554881811142,
-0.0002562470908742398,
-0.1075483113527298,
-0.1798238605260849,
-0.01776859723031521,
0.02420281060039997,
-0.049744103103876114,
-0.010194470174610615,
-0.11160378903150558,
-0.09753140062093735,
0.04774530604481697,
-0.02273140475153923,
0.035124897956848145,
-0.07945280522108078,
0.10853233933448792,
0.08009723573923111,
0.08689231425523758,
-0.04060021787881851,
0.02632182277739048,
-0.09564634412527084,
0.032809577882289886,
-0.17648452520370483,
0.0747283399105072,
-0.05502145364880562,
0.034581489861011505,
-0.12023770809173584,
-0.06769201904535294,
0.019593888893723488,
-0.02207692712545395,
0.06984969973564148,
0.1422162652015686,
-0.18988460302352905,
-0.05618421360850334,
0.19592706859111786,
-0.12054263800382614,
-0.14252912998199463,
0.12800954282283783,
-0.0358167365193367,
0.03821597248315811,
0.0692887231707573,
0.22372925281524658,
0.03177893906831741,
-0.1055835485458374,
-0.039934009313583374,
-0.06312856823205948,
0.0833640992641449,
-0.03678765147924423,
0.11158882081508636,
0.0049759685061872005,
-0.0017774647567421198,
0.016245972365140915,
-0.08059127628803253,
0.0324857197701931,
-0.070980004966259,
-0.09989545494318008,
-0.04381024092435837,
-0.10584957152605057,
0.028362542390823364,
0.016560189425945282,
0.0555352084338665,
-0.09853999316692352,
-0.07023223489522934,
0.010988261550664902,
0.10849704593420029,
-0.11659286916255951,
0.012931226752698421,
-0.10408444702625275,
0.09343922883272171,
-0.1132953017950058,
-0.020289508625864983,
-0.15419894456863403,
-0.007004518993198872,
0.053447116166353226,
0.020570576190948486,
0.014475558884441853,
-0.07579518109560013,
0.08239666372537613,
0.07627762109041214,
-0.049803078174591064,
-0.07374439388513565,
-0.004512934945523739,
0.017329076305031776,
-0.06178053468465805,
-0.17273446917533875,
-0.028687167912721634,
-0.05409182608127594,
0.15981686115264893,
-0.1643008291721344,
0.0008592766243964434,
0.009241053834557533,
0.0898156613111496,
0.04330800473690033,
-0.02304905652999878,
0.020005861297249794,
0.048081524670124054,
-0.026066740974783897,
-0.07161523401737213,
0.029428843408823013,
0.015192309394478798,
-0.10333897918462753,
0.021559974178671837,
-0.16676262021064758,
0.14967842400074005,
0.13897953927516937,
0.04214247688651085,
-0.0530753992497921,
0.020138215273618698,
-0.01438229437917471,
-0.04224589094519615,
-0.0544196218252182,
-0.015210306271910667,
0.10279848426580429,
0.007827705703675747,
0.12034579366445541,
-0.10298613458871841,
0.015775298699736595,
0.0648447573184967,
-0.027702417224645615,
-0.028374601155519485,
0.08159936964511871,
0.01134854182600975,
-0.1394982486963272,
0.13100019097328186,
0.1113022193312645,
-0.07144801318645477,
0.12656442821025848,
-0.061711374670267105,
-0.08448085188865662,
-0.0489732064306736,
0.03438488394021988,
0.034671247005462646,
0.13701258599758148,
-0.07972347736358643,
-0.021082036197185516,
0.02087889239192009,
0.022852823138237,
-0.015685906633734703,
-0.19351166486740112,
-0.020694324746727943,
0.01444272045046091,
-0.09511327743530273,
-0.009177693165838718,
0.005948662292212248,
-0.01745542138814926,
0.09474706649780273,
-0.00020246328494977206,
-0.1138736829161644,
0.023384930565953255,
-0.015161545015871525,
-0.08789047598838806,
0.17237284779548645,
-0.09211786091327667,
-0.17366443574428558,
-0.13716208934783936,
-0.07250114530324936,
-0.05598360672593117,
0.03794451430439949,
0.059924956411123276,
-0.06654167175292969,
-0.0401560477912426,
-0.11523009091615677,
-0.04730040207505226,
0.03217366710305214,
0.04507134109735489,
0.050572942942380905,
-0.009264782071113586,
0.06682556122541428,
-0.08176732808351517,
-0.005150897428393364,
-0.014384080655872822,
-0.008558395318686962,
0.0291933324187994,
-0.00009046850027516484,
0.12677450478076935,
0.12186631560325623,
0.005318664480000734,
0.02372398041188717,
-0.03756563737988472,
0.22658653557300568,
-0.06981690227985382,
-0.01889536902308464,
0.12389808893203735,
-0.026990750804543495,
0.045968543738126755,
0.17758749425411224,
0.03098992258310318,
-0.10714825242757797,
0.0018714808393269777,
-0.04990280792117119,
-0.01623891294002533,
-0.19011567533016205,
-0.03400854393839836,
-0.04805027320981026,
0.014483409933745861,
0.10041266679763794,
0.030758772045373917,
0.014088752679526806,
0.0481368824839592,
0.021463293582201004,
0.045564472675323486,
0.003916191402822733,
0.08064043521881104,
0.09727296233177185,
0.07678142189979553,
0.10750129818916321,
-0.0320168100297451,
-0.04853944480419159,
0.03288719430565834,
0.02195957489311695,
0.2010214626789093,
0.029412759467959404,
0.19249524176120758,
0.0008534341468475759,
0.15456104278564453,
0.026087481528520584,
0.07956702262163162,
0.018622564151883125,
0.00969894602894783,
-0.021107744425535202,
-0.0782138854265213,
-0.05399259552359581,
0.05468466505408287,
-0.016200266778469086,
0.060613445937633514,
-0.10495587438344955,
0.021359672769904137,
0.04980064928531647,
0.27155253291130066,
0.08710509538650513,
-0.3681391477584839,
-0.08723757416009903,
0.02056852914392948,
-0.0367141030728817,
-0.019406549632549286,
0.0170770063996315,
0.15361151099205017,
-0.06106509640812874,
0.06824149191379547,
-0.07190810889005661,
0.0635044053196907,
-0.06401772052049637,
0.01933128945529461,
0.02488080784678459,
0.0469856895506382,
0.003011294174939394,
0.03165389224886894,
-0.23973114788532257,
0.2882067561149597,
0.035894788801670074,
0.0940871462225914,
-0.055969949811697006,
-0.003600972704589367,
0.039963994175195694,
-0.006376258563250303,
0.1179264485836029,
-0.02413090690970421,
-0.1115347370505333,
-0.17906704545021057,
-0.13471442461013794,
0.049174726009368896,
0.10534682124853134,
-0.006670607253909111,
0.11574938893318176,
-0.013927025720477104,
-0.044018957763910294,
0.04482569545507431,
-0.02398107573390007,
-0.07844943553209305,
-0.07496035844087601,
0.009061452001333237,
0.1137252226471901,
0.044010818004608154,
-0.04929763451218605,
-0.09600932151079178,
-0.08880262821912766,
0.09077800065279007,
0.005459045059978962,
-0.006402952596545219,
-0.10512150824069977,
0.01749139279127121,
0.15044106543064117,
-0.09205936640501022,
0.0522707961499691,
0.008970644325017929,
0.11001493781805038,
0.027097104117274284,
-0.04969252645969391,
0.09083406627178192,
-0.06220528855919838,
-0.17888526618480682,
-0.05069049820303917,
0.13796508312225342,
-0.007938303053379059,
0.04343928396701813,
0.02035207860171795,
0.05169060453772545,
-0.005406050942838192,
-0.06730544567108154,
0.03292584419250488,
0.025866858661174774,
0.04112234711647034,
0.01931679993867874,
-0.012161804363131523,
-0.08930954337120056,
-0.09294039756059647,
-0.023337464779615402,
0.14976347982883453,
0.29836949706077576,
-0.06594888865947723,
0.018047410994768143,
0.08729980140924454,
-0.01738031581044197,
-0.15297392010688782,
-0.004694323055446148,
0.04334993660449982,
0.043911583721637726,
-0.004424355458468199,
-0.12294171005487442,
0.04545203968882561,
0.06142594665288925,
-0.04549773037433624,
0.07565579563379288,
-0.24772465229034424,
-0.12760968506336212,
0.08769885450601578,
0.13305044174194336,
0.12336176633834839,
-0.15242372453212738,
-0.0671047791838646,
-0.022867273539304733,
-0.10795829445123672,
0.10292781889438629,
-0.07123148441314697,
0.133108988404274,
-0.0024851039052009583,
0.06633006781339645,
0.0072144437581300735,
-0.05155085772275925,
0.1505575180053711,
0.02278006263077259,
0.052734047174453735,
-0.022657299414277077,
-0.01647016778588295,
0.0461294911801815,
-0.07523658126592636,
0.0686757043004036,
-0.08497676253318787,
0.05090241879224777,
-0.06111409515142441,
-0.024924645200371742,
-0.06217329576611519,
-0.006238694768399,
0.003862778190523386,
-0.035174354910850525,
-0.010881854221224785,
0.036833688616752625,
0.05839431658387184,
0.003138473490253091,
0.13216084241867065,
0.011720972135663033,
0.08360262960195541,
0.14748983085155487,
0.08894938975572586,
-0.03929465636610985,
0.016147635877132416,
-0.006373990327119827,
-0.056030645966529846,
0.05357246473431587,
-0.13212881982326508,
0.048056621104478836,
0.09713246673345566,
0.018338387832045555,
0.15946725010871887,
0.04713453724980354,
-0.049633853137493134,
0.03863062337040901,
0.06935584545135498,
-0.1588054746389389,
-0.11127293854951859,
0.002818298526108265,
-0.011274590156972408,
-0.11128677427768707,
0.04898195341229439,
0.1381271481513977,
-0.07052746415138245,
-0.006605560891330242,
-0.01688520982861519,
0.021399231627583504,
-0.03906308487057686,
0.19976428151130676,
0.042578306049108505,
0.051872946321964264,
-0.10962759703397751,
0.08047158271074295,
0.05671453848481178,
-0.08795212209224701,
0.048727862536907196,
0.03940500319004059,
-0.11479135602712631,
-0.022992126643657684,
-0.0004998566000722349,
0.14125654101371765,
0.0035943782422691584,
-0.07564717531204224,
-0.13771472871303558,
-0.08820784837007523,
0.034642357379198074,
0.17745783925056458,
0.06828415393829346,
0.03635571151971817,
-0.017873700708150864,
-0.0021263936068862677,
-0.10390230268239975,
0.09364498406648636,
0.07278639823198318,
0.07531937956809998,
-0.15001094341278076,
0.08043059706687927,
-0.008228888735175133,
0.02778775990009308,
-0.020601503551006317,
0.017201902344822884,
-0.10918700695037842,
0.005555889569222927,
-0.09797751903533936,
0.05703280121088028,
-0.07829289883375168,
-0.016213063150644302,
-0.001668226788751781,
-0.08194150030612946,
-0.0614941343665123,
0.011479169130325317,
-0.08735302090644836,
-0.026110248640179634,
0.0031852193642407656,
0.04301946610212326,
-0.1362219750881195,
-0.037530720233917236,
0.023305969312787056,
-0.09734237939119339,
0.08377251774072647,
0.08612602949142456,
-0.019566241651773453,
0.0471237450838089,
-0.09531887620687485,
-0.022078856825828552,
0.08248338848352432,
0.002284811343997717,
0.05076700448989868,
-0.14360865950584412,
-0.014379849657416344,
0.03114495798945427,
0.05009199678897858,
0.02169564925134182,
0.14792414009571075,
-0.09708547592163086,
0.007087590638548136,
-0.06750298291444778,
-0.01226793136447668,
-0.05647427588701248,
0.021549979224801064,
0.1419900804758072,
0.003237504046410322,
0.18263021111488342,
-0.09533888101577759,
0.02309240773320198,
-0.19770433008670807,
0.0011794057209044695,
-0.037223946303129196,
-0.12563908100128174,
-0.1474243700504303,
-0.02679671160876751,
0.07802870124578476,
-0.06195024773478508,
0.09478539228439331,
-0.061015862971544266,
0.06891656666994095,
0.013422333635389805,
-0.059194158762693405,
-0.0002872312325052917,
0.039763711392879486,
0.24805356562137604,
0.05787404626607895,
-0.03551553189754486,
0.07833780348300934,
0.010306895710527897,
0.09326206892728806,
0.12546029686927795,
0.12390127032995224,
0.1577652096748352,
0.0332462415099144,
0.14267133176326752,
0.08429278433322906,
-0.02419804222881794,
-0.11930330842733383,
0.05745382606983185,
-0.06907464563846588,
0.0891786590218544,
0.025618888437747955,
0.2082347273826599,
0.0992557555437088,
-0.16331878304481506,
0.004503987729549408,
-0.03784571588039398,
-0.08472839742898941,
-0.09582261741161346,
-0.059616316109895706,
-0.13071000576019287,
-0.145082488656044,
0.011022666469216347,
-0.10669862478971481,
0.03408895432949066,
0.07192801684141159,
0.014300736598670483,
0.00031173089519143105,
0.1421138048171997,
0.01548685971647501,
0.02951667457818985,
0.09437806159257889,
0.008936704136431217,
-0.03922872990369797,
0.000204350333660841,
-0.1039719432592392,
0.024479318410158157,
0.005724349990487099,
0.05637577548623085,
-0.020926199853420258,
-0.025162355974316597,
0.06878668814897537,
-0.025673983618617058,
-0.12570248544216156,
0.010692129842936993,
0.020830895751714706,
0.06029254198074341,
0.04322771355509758,
0.05572039261460304,
-0.01665867306292057,
0.025149133056402206,
0.20819544792175293,
-0.08891750127077103,
-0.07642124593257904,
-0.13295888900756836,
0.15144409239292145,
-0.015287509188055992,
-0.007558114361017942,
0.008901854045689106,
-0.10590888559818268,
0.0015432839281857014,
0.19374702870845795,
0.14979790151119232,
-0.07489100843667984,
-0.0008590759825892746,
-0.02723229117691517,
-0.006055474281311035,
-0.037863317877054214,
0.06658100336790085,
0.0783713087439537,
0.03597036749124527,
-0.0600556954741478,
-0.06273619830608368,
-0.058525312691926956,
-0.040868259966373444,
-0.022366272285580635,
0.039740413427352905,
-0.03190816193819046,
-0.02250455692410469,
-0.0508488304913044,
0.0784219354391098,
-0.08331906050443649,
-0.09764186292886734,
0.007101943250745535,
-0.21792453527450562,
-0.17312918603420258,
-0.0017981445416808128,
0.07537296414375305,
0.03504268825054169,
0.026391860097646713,
-0.03384264186024666,
0.026234446093440056,
0.05435304343700409,
-0.013486234471201897,
-0.05765185505151749,
-0.05856357514858246,
0.04278075695037842,
-0.0805937647819519,
0.17502613365650177,
-0.004664253909140825,
0.06591945886611938,
0.1038215160369873,
0.08136299252510071,
-0.10803729295730591,
0.1037498489022255,
0.061408985406160355,
-0.07544513791799545,
0.055387768894433975,
0.15178146958351135,
-0.05586624890565872,
0.1427116096019745,
0.05213762819766998,
-0.10419749468564987,
0.0006046254420652986,
0.009985820390284061,
-0.028121497482061386,
-0.07477736473083496,
-0.06569526344537735,
-0.046388279646635056,
0.1470259577035904,
0.13528913259506226,
-0.06575675308704376,
0.0014259933959692717,
-0.0165769774466753,
0.05587230995297432,
0.06179123744368553,
0.020786870270967484,
-0.06165136396884918,
-0.28427931666374207,
-0.017122957855463028,
0.037562306970357895,
0.021548131480813026,
-0.24393148720264435,
-0.08912845700979233,
-0.009960807859897614,
-0.04589134827256203,
-0.07417234033346176,
0.09397215396165848,
0.07842769473791122,
0.030639737844467163,
-0.054052744060754776,
-0.050065089017152786,
-0.028803808614611626,
0.17257104814052582,
-0.16336017847061157,
-0.11510152369737625
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | beshoy25/llama_finetuning | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:36:25+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers | # English to Lao Translation Model
Welcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for English to Lao translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.
Built on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.
With a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.
## How to use
### On GPU
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("minhtoan/t5-translate-english-lao")
model = AutoModelForSeq2SeqLM.from_pretrained("minhtoan/t5-translate-english-lao")
model.cuda()
src = "I want to buy a new book"
tokenized_text = tokenizer.encode(src, return_tensors="pt").cuda()
model.eval()
translate_ids = model.generate(tokenized_text, max_length=200)
output = tokenizer.decode(translate_ids[0], skip_special_tokens=True)
output
```
'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'
### On CPU
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("minhtoan/t5-translate-english-lao")
model = AutoModelForSeq2SeqLM.from_pretrained("minhtoan/t5-translate-english-lao")
src = "I want to buy a new book"
input_ids = tokenizer(src, max_length=140, return_tensors="pt", padding="max_length", truncation=True).input_ids
outputs = model.generate(input_ids=input_ids, max_new_tokens=200)
output = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
output
```
'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'
## Author
`
Phan Minh Toan
` | {"language": ["en", "lo"], "license": "mit", "library_name": "transformers", "tags": ["translation"], "widget": [{"text": "i love you so much"}], "inference": {"parameters": {"max_length": 200}}, "pipeline_tag": "translation"} | translation | minhtoan/t5-translate-english-lao | [
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"translation",
"en",
"lo",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T15:38:23+00:00 | [] | [
"en",
"lo"
] | TAGS
#transformers #pytorch #mt5 #text2text-generation #translation #en #lo #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # English to Lao Translation Model
Welcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for English to Lao translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.
Built on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.
With a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.
## How to use
### On GPU
'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'
### On CPU
'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'
## Author
'
Phan Minh Toan
' | [
"# English to Lao Translation Model\nWelcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for English to Lao translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.\n\nBuilt on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.\n\nWith a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.",
"## How to use",
"### On GPU\n\n'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'",
"### On CPU\n\n'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'",
"## Author\n'\nPhan Minh Toan \n'"
] | [
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #translation #en #lo #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# English to Lao Translation Model\nWelcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for English to Lao translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.\n\nBuilt on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.\n\nWith a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.",
"## How to use",
"### On GPU\n\n'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'",
"### On CPU\n\n'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'",
"## Author\n'\nPhan Minh Toan \n'"
] | [
61,
308,
4,
13,
13,
8
] | [
"passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #translation #en #lo #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# English to Lao Translation Model\nWelcome to the forefront of linguistic innovation with our groundbreaking T5 language model designed specifically for English to Lao translation. In a rapidly globalizing world where effective communication is paramount, our T5 model stands as a beacon of excellence, offering unparalleled accuracy, fluency, and efficiency in bridging the language gap between Lao and English.\n\nBuilt on state-of-the-art deep learning architecture and trained on vast datasets of Lao and English texts, our language model (LLM) harnesses the power of transformer-based technology to deliver seamless and precise translations. Whether you're a business expanding into Laotian markets, a researcher seeking to access Lao-language resources, or an individual connecting with Lao-speaking communities, our T5 model is your ultimate solution for unlocking linguistic barriers and fostering meaningful cross-cultural exchanges.\n\nWith a commitment to quality and innovation, our translation model not only translates words but also preserves context, tone, and cultural nuances, ensuring that the essence of the original message remains intact in every translated sentence. Whether it's documents, websites, or multimedia content, our LLM model offers unmatched versatility and reliability, empowering users to communicate effortlessly across languages and borders.## How to use### On GPU\n\n'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'### On CPU\n\n'ຂ້ອຍຢາກຊື້ປຶ້ມໃຫມ່'## Author\n'\nPhan Minh Toan \n'"
] | [
0.04032815247774124,
-0.0574810728430748,
-0.005437284708023071,
0.102484792470932,
0.04480501264333725,
-0.060072530061006546,
-0.00031436156132258475,
0.08240621536970139,
0.07020176947116852,
0.0009859716519713402,
-0.06902895867824554,
-0.034037113189697266,
0.03210055083036423,
0.05225265398621559,
0.028037309646606445,
-0.30318236351013184,
0.027310868725180626,
-0.058313220739364624,
-0.0026935499627143145,
0.08368851989507675,
0.15738338232040405,
-0.0027287041302770376,
0.15338703989982605,
0.03822356462478638,
0.010368135757744312,
0.017595387995243073,
-0.07040036469697952,
-0.0480390228331089,
0.1046275869011879,
0.05216965079307556,
0.05853307247161865,
0.021475639194250107,
-0.022229572758078575,
-0.1569669097661972,
0.003452049335464835,
0.04922330752015114,
-0.044422175735235214,
-0.014156576246023178,
-0.003810287918895483,
-0.0897369384765625,
0.17142100632190704,
-0.023297948762774467,
0.05379980430006981,
0.06632480770349503,
-0.14912918210029602,
-0.0707528367638588,
-0.04691409319639206,
-0.07308607548475266,
0.030443908646702766,
0.04328130558133125,
-0.05774594470858574,
0.11330088973045349,
-0.13474740087985992,
0.07105831801891327,
0.03966745361685753,
-0.3444960117340088,
0.0010172894690185785,
-0.00568400276824832,
0.02312127873301506,
0.1320667564868927,
-0.07769986242055893,
0.06668990105390549,
-0.026612669229507446,
0.020712850615382195,
0.05021166428923607,
-0.05172806605696678,
0.012926013208925724,
-0.034111060202121735,
-0.15474556386470795,
0.03860117122530937,
0.25017890334129333,
0.0009504416957497597,
-0.056331656873226166,
-0.07119980454444885,
-0.06151771917939186,
0.08317214995622635,
0.0017621474107727408,
-0.1236763745546341,
0.07248936593532562,
0.025674743577837944,
0.16391389071941376,
-0.11986580491065979,
-0.07267746329307556,
0.020959794521331787,
-0.2062966376543045,
0.0524931363761425,
0.08940671384334564,
0.016538875177502632,
0.013368104584515095,
0.004890989046543837,
-0.14398615062236786,
-0.014542807824909687,
-0.07891800254583359,
-0.0786537304520607,
-0.0673973485827446,
-0.0010610700119286776,
0.029102465137839317,
-0.013747619464993477,
0.0260026752948761,
0.0996827632188797,
-0.09193585813045502,
-0.009446355514228344,
-0.09920158237218857,
0.043467678129673004,
0.07048514485359192,
0.05467074736952782,
-0.0750172808766365,
-0.07143734395503998,
-0.010994283482432365,
-0.09084572643041611,
-0.04911110922694206,
0.021675972267985344,
-0.05338394641876221,
-0.042567260563373566,
-0.04074791818857193,
0.03291908651590347,
0.02129635028541088,
0.07645821571350098,
-0.03236370161175728,
-0.014499620534479618,
0.04232754558324814,
-0.06596212834119797,
-0.05521519482135773,
-0.061895933002233505,
-0.06464189291000366,
0.11890207231044769,
0.06955263018608093,
0.07428988069295883,
-0.07585921883583069,
0.021904001012444496,
-0.04210195690393448,
-0.013648605905473232,
-0.05577808618545532,
-0.09850489348173141,
0.039778947830200195,
-0.04241722822189331,
-0.010176360607147217,
-0.12564896047115326,
-0.0823805034160614,
-0.001070602098479867,
0.03144700452685356,
-0.07139623910188675,
-0.04070976376533508,
-0.01981993019580841,
0.0031530081760138273,
-0.0317300409078598,
0.033948641270399094,
-0.09515392035245895,
-0.022754546254873276,
-0.021311312913894653,
-0.013396949507296085,
0.08245497196912766,
-0.08715661615133286,
0.05126245692372322,
-0.12307202816009521,
0.028226379305124283,
-0.20033366978168488,
0.09955091774463654,
-0.0830206498503685,
-0.0432804599404335,
-0.09988336265087128,
-0.022803999483585358,
-0.048366978764534,
0.029080281034111977,
-0.07927994430065155,
0.07084879279136658,
-0.08028008788824081,
-0.01101240236312151,
0.20828361809253693,
-0.10775172710418701,
-0.06617235392332077,
0.20695611834526062,
-0.002746489830315113,
0.1593155860900879,
0.09865418821573257,
0.2196175903081894,
0.06135640665888786,
-0.13427740335464478,
-0.009750328958034515,
0.04120966047048569,
-0.1470542997121811,
0.1220187097787857,
0.1538376808166504,
-0.019459348171949387,
0.023156724870204926,
0.02023627609014511,
0.0021531980019062757,
-0.029864436015486717,
-0.04688087850809097,
-0.037832312285900116,
0.017881812527775764,
0.010519768111407757,
0.13827021420001984,
-0.056665025651454926,
-0.005275915376842022,
-0.007986927404999733,
-0.09568408131599426,
0.1018439531326294,
0.006479610688984394,
-0.0036402172408998013,
0.05606130138039589,
-0.15615202486515045,
0.009205714799463749,
0.12281575053930283,
0.005341493524610996,
-0.07510358095169067,
-0.05164065957069397,
0.060831066220998764,
-0.10074736177921295,
0.1458897888660431,
0.10080820322036743,
0.00510877650231123,
0.046310361474752426,
-0.07710129767656326,
0.07698410004377365,
-0.06946706026792526,
0.0018082194728776813,
0.006528117693960667,
-0.12623214721679688,
0.11726895719766617,
-0.026006320491433144,
0.05596845969557762,
0.01432394701987505,
-0.01354310568422079,
0.05330004543066025,
0.0706344023346901,
0.04861101508140564,
0.05691944435238838,
-0.05169817805290222,
0.054437726736068726,
0.016375605016946793,
0.028248723596334457,
0.04690435156226158,
-0.033769138157367706,
0.01173913013190031,
0.2254176139831543,
-0.04787476733326912,
-0.0014307309174910188,
0.07743433862924576,
-0.043316248804330826,
-0.08553951233625412,
-0.08528812974691391,
-0.0046323020942509174,
-0.08346516638994217,
0.026471149176359177,
-0.03860188648104668,
0.1883729100227356,
0.011341872625052929,
0.06447689235210419,
-0.15414123237133026,
0.023302549496293068,
0.0010560526279732585,
-0.09479909390211105,
-0.00866131205111742,
0.05547155439853668,
-0.03634651377797127,
-0.2461196631193161,
0.10113536566495895,
0.08330710977315903,
0.056685831397771835,
0.14041687548160553,
0.024166889488697052,
-0.024629419669508934,
-0.0816957950592041,
0.08923225104808807,
-0.028219325467944145,
0.0017654829425737262,
-0.09251359850168228,
0.020766515284776688,
0.024701645597815514,
0.08154337853193283,
0.042758312076330185,
0.014558485709130764,
0.03259928151965141,
0.03187084570527077,
0.005263578612357378,
0.0011689233360812068,
0.10755881667137146,
0.0020542715210467577,
0.0795053094625473,
0.015261210501194,
0.05787573754787445,
-0.02838681824505329,
0.004931713454425335,
-0.016291677951812744,
0.09828265756368637,
-0.15784549713134766,
-0.332923024892807,
-0.06616035848855972,
-0.0400528647005558,
-0.07233019918203354,
-0.06120916083455086,
0.05189637094736099,
-0.019525978714227676,
-0.05044804513454437,
-0.046032119542360306,
0.0912773460149765,
-0.08183559775352478,
-0.14174219965934753,
-0.1420443207025528,
0.07578939944505692,
-0.08840560168027878,
-0.07187004387378693,
-0.035309191793203354,
-0.00015072962560225278,
-0.1694163978099823,
0.043299365788698196,
-0.0995868518948555,
0.03619235381484032,
-0.012087160721421242,
0.009140532463788986,
-0.04203369840979576,
-0.05749271437525749,
0.16572795808315277,
-0.058443088084459305,
0.0942741185426712,
0.1380079835653305,
-0.010810949839651585,
0.10102397203445435,
0.1784140020608902,
-0.007612958550453186,
-0.02518407627940178,
0.04702287167310715,
0.08478870987892151,
-0.02419714443385601,
-0.21371829509735107,
-0.12473604083061218,
-0.06994079798460007,
0.017811503261327744,
-0.09569873660802841,
0.007951679639518261,
0.05002359300851822,
-0.003000004217028618,
0.0048034777864813805,
-0.005146080162376165,
0.04509774595499039,
0.044423434883356094,
0.2829914689064026,
-0.06186564639210701,
0.07965712249279022,
-0.04006287828087807,
0.0029939967207610607,
0.10486667603254318,
0.12413384765386581,
0.2622171640396118,
0.040632642805576324,
0.198733851313591,
0.14972341060638428,
0.19691337645053864,
0.08630072325468063,
-0.0025228343438357115,
-0.021931877359747887,
0.03805523365736008,
-0.05030282586812973,
-0.041384655982255936,
0.093802809715271,
0.05888260900974274,
0.12533195316791534,
-0.05818437039852142,
0.004800244700163603,
0.13799521327018738,
0.0896177664399147,
0.1908169686794281,
-0.056033600121736526,
0.018088549375534058,
0.03619368001818657,
-0.03671225160360336,
-0.03363468870520592,
-0.10255621373653412,
0.06469755619764328,
0.07635717839002609,
-0.18359601497650146,
0.02643694169819355,
0.024791447445750237,
0.06454707682132721,
-0.05869666486978531,
-0.005396614316850901,
-0.06634698063135147,
0.014328141696751118,
0.006814558524638414,
0.0900229960680008,
-0.24505043029785156,
0.2228533923625946,
0.01642540469765663,
-0.015257835388183594,
-0.019415773451328278,
-0.015333122573792934,
-0.01928890496492386,
0.13170647621154785,
0.18239246308803558,
0.04865597188472748,
0.023899486288428307,
0.03446536883711815,
-0.04824921488761902,
0.001846920931711793,
0.12902022898197174,
-0.006259066518396139,
0.029062019661068916,
-0.0517374686896801,
-0.003472188487648964,
-0.05127032846212387,
0.09596796333789825,
-0.15529155731201172,
-0.13725963234901428,
0.05851011350750923,
-0.1145825907588005,
-0.04241565987467766,
-0.023589275777339935,
-0.00306724919937551,
0.006305536720901728,
0.0346548929810524,
-0.06964515894651413,
-0.10541047155857086,
-0.09211855381727219,
-0.11991613358259201,
0.08321317285299301,
-0.09016449749469757,
0.05463048815727234,
-0.0904160663485527,
0.009537044912576675,
-0.015731194987893105,
-0.035290297120809555,
0.07762522995471954,
-0.008939456194639206,
-0.06967782974243164,
0.0046850768849253654,
0.10052163153886795,
0.05528125539422035,
0.018051324412226677,
0.031245578080415726,
0.011831963434815407,
-0.005835267249494791,
-0.09631790965795517,
-0.029713373631238937,
-0.009271097369492054,
-0.04334500432014465,
0.11239646375179291,
-0.15112991631031036,
-0.033202819526195526,
-0.09885720163583755,
-0.09420675784349442,
0.12912197411060333,
0.13349370658397675,
-0.0518474355340004,
0.18834832310676575,
0.1902584433555603,
-0.09993844479322433,
-0.20819388329982758,
-0.05109212547540665,
0.043120257556438446,
-0.04491071030497551,
0.107125423848629,
-0.18695227801799774,
-0.005454949103295803,
0.03174113482236862,
0.0005794548778794706,
-0.11455298960208893,
-0.13575606048107147,
-0.10644461959600449,
0.01343556772917509,
-0.0717068687081337,
0.038374703377485275,
-0.02118362858891487,
-0.015845071524381638,
0.02691899798810482,
-0.11985401809215546,
0.10401209443807602,
-0.045788586139678955,
0.05631643906235695,
0.015259791165590286,
0.0928567498922348,
0.02219305746257305,
0.05267411097884178,
0.0992167666554451,
-0.08353213965892792,
0.001445584581233561,
-0.09088513255119324,
-0.051760684698820114,
0.08561538904905319,
0.01815224438905716,
0.10199636220932007,
-0.00646184990182519,
0.014713333919644356,
0.02795925736427307,
-0.07536569237709045,
-0.06512054800987244,
0.0869741141796112,
-0.059748079627752304,
-0.09698106348514557,
-0.08967911452054977,
0.13609398901462555,
0.02044363133609295,
-0.008978291414678097,
-0.06067054346203804,
-0.050121236592531204,
-0.036029037088155746,
0.16422390937805176,
0.17158959805965424,
0.012335438281297684,
-0.05502232909202576,
-0.0413631908595562,
-0.0093751335516572,
0.05718347802758217,
-0.0973154604434967,
0.00695073464885354,
0.1252044290304184,
-0.06441988050937653,
0.09974510967731476,
-0.017001453787088394,
-0.16903948783874512,
0.02480076253414154,
0.06857015937566757,
-0.011425234377384186,
-0.3090636730194092,
-0.06343374401330948,
-0.03657188639044762,
0.052881840616464615,
-0.058943867683410645,
0.11758744716644287,
-0.08731631934642792,
-0.006022385787218809,
-0.005617735907435417,
0.03656451404094696,
-0.06269629299640656,
-0.05868766829371452,
-0.05399912968277931,
0.0017896848730742931,
-0.03981710597872734,
0.15234075486660004,
0.12764309346675873,
-0.11419084668159485,
0.007031935732811689,
0.09714534133672714,
-0.07085725665092468,
-0.08572433888912201,
-0.05690363049507141,
0.07878170907497406,
-0.06256765872240067,
-0.03471633791923523,
0.09250470250844955,
-0.0866926982998848,
-0.04586143046617508,
0.2127276062965393,
0.027812570333480835,
0.02000967413187027,
0.022340059280395508,
-0.00943202804774046,
0.033688608556985855,
0.06648880988359451,
-0.007113903295248747,
-0.050037842243909836,
-0.03519017994403839,
-0.014126141555607319,
0.06413834542036057,
0.10467670112848282,
-0.027958843857049942,
-0.10315285623073578,
-0.0961521565914154,
-0.01691650040447712,
-0.07530717551708221,
-0.001761100022122264,
-0.04368335008621216,
-0.028020622208714485,
-0.007232971023768187,
-0.07155939191579819,
-0.053743839263916016,
0.017001423984766006,
-0.08319559693336487,
0.005067354068160057,
0.036964111030101776,
0.1440373659133911,
-0.09009325504302979,
-0.03891783952713013,
0.1619635373353958,
0.0233378354460001,
0.07590749859809875,
-0.018817713484168053,
-0.05574922636151314,
0.057958200573921204,
0.002070263959467411,
0.04549857974052429,
-0.04527915641665459,
0.06908906251192093,
0.04835478588938713,
-0.06329528242349625,
-0.022894972935318947,
-0.023783734068274498,
0.01108943298459053,
-0.014583277516067028,
0.17067940533161163,
-0.09713616222143173,
0.010015717707574368,
-0.002034790115430951,
-0.09874051809310913,
-0.07233773171901703,
-0.011043896898627281,
0.015958577394485474,
0.06620096415281296,
0.03238406777381897,
-0.018435915932059288,
0.03418232500553131,
-0.027363812550902367,
0.013564687222242355,
0.03111138381063938,
0.008049294352531433,
0.09555217623710632,
-0.08095245063304901,
-0.009973704814910889,
0.0001765650522429496,
0.2196737676858902,
0.07174253463745117,
-0.04700960963964462,
0.03677830100059509,
-0.02715437300503254,
-0.17570246756076813,
-0.0029654433019459248,
-0.09748203307390213,
0.01654365286231041,
-0.015322541818022728,
-0.046623535454273224,
-0.025304488837718964,
-0.14792907238006592,
-0.0980885922908783,
0.06338807940483093,
0.20868462324142456,
0.08232977986335754,
-0.0208517387509346,
0.08291386812925339,
0.03254581615328789,
-0.09157998114824295,
-0.08137688785791397,
0.02107575722038746,
0.03651827201247215,
-0.10634127259254456,
0.0348568856716156,
0.13388732075691223,
-0.10913115739822388,
0.18608598411083221,
0.020181581377983093,
-0.06037844717502594,
-0.1271209865808487,
-0.14112532138824463,
-0.04809272289276123,
-0.0414920337498188,
-0.015020743012428284,
-0.11027038842439651,
0.07778220623731613,
0.0022862344048917294,
0.10017973929643631,
-0.0544007308781147,
0.07451067864894867,
-0.2451762855052948,
-0.16010072827339172,
0.049349743872880936,
0.0020919644739478827,
0.09523570537567139,
0.0036884520668536425,
0.030687671154737473,
-0.07567663490772247,
0.07652658969163895,
-0.06870146095752716,
0.05640764907002449,
-0.0733804702758789,
0.0017782626673579216,
-0.07770760357379913,
-0.01172791887074709,
-0.021327197551727295,
0.023198584094643593,
-0.0685487762093544,
0.08470134437084198,
0.02074986882507801,
-0.0729193314909935,
0.014076242223381996,
0.25453880429267883,
0.038073476403951645,
-0.10483192652463913,
-0.1508076936006546,
0.10427863895893097,
-0.013053039088845253,
0.031202102079987526,
-0.0045241969637572765,
0.013035593554377556,
-0.07986146956682205,
0.18746758997440338,
0.15349693596363068,
-0.046538494527339935,
0.0027562042232602835,
0.02426222711801529,
0.045322537422180176,
0.06231782212853432,
0.10842856764793396,
0.02375979721546173,
0.29469233751296997,
-0.07580891251564026,
0.0937218964099884,
-0.0978214219212532,
0.011282647959887981,
-0.10025031864643097,
0.05955502763390541,
0.012755072675645351,
0.0012140370672568679,
-0.03926004841923714,
0.11811795830726624,
-0.04814964905381203,
-0.07107244431972504,
-0.03242158144712448,
-0.06803639978170395,
-0.07006987929344177,
-0.06491875648498535,
0.0008846147102303803,
0.02402486652135849,
0.011648970656096935,
0.0437321737408638,
-0.012919781729578972,
0.05128418654203415,
0.09373297542333603,
-0.07288447767496109,
0.0038407561369240284,
0.15072287619113922,
0.027943018823862076,
0.07956495881080627,
0.008979503065347672,
0.0835682600736618,
0.07063112407922745,
0.06074550375342369,
-0.043256551027297974,
0.05704210698604584,
0.011494441889226437,
-0.05730817839503288,
-0.011409838683903217,
0.044069550931453705,
0.020729809999465942,
0.027558449655771255,
0.05608179420232773,
0.007916884496808052,
0.10216601938009262,
0.02714013308286667,
-0.01862412318587303,
-0.04360880330204964,
0.14332039654254913,
-0.169218972325325,
0.08804518729448318,
0.16596487164497375,
0.015559131279587746,
0.020605245605111122,
-0.07874234020709991,
-0.012603845447301865,
0.02411947213113308,
-0.06825868785381317,
-0.06817207485437393,
-0.040132638067007065,
0.001661549205891788,
0.06602552533149719,
0.04345661774277687,
-0.06616514176130295,
-0.05620790272951126,
0.001473830547183752,
0.06869445741176605,
-0.0678674727678299,
0.08753033727407455,
0.11965689808130264,
0.003099750727415085,
0.0016190337482839823,
-0.21893054246902466,
0.0033846369478851557,
0.06823937594890594,
-0.09658733755350113,
-0.012648072093725204
] |
null | null | null |
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain).
# Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_path = "PATH_TO_THIS_REPO"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
torch_dtype='auto'
).eval()
# Prompt content: "hi"
messages = [
{"role": "user", "content": "hi"}
]
input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt')
output_ids = model.generate(input_ids.to('cuda'))
response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True)
# Model response: "Hello! How can I assist you today?"
print(response)
``` | {"license": "other", "tags": ["autotrain", "text-generation"], "widget": [{"text": "I love AutoTrain because "}]} | text-generation | Tonystark1/Ultimus-Zephyr-7B | [
"safetensors",
"autotrain",
"text-generation",
"conversational",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:39:28+00:00 | [] | [] | TAGS
#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us
|
# Model Trained Using AutoTrain
This model was trained using AutoTrain. For more information, please visit AutoTrain.
# Usage
| [
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
"TAGS\n#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n",
"# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.",
"# Usage"
] | [
37,
29,
3
] | [
"passage: TAGS\n#safetensors #autotrain #text-generation #conversational #license-other #endpoints_compatible #region-us \n# Model Trained Using AutoTrain\n\nThis model was trained using AutoTrain. For more information, please visit AutoTrain.# Usage"
] | [
-0.02089853025972843,
0.03890561684966087,
-0.000762980489525944,
0.037646014243364334,
0.12435931712388992,
-0.03151287883520126,
0.23112058639526367,
0.04494147002696991,
-0.0575568825006485,
-0.09741601347923279,
0.18740901350975037,
0.17386218905448914,
-0.04334506019949913,
0.18782994151115417,
-0.03842408210039139,
-0.23926758766174316,
0.025883177295327187,
-0.0299287848174572,
0.14973880350589752,
0.12130317836999893,
0.15229710936546326,
-0.0829242467880249,
0.05421588197350502,
0.0457366518676281,
-0.19744595885276794,
0.02559680864214897,
0.07502555847167969,
-0.12002695351839066,
0.1892649233341217,
0.040962137281894684,
0.11825616657733917,
0.03324944153428078,
0.1392887830734253,
-0.1323491781949997,
0.01648798957467079,
0.004352208226919174,
-0.015311143361032009,
0.05287393927574158,
0.06082003563642502,
-0.034274082630872726,
0.09492087364196777,
0.19268183410167694,
0.12143059074878693,
0.05840236321091652,
-0.11065401881933212,
0.010359742678701878,
-0.02585293911397457,
0.015595678240060806,
0.12488947808742523,
0.121797576546669,
-0.02974177710711956,
0.2112775444984436,
-0.15929573774337769,
0.0785667672753334,
-0.11720649152994156,
-0.27605608105659485,
-0.007311069872230291,
0.2076014280319214,
0.06324941664934158,
-0.01046263799071312,
-0.13386328518390656,
0.06509426236152649,
0.1174032911658287,
-0.009732136502861977,
0.052042946219444275,
-0.01771010085940361,
-0.05808677524328232,
-0.008316196501255035,
-0.07604839652776718,
0.004176823887974024,
0.2025483250617981,
-0.06435471028089523,
-0.025879809632897377,
-0.1353462189435959,
-0.023601124063134193,
0.04423265904188156,
0.00368077983148396,
-0.10752057284116745,
-0.027382109314203262,
0.10084833204746246,
-0.02734971046447754,
-0.029397934675216675,
-0.1505003720521927,
-0.052210669964551926,
-0.08283388614654541,
0.030309928581118584,
0.0009279148071072996,
0.005750878248363733,
-0.10405394434928894,
0.10598764568567276,
-0.014304609969258308,
-0.09590446949005127,
0.050552137196063995,
-0.10984646528959274,
0.032756756991147995,
-0.11620049923658371,
-0.022093212231993675,
-0.08695599436759949,
0.015334513038396835,
0.21623161435127258,
0.16516101360321045,
-0.003946542274206877,
-0.08353158086538315,
0.03163360059261322,
0.032285887748003006,
0.09010306745767593,
0.07819008082151413,
-0.03263101354241371,
0.06596504896879196,
-0.04041123762726784,
-0.023562058806419373,
-0.026206638664007187,
-0.185186967253685,
0.04729154333472252,
0.006137077696621418,
0.06225769594311714,
-0.07368145138025284,
0.0758923590183258,
-0.02453492395579815,
0.05138348415493965,
0.03385981172323227,
-0.024239709600806236,
0.033983007073402405,
-0.03501613065600395,
0.015362166799604893,
-0.10241638869047165,
0.031124519184231758,
0.13060276210308075,
0.041950587183237076,
0.10722701251506805,
-0.0850663036108017,
-0.03558005392551422,
-0.10486439615488052,
-0.04084291309118271,
0.007949413731694221,
0.032330259680747986,
0.054881513118743896,
-0.20490533113479614,
-0.2844090461730957,
-0.034244854003190994,
0.052770666778087616,
-0.01975797861814499,
-0.07832197844982147,
-0.08976242691278458,
0.02668369561433792,
0.05969720333814621,
-0.03685269504785538,
0.04373543709516525,
-0.022354818880558014,
0.035809289664030075,
-0.0757109671831131,
-0.0067244102247059345,
-0.05800308659672737,
0.007987656630575657,
-0.1394086480140686,
-0.03892948850989342,
-0.01018267311155796,
0.01908150501549244,
-0.03469295799732208,
0.16121862828731537,
-0.010288888588547707,
0.05076303705573082,
-0.05012427642941475,
0.0520540215075016,
0.0038348138332366943,
0.15402163565158844,
-0.12805858254432678,
0.004590215627104044,
0.16217437386512756,
-0.10571835935115814,
-0.11733518540859222,
0.10878685116767883,
-0.11078933626413345,
0.2556385099887848,
0.1126617044210434,
0.14406165480613708,
0.0280612725764513,
-0.12442860752344131,
0.12669576704502106,
0.03417041152715683,
-0.09001672267913818,
-0.027209481224417686,
0.0015774862840771675,
-0.029457205906510353,
-0.21803908050060272,
0.024427056312561035,
0.13007183372974396,
0.07568662613630295,
-0.038225483149290085,
-0.08753399550914764,
-0.013979305513203144,
-0.05888194218277931,
0.05481130629777908,
0.00985832791775465,
0.11558723449707031,
-0.08033457398414612,
-0.03330337256193161,
0.02695239707827568,
0.04780461639165878,
0.07386761158704758,
-0.06066657975316048,
-0.07480321824550629,
-0.03438110277056694,
-0.00005651484752888791,
-0.004678141791373491,
-0.06730625778436661,
-0.0526479035615921,
-0.017854172736406326,
0.14683830738067627,
0.04623232036828995,
0.09310559928417206,
0.03057941049337387,
0.04193659499287605,
-0.01995823159813881,
0.009528989903628826,
0.16668112576007843,
0.04636063799262047,
-0.1251319795846939,
-0.09489064663648605,
0.1198563277721405,
-0.07429909706115723,
0.1495225876569748,
-0.2573336362838745,
0.02191506139934063,
-0.1137506514787674,
0.08119326084852219,
-0.015024850144982338,
0.06582725048065186,
-0.07824977487325668,
0.01642789877951145,
-0.08536693453788757,
0.0042993673123419285,
0.06477862596511841,
0.05614956095814705,
-0.026179833337664604,
0.14061102271080017,
-0.15953490138053894,
0.20964255928993225,
0.1161319687962532,
-0.10498357564210892,
-0.11012911051511765,
-0.10380077362060547,
0.004991353023797274,
-0.005274149589240551,
-0.11000026762485504,
-0.0012808284955099225,
0.11501315236091614,
-0.051325228065252304,
0.184207946062088,
-0.02479202300310135,
-0.027814652770757675,
-0.022695103660225868,
-0.08917387574911118,
-0.004993697162717581,
-0.013311133719980717,
0.0878831148147583,
-0.22586707770824432,
0.1341700702905655,
0.12997865676879883,
-0.011201041750609875,
0.1878158301115036,
0.02932732366025448,
0.028099095448851585,
0.004460213240236044,
-0.03533336520195007,
-0.010984709486365318,
0.02327060140669346,
-0.05687986686825752,
-0.01642347313463688,
0.013465014286339283,
0.010788206942379475,
0.028979692608118057,
-0.1271466314792633,
-0.04724383354187012,
0.014977987855672836,
0.056155066937208176,
0.016029085963964462,
0.05752420425415039,
-0.08498586714267731,
0.06746458262205124,
-0.025121653452515602,
-0.13671542704105377,
0.11770213395357132,
0.01172768697142601,
-0.12705263495445251,
0.17182578146457672,
-0.09404783695936203,
-0.196224644780159,
-0.17304284870624542,
-0.13585984706878662,
0.026043228805065155,
0.08839208632707596,
0.06914421916007996,
-0.06822904944419861,
-0.06807959824800491,
-0.004135052673518658,
-0.12654997408390045,
0.019381104037165642,
-0.03188987448811531,
-0.09604258090257645,
0.057193055748939514,
-0.009717279113829136,
-0.11798624694347382,
-0.05032327026128769,
0.00789867714047432,
-0.06308624148368835,
0.0605158731341362,
-0.03089403733611107,
0.054746001958847046,
0.1381448656320572,
-0.011948119848966599,
0.023544736206531525,
-0.0395624041557312,
0.17897886037826538,
-0.08672381937503815,
-0.0006116208387538791,
0.09763624519109726,
-0.048962898552417755,
0.028884489089250565,
0.2265005260705948,
0.03182725980877876,
-0.06495069712400436,
0.07192723453044891,
-0.035681869834661484,
-0.05174829810857773,
-0.19448144733905792,
-0.11049490422010422,
-0.010373943485319614,
-0.010003382340073586,
0.0674663707613945,
0.04859880357980728,
0.2720578908920288,
0.12234988063573837,
0.059470195323228836,
0.016185441985726357,
0.04209032282233238,
0.08999012410640717,
0.13016381859779358,
-0.04774774983525276,
0.17109765112400055,
-0.06409438699483871,
-0.16133272647857666,
0.044327691197395325,
-0.027926357463002205,
0.051227767020463943,
0.17565013468265533,
-0.03614453971385956,
0.047351136803627014,
0.11210278421640396,
0.12826228141784668,
0.1061127632856369,
0.07705885171890259,
-0.06504974514245987,
-0.010043035261332989,
0.00019683393475133926,
-0.05370469391345978,
0.14862267673015594,
-0.023733152076601982,
-0.06846705824136734,
-0.031645484268665314,
0.010693936608731747,
0.04905892163515091,
0.049152228981256485,
0.03127843141555786,
-0.2666167616844177,
0.03436502441763878,
0.046095263212919235,
-0.06547010689973831,
-0.11317573487758636,
0.09948568791151047,
-0.021655220538377762,
-0.18608878552913666,
0.017802411690354347,
-0.025920318439602852,
0.09116440266370773,
0.04311057925224304,
0.05799582228064537,
-0.09219425916671753,
-0.0708162784576416,
-0.05113530531525612,
0.15323954820632935,
-0.35677093267440796,
0.21487660706043243,
-0.014043435454368591,
0.0690545067191124,
-0.11276184022426605,
0.0014416693011298776,
0.07986348122358322,
0.16165494918823242,
0.11833548545837402,
-0.05488691106438637,
-0.16898946464061737,
-0.09826766699552536,
-0.08969532698392868,
-0.007673082873225212,
0.013347413390874863,
0.003650940954685211,
-0.005118653643876314,
-0.11486039310693741,
-0.0005021608667448163,
0.04620593041181564,
-0.010058995336294174,
-0.1808961033821106,
-0.15823762118816376,
-0.02242000214755535,
0.044828031212091446,
0.10119049996137619,
-0.033685166388750076,
-0.051781389862298965,
-0.06033768132328987,
0.15737107396125793,
0.04368119686841965,
0.012251429259777069,
-0.12371376901865005,
-0.05173582211136818,
-0.06613845378160477,
-0.022030174732208252,
0.07524938881397247,
0.009389028884470463,
0.12098590284585953,
-0.09848834574222565,
-0.05622165650129318,
0.10000088065862656,
-0.12879306077957153,
-0.044098254293203354,
-0.12273328751325607,
0.050619933754205704,
-0.026867562904953957,
-0.004624411929398775,
0.12226194888353348,
0.04077878221869469,
-0.07747189700603485,
-0.06510289013385773,
-0.02182580530643463,
-0.02168603427708149,
0.040108900517225266,
-0.11854132264852524,
-0.10533714294433594,
-0.144134521484375,
-0.03266002982854843,
-0.12010640650987625,
0.22031773626804352,
0.1510319709777832,
-0.0889979898929596,
0.16045299172401428,
0.21687199175357819,
-0.09459521621465683,
-0.28949886560440063,
-0.06218516454100609,
-0.05762689933180809,
0.0012655822793021798,
0.056375544518232346,
-0.09276837855577469,
0.08377362787723541,
-0.004379333462566137,
-0.0921919122338295,
-0.03929101675748825,
-0.10597379505634308,
-0.1628357619047165,
0.24811773002147675,
-0.00695221871137619,
0.216319277882576,
-0.06675629317760468,
-0.04963424429297447,
-0.11837507039308548,
0.03226492181420326,
0.05033990368247032,
-0.08250661194324493,
0.04896571487188339,
0.05970872566103935,
0.07762710750102997,
0.03615579381585121,
-0.04023800045251846,
0.0499248206615448,
-0.07690990716218948,
0.07372726500034332,
-0.17243541777133942,
-0.051966533064842224,
0.0291034784168005,
-0.02003716491162777,
0.11406885087490082,
-0.03866045922040939,
0.04375878721475601,
-0.05661903694272041,
-0.07238272577524185,
0.012632071040570736,
0.06424806267023087,
-0.0111227473244071,
-0.12185013294219971,
0.0070838648825883865,
-0.003560643410310149,
0.004385150969028473,
-0.06248250603675842,
0.016781898215413094,
-0.031206920742988586,
0.15563493967056274,
0.15905016660690308,
0.2279939204454422,
-0.06940897554159164,
0.057850778102874756,
-0.026937630027532578,
-0.12084269523620605,
0.07881549000740051,
-0.060470253229141235,
0.010923074558377266,
0.05394923686981201,
-0.05505755916237831,
0.16708660125732422,
0.053299445658922195,
-0.0007490343996323645,
-0.015869995579123497,
0.15427231788635254,
-0.17436520755290985,
0.028647977858781815,
-0.08862833678722382,
0.15710654854774475,
0.04452139511704445,
-0.029634831473231316,
0.10007839649915695,
-0.07933120429515839,
-0.029322272166609764,
0.006951325573027134,
0.017015496268868446,
-0.03554573282599449,
0.05849390849471092,
0.046525198966264725,
0.024086007848381996,
-0.06793931126594543,
0.026535160839557648,
0.07079220563173294,
0.0025835877750068903,
0.04738464578986168,
0.013694006018340588,
-0.09493011981248856,
-0.1037706807255745,
0.031061364337801933,
0.2576681077480316,
-0.1639707237482071,
-0.08702236413955688,
0.009577915072441101,
-0.10157066583633423,
-0.0026154285296797752,
0.07413817942142487,
0.06880449503660202,
0.03655710443854332,
-0.042900752276182175,
-0.013874638825654984,
-0.11066316813230515,
0.0910448282957077,
-0.015328219160437584,
0.0348287932574749,
-0.14798195660114288,
0.07496067136526108,
-0.03132447972893715,
-0.008997730910778046,
-0.08787791430950165,
-0.033700209110975266,
-0.12531232833862305,
0.030435124412178993,
-0.08465003967285156,
-0.04313739016652107,
-0.05273820459842682,
-0.010747137479484081,
0.0678463876247406,
-0.010134257376194,
-0.017098618671298027,
-0.024644924327731133,
-0.08711723238229752,
0.032871875911951065,
0.004344973247498274,
0.04483238607645035,
-0.04674182087182999,
-0.01993880234658718,
0.037311747670173645,
-0.000004001267825515242,
0.06050976738333702,
0.022565992549061775,
-0.007758983410894871,
0.03770044445991516,
-0.15966764092445374,
0.01916838437318802,
0.06271649152040482,
0.0006143683567643166,
0.016977902501821518,
-0.03355167806148529,
-0.0018841095734387636,
0.0999053344130516,
0.030659453943371773,
0.03639167547225952,
0.01731853187084198,
-0.0949004739522934,
0.037301186472177505,
0.10677090287208557,
-0.14946091175079346,
-0.022807510569691658,
-0.05471193790435791,
-0.011145985685288906,
-0.057102054357528687,
0.22019965946674347,
-0.11838836222887039,
0.04698079079389572,
-0.032419852912425995,
0.03750695660710335,
-0.0519956611096859,
-0.10454028844833374,
-0.10880608856678009,
-0.10406296700239182,
-0.036173172295093536,
-0.0017616144614294171,
0.2634603977203369,
0.14614185690879822,
-0.007627400569617748,
0.04732783883810043,
0.06023077666759491,
0.09986170381307602,
-0.0000392909932998009,
0.1907200664281845,
0.09213747829198837,
-0.004819431807845831,
-0.12899689376354218,
0.07417719066143036,
0.025308500975370407,
-0.10945913195610046,
0.0014507247833535075,
0.0060352059081196785,
-0.07921634614467621,
0.04549342021346092,
0.061475154012441635,
-0.049655646085739136,
-0.10908256471157074,
-0.1897570788860321,
-0.11767365038394928,
0.014547701925039291,
-0.1141902431845665,
0.006054932717233896,
0.18083947896957397,
-0.06133390590548515,
-0.022032413631677628,
-0.09275112301111221,
-0.0474187396466732,
-0.2181331366300583,
-0.15545961260795593,
-0.10639044642448425,
-0.08368334919214249,
0.04896046221256256,
-0.020269649103283882,
0.05286030098795891,
0.018245011568069458,
0.03993610292673111,
-0.06763483583927155,
0.08721300959587097,
-0.10831692814826965,
0.004784486256539822,
-0.009881925769150257,
-0.04393337666988373,
0.01711859367787838,
-0.19800134003162384,
-0.01726091466844082,
-0.14271385967731476,
-0.025886263698339462,
-0.02414889633655548,
-0.03923075646162033,
0.0015599187463521957,
-0.00659944349899888,
-0.022216126322746277,
-0.007123332936316729,
-0.010187787935137749,
0.03588121011853218,
0.030142245814204216,
0.06735268235206604,
0.01930520497262478,
0.021639658138155937,
0.03718075901269913,
0.2173466682434082,
-0.03672509640455246,
-0.18076519668102264,
-0.13255588710308075,
0.22741390764713287,
0.023755958303809166,
0.12003876268863678,
-0.07047237455844879,
-0.003944313619285822,
0.0649246871471405,
0.3151680529117584,
0.27447304129600525,
-0.04221269488334656,
0.012944314628839493,
-0.03759029880166054,
-0.008687055669724941,
-0.0077759926207363605,
0.17214618623256683,
0.0111585957929492,
0.18692266941070557,
-0.061342377215623856,
0.057751890271902084,
-0.007795935031026602,
-0.07976683229207993,
-0.05004684627056122,
0.1371750831604004,
-0.034483592957258224,
-0.013111086562275887,
-0.017309419810771942,
0.08474326133728027,
-0.06475097686052322,
0.1650533229112625,
-0.12438745051622391,
-0.03197024017572403,
-0.04968215525150299,
0.050263699144124985,
0.1181311383843422,
-0.009911769069731236,
0.03671935200691223,
-0.030859731137752533,
-0.025431539863348007,
0.018659215420484543,
-0.03971736878156662,
-0.08324228972196579,
-0.040832240134477615,
0.07943736016750336,
0.018289517611265182,
0.24940812587738037,
-0.016860337927937508,
0.06924241781234741,
0.07830806821584702,
-0.0007601219112984836,
-0.08936040103435516,
0.1169457733631134,
0.010533611290156841,
-0.053996723145246506,
0.1200164407491684,
-0.016792241483926773,
0.008844620548188686,
-0.001643515657633543,
-0.006236417684704065,
-0.18588665127754211,
0.14857490360736847,
-0.09602080285549164,
-0.0948827937245369,
-0.05673005431890488,
0.13433516025543213,
-0.02555198408663273,
0.16195133328437805,
0.05283422768115997,
-0.02981109544634819,
0.0056883953511714935,
-0.020765170454978943,
0.06717022508382797,
-0.002720105228945613,
-0.10159162431955338,
-0.03101331554353237,
-0.19819441437721252,
-0.01870795525610447,
0.10115032643079758,
-0.025165937840938568,
-0.23734821379184723,
-0.07709009200334549,
-0.06396035850048065,
-0.031772181391716,
-0.12610237300395966,
0.06999877095222473,
0.20647278428077698,
0.019630368798971176,
-0.009499672800302505,
-0.12196175009012222,
-0.011895264498889446,
0.02409667894244194,
-0.028847014531493187,
-0.10832608491182327
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | tavalenzuelag/mistral-7b-e2e-mod-6 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:40:28+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
# Uploaded model
- **Developed by:** smotoc
- **License:** apache-2.0
- **Finetuned from model :** unsloth/mistral-7b-bnb-4bit
This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
| {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "mistral", "gguf"], "base_model": "unsloth/mistral-7b-bnb-4bit"} | null | smotoc/foxy_7B_unsloth_temp | [
"transformers",
"safetensors",
"gguf",
"text-generation-inference",
"unsloth",
"mistral",
"en",
"base_model:unsloth/mistral-7b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:40:43+00:00 | [] | [
"en"
] | TAGS
#transformers #safetensors #gguf #text-generation-inference #unsloth #mistral #en #base_model-unsloth/mistral-7b-bnb-4bit #license-apache-2.0 #endpoints_compatible #region-us
|
# Uploaded model
- Developed by: smotoc
- License: apache-2.0
- Finetuned from model : unsloth/mistral-7b-bnb-4bit
This mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.
<img src="URL width="200"/>
| [
"# Uploaded model\n\n- Developed by: smotoc\n- License: apache-2.0\n- Finetuned from model : unsloth/mistral-7b-bnb-4bit\n\nThis mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.\n\n<img src=\"URL width=\"200\"/>"
] | [
"TAGS\n#transformers #safetensors #gguf #text-generation-inference #unsloth #mistral #en #base_model-unsloth/mistral-7b-bnb-4bit #license-apache-2.0 #endpoints_compatible #region-us \n",
"# Uploaded model\n\n- Developed by: smotoc\n- License: apache-2.0\n- Finetuned from model : unsloth/mistral-7b-bnb-4bit\n\nThis mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.\n\n<img src=\"URL width=\"200\"/>"
] | [
69,
79
] | [
"passage: TAGS\n#transformers #safetensors #gguf #text-generation-inference #unsloth #mistral #en #base_model-unsloth/mistral-7b-bnb-4bit #license-apache-2.0 #endpoints_compatible #region-us \n# Uploaded model\n\n- Developed by: smotoc\n- License: apache-2.0\n- Finetuned from model : unsloth/mistral-7b-bnb-4bit\n\nThis mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.\n\n<img src=\"URL width=\"200\"/>"
] | [
-0.09073734283447266,
0.06485948711633682,
-0.0037488050293177366,
0.08022642880678177,
0.047991205006837845,
0.011135050095617771,
0.10988201946020126,
0.07242579013109207,
0.06011021137237549,
-0.0387791283428669,
0.12636283040046692,
0.08979949355125427,
-0.0036079229321330786,
0.040191780775785446,
-0.032321106642484665,
-0.14021086692810059,
0.12435491383075714,
-0.0728936642408371,
-0.07775774598121643,
0.05062108114361763,
0.0768696516752243,
-0.0072922674007713795,
0.08691861480474472,
-0.06779509037733078,
-0.025861727073788643,
-0.007638489827513695,
-0.024174919351935387,
-0.013568161055445671,
-0.016541369259357452,
0.05451444536447525,
-0.04680052772164345,
0.08709421008825302,
0.05802164599299431,
-0.06158967316150665,
0.044334422796964645,
0.05254489555954933,
-0.008831317536532879,
0.07752035558223724,
0.0014878256479278207,
0.06471071392297745,
0.10192239284515381,
-0.027509761974215508,
-0.07708260416984558,
0.05364599823951721,
-0.014294321648776531,
-0.1022665947675705,
-0.06879347562789917,
0.0969560369849205,
0.018906502053141594,
0.05426344275474548,
0.020562734454870224,
0.10597603023052216,
-0.07594265788793564,
0.06677796691656113,
0.1468590795993805,
-0.23873156309127808,
-0.06689462810754776,
0.1537172645330429,
0.004258368164300919,
0.05074722692370415,
-0.04930337145924568,
0.023929132148623466,
0.032174915075302124,
-0.0018727283459156752,
0.027918146923184395,
-0.04798970744013786,
-0.05468827858567238,
0.037673220038414,
-0.11849042028188705,
-0.0024227024987339973,
0.201235830783844,
0.09383370727300644,
-0.07496679574251175,
0.02121240831911564,
-0.140515998005867,
0.10032782703638077,
-0.052975911647081375,
0.024010630324482918,
0.049812983721494675,
0.10299620777368546,
-0.043762825429439545,
-0.08623871952295303,
-0.048045363277196884,
-0.06326711922883987,
-0.07875124365091324,
0.07621503621339798,
0.0582118034362793,
0.1091194674372673,
-0.021382058039307594,
0.07956407964229584,
-0.025303469970822334,
-0.12833598256111145,
-0.0811346247792244,
-0.09134925901889801,
0.09105756133794785,
0.06599336862564087,
-0.039857275784015656,
0.03886795789003372,
0.14931751787662506,
0.2607046663761139,
0.034496404230594635,
0.05820830911397934,
0.04790235683321953,
0.037888601422309875,
-0.07053811848163605,
0.02950519695878029,
-0.14625196158885956,
-0.06156009063124657,
0.1545545905828476,
0.051306843757629395,
0.08398241549730301,
-0.009676926769316196,
-0.0755959153175354,
-0.046818796545267105,
0.01821276731789112,
0.017138376832008362,
0.042612429708242416,
0.12816712260246277,
0.04091399163007736,
-0.021794836968183517,
0.22603662312030792,
-0.06260039657354355,
-0.03596815839409828,
0.022585788741707802,
-0.052050165832042694,
0.086605504155159,
0.13203100860118866,
0.0042468952015042305,
-0.033257484436035156,
-0.0801447331905365,
-0.044322419911623,
-0.00838660541921854,
0.005213659722357988,
-0.04381372779607773,
0.07033174484968185,
-0.014147483743727207,
0.012753087095916271,
-0.1484278440475464,
-0.26753726601600647,
0.04182913899421692,
0.122501902282238,
-0.023608209565281868,
0.008882073685526848,
-0.030412567779421806,
-0.06696992367506027,
0.018854806199669838,
-0.029420385137200356,
-0.0029628558550029993,
-0.09030025452375412,
0.021659163758158684,
-0.054603833705186844,
0.051572274416685104,
-0.1950291097164154,
0.00772177055478096,
-0.13684380054473877,
0.01180289126932621,
-0.1145244613289833,
0.05374019965529442,
-0.093045674264431,
0.13275307416915894,
-0.07998768240213394,
-0.007767742499709129,
-0.08046065270900726,
-0.006394652184098959,
0.03353819251060486,
0.19235435128211975,
-0.09467560052871704,
0.03640739992260933,
0.1311134248971939,
-0.08949989080429077,
-0.12508371472358704,
0.16564159095287323,
0.02007448300719261,
0.004338916856795549,
0.07451160252094269,
0.09424944221973419,
0.11043690145015717,
-0.08498258888721466,
0.04108073189854622,
0.17256534099578857,
-0.042007338255643845,
-0.06286412477493286,
0.0935162901878357,
0.017537232488393784,
-0.14180895686149597,
0.09089484065771103,
-0.09859734773635864,
0.11615247279405594,
0.00009662442607805133,
-0.04446592926979065,
-0.12629668414592743,
-0.07595711946487427,
0.018579009920358658,
-0.02021029405295849,
0.0077751693315804005,
-0.016145631670951843,
-0.05409941449761391,
0.021818941459059715,
0.15007495880126953,
-0.0589325912296772,
0.019243493676185608,
-0.04945991560816765,
0.07241858541965485,
-0.1109626516699791,
0.08503961563110352,
-0.07040730118751526,
-0.0157752875238657,
-0.012312076985836029,
-0.04739023372530937,
0.054699502885341644,
0.01862042397260666,
0.09251396358013153,
0.02967991679906845,
-0.0419125072658062,
-0.014032389037311077,
0.10010462999343872,
0.014874937944114208,
-0.0439131073653698,
-0.12480881065130234,
0.025989219546318054,
-0.02608475275337696,
0.14723798632621765,
-0.07537669688463211,
0.06028730049729347,
-0.03872079402208328,
0.06660766899585724,
-0.013773690909147263,
0.021575579419732094,
0.056100212037563324,
-0.12179797887802124,
-0.0034974021837115288,
-0.10398320853710175,
0.07292281091213226,
0.07572886347770691,
-0.08434890955686569,
0.1052224189043045,
-0.051743924617767334,
0.12808437645435333,
0.1603950411081314,
0.04576335474848747,
0.07024643570184708,
0.008966737426817417,
-0.032392650842666626,
-0.022440779954195023,
0.07316834479570389,
-0.0022556178737431765,
-0.04846952483057976,
-0.02673877216875553,
0.10232343524694443,
-0.10902124643325806,
-0.02652878873050213,
0.005001960787922144,
-0.07927267998456955,
0.010496443137526512,
0.04505718871951103,
0.04128969460725784,
-0.20234942436218262,
0.0705171599984169,
0.2494184523820877,
-0.06070324033498764,
0.11549630016088486,
-0.06835178285837173,
-0.053414300084114075,
0.010013608261942863,
0.05659085884690285,
-0.019729895517230034,
0.02605547569692135,
-0.09722447395324707,
0.0558059960603714,
0.04073783755302429,
0.008484299294650555,
0.04292581230401993,
-0.10735554248094559,
0.007692236453294754,
-0.03950345516204834,
-0.09136825799942017,
-0.02809501253068447,
0.07124310731887817,
-0.0844593197107315,
0.05284437909722328,
-0.05233987420797348,
-0.090587317943573,
0.056327130645513535,
0.03483490273356438,
-0.059026435017585754,
0.13263361155986786,
-0.09990336745977402,
-0.07520771026611328,
-0.1554654836654663,
-0.005018540658056736,
-0.15382742881774902,
0.0006956326542422175,
0.07375198602676392,
-0.02687772922217846,
-0.0747746154665947,
-0.14689594507217407,
-0.08809564262628555,
0.0880221426486969,
0.025942694395780563,
0.07145301252603531,
0.021418508142232895,
0.09207689762115479,
-0.12135017663240433,
-0.012371210381388664,
0.004120517522096634,
-0.07579642534255981,
0.023777365684509277,
-0.07031088322401047,
0.07605468481779099,
0.100066639482975,
0.03100903332233429,
-0.011025220155715942,
0.06516610085964203,
0.18232865631580353,
0.045183226466178894,
0.08525479584932327,
0.18680691719055176,
0.025929046794772148,
0.07671015709638596,
0.10501625388860703,
-0.00020982461865060031,
-0.04883745312690735,
0.015972284600138664,
-0.02555960975587368,
-0.04600955545902252,
-0.1469797044992447,
-0.014180346392095089,
-0.07194319367408752,
0.0668332427740097,
0.02597099356353283,
0.09130985289812088,
-0.032908204942941666,
0.1402636021375656,
-0.10170557349920273,
0.0449705496430397,
0.12967512011528015,
0.0659472718834877,
0.009303512051701546,
0.016758129000663757,
0.05034618452191353,
-0.1301170438528061,
0.03271382302045822,
0.14402854442596436,
0.012697187252342701,
0.12227509170770645,
-0.03924565389752388,
0.0143468352034688,
0.06055156886577606,
0.1484806388616562,
0.010214411653578281,
0.11840295791625977,
-0.03227197378873825,
-0.011201245710253716,
-0.0837680771946907,
-0.0764722004532814,
-0.02840019203722477,
0.029771629720926285,
-0.15988004207611084,
-0.006532828789204359,
0.03126173093914986,
0.12799857556819916,
0.12264695763587952,
0.16889260709285736,
0.06387018412351608,
-0.2487042099237442,
-0.0965900719165802,
0.07252063602209091,
0.05351516976952553,
-0.009770955890417099,
0.050094153732061386,
0.06399746984243393,
0.024782005697488785,
0.08265529572963715,
-0.04092234745621681,
0.09717672318220139,
0.10543130338191986,
0.03436778858304024,
0.017331790179014206,
0.1759984940290451,
0.03101508319377899,
0.08612943440675735,
-0.19013790786266327,
0.03787776455283165,
0.011473264545202255,
0.047207754105329514,
-0.0337478406727314,
0.005441403482109308,
0.14995108544826508,
0.19620203971862793,
0.075168177485466,
0.03550708666443825,
-0.03900633379817009,
0.028229760006070137,
-0.16283278167247772,
0.07709631323814392,
0.0034083197824656963,
0.027528060600161552,
0.03864768147468567,
-0.08856512606143951,
-0.05513917654752731,
0.04267800599336624,
0.08153519779443741,
-0.13977199792861938,
-0.07720372825860977,
-0.019453397020697594,
0.06108400225639343,
-0.06475253403186798,
-0.07767752557992935,
0.03208230435848236,
-0.09325236082077026,
0.12912479043006897,
0.032602302730083466,
-0.06646429002285004,
-0.07383444905281067,
-0.056502651423215866,
0.10576154291629791,
-0.07713192701339722,
0.014321833848953247,
-0.07465938478708267,
-0.002980181947350502,
0.029429275542497635,
-0.20701833069324493,
0.059617940336465836,
-0.12405465543270111,
-0.07004364579916,
0.019683878868818283,
0.007150308229029179,
-0.08969880640506744,
-0.015248440206050873,
0.008259961381554604,
-0.0398486964404583,
-0.07221176475286484,
-0.13755707442760468,
-0.0764327272772789,
0.17838558554649353,
-0.036647018045186996,
0.03891316056251526,
-0.07035698741674423,
-0.09427668154239655,
0.03308865427970886,
0.01587609201669693,
-0.0054944418370723724,
0.2126789391040802,
-0.022181198000907898,
0.040641337633132935,
0.28266119956970215,
-0.0072309705428779125,
-0.30207720398902893,
-0.10969895124435425,
-0.11626630276441574,
-0.08523907512426376,
-0.06700556725263596,
-0.029630109667778015,
0.14474453032016754,
0.06729910522699356,
-0.05077724531292915,
0.08054499328136444,
-0.22953657805919647,
-0.09605090320110321,
0.13016930222511292,
0.047243405133485794,
0.3216591775417328,
-0.13327282667160034,
-0.057928744703531265,
-0.13755367696285248,
-0.22692130506038666,
0.022588055580854416,
-0.27852368354797363,
0.07576199620962143,
-0.012118657119572163,
-0.060023289173841476,
-0.032676778733730316,
-0.05733204632997513,
0.128243088722229,
-0.018059642985463142,
0.11548397690057755,
-0.10826143622398376,
0.1142175942659378,
0.1303485929965973,
-0.0852547138929367,
0.22510556876659393,
-0.17714780569076538,
0.11108627915382385,
0.049339473247528076,
-0.008399272337555885,
-0.019310258328914642,
0.00690833106637001,
0.01720815896987915,
-0.03167569264769554,
-0.061363451182842255,
-0.020279591903090477,
0.03890672326087952,
-0.0011253264965489507,
0.11969151347875595,
0.04397323355078697,
-0.07856310158967972,
0.1799895316362381,
0.002150307409465313,
-0.19731315970420837,
0.06267164647579193,
-0.06631157547235489,
-0.051317255944013596,
0.08039046823978424,
-0.20267967879772186,
0.05430063605308533,
0.02739952690899372,
-0.043498020619153976,
0.05485006421804428,
0.021002186462283134,
0.019208962097764015,
-0.010522070340812206,
0.04706382378935814,
-0.1282617747783661,
-0.09725747257471085,
-0.04316909983754158,
-0.020872991532087326,
-0.05230532959103584,
0.11901336908340454,
0.14600113034248352,
-0.07164143025875092,
0.02893749438226223,
-0.003209911985322833,
0.04858538508415222,
-0.056962061673402786,
0.07114484161138535,
0.05920301005244255,
-0.03147534281015396,
-0.11638923734426498,
0.1892966777086258,
-0.04255680367350578,
0.03995633125305176,
-0.02791173756122589,
0.04695780202746391,
-0.20027147233486176,
-0.11093110591173172,
-0.036570195108652115,
0.07284019142389297,
-0.07039488106966019,
-0.045673247426748276,
-0.0801481083035469,
-0.05184655264019966,
0.03714932128787041,
-0.00411624601110816,
0.07419580221176147,
0.019241182133555412,
0.003246768144890666,
0.003850518027320504,
-0.017644697800278664,
0.053853701800107956,
0.06926816701889038,
0.05605730786919594,
-0.1878267079591751,
-0.03332631289958954,
-0.01947217620909214,
0.010130278766155243,
-0.05042608454823494,
0.03272939473390579,
-0.08151920139789581,
-0.0229108277708292,
-0.2974180281162262,
0.0668555423617363,
-0.08868566900491714,
0.04577907174825668,
-0.007397927809506655,
-0.05813651904463768,
-0.034154266119003296,
0.08335866034030914,
-0.045935314148664474,
-0.0369032621383667,
-0.009261644445359707,
0.01061952393501997,
-0.09509536623954773,
-0.05304750055074692,
0.010328748263418674,
-0.06332366168498993,
0.06519795954227448,
0.11247023940086365,
-0.11530674248933792,
0.047307390719652176,
-0.2577924132347107,
-0.10572075843811035,
0.06144891306757927,
0.03341138735413551,
-0.030940786004066467,
0.06895487010478973,
-0.0032651128713041544,
0.05420561134815216,
0.03221932053565979,
-0.04077920690178871,
0.10242040455341339,
-0.037979889661073685,
-0.05049901083111763,
-0.09808211773633957,
0.024536916986107826,
-0.059459127485752106,
-0.05690079554915428,
0.13833454251289368,
0.08246462792158127,
0.16391755640506744,
-0.05436510965228081,
-0.07251580059528351,
-0.1708066612482071,
-0.010773427784442902,
0.03970620408654213,
-0.13703791797161102,
-0.14597363770008087,
-0.04027874395251274,
0.016842151060700417,
-0.01725594699382782,
0.12148132175207138,
-0.04392300546169281,
-0.111797995865345,
-0.018491309136152267,
0.06240320950746536,
0.06477930396795273,
-0.01332712359726429,
0.24748846888542175,
0.030599217861890793,
0.02705036662518978,
-0.11022629588842392,
0.008790743537247181,
0.12283848226070404,
-0.04781937599182129,
-0.006835629232227802,
0.13386447727680206,
-0.016805877909064293,
0.16410894691944122,
0.010669870302081108,
0.071599081158638,
0.019070491194725037,
0.10175822675228119,
-0.029325243085622787,
0.08963417261838913,
-0.041905734688043594,
0.0871119275689125,
0.19933746755123138,
-0.04859602451324463,
-0.046182192862033844,
-0.023899637162685394,
-0.02056431584060192,
-0.13630937039852142,
-0.20304206013679504,
-0.11331785470247269,
-0.18400093913078308,
0.005142166279256344,
-0.051005054265260696,
0.002805712865665555,
0.08496032655239105,
0.023769691586494446,
0.023024525493383408,
0.06774240732192993,
-0.051614902913570404,
-0.054099421948194504,
0.03502710536122322,
-0.047996554523706436,
-0.12949787080287933,
0.16292591392993927,
-0.08239670097827911,
0.1290668249130249,
-0.022769581526517868,
0.020180467516183853,
0.023535018786787987,
0.0959106832742691,
0.0816233903169632,
-0.052182842046022415,
-0.06474878638982773,
-0.0615500770509243,
0.06743238121271133,
-0.019099490717053413,
0.07840829342603683,
0.06152671203017235,
-0.022138988599181175,
0.06056194379925728,
0.1808374971151352,
-0.07798463106155396,
-0.1876947283744812,
-0.12484414875507355,
0.06839922070503235,
-0.04011328145861626,
0.04718911275267601,
0.006012399680912495,
-0.05711129307746887,
0.035840097814798355,
0.1807534098625183,
0.186955526471138,
-0.10383448004722595,
-0.000173057327629067,
-0.02163984812796116,
-0.006837136577814817,
-0.04766799509525299,
0.13572508096694946,
0.1264045685529709,
0.022497689351439476,
-0.0033494955860078335,
-0.020566798746585846,
0.005698440596461296,
-0.024464121088385582,
-0.12919549643993378,
0.007727000396698713,
-0.08947039395570755,
-0.032542597502470016,
-0.02070460468530655,
0.005321551579982042,
-0.05318712443113327,
-0.07316798716783524,
-0.07420849800109863,
0.05546111986041069,
0.0014533302746713161,
-0.08053627610206604,
0.029041483998298645,
0.09253188222646713,
-0.0035081710666418076,
-0.09212682396173477,
0.07598704099655151,
0.12672656774520874,
-0.052302051335573196,
-0.12977898120880127,
-0.024899646639823914,
0.023535657674074173,
0.032839104533195496,
0.16266651451587677,
0.061057377606630325,
0.018872255459427834,
0.07097480446100235,
-0.04878688231110573,
-0.1785886436700821,
0.1186685785651207,
-0.04147284850478172,
-0.03092516027390957,
0.01824607513844967,
-0.03948238492012024,
-0.07362470775842667,
0.008962979540228844,
0.029853476211428642,
0.029658885672688484,
-0.04470803216099739,
0.10296875238418579,
-0.08583588153123856,
-0.08185906708240509,
0.006692495662719011,
-0.08545254915952682,
0.0975273996591568,
0.050012994557619095,
-0.043560415506362915,
-0.010937628336250782,
-0.08346528559923172,
0.08518145233392715,
0.018462102860212326,
-0.1432017683982849,
0.04062549024820328,
-0.01825881376862526,
-0.0011804829118773341,
0.017800403758883476,
0.06551897525787354,
-0.1411486119031906,
-0.030061420053243637,
-0.1090436577796936,
-0.0164520014077425,
-0.059693824499845505,
0.08420125395059586,
0.13707667589187622,
0.04151008278131485,
-0.01662694476544857,
-0.11354578286409378,
-0.02491127699613571,
0.05246390774846077,
-0.03943437710404396,
-0.09345787763595581
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# sparse_sparse_80_percent_pretraining_warmup_20K_0_2_steps_5k
This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the openwebtext dataset.
It achieves the following results on the evaluation set:
- Loss: 4.9832
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 16
- seed: 0
- distributed_type: multi-GPU
- num_devices: 3
- gradient_accumulation_steps: 2
- total_train_batch_size: 48
- total_eval_batch_size: 48
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 5000
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.2964 | 0.02 | 50 | 1.2517 |
| 1.1086 | 0.05 | 100 | 1.0714 |
| 0.9727 | 0.07 | 150 | 0.9857 |
| 0.9326 | 0.1 | 200 | 0.9357 |
| 0.8944 | 0.12 | 250 | 0.8988 |
| 0.872 | 0.15 | 300 | 0.8700 |
| 0.8523 | 0.17 | 350 | 0.8516 |
| 0.8369 | 0.19 | 400 | 0.8358 |
| 0.8372 | 0.22 | 450 | 0.8226 |
| 0.8221 | 0.24 | 500 | 0.8116 |
| 0.8093 | 0.27 | 550 | 0.8020 |
| 0.804 | 0.29 | 600 | 0.7937 |
| 0.8111 | 0.32 | 650 | 0.7935 |
| 0.7949 | 0.34 | 700 | 0.7872 |
| 0.7947 | 0.36 | 750 | 0.7815 |
| 0.8045 | 0.39 | 800 | 0.7771 |
| 0.7706 | 0.41 | 850 | 0.7724 |
| 0.7669 | 0.44 | 900 | 0.7683 |
| 0.7691 | 0.46 | 950 | 0.7825 |
| 0.7737 | 0.48 | 1000 | 0.7779 |
| 0.7595 | 0.51 | 1050 | 0.7748 |
| 0.7672 | 0.53 | 1100 | 0.7709 |
| 0.7725 | 0.56 | 1150 | 0.7681 |
| 0.7551 | 0.58 | 1200 | 0.7658 |
| 0.8035 | 0.61 | 1250 | 0.8159 |
| 0.804 | 0.63 | 1300 | 0.8068 |
| 0.8074 | 0.65 | 1350 | 0.8016 |
| 0.7801 | 0.68 | 1400 | 0.7982 |
| 0.7842 | 0.7 | 1450 | 0.7951 |
| 0.7938 | 0.73 | 1500 | 0.7907 |
| 0.8625 | 0.75 | 1550 | 0.8568 |
| 0.8467 | 0.78 | 1600 | 0.8443 |
| 0.8216 | 0.8 | 1650 | 0.8379 |
| 0.8334 | 0.82 | 1700 | 0.8332 |
| 0.8287 | 0.85 | 1750 | 0.8292 |
| 0.8251 | 0.87 | 1800 | 0.8250 |
| 0.8969 | 0.9 | 1850 | 0.8790 |
| 0.8619 | 0.92 | 1900 | 0.8696 |
| 0.8566 | 0.95 | 1950 | 0.8645 |
| 0.8633 | 0.97 | 2000 | 0.8599 |
| 0.8622 | 0.99 | 2050 | 0.8558 |
| 0.8336 | 1.02 | 2100 | 0.8520 |
| 0.918 | 1.04 | 2150 | 0.9045 |
| 0.8755 | 1.07 | 2200 | 0.8960 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["openwebtext"], "base_model": "mistralai/Mistral-7B-Instruct-v0.1", "model-index": [{"name": "sparse_sparse_80_percent_pretraining_warmup_20K_0_2_steps_5k", "results": []}]} | text-generation | thrunlab/sparse_sparse_80_percent_pretraining_warmup_20K_0_2_steps_5k | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"dataset:openwebtext",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T15:42:07+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #generated_from_trainer #dataset-openwebtext #base_model-mistralai/Mistral-7B-Instruct-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| sparse\_sparse\_80\_percent\_pretraining\_warmup\_20K\_0\_2\_steps\_5k
======================================================================
This model is a fine-tuned version of mistralai/Mistral-7B-Instruct-v0.1 on the openwebtext dataset.
It achieves the following results on the evaluation set:
* Loss: 4.9832
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 8
* eval\_batch\_size: 16
* seed: 0
* distributed\_type: multi-GPU
* num\_devices: 3
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 48
* total\_eval\_batch\_size: 48
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 5000
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.1+cu121
* Datasets 2.15.0
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 16\n* seed: 0\n* distributed\\_type: multi-GPU\n* num\\_devices: 3\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 48\n* total\\_eval\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 5000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #generated_from_trainer #dataset-openwebtext #base_model-mistralai/Mistral-7B-Instruct-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 16\n* seed: 0\n* distributed\\_type: multi-GPU\n* num\\_devices: 3\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 48\n* total\\_eval\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 5000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
88,
159,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #generated_from_trainer #dataset-openwebtext #base_model-mistralai/Mistral-7B-Instruct-v0.1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 16\n* seed: 0\n* distributed\\_type: multi-GPU\n* num\\_devices: 3\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 48\n* total\\_eval\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 5000### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.1+cu121\n* Datasets 2.15.0\n* Tokenizers 0.15.0"
] | [
-0.11656016856431961,
0.16379408538341522,
-0.004077443387359381,
0.08532476425170898,
0.10991210490465164,
0.04238564893603325,
0.14270780980587006,
0.15328873693943024,
-0.049688082188367844,
0.12170975655317307,
0.11878757923841476,
0.054467685520648956,
0.06454618275165558,
0.1988765448331833,
-0.03582637757062912,
-0.23163588345050812,
0.0379544235765934,
-0.04284830763936043,
-0.10081440955400467,
0.10722321271896362,
0.08271297067403793,
-0.10382620990276337,
0.0714406818151474,
-0.03363863378763199,
-0.09783557802438736,
-0.047938305884599686,
-0.03157222270965576,
-0.03805292397737503,
0.09599096328020096,
0.036135319620370865,
0.07293878495693207,
0.019058257341384888,
0.07875421643257141,
-0.2275000363588333,
-0.002402333077043295,
0.06736457347869873,
0.020258435979485512,
0.08187409490346909,
0.09603425115346909,
0.008121940307319164,
0.045354053378105164,
-0.09865104407072067,
0.04992295801639557,
0.024884801357984543,
-0.11605929583311081,
-0.23408812284469604,
-0.09771706908941269,
0.042095303535461426,
0.12788702547550201,
0.04553931951522827,
-0.015471848659217358,
0.05803751200437546,
-0.02522742748260498,
0.0752272978425026,
0.22748242318630219,
-0.2378915399312973,
-0.05899666249752045,
0.0471498966217041,
0.02969404309988022,
0.08615707606077194,
-0.08477199822664261,
-0.029737349599599838,
0.040043704211711884,
0.010626662522554398,
0.10913966596126556,
0.0229544248431921,
0.053107697516679764,
0.005381750874221325,
-0.14103296399116516,
-0.07794780284166336,
0.16395805776119232,
0.05835128203034401,
-0.020853839814662933,
-0.12366568297147751,
-0.06845119595527649,
-0.14158661663532257,
-0.028752779588103294,
0.02115817926824093,
0.024934491142630577,
-0.023907840251922607,
-0.044440437108278275,
0.028390031307935715,
-0.055831409990787506,
-0.07601381093263626,
0.03085310384631157,
0.08209604769945145,
0.05898350104689598,
-0.0034324147272855043,
0.0028118700720369816,
0.10714972764253616,
0.0032457425259053707,
-0.17250709235668182,
-0.0038386182859539986,
0.008564602583646774,
-0.06968230754137039,
-0.02422102354466915,
-0.019954465329647064,
0.03626043722033501,
0.05973164364695549,
0.18018807470798492,
-0.03723593056201935,
0.0622321180999279,
0.033251143991947174,
0.007902502082288265,
-0.059433355927467346,
0.1319292187690735,
-0.07162155210971832,
-0.09787070006132126,
-0.01763439178466797,
0.12629905343055725,
0.03584926947951317,
-0.0051882206462323666,
-0.0667639970779419,
0.015381532721221447,
0.10774360597133636,
0.05737188830971718,
0.01135832816362381,
0.040657345205545425,
-0.0730242133140564,
-0.022909311577677727,
0.10265861451625824,
-0.11739765852689743,
0.028535593301057816,
0.02760518155992031,
-0.06475847214460373,
-0.05350009724497795,
-0.004456894472241402,
-0.0076165031641721725,
-0.004458244889974594,
0.07778614014387131,
-0.07076293975114822,
-0.03484645113348961,
-0.06783971935510635,
-0.08808097243309021,
0.030692754313349724,
-0.06175718829035759,
-0.00910209771245718,
-0.05616101995110512,
-0.19038125872612,
-0.034117817878723145,
0.04828738048672676,
-0.07200899720191956,
-0.06819578260183334,
-0.05432838946580887,
-0.0922812819480896,
0.049313463270664215,
-0.012289597652852535,
0.11352665722370148,
-0.06474778056144714,
0.08184491097927094,
0.022505979984998703,
0.0724097192287445,
0.09281325340270996,
0.0396868996322155,
-0.05582466721534729,
0.07458990067243576,
-0.16792964935302734,
0.08034499734640121,
-0.08560676872730255,
0.06549707800149918,
-0.1371804028749466,
-0.07362518459558487,
0.027667419984936714,
-0.024128608405590057,
0.08446690440177917,
0.14349418878555298,
-0.1635718196630478,
-0.051609545946121216,
0.20322223007678986,
-0.09749026596546173,
-0.14215828478336334,
0.13132046163082123,
-0.024951644241809845,
-0.039706889539957047,
0.030291831120848656,
0.14625176787376404,
0.11700547486543655,
-0.08988720923662186,
-0.009382914751768112,
-0.02653803676366806,
0.11602783203125,
0.019499707967042923,
0.11042840778827667,
-0.02835834212601185,
0.03047245182096958,
0.010361094027757645,
-0.0749628096818924,
0.04121367633342743,
-0.08295394480228424,
-0.08384202420711517,
-0.04729863256216049,
-0.0892333909869194,
0.026984654366970062,
0.033520568162202835,
0.025307849049568176,
-0.08578789979219437,
-0.11033709347248077,
-0.016274504363536835,
0.11457207053899765,
-0.09306685626506805,
0.0067903585731983185,
-0.04772469028830528,
0.10001998394727707,
-0.05374731495976448,
0.00628664018586278,
-0.14782416820526123,
-0.09026231616735458,
0.06338565796613693,
-0.043714489787817,
-0.041551943868398666,
-0.0346517413854599,
0.06245418265461922,
0.08510756492614746,
-0.061035968363285065,
-0.06789367645978928,
-0.005703662056475878,
-0.006505467463284731,
-0.07502903789281845,
-0.21867696940898895,
-0.05110830068588257,
-0.04637559875845909,
0.12353923916816711,
-0.2156403511762619,
0.029004277661442757,
0.023444199934601784,
0.12557822465896606,
0.030009523034095764,
-0.04381949082016945,
0.01118833664804697,
0.018282709643244743,
-0.04674914851784706,
-0.09644410759210587,
0.027079399675130844,
-0.008392623625695705,
-0.09942280501127243,
-0.006929260212928057,
-0.13942216336727142,
0.11413680762052536,
0.0846656933426857,
0.05684342235326767,
-0.07858840376138687,
-0.033145904541015625,
-0.05129339173436165,
-0.05836218222975731,
-0.043161943554878235,
-0.03210033103823662,
0.09414612501859665,
0.012455975636839867,
0.11528845876455307,
-0.09286656230688095,
-0.0636136457324028,
0.02342432178556919,
-0.004071634262800217,
-0.020146267488598824,
0.13043944537639618,
0.017267465591430664,
-0.08175993710756302,
0.1384475976228714,
0.1417538821697235,
-0.03214219957590103,
0.10637883096933365,
-0.08546686172485352,
-0.07959691435098648,
-0.04875907301902771,
0.051897481083869934,
0.03231857717037201,
0.09326764941215515,
-0.07278865575790405,
0.0037839855067431927,
0.03268679231405258,
0.03645150363445282,
0.0042569334618747234,
-0.16120103001594543,
-0.008664017543196678,
0.039033666253089905,
-0.08918919414281845,
0.041558749973773956,
-0.020495496690273285,
-0.017367292195558548,
0.09274828433990479,
0.0043809739872813225,
-0.05099562183022499,
0.008070142939686775,
-0.029726572334766388,
-0.09065308421850204,
0.19505326449871063,
-0.0983637198805809,
-0.09739913791418076,
-0.1288929432630539,
-0.01903962716460228,
-0.02509784884750843,
0.002760221716016531,
0.02927287667989731,
-0.07486200332641602,
-0.0508078895509243,
-0.1095716580748558,
0.006082449574023485,
0.004769469145685434,
0.025035057216882706,
0.055679868906736374,
0.004255796782672405,
0.07037001848220825,
-0.0897197499871254,
0.009227773174643517,
0.012545285746455193,
-0.049436554312705994,
0.02888103947043419,
0.01477702334523201,
0.09224674850702286,
0.12933240830898285,
0.037890005856752396,
0.01448868215084076,
-0.027072494849562645,
0.2163587510585785,
-0.07256437093019485,
0.017979176715016365,
0.10172666609287262,
-0.0026622998993843794,
0.07401334494352341,
0.18621572852134705,
0.03244045749306679,
-0.09137023985385895,
0.021909503266215324,
0.010592438280582428,
-0.01732444018125534,
-0.22732026875019073,
-0.034070271998643875,
-0.03897798806428909,
0.038345564156770706,
0.11384660005569458,
0.03613540530204773,
-0.01426529698073864,
0.04939939081668854,
-0.037165362387895584,
0.018877992406487465,
0.03865225985646248,
0.07785037904977798,
0.03432077541947365,
0.04889405891299248,
0.10740687698125839,
-0.03050827793776989,
-0.008510625921189785,
0.053470443934202194,
0.016580430790781975,
0.2652735710144043,
-0.0296246986836195,
0.21637769043445587,
0.030552919954061508,
0.1340404450893402,
-0.008211997337639332,
0.0563652329146862,
-0.006699762772768736,
0.016362452879548073,
-0.007146984338760376,
-0.059522248804569244,
-0.011810870841145515,
0.04507344216108322,
-0.025425616651773453,
0.04043254628777504,
-0.09059594571590424,
0.02901514060795307,
0.060108959674835205,
0.2621021270751953,
0.08708120137453079,
-0.3127545714378357,
-0.0804901197552681,
0.055672239512205124,
-0.0447428934276104,
-0.01765087991952896,
0.012127751484513283,
0.1498565524816513,
-0.07551149278879166,
0.07731179893016815,
-0.0653330534696579,
0.07843701541423798,
-0.06627985835075378,
0.0003724681737367064,
0.09205770492553711,
0.08330332487821579,
-0.0006558923050761223,
0.05676953122019768,
-0.23530228435993195,
0.2715862989425659,
-0.002501231851056218,
0.04268483445048332,
-0.05271999537944794,
0.052493564784526825,
0.012717178091406822,
0.011797581799328327,
0.09065801650285721,
-0.006843642797321081,
-0.1295480877161026,
-0.15503086149692535,
-0.1245848536491394,
0.005925468634814024,
0.10193698108196259,
-0.09669341892004013,
0.11191661655902863,
-0.03115115500986576,
-0.03249078616499901,
0.045747846364974976,
-0.02122427150607109,
-0.07327234745025635,
-0.1002935916185379,
0.027365559712052345,
-0.020128915086388588,
0.0009331672918051481,
-0.09029824286699295,
-0.08252006769180298,
-0.10648871958255768,
0.143052875995636,
-0.10542364418506622,
-0.0572398379445076,
-0.12316898256540298,
0.05486222356557846,
0.1591663807630539,
-0.0916440337896347,
0.05369146913290024,
0.0013649949105456471,
0.10305172950029373,
0.028518136590719223,
-0.038359012454748154,
0.10099012404680252,
-0.08377286046743393,
-0.2366519719362259,
-0.07532399892807007,
0.1267174631357193,
0.01812615990638733,
0.04721466824412346,
-0.03354458883404732,
0.03935961797833443,
-0.0039962539449334145,
-0.08196719735860825,
0.03569839149713516,
0.0829230397939682,
0.08010827749967575,
0.052125170826911926,
-0.03887658193707466,
-0.04213317483663559,
-0.04298354312777519,
-0.0595928356051445,
0.09260943531990051,
0.3056355118751526,
-0.09556896239519119,
0.03825224190950394,
0.06243494898080826,
-0.06283563375473022,
-0.1648954600095749,
-0.025849517434835434,
0.08160557597875595,
0.020582009106874466,
0.020190512761473656,
-0.16499583423137665,
0.04527932405471802,
0.09177708625793457,
-0.036855485290288925,
0.07254023104906082,
-0.3340925872325897,
-0.13598331809043884,
0.052967946976423264,
0.1018812358379364,
-0.0016835861606523395,
-0.18744173645973206,
-0.07739527523517609,
-0.010546671226620674,
-0.10894934087991714,
0.06833852082490921,
-0.011342419311404228,
0.09563088417053223,
-0.019671309739351273,
0.009691144339740276,
0.014838687144219875,
-0.06832179427146912,
0.17374379932880402,
0.02984474040567875,
0.07127919793128967,
-0.04750865697860718,
0.024002479389309883,
0.06519678235054016,
-0.09119895845651627,
0.030553987249732018,
-0.10337301343679428,
0.05777513235807419,
-0.11927832663059235,
-0.014125300571322441,
-0.05060254782438278,
-0.011303492821753025,
-0.04624072462320328,
-0.03236470744013786,
-0.044579487293958664,
0.05506351590156555,
0.09491471946239471,
-0.014567963778972626,
0.11244439333677292,
0.033005621284246445,
0.11788281798362732,
0.12552779912948608,
0.05923870578408241,
0.02697487361729145,
-0.07429762184619904,
-0.0072434814646840096,
-0.011664967983961105,
0.029062431305646896,
-0.17539067566394806,
0.03987504541873932,
0.12395963072776794,
0.019765034317970276,
0.13623861968517303,
0.04358178749680519,
-0.05914101377129555,
-0.003092625178396702,
0.06351309269666672,
-0.12414401024580002,
-0.14655590057373047,
0.005663441028445959,
0.004943953361362219,
-0.1665484756231308,
0.013624902814626694,
0.10313097387552261,
-0.032585322856903076,
-0.002465645782649517,
-0.0132296122610569,
0.07306598126888275,
0.0033858574461191893,
0.2229466438293457,
0.028973249718546867,
0.08637838065624237,
-0.08554421365261078,
0.11054029315710068,
0.06711897999048233,
-0.08436429500579834,
0.028679924085736275,
0.07524824887514114,
-0.09041529148817062,
-0.02333977073431015,
0.06502910703420639,
0.11038488894701004,
0.03128427639603615,
-0.02983788214623928,
-0.10593388974666595,
-0.10926967114210129,
0.08296280354261398,
0.08558562397956848,
0.06169845163822174,
0.0476226843893528,
0.012796872295439243,
0.004376816097646952,
-0.0781843513250351,
0.1499299257993698,
0.10819374024868011,
0.0839034914970398,
-0.13867329061031342,
0.06548599153757095,
-0.019677329808473587,
-0.011013406328856945,
-0.007446495816111565,
0.04621638357639313,
-0.12133047729730606,
-0.01901458390057087,
-0.10253571718931198,
0.03831956908106804,
-0.06866738945245743,
-0.004799835383892059,
-0.012203509919345379,
-0.05656716600060463,
-0.026812106370925903,
0.008432713337242603,
-0.0784534439444542,
-0.05142512544989586,
-0.03107856959104538,
0.07815011590719223,
-0.12199244648218155,
-0.04038058593869209,
0.03346898406744003,
-0.11842291802167892,
0.11309804767370224,
0.01813322678208351,
0.03298325464129448,
-0.008080684579908848,
-0.09204982966184616,
0.029514215886592865,
0.018775975331664085,
0.003727721981704235,
0.020716894418001175,
-0.1702912300825119,
-0.00001727956441754941,
-0.03100806288421154,
-0.012410074472427368,
0.006251998245716095,
0.04430883005261421,
-0.12304636836051941,
0.028766531497240067,
-0.028815466910600662,
-0.03692779317498207,
-0.05020317807793617,
0.03887200728058815,
0.0967399924993515,
-0.016690034419298172,
0.15519526600837708,
-0.0713822990655899,
0.04972141608595848,
-0.23408423364162445,
-0.015018682926893234,
0.009195269085466862,
-0.0912969782948494,
-0.11248823255300522,
-0.008327464573085308,
0.07819190621376038,
-0.05772589147090912,
0.12161564081907272,
-0.057929687201976776,
0.01728908345103264,
0.023654405027627945,
-0.0265805721282959,
0.06714893132448196,
0.06317329406738281,
0.17122100293636322,
0.02567688189446926,
-0.03388562798500061,
0.04067119583487511,
-0.008475316688418388,
0.07125715166330338,
0.05074431002140045,
0.13530923426151276,
0.15290655195713043,
0.008260758593678474,
0.05987416207790375,
0.034506335854530334,
-0.12561924755573273,
-0.13622254133224487,
0.10195820033550262,
-0.0698748528957367,
0.12258755415678024,
-0.022098852321505547,
0.1597266048192978,
0.0923357754945755,
-0.21413755416870117,
0.03302774205803871,
-0.032805830240249634,
-0.08189897239208221,
-0.09951833635568619,
-0.11245962232351303,
-0.091365285217762,
-0.14368847012519836,
-0.00020707320072688162,
-0.11933925747871399,
0.029038632288575172,
0.10446209460496902,
0.01976086013019085,
0.017539622262120247,
0.10659893602132797,
0.0679725706577301,
0.027143454179167747,
0.05882509797811508,
0.03662518039345741,
-0.010736022144556046,
-0.002979520708322525,
-0.09273263812065125,
0.019465062767267227,
-0.0048275478184223175,
0.045503661036491394,
-0.01800885610282421,
-0.027332700788974762,
0.06283251196146011,
0.001948708202689886,
-0.09792061895132065,
0.011241229251027107,
-0.009913505986332893,
0.018717758357524872,
0.06284357607364655,
0.03131859004497528,
0.0034496996086090803,
-0.009801676496863365,
0.19753915071487427,
-0.07496194541454315,
-0.05619117245078087,
-0.12093672156333923,
0.1756768375635147,
-0.015364790335297585,
-0.006588879972696304,
0.04856251925230026,
-0.07689137011766434,
-0.007480607833713293,
0.13550332188606262,
0.17075420916080475,
-0.037327248603105545,
-0.01896485686302185,
0.01601439155638218,
-0.017784956842660904,
-0.0031472237315028906,
0.08735035359859467,
0.09257761389017105,
0.036909595131874084,
-0.04689827933907509,
-0.017958300188183784,
-0.012710925191640854,
-0.04423641413450241,
-0.053057607263326645,
0.05631188675761223,
-0.006705369334667921,
0.008036484010517597,
-0.025701792910695076,
0.0770920068025589,
-0.03561215475201607,
-0.09589699655771255,
0.04836960509419441,
-0.19936534762382507,
-0.1801488697528839,
-0.024267634376883507,
0.09090232104063034,
0.014764802530407906,
0.04464123398065567,
0.010521117597818375,
-0.007283895276486874,
0.10556768625974655,
-0.022920651361346245,
-0.05846785753965378,
-0.08201988786458969,
0.042674802243709564,
-0.09617818892002106,
0.21137671172618866,
-0.01748170703649521,
0.03195487707853317,
0.13012008368968964,
0.004630686715245247,
-0.1501578837633133,
0.029257306829094887,
0.09181646257638931,
-0.05078629404306412,
0.03793061152100563,
0.13561499118804932,
-0.04513341560959816,
0.11594486981630325,
0.0682762861251831,
-0.06783126294612885,
-0.014737872406840324,
-0.00967022031545639,
-0.03304682672023773,
-0.07335896044969559,
-0.017666520550847054,
-0.04138462617993355,
0.16032768785953522,
0.18800924718379974,
-0.07423518598079681,
-0.004600682761520147,
-0.023738887161016464,
0.025039827451109886,
0.026792176067829132,
0.12150301784276962,
-0.007742081303149462,
-0.2681795358657837,
0.024162529036402702,
0.04263226315379143,
0.05487583950161934,
-0.20372052490711212,
-0.10033112019300461,
0.017686516046524048,
-0.032235171645879745,
-0.1033378466963768,
0.13474838435649872,
0.07881329208612442,
0.03750460222363472,
-0.052224863320589066,
-0.08868247270584106,
-0.06627457588911057,
0.1617753952741623,
-0.1514575183391571,
-0.08004269748926163
] |
null | null | null | ips rssi coor prediction | {} | null | naphatkmitl/ips-rssi | [
"region:us"
] | 2024-02-08T15:49:19+00:00 | [] | [] | TAGS
#region-us
| ips rssi coor prediction | [] | [
"TAGS\n#region-us \n"
] | [
6
] | [
"passage: TAGS\n#region-us \n"
] | [
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5
This model is a fine-tuned version of [nvidia/mit-b5](https://huggingface.co/nvidia/mit-b5) on the sam1120/safety-utcustom-TRAIN dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0472
- Mean Iou: 0.8728
- Mean Accuracy: 0.9195
- Overall Accuracy: 0.9919
- Accuracy Unlabeled: nan
- Accuracy Safe: 0.8426
- Accuracy Unsafe: 0.9964
- Iou Unlabeled: nan
- Iou Safe: 0.7540
- Iou Unsafe: 0.9917
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 2000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Accuracy Unlabeled | Accuracy Safe | Accuracy Unsafe | Iou Unlabeled | Iou Safe | Iou Unsafe |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:------------------:|:-------------:|:---------------:|:-------------:|:--------:|:----------:|
| 1.2199 | 2.0 | 20 | 1.1474 | 0.0765 | 0.4399 | 0.2168 | nan | 0.6770 | 0.2028 | 0.0 | 0.0279 | 0.2014 |
| 1.1542 | 4.0 | 40 | 1.0616 | 0.1558 | 0.6082 | 0.4365 | nan | 0.7908 | 0.4257 | 0.0 | 0.0436 | 0.4237 |
| 1.0324 | 6.0 | 60 | 0.9140 | 0.2569 | 0.6886 | 0.7091 | nan | 0.6667 | 0.7104 | 0.0 | 0.0672 | 0.7036 |
| 0.8058 | 8.0 | 80 | 0.7629 | 0.2970 | 0.6780 | 0.8115 | nan | 0.5360 | 0.8199 | 0.0 | 0.0823 | 0.8086 |
| 0.681 | 10.0 | 100 | 0.5545 | 0.3510 | 0.6964 | 0.9135 | nan | 0.4657 | 0.9271 | 0.0 | 0.1407 | 0.9123 |
| 0.5248 | 12.0 | 120 | 0.4153 | 0.3738 | 0.6747 | 0.9462 | nan | 0.3861 | 0.9633 | 0.0 | 0.1757 | 0.9456 |
| 0.3372 | 14.0 | 140 | 0.3050 | 0.3896 | 0.6548 | 0.9628 | nan | 0.3276 | 0.9821 | 0.0 | 0.2064 | 0.9624 |
| 0.2818 | 16.0 | 160 | 0.2346 | 0.4313 | 0.7331 | 0.9703 | nan | 0.4810 | 0.9852 | 0.0 | 0.3239 | 0.9699 |
| 0.2081 | 18.0 | 180 | 0.1802 | 0.4762 | 0.7973 | 0.9782 | nan | 0.6050 | 0.9896 | 0.0 | 0.4509 | 0.9778 |
| 0.141 | 20.0 | 200 | 0.1362 | 0.4968 | 0.7945 | 0.9830 | nan | 0.5943 | 0.9948 | 0.0 | 0.5078 | 0.9827 |
| 0.0963 | 22.0 | 220 | 0.1032 | 0.7409 | 0.7652 | 0.9841 | nan | 0.5326 | 0.9979 | nan | 0.4978 | 0.9839 |
| 0.0842 | 24.0 | 240 | 0.0770 | 0.7840 | 0.8200 | 0.9863 | nan | 0.6432 | 0.9968 | nan | 0.5818 | 0.9861 |
| 0.0702 | 26.0 | 260 | 0.0669 | 0.7836 | 0.8193 | 0.9863 | nan | 0.6417 | 0.9968 | nan | 0.5812 | 0.9861 |
| 0.0706 | 28.0 | 280 | 0.0671 | 0.8065 | 0.8593 | 0.9872 | nan | 0.7234 | 0.9953 | nan | 0.6261 | 0.9870 |
| 0.0747 | 30.0 | 300 | 0.0551 | 0.7808 | 0.7980 | 0.9870 | nan | 0.5971 | 0.9988 | nan | 0.5748 | 0.9867 |
| 0.057 | 32.0 | 320 | 0.0492 | 0.8267 | 0.8736 | 0.9888 | nan | 0.7511 | 0.9961 | nan | 0.6648 | 0.9886 |
| 0.0435 | 34.0 | 340 | 0.0507 | 0.7956 | 0.8134 | 0.9878 | nan | 0.6280 | 0.9988 | nan | 0.6035 | 0.9876 |
| 0.0326 | 36.0 | 360 | 0.0418 | 0.8422 | 0.8895 | 0.9898 | nan | 0.7830 | 0.9961 | nan | 0.6947 | 0.9896 |
| 0.0262 | 38.0 | 380 | 0.0420 | 0.8280 | 0.8550 | 0.9895 | nan | 0.7120 | 0.9979 | nan | 0.6667 | 0.9893 |
| 0.0268 | 40.0 | 400 | 0.0392 | 0.8407 | 0.8822 | 0.9899 | nan | 0.7676 | 0.9967 | nan | 0.6918 | 0.9897 |
| 0.0395 | 42.0 | 420 | 0.0466 | 0.5436 | 0.8370 | 0.9889 | nan | 0.6755 | 0.9984 | 0.0 | 0.6422 | 0.9887 |
| 0.0279 | 44.0 | 440 | 0.0439 | 0.8321 | 0.8946 | 0.9887 | nan | 0.7945 | 0.9947 | nan | 0.6758 | 0.9885 |
| 0.0468 | 46.0 | 460 | 0.0360 | 0.8480 | 0.8894 | 0.9904 | nan | 0.7822 | 0.9967 | nan | 0.7059 | 0.9901 |
| 0.0233 | 48.0 | 480 | 0.0376 | 0.8507 | 0.8962 | 0.9905 | nan | 0.7960 | 0.9964 | nan | 0.7113 | 0.9902 |
| 0.0288 | 50.0 | 500 | 0.0386 | 0.8404 | 0.8845 | 0.9898 | nan | 0.7725 | 0.9964 | nan | 0.6913 | 0.9896 |
| 0.0266 | 52.0 | 520 | 0.0361 | 0.8455 | 0.8768 | 0.9905 | nan | 0.7560 | 0.9976 | nan | 0.7008 | 0.9902 |
| 0.0241 | 54.0 | 540 | 0.0367 | 0.8504 | 0.9039 | 0.9902 | nan | 0.8122 | 0.9957 | nan | 0.7109 | 0.9900 |
| 0.0239 | 56.0 | 560 | 0.0414 | 0.8401 | 0.8809 | 0.9899 | nan | 0.7650 | 0.9967 | nan | 0.6906 | 0.9896 |
| 0.0221 | 58.0 | 580 | 0.0375 | 0.8536 | 0.8971 | 0.9907 | nan | 0.7977 | 0.9966 | nan | 0.7167 | 0.9905 |
| 0.0324 | 60.0 | 600 | 0.0390 | 0.8566 | 0.9017 | 0.9908 | nan | 0.8069 | 0.9964 | nan | 0.7226 | 0.9906 |
| 0.0264 | 62.0 | 620 | 0.0405 | 0.8506 | 0.9088 | 0.9901 | nan | 0.8224 | 0.9952 | nan | 0.7114 | 0.9899 |
| 0.0146 | 64.0 | 640 | 0.0356 | 0.8627 | 0.9158 | 0.9911 | nan | 0.8358 | 0.9958 | nan | 0.7346 | 0.9909 |
| 0.0252 | 66.0 | 660 | 0.0310 | 0.8667 | 0.9010 | 0.9917 | nan | 0.8046 | 0.9974 | nan | 0.7418 | 0.9915 |
| 0.0155 | 68.0 | 680 | 0.0359 | 0.8567 | 0.9056 | 0.9908 | nan | 0.8152 | 0.9961 | nan | 0.7229 | 0.9905 |
| 0.0169 | 70.0 | 700 | 0.0490 | 0.8476 | 0.9182 | 0.9896 | nan | 0.8422 | 0.9941 | nan | 0.7058 | 0.9894 |
| 0.0142 | 72.0 | 720 | 0.0357 | 0.8442 | 0.8771 | 0.9903 | nan | 0.7568 | 0.9974 | nan | 0.6982 | 0.9901 |
| 0.0244 | 74.0 | 740 | 0.0400 | 0.8523 | 0.9070 | 0.9903 | nan | 0.8183 | 0.9956 | nan | 0.7146 | 0.9901 |
| 0.016 | 76.0 | 760 | 0.0302 | 0.8644 | 0.9037 | 0.9915 | nan | 0.8105 | 0.9970 | nan | 0.7376 | 0.9913 |
| 0.0137 | 78.0 | 780 | 0.0325 | 0.8664 | 0.9118 | 0.9915 | nan | 0.8271 | 0.9965 | nan | 0.7415 | 0.9913 |
| 0.0115 | 80.0 | 800 | 0.0347 | 0.8678 | 0.9162 | 0.9915 | nan | 0.8362 | 0.9962 | nan | 0.7443 | 0.9913 |
| 0.0117 | 82.0 | 820 | 0.0320 | 0.8697 | 0.9084 | 0.9918 | nan | 0.8197 | 0.9971 | nan | 0.7478 | 0.9916 |
| 0.0108 | 84.0 | 840 | 0.0348 | 0.8691 | 0.9192 | 0.9916 | nan | 0.8423 | 0.9961 | nan | 0.7468 | 0.9913 |
| 0.0101 | 86.0 | 860 | 0.0342 | 0.8683 | 0.9081 | 0.9917 | nan | 0.8192 | 0.9970 | nan | 0.7452 | 0.9915 |
| 0.0085 | 88.0 | 880 | 0.0441 | 0.8639 | 0.9214 | 0.9911 | nan | 0.8474 | 0.9954 | nan | 0.7370 | 0.9908 |
| 0.009 | 90.0 | 900 | 0.0428 | 0.8619 | 0.9086 | 0.9912 | nan | 0.8209 | 0.9963 | nan | 0.7330 | 0.9909 |
| 0.009 | 92.0 | 920 | 0.0444 | 0.8620 | 0.9089 | 0.9912 | nan | 0.8215 | 0.9963 | nan | 0.7331 | 0.9909 |
| 0.0089 | 94.0 | 940 | 0.0410 | 0.8645 | 0.9147 | 0.9913 | nan | 0.8334 | 0.9961 | nan | 0.7380 | 0.9910 |
| 0.0091 | 96.0 | 960 | 0.0418 | 0.8663 | 0.9155 | 0.9914 | nan | 0.8349 | 0.9962 | nan | 0.7413 | 0.9912 |
| 0.0079 | 98.0 | 980 | 0.0398 | 0.8629 | 0.9085 | 0.9913 | nan | 0.8205 | 0.9965 | nan | 0.7348 | 0.9910 |
| 0.0084 | 100.0 | 1000 | 0.0497 | 0.8553 | 0.9109 | 0.9905 | nan | 0.8262 | 0.9955 | nan | 0.7204 | 0.9903 |
| 0.0088 | 102.0 | 1020 | 0.0399 | 0.8558 | 0.9058 | 0.9907 | nan | 0.8156 | 0.9960 | nan | 0.7212 | 0.9905 |
| 0.0089 | 104.0 | 1040 | 0.0388 | 0.8678 | 0.9225 | 0.9914 | nan | 0.8494 | 0.9957 | nan | 0.7444 | 0.9912 |
| 0.008 | 106.0 | 1060 | 0.0449 | 0.8622 | 0.9225 | 0.9909 | nan | 0.8498 | 0.9952 | nan | 0.7337 | 0.9907 |
| 0.0084 | 108.0 | 1080 | 0.0429 | 0.8687 | 0.9233 | 0.9914 | nan | 0.8510 | 0.9957 | nan | 0.7462 | 0.9912 |
| 0.0084 | 110.0 | 1100 | 0.0405 | 0.8687 | 0.9169 | 0.9916 | nan | 0.8375 | 0.9963 | nan | 0.7460 | 0.9914 |
| 0.007 | 112.0 | 1120 | 0.0544 | 0.8620 | 0.9180 | 0.9910 | nan | 0.8404 | 0.9956 | nan | 0.7333 | 0.9907 |
| 0.0079 | 114.0 | 1140 | 0.0501 | 0.8602 | 0.9176 | 0.9908 | nan | 0.8399 | 0.9954 | nan | 0.7299 | 0.9906 |
| 0.0084 | 116.0 | 1160 | 0.0508 | 0.8605 | 0.9212 | 0.9908 | nan | 0.8473 | 0.9951 | nan | 0.7304 | 0.9905 |
| 0.0113 | 118.0 | 1180 | 0.0511 | 0.8601 | 0.9228 | 0.9907 | nan | 0.8507 | 0.9950 | nan | 0.7298 | 0.9905 |
| 0.0076 | 120.0 | 1200 | 0.0556 | 0.8602 | 0.9264 | 0.9906 | nan | 0.8582 | 0.9947 | nan | 0.7299 | 0.9904 |
| 0.0081 | 122.0 | 1220 | 0.0471 | 0.8665 | 0.9256 | 0.9912 | nan | 0.8559 | 0.9953 | nan | 0.7420 | 0.9910 |
| 0.0054 | 124.0 | 1240 | 0.0504 | 0.8652 | 0.9174 | 0.9913 | nan | 0.8389 | 0.9959 | nan | 0.7394 | 0.9910 |
| 0.0054 | 126.0 | 1260 | 0.0502 | 0.8666 | 0.9209 | 0.9913 | nan | 0.8461 | 0.9957 | nan | 0.7420 | 0.9911 |
| 0.0092 | 128.0 | 1280 | 0.0540 | 0.8642 | 0.9223 | 0.9911 | nan | 0.8492 | 0.9954 | nan | 0.7376 | 0.9908 |
| 0.007 | 130.0 | 1300 | 0.0533 | 0.8637 | 0.9207 | 0.9911 | nan | 0.8459 | 0.9955 | nan | 0.7366 | 0.9908 |
| 0.0063 | 132.0 | 1320 | 0.0526 | 0.8644 | 0.9259 | 0.9910 | nan | 0.8567 | 0.9951 | nan | 0.7380 | 0.9908 |
| 0.0101 | 134.0 | 1340 | 0.0462 | 0.8653 | 0.9067 | 0.9915 | nan | 0.8166 | 0.9968 | nan | 0.7393 | 0.9913 |
| 0.0183 | 136.0 | 1360 | 0.0516 | 0.5675 | 0.9024 | 0.9904 | nan | 0.8089 | 0.9959 | 0.0 | 0.7123 | 0.9901 |
| 0.0102 | 138.0 | 1380 | 0.0388 | 0.8366 | 0.8716 | 0.9898 | nan | 0.7460 | 0.9972 | nan | 0.6837 | 0.9896 |
| 0.0277 | 140.0 | 1400 | 0.0649 | 0.8159 | 0.9590 | 0.9850 | nan | 0.9313 | 0.9866 | nan | 0.6472 | 0.9846 |
| 0.0169 | 142.0 | 1420 | 0.0340 | 0.8444 | 0.9148 | 0.9894 | nan | 0.8355 | 0.9941 | nan | 0.6996 | 0.9891 |
| 0.0359 | 144.0 | 1440 | 0.0314 | 0.8667 | 0.8987 | 0.9918 | nan | 0.7997 | 0.9976 | nan | 0.7419 | 0.9916 |
| 0.0117 | 146.0 | 1460 | 0.0307 | 0.8517 | 0.8869 | 0.9908 | nan | 0.7765 | 0.9973 | nan | 0.7129 | 0.9905 |
| 0.0097 | 148.0 | 1480 | 0.0323 | 0.8757 | 0.9070 | 0.9924 | nan | 0.8162 | 0.9977 | nan | 0.7593 | 0.9922 |
| 0.0063 | 150.0 | 1500 | 0.0302 | 0.8808 | 0.9155 | 0.9926 | nan | 0.8335 | 0.9975 | nan | 0.7692 | 0.9924 |
| 0.0085 | 152.0 | 1520 | 0.0352 | 0.8697 | 0.9113 | 0.9918 | nan | 0.8258 | 0.9968 | nan | 0.7478 | 0.9916 |
| 0.0078 | 154.0 | 1540 | 0.0428 | 0.8649 | 0.9190 | 0.9912 | nan | 0.8423 | 0.9957 | nan | 0.7388 | 0.9910 |
| 0.0056 | 156.0 | 1560 | 0.0340 | 0.8709 | 0.9170 | 0.9918 | nan | 0.8376 | 0.9965 | nan | 0.7502 | 0.9916 |
| 0.0063 | 158.0 | 1580 | 0.0359 | 0.8661 | 0.9201 | 0.9913 | nan | 0.8445 | 0.9958 | nan | 0.7412 | 0.9911 |
| 0.0083 | 160.0 | 1600 | 0.0375 | 0.8684 | 0.9186 | 0.9915 | nan | 0.8410 | 0.9961 | nan | 0.7456 | 0.9913 |
| 0.0065 | 162.0 | 1620 | 0.0370 | 0.8699 | 0.9210 | 0.9916 | nan | 0.8459 | 0.9960 | nan | 0.7484 | 0.9914 |
| 0.0063 | 164.0 | 1640 | 0.0388 | 0.8699 | 0.9228 | 0.9916 | nan | 0.8498 | 0.9959 | nan | 0.7484 | 0.9913 |
| 0.0056 | 166.0 | 1660 | 0.0386 | 0.8702 | 0.9238 | 0.9916 | nan | 0.8517 | 0.9958 | nan | 0.7491 | 0.9914 |
| 0.0049 | 168.0 | 1680 | 0.0394 | 0.8703 | 0.9199 | 0.9917 | nan | 0.8436 | 0.9962 | nan | 0.7491 | 0.9914 |
| 0.0054 | 170.0 | 1700 | 0.0400 | 0.8704 | 0.9195 | 0.9917 | nan | 0.8428 | 0.9962 | nan | 0.7494 | 0.9915 |
| 0.0046 | 172.0 | 1720 | 0.0398 | 0.8728 | 0.9187 | 0.9919 | nan | 0.8410 | 0.9965 | nan | 0.7539 | 0.9917 |
| 0.0058 | 174.0 | 1740 | 0.0402 | 0.8711 | 0.9166 | 0.9918 | nan | 0.8367 | 0.9965 | nan | 0.7507 | 0.9916 |
| 0.005 | 176.0 | 1760 | 0.0400 | 0.8720 | 0.9196 | 0.9918 | nan | 0.8428 | 0.9963 | nan | 0.7525 | 0.9916 |
| 0.0061 | 178.0 | 1780 | 0.0417 | 0.8714 | 0.9226 | 0.9917 | nan | 0.8492 | 0.9960 | nan | 0.7513 | 0.9915 |
| 0.0061 | 180.0 | 1800 | 0.0407 | 0.8731 | 0.9249 | 0.9918 | nan | 0.8538 | 0.9960 | nan | 0.7545 | 0.9916 |
| 0.0065 | 182.0 | 1820 | 0.0420 | 0.8712 | 0.9235 | 0.9917 | nan | 0.8511 | 0.9959 | nan | 0.7509 | 0.9914 |
| 0.0045 | 184.0 | 1840 | 0.0421 | 0.8718 | 0.9250 | 0.9917 | nan | 0.8541 | 0.9959 | nan | 0.7522 | 0.9915 |
| 0.0056 | 186.0 | 1860 | 0.0435 | 0.8703 | 0.9157 | 0.9917 | nan | 0.8349 | 0.9965 | nan | 0.7490 | 0.9915 |
| 0.0059 | 188.0 | 1880 | 0.0436 | 0.8707 | 0.9191 | 0.9917 | nan | 0.8419 | 0.9963 | nan | 0.7499 | 0.9915 |
| 0.0042 | 190.0 | 1900 | 0.0436 | 0.8707 | 0.9210 | 0.9917 | nan | 0.8458 | 0.9961 | nan | 0.7499 | 0.9915 |
| 0.006 | 192.0 | 1920 | 0.0426 | 0.8697 | 0.9193 | 0.9916 | nan | 0.8425 | 0.9962 | nan | 0.7480 | 0.9914 |
| 0.0053 | 194.0 | 1940 | 0.0447 | 0.8697 | 0.9199 | 0.9916 | nan | 0.8437 | 0.9961 | nan | 0.7480 | 0.9914 |
| 0.0044 | 196.0 | 1960 | 0.0441 | 0.8710 | 0.9238 | 0.9916 | nan | 0.8516 | 0.9959 | nan | 0.7505 | 0.9914 |
| 0.0049 | 198.0 | 1980 | 0.0453 | 0.8693 | 0.9219 | 0.9915 | nan | 0.8479 | 0.9959 | nan | 0.7473 | 0.9913 |
| 0.0059 | 200.0 | 2000 | 0.0444 | 0.8726 | 0.9233 | 0.9918 | nan | 0.8506 | 0.9961 | nan | 0.7537 | 0.9916 |
| 0.005 | 202.0 | 2020 | 0.0447 | 0.8717 | 0.9256 | 0.9917 | nan | 0.8555 | 0.9958 | nan | 0.7519 | 0.9914 |
| 0.005 | 204.0 | 2040 | 0.0451 | 0.8711 | 0.9227 | 0.9917 | nan | 0.8494 | 0.9960 | nan | 0.7507 | 0.9915 |
| 0.0043 | 206.0 | 2060 | 0.0458 | 0.8707 | 0.9220 | 0.9916 | nan | 0.8480 | 0.9960 | nan | 0.7499 | 0.9914 |
| 0.0043 | 208.0 | 2080 | 0.0462 | 0.8696 | 0.9221 | 0.9916 | nan | 0.8482 | 0.9959 | nan | 0.7479 | 0.9913 |
| 0.0062 | 210.0 | 2100 | 0.0451 | 0.8715 | 0.9211 | 0.9917 | nan | 0.8461 | 0.9962 | nan | 0.7515 | 0.9915 |
| 0.0056 | 212.0 | 2120 | 0.0470 | 0.8706 | 0.9213 | 0.9917 | nan | 0.8466 | 0.9961 | nan | 0.7498 | 0.9914 |
| 0.0049 | 214.0 | 2140 | 0.0480 | 0.8679 | 0.9229 | 0.9914 | nan | 0.8500 | 0.9957 | nan | 0.7447 | 0.9912 |
| 0.0038 | 216.0 | 2160 | 0.0474 | 0.8700 | 0.9194 | 0.9916 | nan | 0.8427 | 0.9962 | nan | 0.7486 | 0.9914 |
| 0.0043 | 218.0 | 2180 | 0.0472 | 0.8693 | 0.9231 | 0.9915 | nan | 0.8503 | 0.9958 | nan | 0.7474 | 0.9913 |
| 0.005 | 220.0 | 2200 | 0.0471 | 0.8695 | 0.9156 | 0.9917 | nan | 0.8348 | 0.9964 | nan | 0.7475 | 0.9915 |
| 0.0041 | 222.0 | 2220 | 0.0472 | 0.8719 | 0.9187 | 0.9918 | nan | 0.8411 | 0.9964 | nan | 0.7522 | 0.9916 |
| 0.0041 | 224.0 | 2240 | 0.0471 | 0.8720 | 0.9219 | 0.9918 | nan | 0.8477 | 0.9962 | nan | 0.7525 | 0.9916 |
| 0.005 | 226.0 | 2260 | 0.0479 | 0.8720 | 0.9191 | 0.9918 | nan | 0.8418 | 0.9964 | nan | 0.7524 | 0.9916 |
| 0.0041 | 228.0 | 2280 | 0.0489 | 0.8706 | 0.9183 | 0.9917 | nan | 0.8403 | 0.9963 | nan | 0.7498 | 0.9915 |
| 0.005 | 230.0 | 2300 | 0.0472 | 0.8728 | 0.9195 | 0.9919 | nan | 0.8426 | 0.9964 | nan | 0.7540 | 0.9917 |
### Framework versions
- Transformers 4.30.2
- Pytorch 2.0.1+cu117
- Datasets 2.13.1
- Tokenizers 0.13.3
| {"license": "other", "tags": ["vision", "image-segmentation", "generated_from_trainer"], "model-index": [{"name": "INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5", "results": []}]} | image-segmentation | sam1120/INTERNAL_BEST-safety-utcustom-train-SF-RGB-b5 | [
"transformers",
"pytorch",
"tensorboard",
"segformer",
"vision",
"image-segmentation",
"generated_from_trainer",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:49:28+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us
| INTERNAL\_BEST-safety-utcustom-train-SF-RGB-b5
==============================================
This model is a fine-tuned version of nvidia/mit-b5 on the sam1120/safety-utcustom-TRAIN dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0472
* Mean Iou: 0.8728
* Mean Accuracy: 0.9195
* Overall Accuracy: 0.9919
* Accuracy Unlabeled: nan
* Accuracy Safe: 0.8426
* Accuracy Unsafe: 0.9964
* Iou Unlabeled: nan
* Iou Safe: 0.7540
* Iou Unsafe: 0.9917
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.05
* num\_epochs: 2000
### Training results
### Framework versions
* Transformers 4.30.2
* Pytorch 2.0.1+cu117
* Datasets 2.13.1
* Tokenizers 0.13.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 2000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 2000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
48,
116,
4,
33
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 2000### Training results### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
-0.11039795726537704,
0.04744671657681465,
-0.0022476441226899624,
0.11642475426197052,
0.15873928368091583,
0.03221344202756882,
0.11706838011741638,
0.12144872546195984,
-0.10583102703094482,
0.036056146025657654,
0.10768933594226837,
0.13619773089885712,
0.02229778654873371,
0.09303560853004456,
-0.02215495891869068,
-0.31093746423721313,
-0.028257956728339195,
0.03592313453555107,
-0.0803816020488739,
0.12605319917201996,
0.06637299060821533,
-0.16484282910823822,
0.08944474160671234,
-0.0033625192008912563,
-0.2133309692144394,
0.013505863957107067,
-0.004936042707413435,
-0.03460860997438431,
0.15876947343349457,
0.015168293379247189,
0.11607424914836884,
0.009784641675651073,
0.11428932845592499,
-0.19983206689357758,
0.016943177208304405,
0.0571187399327755,
-0.0001794834970496595,
0.0716172605752945,
0.06793700903654099,
-0.001814178191125393,
0.15837696194648743,
-0.10354293882846832,
0.0663604810833931,
-0.0010239859111607075,
-0.14045973122119904,
-0.21935617923736572,
-0.07267621159553528,
0.03184716776013374,
0.0764131098985672,
0.0976744219660759,
-0.003560164012014866,
0.10708186775445938,
-0.08856664597988129,
0.11172434687614441,
0.26642942428588867,
-0.2354210466146469,
-0.08783552795648575,
0.05001935362815857,
-0.00018788778106682003,
0.06498067826032639,
-0.13786634802818298,
0.008686408400535583,
0.02689722552895546,
0.05047990754246712,
0.12038614600896835,
-0.03297187387943268,
-0.10060591995716095,
0.024369625374674797,
-0.13509029150009155,
-0.043476831167936325,
0.0536707267165184,
0.04583155736327171,
-0.020856911316514015,
-0.038935337215662,
-0.06900931149721146,
-0.18948982656002045,
-0.06623955070972443,
0.007696798071265221,
0.06576761603355408,
-0.05963996425271034,
-0.11131437867879868,
-0.015498701483011246,
-0.11570382118225098,
-0.08977333456277847,
-0.05045868828892708,
0.12804701924324036,
0.037543512880802155,
0.02634645812213421,
-0.03757878392934799,
0.13383528590202332,
-0.027294503524899483,
-0.1460777372121811,
0.018934326246380806,
0.030994918197393417,
-0.041953228414058685,
-0.022967787459492683,
-0.05178068205714226,
-0.06815198063850403,
-0.012979269027709961,
0.09855391830205917,
-0.07312262803316116,
0.06382548809051514,
0.03874092176556587,
0.053921520709991455,
-0.11330197006464005,
0.18255245685577393,
-0.07002454996109009,
-0.00958038866519928,
-0.03809827193617821,
0.05514763295650482,
0.0021887896582484245,
-0.01759055070579052,
-0.096625916659832,
-0.001428210991434753,
0.07594917714595795,
-0.00828338973224163,
-0.0809297114610672,
0.07176846265792847,
-0.03241967037320137,
-0.01376025564968586,
-0.013480735011398792,
-0.07766608893871307,
0.044034503400325775,
0.0009604717488400638,
-0.09313987195491791,
-0.028784554451704025,
0.053447872400283813,
0.01212509348988533,
0.005853558890521526,
0.1694091409444809,
-0.08226628601551056,
0.061719734221696854,
-0.11300858110189438,
-0.09128659963607788,
0.0013854372082278132,
-0.08453445881605148,
0.02893376536667347,
-0.07298984378576279,
-0.16190937161445618,
-0.009926621802151203,
0.07707449048757553,
-0.04111609607934952,
0.009327887557446957,
-0.04817066341638565,
-0.08660968393087387,
0.003711612429469824,
-0.011752074584364891,
0.16988709568977356,
-0.06012720614671707,
0.127560555934906,
0.02694893069565296,
0.07496103644371033,
-0.06049863249063492,
0.04179200902581215,
-0.08750265091657639,
0.017294898629188538,
-0.21855494379997253,
0.04203769192099571,
-0.04501528665423393,
0.055102694779634476,
-0.06411315500736237,
-0.12687048316001892,
-0.009794777259230614,
-0.0008681196486577392,
0.09206502884626389,
0.11360625922679901,
-0.23066140711307526,
-0.07384594529867172,
0.14980289340019226,
-0.07754012197256088,
-0.0961260125041008,
0.1123349592089653,
-0.06431455910205841,
0.019794268533587456,
0.0610935315489769,
0.1952345222234726,
0.054146409034729004,
-0.13106416165828705,
0.02034563571214676,
-0.016324391588568687,
0.04718637093901634,
-0.028559867292642593,
0.05026523768901825,
0.019828172400593758,
0.09770243614912033,
0.023446358740329742,
-0.06155926361680031,
0.061296042054891586,
-0.12458213418722153,
-0.10067212581634521,
-0.02884479984641075,
-0.08842127025127411,
0.04155765846371651,
0.09469661861658096,
0.06984256953001022,
-0.09720654040575027,
-0.0768585205078125,
0.06891264021396637,
0.07224402576684952,
-0.07037514448165894,
0.04023918882012367,
-0.05314956605434418,
0.039439573884010315,
-0.013291528448462486,
-0.03367701172828674,
-0.18073177337646484,
-0.01758882775902748,
-0.016976816579699516,
0.02655920200049877,
0.024791939184069633,
0.021945063024759293,
0.09583001583814621,
0.08227185159921646,
-0.07642602175474167,
-0.026746461167931557,
-0.06064150482416153,
-0.0023098145611584187,
-0.12107361108064651,
-0.2229137122631073,
-0.04262123256921768,
-0.0064994958229362965,
0.08323407918214798,
-0.20762985944747925,
0.024949025362730026,
0.018157292157411575,
0.09390529245138168,
0.017895719036459923,
-0.03544824942946434,
-0.05120993033051491,
0.06968764960765839,
-0.01173117384314537,
-0.0633629783987999,
0.0699262022972107,
-0.007956593297421932,
-0.07063284516334534,
-0.054204218089580536,
-0.10104572772979736,
0.1578836441040039,
0.13089947402477264,
-0.1499118208885193,
-0.09444820135831833,
-0.009991341270506382,
-0.07045866549015045,
-0.03737713024020195,
-0.046092208474874496,
0.03845122456550598,
0.18351776897907257,
0.0010273267980664968,
0.12999527156352997,
-0.06191081553697586,
-0.03393072634935379,
0.03107137605547905,
-0.026572642847895622,
0.02832629345357418,
0.1240646168589592,
0.11693868786096573,
-0.05581090971827507,
0.12463817745447159,
0.11490742862224579,
-0.08590645343065262,
0.14578281342983246,
-0.03379332646727562,
-0.08348795026540756,
-0.017608463764190674,
-0.009553159587085247,
-0.010164118371903896,
0.17649802565574646,
-0.15099486708641052,
-0.011772021651268005,
0.00009047713683685288,
0.01603374071419239,
0.020330222323536873,
-0.2510543167591095,
-0.052880480885505676,
0.033667512238025665,
-0.04750606417655945,
-0.0047578574158251286,
-0.018267018720507622,
0.0007847308297641575,
0.10990884155035019,
-0.011743415147066116,
-0.0707528367638588,
0.0006403829902410507,
-0.012067218311131,
-0.052751198410987854,
0.21168175339698792,
-0.05173778906464577,
-0.13159513473510742,
-0.0885889008641243,
-0.06462807208299637,
-0.042105533182621,
0.0021549484226852655,
0.05858109891414642,
-0.10017707943916321,
-0.015306616201996803,
-0.05771737918257713,
0.016568094491958618,
0.004512196406722069,
0.042023904621601105,
-0.007302457466721535,
-0.0072867246344685555,
0.0550902783870697,
-0.09633186459541321,
-0.011082724668085575,
-0.07050994783639908,
-0.05173591524362564,
0.055073924362659454,
0.06347458809614182,
0.1402617245912552,
0.13285328447818756,
-0.025762025266885757,
0.02400490827858448,
-0.032124049961566925,
0.24582955241203308,
-0.09381268173456192,
-0.023579347878694534,
0.1265234500169754,
-0.013277471996843815,
0.05579666048288345,
0.1073918342590332,
0.08331019431352615,
-0.10248076915740967,
-0.008128278888761997,
0.0588558129966259,
-0.054230038076639175,
-0.15866181254386902,
-0.01894269324839115,
-0.059334900230169296,
-0.027850870043039322,
0.08315294981002808,
0.031168291345238686,
0.004031982738524675,
0.05985315144062042,
0.04642793908715248,
0.041795261204242706,
-0.022746169939637184,
0.047721017152071,
0.0799475833773613,
0.02813779003918171,
0.10753782093524933,
-0.04294651374220848,
-0.06815076619386673,
0.032316967844963074,
0.00038642968866042793,
0.2555084228515625,
-0.02128446102142334,
0.08407852053642273,
0.07836655527353287,
0.16113661229610443,
-0.010313188657164574,
0.0528581365942955,
-0.01937446929514408,
-0.06486015021800995,
-0.01570998691022396,
-0.05026211217045784,
-0.02101694792509079,
0.011679807677865028,
-0.04989500343799591,
0.04570799320936203,
-0.12631703913211823,
-0.0011442588875070214,
0.06363855302333832,
0.25248244404792786,
0.02340034209191799,
-0.3153628408908844,
-0.06479998677968979,
-0.00789084006100893,
-0.010065983980894089,
-0.014185087755322456,
0.004779232665896416,
0.15509340167045593,
-0.0851844996213913,
0.052648045122623444,
-0.08915534615516663,
0.08132658153772354,
-0.03822542354464531,
0.047016408294439316,
0.0766066163778305,
0.06381877511739731,
0.00042745255632326007,
0.05666950345039368,
-0.28766730427742004,
0.31313973665237427,
0.0003204481618013233,
0.08762416988611221,
-0.07457481324672699,
-0.027351398020982742,
0.03285415098071098,
0.07256554067134857,
0.0982247143983841,
-0.015890469774603844,
-0.020167596638202667,
-0.20063874125480652,
-0.0239197239279747,
0.027335146442055702,
0.13365782797336578,
-0.019925694912672043,
0.09950661659240723,
-0.003986073657870293,
-0.0015372736379504204,
0.07296212762594223,
0.00020904303528368473,
-0.039361435920000076,
-0.08727344870567322,
-0.018291009590029716,
-0.017106276005506516,
-0.05585135892033577,
-0.06191989779472351,
-0.10959021747112274,
-0.111663818359375,
0.11509302258491516,
0.01673644222319126,
-0.017509063705801964,
-0.12324706465005875,
0.10674988478422165,
0.09161435812711716,
-0.07603662461042404,
0.04534578695893288,
0.027951404452323914,
0.056470900774002075,
0.0336063988506794,
-0.06021305173635483,
0.10610564798116684,
-0.060927823185920715,
-0.1610373854637146,
-0.05853314697742462,
0.0983576774597168,
0.049973923712968826,
0.05564684793353081,
-0.022950604557991028,
0.018485868349671364,
-0.021152082830667496,
-0.09320811927318573,
0.057859867811203,
-0.039624590426683426,
0.07185231149196625,
0.013807416893541813,
-0.025967508554458618,
0.03855977952480316,
-0.050764571875333786,
-0.011904506012797356,
0.14953769743442535,
0.2737651467323303,
-0.09304234385490417,
0.017465349286794662,
0.021391578018665314,
-0.06851085275411606,
-0.19840319454669952,
0.07680352032184601,
0.06565743684768677,
-0.0011194144608452916,
0.08330640941858292,
-0.16989918053150177,
0.09515079855918884,
0.10188671201467514,
-0.004943953361362219,
0.11534547060728073,
-0.3726079761981964,
-0.12947891652584076,
0.08375634998083115,
0.19381697475910187,
0.07198338955640793,
-0.1594688892364502,
-0.0046258000656962395,
0.0010852550622075796,
-0.14283254742622375,
0.08943846076726913,
-0.06327912211418152,
0.13833099603652954,
-0.02771286480128765,
0.0940447449684143,
0.018642539158463478,
-0.06861840933561325,
0.1222112774848938,
-0.002786334604024887,
0.13681331276893616,
-0.06807814538478851,
-0.039261333644390106,
0.0586407445371151,
-0.03934194892644882,
-0.0047933380119502544,
-0.04196548834443092,
0.029967980459332466,
-0.05709642544388771,
-0.011066216044127941,
-0.11041831225156784,
0.01826964132487774,
-0.03854544460773468,
-0.0703996941447258,
-0.04854719340801239,
0.04861472174525261,
0.050289615988731384,
-0.0040628984570503235,
0.14784713089466095,
-0.007251633796840906,
0.11816180497407913,
0.05520064756274223,
0.05474237725138664,
-0.05149982124567032,
-0.11295336484909058,
-0.022892465814948082,
0.010504991747438908,
0.04530028998851776,
-0.1402750015258789,
0.014277891255915165,
0.1556268334388733,
0.053342923521995544,
0.1194651871919632,
0.08736862987279892,
-0.033654600381851196,
0.024498915299773216,
0.07077181339263916,
-0.16369350254535675,
-0.1145637035369873,
0.0009689911967143416,
-0.06781125068664551,
-0.07890193909406662,
0.057507868856191635,
0.07062981277704239,
-0.0786101371049881,
0.016481200233101845,
-0.00809318758547306,
0.01021494995802641,
-0.06635776907205582,
0.21554182469844818,
0.05092667415738106,
0.040452536195516586,
-0.10052235424518585,
0.0811816081404686,
0.019416682422161102,
-0.08774058520793915,
-0.0044161491096019745,
0.08888985961675644,
-0.06762023270130157,
-0.02514037676155567,
0.07541482150554657,
0.18231894075870514,
-0.06616143137216568,
-0.023494165390729904,
-0.15268397331237793,
-0.10547735542058945,
0.07091081887483597,
0.17830653488636017,
0.0985909104347229,
-0.0025373853277415037,
-0.058326270431280136,
0.04474443569779396,
-0.11617304384708405,
0.08369894325733185,
0.029952336102724075,
0.07852999866008759,
-0.1441047042608261,
0.18565654754638672,
0.015129580162465572,
0.051565639674663544,
-0.027344724163413048,
0.029989317059516907,
-0.11526460945606232,
0.044853322207927704,
-0.10266905277967453,
-0.03684595599770546,
-0.015640979632735252,
-0.004554261919111013,
-0.0163115207105875,
-0.0639360249042511,
-0.065290167927742,
0.008323058485984802,
-0.12601767480373383,
-0.021701935678720474,
0.03972410410642624,
0.024704769253730774,
-0.12638932466506958,
-0.040801990777254105,
0.024393746629357338,
-0.061448175460100174,
0.056747689843177795,
0.04286917299032211,
0.010980903171002865,
0.06431291252374649,
-0.18126338720321655,
-0.01691504754126072,
0.06855057924985886,
-0.005633178167045116,
0.05763586610555649,
-0.03690208122134209,
-0.021066179499030113,
-0.024092627689242363,
0.07868292927742004,
0.015181015245616436,
0.05821036919951439,
-0.13905112445354462,
0.006561601534485817,
-0.03390803188085556,
-0.08482298254966736,
-0.06139477714896202,
0.052336081862449646,
0.06355778127908707,
0.04373735561966896,
0.1587870866060257,
-0.0788104310631752,
0.05393371731042862,
-0.22375433146953583,
-0.014254114590585232,
0.0028350637294352055,
-0.10679230093955994,
-0.07875102013349533,
-0.07085923850536346,
0.08447269350290298,
-0.07374872267246246,
0.11508938670158386,
0.04286443814635277,
0.054070476442575455,
0.028167294338345528,
-0.029448868706822395,
0.00537752965465188,
0.030628466978669167,
0.21777144074440002,
0.012064973823726177,
-0.03595554083585739,
0.09052971005439758,
0.08115328103303909,
0.09367207437753677,
0.1371016800403595,
0.2314867228269577,
0.1593104600906372,
-0.020153434947133064,
0.09074719250202179,
0.05274322256445885,
-0.06430623680353165,
-0.16376198828220367,
0.03221311420202255,
-0.04852147400379181,
0.09331445395946503,
-0.06085919961333275,
0.20707738399505615,
0.08215400576591492,
-0.18980321288108826,
0.07292602211236954,
-0.04163398966193199,
-0.09925026446580887,
-0.08032315969467163,
-0.05119328573346138,
-0.06729913502931595,
-0.1385776251554489,
0.025476641952991486,
-0.10144312679767609,
0.04486292600631714,
0.14378704130649567,
0.012379438616335392,
-0.016464857384562492,
0.2148275524377823,
0.0255566518753767,
0.036095958203077316,
0.062022287398576736,
0.022630270570516586,
-0.025178873911499977,
-0.09520312398672104,
-0.059654321521520615,
0.012315134517848492,
-0.025182345882058144,
0.02148197405040264,
-0.0722140520811081,
-0.07633329927921295,
0.022971006110310555,
0.009204497560858727,
-0.08886487782001495,
0.022456923499703407,
0.010802611708641052,
0.09580674022436142,
0.033918000757694244,
0.007072628941386938,
0.017136337235569954,
-0.03245919942855835,
0.24305817484855652,
-0.0934789851307869,
-0.08214910328388214,
-0.07721929252147675,
0.21824295818805695,
0.03131433576345444,
-0.008168509230017662,
0.012119702994823456,
-0.08289958536624908,
0.006131164729595184,
0.2376887947320938,
0.1785605549812317,
-0.12377935647964478,
-0.009784313850104809,
0.00957542285323143,
0.000780968985054642,
-0.026119301095604897,
0.11798909306526184,
0.11952889710664749,
0.07035262137651443,
-0.1136026680469513,
-0.043439075350761414,
-0.0534878708422184,
-0.020917490124702454,
-0.03360297158360481,
0.06520260125398636,
0.06334725767374039,
0.02183368243277073,
-0.07287447154521942,
0.07714451104402542,
-0.06122518330812454,
-0.14801868796348572,
0.11057554930448532,
-0.23360992968082428,
-0.15744543075561523,
-0.005119433160871267,
0.12600183486938477,
0.0011233910918235779,
0.055905986577272415,
-0.04084174335002899,
0.005440291948616505,
0.04504118859767914,
-0.0022483603097498417,
-0.07640732079744339,
-0.10483697801828384,
0.08160489797592163,
-0.1169019341468811,
0.22620800137519836,
-0.059427279978990555,
0.04070629924535751,
0.11164657771587372,
0.06322084367275238,
-0.05326419696211815,
0.05294865742325783,
0.04103056341409683,
-0.12789131700992584,
-0.0065833148546516895,
0.13375353813171387,
-0.03740512952208519,
0.04825790598988533,
0.02966313436627388,
-0.12416685372591019,
0.030161796137690544,
-0.05209925025701523,
-0.04339206963777542,
-0.03176753595471382,
-0.048687104135751724,
-0.059971194714307785,
0.11479593813419342,
0.20588545501232147,
-0.007781834341585636,
0.022810213267803192,
-0.09105495363473892,
0.018899528309702873,
0.06656836718320847,
0.048206478357315063,
-0.07291742414236069,
-0.2203298658132553,
0.010881823487579823,
0.0703558549284935,
-0.04031459987163544,
-0.2001783698797226,
-0.11582236737012863,
0.04157976806163788,
-0.06066508963704109,
-0.06930217146873474,
0.08375546336174011,
0.08766098320484161,
0.058806829154491425,
-0.05177832767367363,
-0.11009946465492249,
-0.05717457830905914,
0.1673433631658554,
-0.14963385462760925,
-0.07691185921430588
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# INTERNAL_BEST-safety-utcustom-train-SF-RGBD-b5
This model is a fine-tuned version of [nvidia/mit-b5](https://huggingface.co/nvidia/mit-b5) on the sam1120/safety-utcustom-TRAIN dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0506
- Mean Iou: 0.8519
- Mean Accuracy: 0.9125
- Overall Accuracy: 0.9902
- Accuracy Unlabeled: nan
- Accuracy Safe: 0.8300
- Accuracy Unsafe: 0.9950
- Iou Unlabeled: nan
- Iou Safe: 0.7138
- Iou Unsafe: 0.9899
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 2000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Accuracy Unlabeled | Accuracy Safe | Accuracy Unsafe | Iou Unlabeled | Iou Safe | Iou Unsafe |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:------------------:|:-------------:|:---------------:|:-------------:|:--------:|:----------:|
| 1.4012 | 2.0 | 20 | 1.1653 | 0.0379 | 0.0923 | 0.0926 | nan | 0.0920 | 0.0926 | 0.0 | 0.0214 | 0.0923 |
| 1.2461 | 4.0 | 40 | 0.9899 | 0.2255 | 0.3670 | 0.6307 | nan | 0.0868 | 0.6473 | 0.0 | 0.0379 | 0.6386 |
| 1.0596 | 6.0 | 60 | 0.7738 | 0.2703 | 0.4188 | 0.7941 | nan | 0.0199 | 0.8177 | 0.0 | 0.0143 | 0.7967 |
| 0.8267 | 8.0 | 80 | 0.6767 | 0.2902 | 0.4512 | 0.8507 | nan | 0.0265 | 0.8758 | 0.0 | 0.0183 | 0.8522 |
| 0.7282 | 10.0 | 100 | 0.5637 | 0.3086 | 0.4776 | 0.9098 | nan | 0.0183 | 0.9370 | 0.0 | 0.0149 | 0.9110 |
| 0.5124 | 12.0 | 120 | 0.4667 | 0.3254 | 0.5053 | 0.9512 | nan | 0.0314 | 0.9792 | 0.0 | 0.0247 | 0.9516 |
| 0.3126 | 14.0 | 140 | 0.3585 | 0.3325 | 0.5147 | 0.9662 | nan | 0.0349 | 0.9945 | 0.0 | 0.0313 | 0.9663 |
| 0.2862 | 16.0 | 160 | 0.2890 | 0.3346 | 0.5168 | 0.9703 | nan | 0.0349 | 0.9988 | 0.0 | 0.0336 | 0.9703 |
| 0.2374 | 18.0 | 180 | 0.2102 | 0.3647 | 0.5637 | 0.9725 | nan | 0.1291 | 0.9982 | 0.0 | 0.1218 | 0.9724 |
| 0.1583 | 20.0 | 200 | 0.1730 | 0.6293 | 0.6574 | 0.9761 | nan | 0.3186 | 0.9961 | nan | 0.2827 | 0.9759 |
| 0.1082 | 22.0 | 220 | 0.1317 | 0.6306 | 0.6566 | 0.9765 | nan | 0.3166 | 0.9966 | nan | 0.2849 | 0.9763 |
| 0.1025 | 24.0 | 240 | 0.1116 | 0.6494 | 0.6766 | 0.9777 | nan | 0.3565 | 0.9967 | nan | 0.3212 | 0.9775 |
| 0.1158 | 26.0 | 260 | 0.0965 | 0.7200 | 0.7978 | 0.9791 | nan | 0.6051 | 0.9905 | nan | 0.4612 | 0.9787 |
| 0.0882 | 28.0 | 280 | 0.0857 | 0.7356 | 0.7857 | 0.9822 | nan | 0.5769 | 0.9946 | nan | 0.4893 | 0.9819 |
| 0.07 | 30.0 | 300 | 0.0829 | 0.6717 | 0.6934 | 0.9799 | nan | 0.3890 | 0.9979 | nan | 0.3637 | 0.9797 |
| 0.0911 | 32.0 | 320 | 0.0677 | 0.7680 | 0.8244 | 0.9843 | nan | 0.6545 | 0.9944 | nan | 0.5521 | 0.9840 |
| 0.0807 | 34.0 | 340 | 0.0696 | 0.7779 | 0.8716 | 0.9834 | nan | 0.7528 | 0.9904 | nan | 0.5727 | 0.9830 |
| 0.0531 | 36.0 | 360 | 0.0611 | 0.7761 | 0.8781 | 0.9829 | nan | 0.7668 | 0.9895 | nan | 0.5698 | 0.9825 |
| 0.0407 | 38.0 | 380 | 0.0567 | 0.7828 | 0.8396 | 0.9854 | nan | 0.6846 | 0.9945 | nan | 0.5805 | 0.9851 |
| 0.0449 | 40.0 | 400 | 0.0639 | 0.7725 | 0.8200 | 0.9851 | nan | 0.6446 | 0.9954 | nan | 0.5602 | 0.9848 |
| 0.0932 | 42.0 | 420 | 0.0503 | 0.7726 | 0.7983 | 0.9861 | nan | 0.5987 | 0.9979 | nan | 0.5593 | 0.9858 |
| 0.0362 | 44.0 | 440 | 0.0634 | 0.7553 | 0.8670 | 0.9805 | nan | 0.7464 | 0.9876 | nan | 0.5306 | 0.9801 |
| 0.0324 | 46.0 | 460 | 0.0501 | 0.8024 | 0.8615 | 0.9867 | nan | 0.7284 | 0.9946 | nan | 0.6184 | 0.9864 |
| 0.036 | 48.0 | 480 | 0.0454 | 0.8010 | 0.8454 | 0.9872 | nan | 0.6947 | 0.9961 | nan | 0.6151 | 0.9869 |
| 0.0356 | 50.0 | 500 | 0.0495 | 0.8061 | 0.8760 | 0.9866 | nan | 0.7585 | 0.9936 | nan | 0.6260 | 0.9863 |
| 0.0333 | 52.0 | 520 | 0.0483 | 0.7743 | 0.8128 | 0.9856 | nan | 0.6292 | 0.9964 | nan | 0.5632 | 0.9853 |
| 0.0277 | 54.0 | 540 | 0.0445 | 0.7714 | 0.7932 | 0.9862 | nan | 0.5880 | 0.9983 | nan | 0.5569 | 0.9859 |
| 0.0298 | 56.0 | 560 | 0.0460 | 0.8034 | 0.8518 | 0.9872 | nan | 0.7078 | 0.9957 | nan | 0.6198 | 0.9869 |
| 0.0256 | 58.0 | 580 | 0.0416 | 0.8181 | 0.8548 | 0.9886 | nan | 0.7126 | 0.9970 | nan | 0.6479 | 0.9883 |
| 0.0336 | 60.0 | 600 | 0.0442 | 0.7957 | 0.8168 | 0.9877 | nan | 0.6351 | 0.9984 | nan | 0.6039 | 0.9875 |
| 0.0283 | 62.0 | 620 | 0.0425 | 0.8141 | 0.8812 | 0.9873 | nan | 0.7684 | 0.9940 | nan | 0.6413 | 0.9870 |
| 0.0198 | 64.0 | 640 | 0.0455 | 0.8059 | 0.8401 | 0.9879 | nan | 0.6830 | 0.9971 | nan | 0.6242 | 0.9876 |
| 0.0181 | 66.0 | 660 | 0.0444 | 0.8144 | 0.8733 | 0.9876 | nan | 0.7519 | 0.9948 | nan | 0.6415 | 0.9873 |
| 0.0188 | 68.0 | 680 | 0.0456 | 0.8179 | 0.8696 | 0.9881 | nan | 0.7436 | 0.9955 | nan | 0.6479 | 0.9878 |
| 0.0165 | 70.0 | 700 | 0.0431 | 0.8208 | 0.8985 | 0.9875 | nan | 0.8040 | 0.9930 | nan | 0.6544 | 0.9872 |
| 0.0184 | 72.0 | 720 | 0.0421 | 0.8165 | 0.8785 | 0.9876 | nan | 0.7625 | 0.9945 | nan | 0.6457 | 0.9874 |
| 0.0336 | 74.0 | 740 | 0.0441 | 0.8081 | 0.8792 | 0.9867 | nan | 0.7650 | 0.9935 | nan | 0.6298 | 0.9864 |
| 0.0165 | 76.0 | 760 | 0.0374 | 0.8200 | 0.8555 | 0.9887 | nan | 0.7139 | 0.9971 | nan | 0.6515 | 0.9885 |
| 0.0127 | 78.0 | 780 | 0.0402 | 0.8222 | 0.8780 | 0.9882 | nan | 0.7608 | 0.9952 | nan | 0.6563 | 0.9880 |
| 0.0152 | 80.0 | 800 | 0.0430 | 0.8230 | 0.8687 | 0.9886 | nan | 0.7413 | 0.9961 | nan | 0.6576 | 0.9883 |
| 0.0143 | 82.0 | 820 | 0.0410 | 0.8087 | 0.8422 | 0.9881 | nan | 0.6873 | 0.9972 | nan | 0.6297 | 0.9878 |
| 0.0134 | 84.0 | 840 | 0.0335 | 0.8429 | 0.8893 | 0.9899 | nan | 0.7823 | 0.9962 | nan | 0.6962 | 0.9897 |
| 0.0122 | 86.0 | 860 | 0.0396 | 0.8312 | 0.8749 | 0.9892 | nan | 0.7534 | 0.9964 | nan | 0.6734 | 0.9890 |
| 0.0126 | 88.0 | 880 | 0.0405 | 0.8341 | 0.8805 | 0.9893 | nan | 0.7649 | 0.9962 | nan | 0.6791 | 0.9891 |
| 0.0121 | 90.0 | 900 | 0.0400 | 0.8390 | 0.8810 | 0.9898 | nan | 0.7654 | 0.9966 | nan | 0.6884 | 0.9895 |
| 0.0104 | 92.0 | 920 | 0.0372 | 0.8453 | 0.8990 | 0.9899 | nan | 0.8024 | 0.9956 | nan | 0.7010 | 0.9896 |
| 0.0128 | 94.0 | 940 | 0.0394 | 0.8411 | 0.8893 | 0.9897 | nan | 0.7825 | 0.9961 | nan | 0.6927 | 0.9895 |
| 0.0124 | 96.0 | 960 | 0.0409 | 0.8395 | 0.8948 | 0.9895 | nan | 0.7943 | 0.9954 | nan | 0.6899 | 0.9892 |
| 0.0095 | 98.0 | 980 | 0.0413 | 0.8258 | 0.8903 | 0.9882 | nan | 0.7863 | 0.9944 | nan | 0.6637 | 0.9880 |
| 0.0147 | 100.0 | 1000 | 0.0468 | 0.8181 | 0.9044 | 0.9870 | nan | 0.8167 | 0.9922 | nan | 0.6496 | 0.9867 |
| 0.0125 | 102.0 | 1020 | 0.0379 | 0.8213 | 0.8961 | 0.9876 | nan | 0.7989 | 0.9933 | nan | 0.6553 | 0.9873 |
| 0.0142 | 104.0 | 1040 | 0.0328 | 0.8449 | 0.9154 | 0.9894 | nan | 0.8366 | 0.9941 | nan | 0.7006 | 0.9892 |
| 0.0101 | 106.0 | 1060 | 0.0428 | 0.8407 | 0.9144 | 0.9891 | nan | 0.8351 | 0.9937 | nan | 0.6927 | 0.9888 |
| 0.0097 | 108.0 | 1080 | 0.0397 | 0.8296 | 0.8847 | 0.9888 | nan | 0.7740 | 0.9953 | nan | 0.6707 | 0.9885 |
| 0.01 | 110.0 | 1100 | 0.0384 | 0.8457 | 0.8935 | 0.9901 | nan | 0.7910 | 0.9961 | nan | 0.7016 | 0.9898 |
| 0.0084 | 112.0 | 1120 | 0.0385 | 0.8421 | 0.8874 | 0.9899 | nan | 0.7784 | 0.9963 | nan | 0.6945 | 0.9896 |
| 0.0086 | 114.0 | 1140 | 0.0413 | 0.8488 | 0.8882 | 0.9905 | nan | 0.7795 | 0.9969 | nan | 0.7074 | 0.9903 |
| 0.0112 | 116.0 | 1160 | 0.0427 | 0.8459 | 0.8942 | 0.9901 | nan | 0.7924 | 0.9961 | nan | 0.7020 | 0.9898 |
| 0.0132 | 118.0 | 1180 | 0.0407 | 0.8510 | 0.9011 | 0.9904 | nan | 0.8062 | 0.9960 | nan | 0.7118 | 0.9901 |
| 0.0084 | 120.0 | 1200 | 0.0432 | 0.8510 | 0.9015 | 0.9903 | nan | 0.8071 | 0.9959 | nan | 0.7118 | 0.9901 |
| 0.008 | 122.0 | 1220 | 0.0431 | 0.8504 | 0.9077 | 0.9901 | nan | 0.8202 | 0.9953 | nan | 0.7109 | 0.9899 |
| 0.0069 | 124.0 | 1240 | 0.0424 | 0.8522 | 0.8982 | 0.9905 | nan | 0.8001 | 0.9963 | nan | 0.7141 | 0.9903 |
| 0.006 | 126.0 | 1260 | 0.0447 | 0.8537 | 0.9114 | 0.9904 | nan | 0.8275 | 0.9953 | nan | 0.7173 | 0.9901 |
| 0.0123 | 128.0 | 1280 | 0.0464 | 0.8529 | 0.9102 | 0.9903 | nan | 0.8250 | 0.9954 | nan | 0.7157 | 0.9901 |
| 0.0073 | 130.0 | 1300 | 0.0441 | 0.8520 | 0.9025 | 0.9904 | nan | 0.8090 | 0.9959 | nan | 0.7139 | 0.9902 |
| 0.0066 | 132.0 | 1320 | 0.0447 | 0.8524 | 0.9086 | 0.9903 | nan | 0.8217 | 0.9954 | nan | 0.7148 | 0.9901 |
| 0.0063 | 134.0 | 1340 | 0.0434 | 0.8546 | 0.9077 | 0.9905 | nan | 0.8197 | 0.9957 | nan | 0.7189 | 0.9903 |
| 0.0068 | 136.0 | 1360 | 0.0475 | 0.8518 | 0.9090 | 0.9902 | nan | 0.8226 | 0.9953 | nan | 0.7135 | 0.9900 |
| 0.0056 | 138.0 | 1380 | 0.0458 | 0.8549 | 0.9122 | 0.9905 | nan | 0.8291 | 0.9954 | nan | 0.7195 | 0.9902 |
| 0.007 | 140.0 | 1400 | 0.0455 | 0.8554 | 0.9126 | 0.9905 | nan | 0.8298 | 0.9954 | nan | 0.7205 | 0.9903 |
| 0.0064 | 142.0 | 1420 | 0.0476 | 0.8542 | 0.9047 | 0.9906 | nan | 0.8133 | 0.9960 | nan | 0.7180 | 0.9903 |
| 0.0065 | 144.0 | 1440 | 0.0437 | 0.8556 | 0.9107 | 0.9906 | nan | 0.8258 | 0.9956 | nan | 0.7210 | 0.9903 |
| 0.005 | 146.0 | 1460 | 0.0455 | 0.8551 | 0.9098 | 0.9905 | nan | 0.8239 | 0.9956 | nan | 0.7198 | 0.9903 |
| 0.005 | 148.0 | 1480 | 0.0458 | 0.8539 | 0.9084 | 0.9905 | nan | 0.8212 | 0.9956 | nan | 0.7175 | 0.9902 |
| 0.0048 | 150.0 | 1500 | 0.0462 | 0.8558 | 0.9041 | 0.9907 | nan | 0.8121 | 0.9962 | nan | 0.7211 | 0.9905 |
| 0.0063 | 152.0 | 1520 | 0.0453 | 0.8560 | 0.9175 | 0.9904 | nan | 0.8400 | 0.9950 | nan | 0.7217 | 0.9902 |
| 0.006 | 154.0 | 1540 | 0.0473 | 0.8531 | 0.9073 | 0.9904 | nan | 0.8190 | 0.9956 | nan | 0.7160 | 0.9902 |
| 0.0043 | 156.0 | 1560 | 0.0448 | 0.8562 | 0.9100 | 0.9906 | nan | 0.8243 | 0.9957 | nan | 0.7220 | 0.9904 |
| 0.0049 | 158.0 | 1580 | 0.0480 | 0.8518 | 0.9137 | 0.9901 | nan | 0.8324 | 0.9949 | nan | 0.7138 | 0.9899 |
| 0.0065 | 160.0 | 1600 | 0.0475 | 0.8556 | 0.9095 | 0.9906 | nan | 0.8233 | 0.9957 | nan | 0.7209 | 0.9903 |
| 0.0052 | 162.0 | 1620 | 0.0479 | 0.8531 | 0.9087 | 0.9904 | nan | 0.8218 | 0.9955 | nan | 0.7161 | 0.9901 |
| 0.0063 | 164.0 | 1640 | 0.0488 | 0.8571 | 0.9115 | 0.9907 | nan | 0.8273 | 0.9956 | nan | 0.7238 | 0.9904 |
| 0.0053 | 166.0 | 1660 | 0.0514 | 0.8515 | 0.9152 | 0.9901 | nan | 0.8357 | 0.9948 | nan | 0.7132 | 0.9898 |
| 0.0046 | 168.0 | 1680 | 0.0476 | 0.8540 | 0.9040 | 0.9906 | nan | 0.8119 | 0.9960 | nan | 0.7177 | 0.9903 |
| 0.0039 | 170.0 | 1700 | 0.0483 | 0.5699 | 0.9121 | 0.9905 | nan | 0.8289 | 0.9954 | 0.0 | 0.7195 | 0.9902 |
| 0.0044 | 172.0 | 1720 | 0.0494 | 0.8550 | 0.9114 | 0.9905 | nan | 0.8273 | 0.9954 | nan | 0.7197 | 0.9902 |
| 0.0051 | 174.0 | 1740 | 0.0503 | 0.8556 | 0.9103 | 0.9906 | nan | 0.8250 | 0.9956 | nan | 0.7208 | 0.9903 |
| 0.0041 | 176.0 | 1760 | 0.0499 | 0.8545 | 0.9118 | 0.9904 | nan | 0.8283 | 0.9954 | nan | 0.7188 | 0.9902 |
| 0.0049 | 178.0 | 1780 | 0.0525 | 0.8541 | 0.9066 | 0.9905 | nan | 0.8174 | 0.9958 | nan | 0.7179 | 0.9903 |
| 0.0048 | 180.0 | 1800 | 0.0496 | 0.8556 | 0.9165 | 0.9904 | nan | 0.8380 | 0.9951 | nan | 0.7210 | 0.9902 |
| 0.008 | 182.0 | 1820 | 0.0487 | 0.8528 | 0.9085 | 0.9904 | nan | 0.8215 | 0.9955 | nan | 0.7155 | 0.9901 |
| 0.0041 | 184.0 | 1840 | 0.0506 | 0.8519 | 0.9125 | 0.9902 | nan | 0.8300 | 0.9950 | nan | 0.7138 | 0.9899 |
### Framework versions
- Transformers 4.30.2
- Pytorch 2.0.1+cu117
- Datasets 2.13.1
- Tokenizers 0.13.3
| {"license": "other", "tags": ["vision", "image-segmentation", "generated_from_trainer"], "model-index": [{"name": "INTERNAL_BEST-safety-utcustom-train-SF-RGBD-b5", "results": []}]} | image-segmentation | sam1120/INTERNAL_BEST-safety-utcustom-train-SF-RGBD-b5 | [
"transformers",
"pytorch",
"tensorboard",
"segformer",
"vision",
"image-segmentation",
"generated_from_trainer",
"license:other",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:52:07+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us
| INTERNAL\_BEST-safety-utcustom-train-SF-RGBD-b5
===============================================
This model is a fine-tuned version of nvidia/mit-b5 on the sam1120/safety-utcustom-TRAIN dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0506
* Mean Iou: 0.8519
* Mean Accuracy: 0.9125
* Overall Accuracy: 0.9902
* Accuracy Unlabeled: nan
* Accuracy Safe: 0.8300
* Accuracy Unsafe: 0.9950
* Iou Unlabeled: nan
* Iou Safe: 0.7138
* Iou Unsafe: 0.9899
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.05
* num\_epochs: 2000
### Training results
### Framework versions
* Transformers 4.30.2
* Pytorch 2.0.1+cu117
* Datasets 2.13.1
* Tokenizers 0.13.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 2000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 2000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
48,
116,
4,
33
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #segformer #vision #image-segmentation #generated_from_trainer #license-other #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.05\n* num\\_epochs: 2000### Training results### Framework versions\n\n\n* Transformers 4.30.2\n* Pytorch 2.0.1+cu117\n* Datasets 2.13.1\n* Tokenizers 0.13.3"
] | [
-0.11039795726537704,
0.04744671657681465,
-0.0022476441226899624,
0.11642475426197052,
0.15873928368091583,
0.03221344202756882,
0.11706838011741638,
0.12144872546195984,
-0.10583102703094482,
0.036056146025657654,
0.10768933594226837,
0.13619773089885712,
0.02229778654873371,
0.09303560853004456,
-0.02215495891869068,
-0.31093746423721313,
-0.028257956728339195,
0.03592313453555107,
-0.0803816020488739,
0.12605319917201996,
0.06637299060821533,
-0.16484282910823822,
0.08944474160671234,
-0.0033625192008912563,
-0.2133309692144394,
0.013505863957107067,
-0.004936042707413435,
-0.03460860997438431,
0.15876947343349457,
0.015168293379247189,
0.11607424914836884,
0.009784641675651073,
0.11428932845592499,
-0.19983206689357758,
0.016943177208304405,
0.0571187399327755,
-0.0001794834970496595,
0.0716172605752945,
0.06793700903654099,
-0.001814178191125393,
0.15837696194648743,
-0.10354293882846832,
0.0663604810833931,
-0.0010239859111607075,
-0.14045973122119904,
-0.21935617923736572,
-0.07267621159553528,
0.03184716776013374,
0.0764131098985672,
0.0976744219660759,
-0.003560164012014866,
0.10708186775445938,
-0.08856664597988129,
0.11172434687614441,
0.26642942428588867,
-0.2354210466146469,
-0.08783552795648575,
0.05001935362815857,
-0.00018788778106682003,
0.06498067826032639,
-0.13786634802818298,
0.008686408400535583,
0.02689722552895546,
0.05047990754246712,
0.12038614600896835,
-0.03297187387943268,
-0.10060591995716095,
0.024369625374674797,
-0.13509029150009155,
-0.043476831167936325,
0.0536707267165184,
0.04583155736327171,
-0.020856911316514015,
-0.038935337215662,
-0.06900931149721146,
-0.18948982656002045,
-0.06623955070972443,
0.007696798071265221,
0.06576761603355408,
-0.05963996425271034,
-0.11131437867879868,
-0.015498701483011246,
-0.11570382118225098,
-0.08977333456277847,
-0.05045868828892708,
0.12804701924324036,
0.037543512880802155,
0.02634645812213421,
-0.03757878392934799,
0.13383528590202332,
-0.027294503524899483,
-0.1460777372121811,
0.018934326246380806,
0.030994918197393417,
-0.041953228414058685,
-0.022967787459492683,
-0.05178068205714226,
-0.06815198063850403,
-0.012979269027709961,
0.09855391830205917,
-0.07312262803316116,
0.06382548809051514,
0.03874092176556587,
0.053921520709991455,
-0.11330197006464005,
0.18255245685577393,
-0.07002454996109009,
-0.00958038866519928,
-0.03809827193617821,
0.05514763295650482,
0.0021887896582484245,
-0.01759055070579052,
-0.096625916659832,
-0.001428210991434753,
0.07594917714595795,
-0.00828338973224163,
-0.0809297114610672,
0.07176846265792847,
-0.03241967037320137,
-0.01376025564968586,
-0.013480735011398792,
-0.07766608893871307,
0.044034503400325775,
0.0009604717488400638,
-0.09313987195491791,
-0.028784554451704025,
0.053447872400283813,
0.01212509348988533,
0.005853558890521526,
0.1694091409444809,
-0.08226628601551056,
0.061719734221696854,
-0.11300858110189438,
-0.09128659963607788,
0.0013854372082278132,
-0.08453445881605148,
0.02893376536667347,
-0.07298984378576279,
-0.16190937161445618,
-0.009926621802151203,
0.07707449048757553,
-0.04111609607934952,
0.009327887557446957,
-0.04817066341638565,
-0.08660968393087387,
0.003711612429469824,
-0.011752074584364891,
0.16988709568977356,
-0.06012720614671707,
0.127560555934906,
0.02694893069565296,
0.07496103644371033,
-0.06049863249063492,
0.04179200902581215,
-0.08750265091657639,
0.017294898629188538,
-0.21855494379997253,
0.04203769192099571,
-0.04501528665423393,
0.055102694779634476,
-0.06411315500736237,
-0.12687048316001892,
-0.009794777259230614,
-0.0008681196486577392,
0.09206502884626389,
0.11360625922679901,
-0.23066140711307526,
-0.07384594529867172,
0.14980289340019226,
-0.07754012197256088,
-0.0961260125041008,
0.1123349592089653,
-0.06431455910205841,
0.019794268533587456,
0.0610935315489769,
0.1952345222234726,
0.054146409034729004,
-0.13106416165828705,
0.02034563571214676,
-0.016324391588568687,
0.04718637093901634,
-0.028559867292642593,
0.05026523768901825,
0.019828172400593758,
0.09770243614912033,
0.023446358740329742,
-0.06155926361680031,
0.061296042054891586,
-0.12458213418722153,
-0.10067212581634521,
-0.02884479984641075,
-0.08842127025127411,
0.04155765846371651,
0.09469661861658096,
0.06984256953001022,
-0.09720654040575027,
-0.0768585205078125,
0.06891264021396637,
0.07224402576684952,
-0.07037514448165894,
0.04023918882012367,
-0.05314956605434418,
0.039439573884010315,
-0.013291528448462486,
-0.03367701172828674,
-0.18073177337646484,
-0.01758882775902748,
-0.016976816579699516,
0.02655920200049877,
0.024791939184069633,
0.021945063024759293,
0.09583001583814621,
0.08227185159921646,
-0.07642602175474167,
-0.026746461167931557,
-0.06064150482416153,
-0.0023098145611584187,
-0.12107361108064651,
-0.2229137122631073,
-0.04262123256921768,
-0.0064994958229362965,
0.08323407918214798,
-0.20762985944747925,
0.024949025362730026,
0.018157292157411575,
0.09390529245138168,
0.017895719036459923,
-0.03544824942946434,
-0.05120993033051491,
0.06968764960765839,
-0.01173117384314537,
-0.0633629783987999,
0.0699262022972107,
-0.007956593297421932,
-0.07063284516334534,
-0.054204218089580536,
-0.10104572772979736,
0.1578836441040039,
0.13089947402477264,
-0.1499118208885193,
-0.09444820135831833,
-0.009991341270506382,
-0.07045866549015045,
-0.03737713024020195,
-0.046092208474874496,
0.03845122456550598,
0.18351776897907257,
0.0010273267980664968,
0.12999527156352997,
-0.06191081553697586,
-0.03393072634935379,
0.03107137605547905,
-0.026572642847895622,
0.02832629345357418,
0.1240646168589592,
0.11693868786096573,
-0.05581090971827507,
0.12463817745447159,
0.11490742862224579,
-0.08590645343065262,
0.14578281342983246,
-0.03379332646727562,
-0.08348795026540756,
-0.017608463764190674,
-0.009553159587085247,
-0.010164118371903896,
0.17649802565574646,
-0.15099486708641052,
-0.011772021651268005,
0.00009047713683685288,
0.01603374071419239,
0.020330222323536873,
-0.2510543167591095,
-0.052880480885505676,
0.033667512238025665,
-0.04750606417655945,
-0.0047578574158251286,
-0.018267018720507622,
0.0007847308297641575,
0.10990884155035019,
-0.011743415147066116,
-0.0707528367638588,
0.0006403829902410507,
-0.012067218311131,
-0.052751198410987854,
0.21168175339698792,
-0.05173778906464577,
-0.13159513473510742,
-0.0885889008641243,
-0.06462807208299637,
-0.042105533182621,
0.0021549484226852655,
0.05858109891414642,
-0.10017707943916321,
-0.015306616201996803,
-0.05771737918257713,
0.016568094491958618,
0.004512196406722069,
0.042023904621601105,
-0.007302457466721535,
-0.0072867246344685555,
0.0550902783870697,
-0.09633186459541321,
-0.011082724668085575,
-0.07050994783639908,
-0.05173591524362564,
0.055073924362659454,
0.06347458809614182,
0.1402617245912552,
0.13285328447818756,
-0.025762025266885757,
0.02400490827858448,
-0.032124049961566925,
0.24582955241203308,
-0.09381268173456192,
-0.023579347878694534,
0.1265234500169754,
-0.013277471996843815,
0.05579666048288345,
0.1073918342590332,
0.08331019431352615,
-0.10248076915740967,
-0.008128278888761997,
0.0588558129966259,
-0.054230038076639175,
-0.15866181254386902,
-0.01894269324839115,
-0.059334900230169296,
-0.027850870043039322,
0.08315294981002808,
0.031168291345238686,
0.004031982738524675,
0.05985315144062042,
0.04642793908715248,
0.041795261204242706,
-0.022746169939637184,
0.047721017152071,
0.0799475833773613,
0.02813779003918171,
0.10753782093524933,
-0.04294651374220848,
-0.06815076619386673,
0.032316967844963074,
0.00038642968866042793,
0.2555084228515625,
-0.02128446102142334,
0.08407852053642273,
0.07836655527353287,
0.16113661229610443,
-0.010313188657164574,
0.0528581365942955,
-0.01937446929514408,
-0.06486015021800995,
-0.01570998691022396,
-0.05026211217045784,
-0.02101694792509079,
0.011679807677865028,
-0.04989500343799591,
0.04570799320936203,
-0.12631703913211823,
-0.0011442588875070214,
0.06363855302333832,
0.25248244404792786,
0.02340034209191799,
-0.3153628408908844,
-0.06479998677968979,
-0.00789084006100893,
-0.010065983980894089,
-0.014185087755322456,
0.004779232665896416,
0.15509340167045593,
-0.0851844996213913,
0.052648045122623444,
-0.08915534615516663,
0.08132658153772354,
-0.03822542354464531,
0.047016408294439316,
0.0766066163778305,
0.06381877511739731,
0.00042745255632326007,
0.05666950345039368,
-0.28766730427742004,
0.31313973665237427,
0.0003204481618013233,
0.08762416988611221,
-0.07457481324672699,
-0.027351398020982742,
0.03285415098071098,
0.07256554067134857,
0.0982247143983841,
-0.015890469774603844,
-0.020167596638202667,
-0.20063874125480652,
-0.0239197239279747,
0.027335146442055702,
0.13365782797336578,
-0.019925694912672043,
0.09950661659240723,
-0.003986073657870293,
-0.0015372736379504204,
0.07296212762594223,
0.00020904303528368473,
-0.039361435920000076,
-0.08727344870567322,
-0.018291009590029716,
-0.017106276005506516,
-0.05585135892033577,
-0.06191989779472351,
-0.10959021747112274,
-0.111663818359375,
0.11509302258491516,
0.01673644222319126,
-0.017509063705801964,
-0.12324706465005875,
0.10674988478422165,
0.09161435812711716,
-0.07603662461042404,
0.04534578695893288,
0.027951404452323914,
0.056470900774002075,
0.0336063988506794,
-0.06021305173635483,
0.10610564798116684,
-0.060927823185920715,
-0.1610373854637146,
-0.05853314697742462,
0.0983576774597168,
0.049973923712968826,
0.05564684793353081,
-0.022950604557991028,
0.018485868349671364,
-0.021152082830667496,
-0.09320811927318573,
0.057859867811203,
-0.039624590426683426,
0.07185231149196625,
0.013807416893541813,
-0.025967508554458618,
0.03855977952480316,
-0.050764571875333786,
-0.011904506012797356,
0.14953769743442535,
0.2737651467323303,
-0.09304234385490417,
0.017465349286794662,
0.021391578018665314,
-0.06851085275411606,
-0.19840319454669952,
0.07680352032184601,
0.06565743684768677,
-0.0011194144608452916,
0.08330640941858292,
-0.16989918053150177,
0.09515079855918884,
0.10188671201467514,
-0.004943953361362219,
0.11534547060728073,
-0.3726079761981964,
-0.12947891652584076,
0.08375634998083115,
0.19381697475910187,
0.07198338955640793,
-0.1594688892364502,
-0.0046258000656962395,
0.0010852550622075796,
-0.14283254742622375,
0.08943846076726913,
-0.06327912211418152,
0.13833099603652954,
-0.02771286480128765,
0.0940447449684143,
0.018642539158463478,
-0.06861840933561325,
0.1222112774848938,
-0.002786334604024887,
0.13681331276893616,
-0.06807814538478851,
-0.039261333644390106,
0.0586407445371151,
-0.03934194892644882,
-0.0047933380119502544,
-0.04196548834443092,
0.029967980459332466,
-0.05709642544388771,
-0.011066216044127941,
-0.11041831225156784,
0.01826964132487774,
-0.03854544460773468,
-0.0703996941447258,
-0.04854719340801239,
0.04861472174525261,
0.050289615988731384,
-0.0040628984570503235,
0.14784713089466095,
-0.007251633796840906,
0.11816180497407913,
0.05520064756274223,
0.05474237725138664,
-0.05149982124567032,
-0.11295336484909058,
-0.022892465814948082,
0.010504991747438908,
0.04530028998851776,
-0.1402750015258789,
0.014277891255915165,
0.1556268334388733,
0.053342923521995544,
0.1194651871919632,
0.08736862987279892,
-0.033654600381851196,
0.024498915299773216,
0.07077181339263916,
-0.16369350254535675,
-0.1145637035369873,
0.0009689911967143416,
-0.06781125068664551,
-0.07890193909406662,
0.057507868856191635,
0.07062981277704239,
-0.0786101371049881,
0.016481200233101845,
-0.00809318758547306,
0.01021494995802641,
-0.06635776907205582,
0.21554182469844818,
0.05092667415738106,
0.040452536195516586,
-0.10052235424518585,
0.0811816081404686,
0.019416682422161102,
-0.08774058520793915,
-0.0044161491096019745,
0.08888985961675644,
-0.06762023270130157,
-0.02514037676155567,
0.07541482150554657,
0.18231894075870514,
-0.06616143137216568,
-0.023494165390729904,
-0.15268397331237793,
-0.10547735542058945,
0.07091081887483597,
0.17830653488636017,
0.0985909104347229,
-0.0025373853277415037,
-0.058326270431280136,
0.04474443569779396,
-0.11617304384708405,
0.08369894325733185,
0.029952336102724075,
0.07852999866008759,
-0.1441047042608261,
0.18565654754638672,
0.015129580162465572,
0.051565639674663544,
-0.027344724163413048,
0.029989317059516907,
-0.11526460945606232,
0.044853322207927704,
-0.10266905277967453,
-0.03684595599770546,
-0.015640979632735252,
-0.004554261919111013,
-0.0163115207105875,
-0.0639360249042511,
-0.065290167927742,
0.008323058485984802,
-0.12601767480373383,
-0.021701935678720474,
0.03972410410642624,
0.024704769253730774,
-0.12638932466506958,
-0.040801990777254105,
0.024393746629357338,
-0.061448175460100174,
0.056747689843177795,
0.04286917299032211,
0.010980903171002865,
0.06431291252374649,
-0.18126338720321655,
-0.01691504754126072,
0.06855057924985886,
-0.005633178167045116,
0.05763586610555649,
-0.03690208122134209,
-0.021066179499030113,
-0.024092627689242363,
0.07868292927742004,
0.015181015245616436,
0.05821036919951439,
-0.13905112445354462,
0.006561601534485817,
-0.03390803188085556,
-0.08482298254966736,
-0.06139477714896202,
0.052336081862449646,
0.06355778127908707,
0.04373735561966896,
0.1587870866060257,
-0.0788104310631752,
0.05393371731042862,
-0.22375433146953583,
-0.014254114590585232,
0.0028350637294352055,
-0.10679230093955994,
-0.07875102013349533,
-0.07085923850536346,
0.08447269350290298,
-0.07374872267246246,
0.11508938670158386,
0.04286443814635277,
0.054070476442575455,
0.028167294338345528,
-0.029448868706822395,
0.00537752965465188,
0.030628466978669167,
0.21777144074440002,
0.012064973823726177,
-0.03595554083585739,
0.09052971005439758,
0.08115328103303909,
0.09367207437753677,
0.1371016800403595,
0.2314867228269577,
0.1593104600906372,
-0.020153434947133064,
0.09074719250202179,
0.05274322256445885,
-0.06430623680353165,
-0.16376198828220367,
0.03221311420202255,
-0.04852147400379181,
0.09331445395946503,
-0.06085919961333275,
0.20707738399505615,
0.08215400576591492,
-0.18980321288108826,
0.07292602211236954,
-0.04163398966193199,
-0.09925026446580887,
-0.08032315969467163,
-0.05119328573346138,
-0.06729913502931595,
-0.1385776251554489,
0.025476641952991486,
-0.10144312679767609,
0.04486292600631714,
0.14378704130649567,
0.012379438616335392,
-0.016464857384562492,
0.2148275524377823,
0.0255566518753767,
0.036095958203077316,
0.062022287398576736,
0.022630270570516586,
-0.025178873911499977,
-0.09520312398672104,
-0.059654321521520615,
0.012315134517848492,
-0.025182345882058144,
0.02148197405040264,
-0.0722140520811081,
-0.07633329927921295,
0.022971006110310555,
0.009204497560858727,
-0.08886487782001495,
0.022456923499703407,
0.010802611708641052,
0.09580674022436142,
0.033918000757694244,
0.007072628941386938,
0.017136337235569954,
-0.03245919942855835,
0.24305817484855652,
-0.0934789851307869,
-0.08214910328388214,
-0.07721929252147675,
0.21824295818805695,
0.03131433576345444,
-0.008168509230017662,
0.012119702994823456,
-0.08289958536624908,
0.006131164729595184,
0.2376887947320938,
0.1785605549812317,
-0.12377935647964478,
-0.009784313850104809,
0.00957542285323143,
0.000780968985054642,
-0.026119301095604897,
0.11798909306526184,
0.11952889710664749,
0.07035262137651443,
-0.1136026680469513,
-0.043439075350761414,
-0.0534878708422184,
-0.020917490124702454,
-0.03360297158360481,
0.06520260125398636,
0.06334725767374039,
0.02183368243277073,
-0.07287447154521942,
0.07714451104402542,
-0.06122518330812454,
-0.14801868796348572,
0.11057554930448532,
-0.23360992968082428,
-0.15744543075561523,
-0.005119433160871267,
0.12600183486938477,
0.0011233910918235779,
0.055905986577272415,
-0.04084174335002899,
0.005440291948616505,
0.04504118859767914,
-0.0022483603097498417,
-0.07640732079744339,
-0.10483697801828384,
0.08160489797592163,
-0.1169019341468811,
0.22620800137519836,
-0.059427279978990555,
0.04070629924535751,
0.11164657771587372,
0.06322084367275238,
-0.05326419696211815,
0.05294865742325783,
0.04103056341409683,
-0.12789131700992584,
-0.0065833148546516895,
0.13375353813171387,
-0.03740512952208519,
0.04825790598988533,
0.02966313436627388,
-0.12416685372591019,
0.030161796137690544,
-0.05209925025701523,
-0.04339206963777542,
-0.03176753595471382,
-0.048687104135751724,
-0.059971194714307785,
0.11479593813419342,
0.20588545501232147,
-0.007781834341585636,
0.022810213267803192,
-0.09105495363473892,
0.018899528309702873,
0.06656836718320847,
0.048206478357315063,
-0.07291742414236069,
-0.2203298658132553,
0.010881823487579823,
0.0703558549284935,
-0.04031459987163544,
-0.2001783698797226,
-0.11582236737012863,
0.04157976806163788,
-0.06066508963704109,
-0.06930217146873474,
0.08375546336174011,
0.08766098320484161,
0.058806829154491425,
-0.05177832767367363,
-0.11009946465492249,
-0.05717457830905914,
0.1673433631658554,
-0.14963385462760925,
-0.07691185921430588
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-squad-model3
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 79
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-squad-model3", "results": []}]} | question-answering | varun-v-rao/roberta-base-squad-model3 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-base",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T15:52:12+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us
|
# roberta-base-squad-model3
This model is a fine-tuned version of roberta-base on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 79
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-base-squad-model3\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 79\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us \n",
"# roberta-base-squad-model3\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 79\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
69,
32,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-base #license-mit #endpoints_compatible #region-us \n# roberta-base-squad-model3\n\nThis model is a fine-tuned version of roberta-base on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 79\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.05982813611626625,
0.04306873679161072,
-0.0017067405860871077,
0.07177712023258209,
0.18752436339855194,
0.015194974839687347,
0.14876236021518707,
0.07608761638402939,
-0.1477811485528946,
0.0527522936463356,
0.06592388451099396,
0.08668716996908188,
0.011916406452655792,
0.0980151817202568,
-0.03924936801195145,
-0.2309672087430954,
0.0058240825310349464,
-0.007711314130574465,
-0.12674932181835175,
0.09674058854579926,
0.09953361749649048,
-0.11893812566995621,
0.059988152235746384,
-0.006982858758419752,
-0.21911700069904327,
0.05482267588376999,
0.007261353544890881,
-0.03661186993122101,
0.10629664361476898,
0.02003469318151474,
0.1387123465538025,
0.0026201612781733274,
0.13542120158672333,
-0.20976132154464722,
0.018100555986166,
0.09045007079839706,
0.020586591213941574,
0.0652569830417633,
0.02268800139427185,
-0.010560011491179466,
0.12513180077075958,
-0.1264418214559555,
0.09427095204591751,
0.019911203533411026,
-0.09152065217494965,
-0.16810204088687897,
-0.09180624783039093,
0.02854757010936737,
0.08543497323989868,
0.10356828570365906,
-0.013064676895737648,
0.1790224313735962,
-0.11809517443180084,
0.07817690074443817,
0.1942148357629776,
-0.2776574492454529,
-0.10537773370742798,
0.1083085909485817,
0.04987730458378792,
0.09230993688106537,
-0.11343005299568176,
-0.001814053044654429,
0.05023600906133652,
0.03197425603866577,
0.08003028482198715,
-0.026316560804843903,
-0.08279913663864136,
0.007223043125122786,
-0.14712120592594147,
0.03467695042490959,
0.12646591663360596,
0.054936669766902924,
-0.032834894955158234,
-0.023041050881147385,
-0.042295556515455246,
-0.02923678606748581,
-0.040294598788022995,
-0.05983410030603409,
0.04845251142978668,
-0.04883835092186928,
-0.1352015733718872,
-0.03186315298080444,
-0.08203054219484329,
-0.07088256627321243,
-0.024586623534560204,
0.13542787730693817,
0.04358142614364624,
0.012096544727683067,
-0.0658920630812645,
0.0815838947892189,
-0.014482221566140652,
-0.10429510474205017,
0.015093338675796986,
0.006289603188633919,
-0.07841361314058304,
-0.07552781701087952,
-0.05719148367643356,
-0.060546256601810455,
0.04079953953623772,
0.18300727009773254,
-0.05099696293473244,
0.0570148266851902,
0.005551837384700775,
0.010351891629397869,
-0.0495610348880291,
0.12521858513355255,
-0.06555173546075821,
-0.06770440191030502,
0.015544172376394272,
0.06344304978847504,
0.016617823392152786,
0.008524511940777302,
-0.08085785806179047,
0.013509794138371944,
0.07989577203989029,
0.03548157960176468,
-0.06043671444058418,
0.03959432616829872,
0.0005933359498158097,
-0.007892688736319542,
-0.03310316801071167,
-0.09964019805192947,
0.036282386630773544,
-0.010971594601869583,
-0.0611320324242115,
-0.014537687413394451,
0.01251247525215149,
0.02512916922569275,
0.0188498105853796,
0.1109166070818901,
-0.09836950153112411,
0.014268132857978344,
-0.11237463355064392,
-0.09986132383346558,
-0.0022255207877606153,
-0.06522157043218613,
0.006001413334161043,
-0.08482842147350311,
-0.14127425849437714,
-0.044520702213048935,
0.04570732265710831,
-0.047591473907232285,
-0.008805518969893456,
-0.050804197788238525,
-0.08644243329763412,
-0.003818284021690488,
0.005782625172287226,
0.16327571868896484,
-0.03977980092167854,
0.06897907704114914,
0.0328667014837265,
0.03607945516705513,
-0.028271809220314026,
0.02154279127717018,
-0.08784712105989456,
0.01363381464034319,
-0.18514278531074524,
0.03451735898852348,
-0.09180925041437149,
0.06885898858308792,
-0.10108767449855804,
-0.08832891285419464,
-0.017250362783670425,
0.007572972681373358,
0.07782217860221863,
0.07264237105846405,
-0.16575339436531067,
-0.03712543472647667,
0.1628180891275406,
-0.09034460037946701,
-0.08840196579694748,
0.09103255718946457,
-0.06749986112117767,
0.06722704321146011,
0.07326001673936844,
0.15921196341514587,
0.05354221537709236,
-0.10949014127254486,
-0.005095365922898054,
-0.03954325243830681,
0.04978804290294647,
-0.012100700289011002,
0.03953968361020088,
0.007933295331895351,
0.02901526354253292,
0.0065727210603654385,
-0.06655985862016678,
0.02001980133354664,
-0.10973770171403885,
-0.07358034700155258,
-0.04281932860612869,
-0.09684095531702042,
-0.03322320803999901,
0.053631871938705444,
0.05967114865779877,
-0.09113096445798874,
-0.09239610284566879,
0.15265525877475739,
0.10747160017490387,
-0.06129354238510132,
0.017351631075143814,
-0.08392607420682907,
0.04845672473311424,
-0.06914906203746796,
-0.026599764823913574,
-0.19099563360214233,
-0.11409187316894531,
-0.007230759132653475,
0.019585350528359413,
0.05835367739200592,
0.040536701679229736,
0.0710260197520256,
0.05668581277132034,
-0.052435461431741714,
0.012896011583507061,
-0.066985122859478,
0.00939608458429575,
-0.09596488624811172,
-0.2402079999446869,
-0.022715218365192413,
-0.028782028704881668,
0.17937742173671722,
-0.25497934222221375,
0.020573584362864494,
-0.05233825370669365,
0.13410614430904388,
0.031943611800670624,
-0.03859074041247368,
-0.046441156417131424,
0.06865831464529037,
-0.003743615699931979,
-0.0731247067451477,
0.04307826608419418,
-0.0039095389656722546,
-0.0630926638841629,
-0.082967109978199,
-0.13864175975322723,
0.08335994929075241,
0.10161640495061874,
-0.03819545358419418,
-0.10048295557498932,
0.02707098051905632,
-0.0662553608417511,
-0.0285563413053751,
-0.06694335490465164,
0.021143874153494835,
0.128011554479599,
-0.011606775224208832,
0.12180177122354507,
-0.053575072437524796,
-0.04628333821892738,
0.0017930630128830671,
-0.0339578278362751,
0.03805937618017197,
0.08603733032941818,
0.12786522507667542,
-0.12556616961956024,
0.09112855792045593,
0.08433038741350174,
-0.10295494645833969,
0.13275852799415588,
-0.04205571487545967,
-0.06710551679134369,
-0.029198555275797844,
-0.014580762013792992,
0.007507430389523506,
0.14058291912078857,
-0.08077660202980042,
0.00176548573654145,
-0.0021928732749074697,
0.015325140208005905,
0.04238884896039963,
-0.1826886683702469,
-0.029815837740898132,
0.002791984938085079,
-0.01703752391040325,
0.0052263145335018635,
-0.007533679250627756,
0.017762592062354088,
0.10074426233768463,
0.012430180795490742,
-0.03369025141000748,
-0.004793327301740646,
-0.0073993694968521595,
-0.08023785799741745,
0.20270423591136932,
-0.07946911454200745,
-0.049588676542043686,
-0.08459042012691498,
-0.016028406098484993,
-0.04044133424758911,
-0.030424736440181732,
0.028054719790816307,
-0.1183285042643547,
-0.04755295440554619,
-0.08336853235960007,
0.015217255800962448,
-0.0002164090401493013,
0.001339519745670259,
0.03428603336215019,
0.02675226517021656,
0.08952774107456207,
-0.13169927895069122,
0.012226357124745846,
-0.08282313495874405,
-0.13462868332862854,
0.007755117490887642,
0.05800503119826317,
0.12550030648708344,
0.11861798912286758,
-0.017752408981323242,
0.012213563546538353,
-0.025399483740329742,
0.23772591352462769,
-0.06147699058055878,
-0.012133769690990448,
0.10270804166793823,
0.009158797562122345,
0.029401997104287148,
0.12530122697353363,
0.05731341615319252,
-0.11668737232685089,
0.043762870132923126,
0.10575520992279053,
-0.03299127146601677,
-0.23275747895240784,
-0.02995791658759117,
-0.03698819503188133,
-0.11220885068178177,
0.046115510165691376,
0.04097502678632736,
0.004465142730623484,
0.05619843304157257,
0.026208434253931046,
0.05549211800098419,
-0.024091826751828194,
0.08035513758659363,
0.08608213812112808,
0.049494028091430664,
0.12962330877780914,
-0.057305578142404556,
-0.06358924508094788,
0.03711436688899994,
-0.023988477885723114,
0.30503973364830017,
-0.002532461192458868,
0.01994774118065834,
0.09813669323921204,
0.11539925634860992,
-0.018248172476887703,
0.019159657880663872,
0.00282755377702415,
-0.049410685896873474,
0.00010777969873743132,
-0.055412501096725464,
0.010186728090047836,
0.005696443375200033,
-0.06737077236175537,
0.06079002097249031,
-0.06586495786905289,
0.04819033294916153,
0.054904159158468246,
0.21604718267917633,
0.010756893083453178,
-0.2707166075706482,
-0.07863283157348633,
0.010765635408461094,
-0.028750181198120117,
-0.008664417080581188,
0.006265533622354269,
0.10812205076217651,
-0.09132414311170578,
0.024999625980854034,
-0.05943252518773079,
0.09169182181358337,
0.004575819242745638,
0.009041054174304008,
0.04075273871421814,
0.1731470227241516,
-0.015322847291827202,
0.059509966522455215,
-0.22811634838581085,
0.2402912825345993,
0.009430178441107273,
0.14116331934928894,
-0.02076098509132862,
0.0008077977108769119,
0.021815963089466095,
0.07417578250169754,
0.05321572721004486,
-0.006229955703020096,
-0.05444367229938507,
-0.16340535879135132,
-0.013553631491959095,
0.06897438317537308,
0.1234588772058487,
-0.007289245259016752,
0.1099136471748352,
-0.045639555901288986,
0.011424348689615726,
0.06907931715250015,
-0.05694395676255226,
-0.18361066281795502,
-0.09701766073703766,
-0.03626888245344162,
-0.007855644449591637,
-0.09146659076213837,
-0.0835690051317215,
-0.09359017014503479,
-0.07026371359825134,
0.15552617609500885,
0.026028554886579514,
0.012082982808351517,
-0.11912534385919571,
0.14197832345962524,
0.08940178155899048,
-0.05904008820652962,
0.010666151531040668,
0.013128644786775112,
0.07488708198070526,
0.04173387587070465,
-0.06901683658361435,
0.06912508606910706,
-0.07680122554302216,
-0.13870646059513092,
-0.05983905494213104,
0.08152184635400772,
0.07262206077575684,
0.038473278284072876,
0.0020207585766911507,
0.03219880163669586,
0.017874279990792274,
-0.09239554405212402,
-0.0006282652611844242,
0.0499243289232254,
0.04905201494693756,
0.04348069801926613,
-0.07931993901729584,
-0.01810271292924881,
-0.04528636112809181,
-0.022638635709881783,
0.1120852380990982,
0.2321888655424118,
-0.07640109956264496,
0.015175748616456985,
0.05347859114408493,
-0.08188433945178986,
-0.17704063653945923,
0.12838487327098846,
0.08036991953849792,
-0.002398984506726265,
0.06086068972945213,
-0.14434988796710968,
0.18895475566387177,
0.11566419899463654,
-0.01332371961325407,
0.03192838281393051,
-0.2859286069869995,
-0.13508856296539307,
0.07111819833517075,
0.16122199594974518,
0.09788884967565536,
-0.16014672815799713,
-0.015657886862754822,
-0.03298700973391533,
-0.13933280110359192,
0.10887620598077774,
-0.18265582621097565,
0.1009821742773056,
0.01169423758983612,
0.07287228852510452,
-0.0007004684302955866,
-0.046737827360630035,
0.11916851997375488,
0.021644514054059982,
0.116212397813797,
-0.0556931234896183,
-0.0032651701476424932,
0.12594731152057648,
-0.03655187785625458,
-0.012765675783157349,
-0.021961020305752754,
0.0517372228205204,
-0.07463675737380981,
-0.016013091430068016,
-0.07671066373586655,
0.04418868571519852,
-0.05517920106649399,
-0.059042979031801224,
-0.056442949920892715,
0.0629366859793663,
0.03733690455555916,
-0.023948974907398224,
0.07962941378355026,
-0.01655922457575798,
0.158939391374588,
0.011563591659069061,
0.09201586246490479,
-0.010563313961029053,
-0.021001551300287247,
0.016866186633706093,
-0.020161038264632225,
0.05665047466754913,
-0.14311416447162628,
0.015864616259932518,
0.11055803298950195,
0.03574878349900246,
0.1305205374956131,
0.06045305356383324,
-0.05682194232940674,
0.024995187297463417,
0.07100401818752289,
-0.0973651111125946,
-0.1490953415632248,
0.03857395052909851,
-0.06665915250778198,
-0.11619485169649124,
0.06333611160516739,
0.11413101106882095,
-0.06034037098288536,
-0.0067564514465630054,
-0.014706446789205074,
-0.002693803980946541,
-0.06241661682724953,
0.18247461318969727,
0.06865758448839188,
0.04303070530295372,
-0.08280009031295776,
0.08392832428216934,
0.04641754552721977,
-0.03657548502087593,
0.02519412897527218,
0.07886151969432831,
-0.08222270011901855,
-0.0165229681879282,
0.07506585866212845,
0.23993130028247833,
-0.06981786340475082,
-0.02927202731370926,
-0.1396557092666626,
-0.09162753075361252,
0.02998841553926468,
0.17696383595466614,
0.06617344915866852,
-0.043557506054639816,
-0.031378258019685745,
0.061083078384399414,
-0.16278748214244843,
0.07721013575792313,
0.04111652448773384,
0.07979920506477356,
-0.12164611369371414,
0.10578862577676773,
0.021154481917619705,
-0.005171316210180521,
-0.021883109584450722,
0.03259008377790451,
-0.1394236534833908,
-0.008536159992218018,
-0.133149191737175,
-0.03793416544795036,
-0.0230864305049181,
0.0003025862970389426,
-0.0003073846164625138,
-0.06439591944217682,
-0.08814560621976852,
0.038146715611219406,
-0.07690883427858353,
-0.03534106910228729,
0.03538743406534195,
0.04187673702836037,
-0.14575043320655823,
0.006845186464488506,
0.013855648227036,
-0.07112788408994675,
0.053632497787475586,
0.048049263656139374,
0.04315851256251335,
0.061069488525390625,
-0.2076016217470169,
-0.010480801574885845,
0.02586856298148632,
-0.0007729267235845327,
0.08381034433841705,
-0.0668516755104065,
-0.007588520180433989,
-0.014033201150596142,
0.10791848599910736,
0.024430211633443832,
0.05499587580561638,
-0.10445541888475418,
0.02230774611234665,
-0.08838178217411041,
-0.06631004065275192,
-0.050334859639406204,
0.02597690001130104,
0.08518816530704498,
0.04687626659870148,
0.17710870504379272,
-0.10529573261737823,
0.025811338797211647,
-0.21660065650939941,
-0.030462758615612984,
-0.0018808809109032154,
-0.044440582394599915,
-0.09086085855960846,
-0.04578345641493797,
0.06448393315076828,
-0.06694046407938004,
0.11968870460987091,
0.005745941307395697,
0.07877016067504883,
0.05354125797748566,
-0.04354488477110863,
-0.006123523227870464,
0.021755550056695938,
0.1899280846118927,
0.060736317187547684,
-0.0049407160840928555,
0.05139824375510216,
0.01520830299705267,
0.06488856673240662,
0.025213997811079025,
0.1959882527589798,
0.19365277886390686,
-0.04237733781337738,
0.0559629388153553,
0.0796189084649086,
-0.06904395669698715,
-0.07570726424455643,
0.15284357964992523,
-0.04279276356101036,
0.056032609194517136,
-0.04841955006122589,
0.1433563083410263,
0.15914130210876465,
-0.15210795402526855,
0.017545532435178757,
-0.07628316432237625,
-0.08762981742620468,
-0.11644333600997925,
0.0006946238572709262,
-0.09550732374191284,
-0.13879632949829102,
0.04797559976577759,
-0.12650999426841736,
0.006686023902148008,
0.11507408320903778,
0.019161729142069817,
0.01881016418337822,
0.16918684542179108,
0.004618586506694555,
0.043508898466825485,
0.02189180627465248,
-0.010921338573098183,
-0.027084754779934883,
-0.04849307984113693,
-0.04301420599222183,
0.02774718776345253,
-0.011775324121117592,
0.05763763189315796,
-0.04956197366118431,
-0.03451233729720116,
0.04324039816856384,
-0.02449062094092369,
-0.07381042093038559,
0.03638157993555069,
0.039452578872442245,
0.03531355410814285,
0.04436852037906647,
0.039464399218559265,
-0.010639669373631477,
-0.03101075068116188,
0.2810053825378418,
-0.06101181358098984,
-0.1276000738143921,
-0.13816730678081512,
0.22119183838367462,
0.03160969540476799,
-0.00004841088957618922,
0.018407564610242844,
-0.1060819923877716,
0.017219921573996544,
0.18120728433132172,
0.19735999405384064,
-0.07889597862958908,
-0.0082101970911026,
0.0019513489678502083,
-0.02510656788945198,
-0.07939354330301285,
0.12751026451587677,
0.10460616648197174,
0.03880451247096062,
-0.05553661286830902,
-0.04754449054598808,
-0.03360028564929962,
-0.02347649075090885,
-0.046745866537094116,
0.029035165905952454,
0.06084613874554634,
0.01135625597089529,
-0.027905860915780067,
0.093790702521801,
-0.015698891133069992,
-0.19866369664669037,
0.06660842895507812,
-0.15533478558063507,
-0.1749866008758545,
-0.032249823212623596,
0.08995170891284943,
-0.03137115016579628,
0.05959409847855568,
-0.03631998598575592,
-0.015047550201416016,
0.07580812275409698,
-0.01848040707409382,
-0.031200919300317764,
-0.11201541870832443,
0.09110523760318756,
-0.07723767310380936,
0.20341503620147705,
-0.03923080489039421,
0.06046455726027489,
0.13997507095336914,
0.021860528737306595,
-0.09091348201036453,
0.06844821572303772,
0.04699575528502464,
-0.0975964143872261,
0.015572315081954002,
0.127332866191864,
-0.05332560837268829,
0.11083739995956421,
0.058426376432180405,
-0.14378611743450165,
0.028871502727270126,
-0.07083140313625336,
-0.046201787889003754,
-0.08222654461860657,
0.002493347506970167,
-0.09117963910102844,
0.1429070234298706,
0.20183013379573822,
-0.021076494827866554,
0.04026377946138382,
-0.08954154700040817,
0.034077905118465424,
0.06357626616954803,
0.10241174697875977,
-0.05479804053902626,
-0.24448880553245544,
0.03342647850513458,
0.029378417879343033,
-0.012178878299891949,
-0.28931406140327454,
-0.07603494077920914,
0.03341802582144737,
-0.049059752374887466,
-0.0626588985323906,
0.10629719495773315,
0.12424348294734955,
0.06314436346292496,
-0.05105065926909447,
-0.18038161098957062,
-0.05819676071405411,
0.1673935055732727,
-0.14523111283779144,
-0.06880486756563187
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "roberta-base"} | null | vgaraujov/roberta-base-peft-prefix-tuning | [
"peft",
"arxiv:1910.09700",
"base_model:roberta-base",
"region:us"
] | 2024-02-08T15:55:18+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-roberta-base #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-roberta-base #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
27,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-roberta-base #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.08285140246152878,
0.1870795041322708,
-0.004291990306228399,
0.03514629229903221,
0.09340392798185349,
0.01857364922761917,
0.053081028163433075,
0.11053172498941422,
-0.06043691188097,
0.10639694333076477,
0.047299712896347046,
0.0870538055896759,
0.1015196144580841,
0.18061839044094086,
0.002947032218798995,
-0.2276124507188797,
0.02974306419491768,
-0.10343362390995026,
0.005745808128267527,
0.12418028712272644,
0.15333674848079681,
-0.09362804889678955,
0.08477873355150223,
-0.020357739180326462,
-0.02156938798725605,
-0.019576765596866608,
-0.07358665019273758,
-0.05554095283150673,
0.048622503876686096,
0.07523737102746964,
0.06825371086597443,
0.007452594581991434,
0.07703933119773865,
-0.2747360169887543,
0.0166135486215353,
0.04735533893108368,
-0.015463616698980331,
0.08109390735626221,
0.10252048820257187,
-0.05235685035586357,
0.11247078329324722,
-0.02063601277768612,
0.1311550736427307,
0.06298994272947311,
-0.08531961590051651,
-0.16625629365444183,
-0.07294447720050812,
0.07556130737066269,
0.15525059401988983,
0.06469850242137909,
-0.044183190912008286,
0.15222740173339844,
-0.1419534683227539,
0.0071436683647334576,
0.05259356275200844,
-0.05045810714364052,
-0.08235543221235275,
0.05197382718324661,
0.09099472314119339,
0.07838037610054016,
-0.1432379186153412,
-0.03395979851484299,
0.028172824531793594,
0.024701016023755074,
0.07243921607732773,
0.015565977431833744,
0.12595251202583313,
0.04034217447042465,
-0.14343446493148804,
-0.03859679400920868,
0.17119714617729187,
0.05472014471888542,
-0.052087269723415375,
-0.20044367015361786,
0.0019158272771164775,
-0.06482341885566711,
-0.023990992456674576,
-0.04073500260710716,
0.043680377304553986,
-0.017879242077469826,
0.0689711719751358,
0.01440432108938694,
-0.0927533209323883,
-0.03897998481988907,
0.08427929878234863,
0.03963244706392288,
0.021780110895633698,
-0.027822701260447502,
-0.008993136696517467,
0.13477888703346252,
0.05148020386695862,
-0.11862911283969879,
-0.06310164928436279,
-0.0643397718667984,
-0.055461395531892776,
-0.06098920851945877,
0.034916747361421585,
0.04167434945702553,
0.0713343471288681,
0.22512738406658173,
0.010422354564070702,
0.04918944090604782,
0.057026419788599014,
0.019308358430862427,
0.05940360948443413,
0.0874672532081604,
-0.08103062212467194,
-0.13775435090065002,
-0.04033121094107628,
0.09091638028621674,
-0.010497943498194218,
-0.013952652923762798,
-0.04074086248874664,
0.0549684502184391,
0.029693609103560448,
0.10441268980503082,
0.07389900833368301,
-0.006427030544728041,
-0.09708281606435776,
-0.04174235463142395,
0.21529774367809296,
-0.13807342946529388,
0.032444294542074203,
0.01372698787599802,
-0.03043084777891636,
-0.038185931742191315,
0.01328256819397211,
0.025393111631274223,
-0.022758495062589645,
0.11638803780078888,
-0.07769777625799179,
-0.028315404430031776,
-0.11559724807739258,
-0.016586411744356155,
0.030540017411112785,
0.03489582985639572,
0.0005345405079424381,
-0.023956092074513435,
-0.05926031246781349,
-0.06691323965787888,
0.0796133428812027,
-0.08840637654066086,
-0.07428096979856491,
-0.015857413411140442,
-0.08558916300535202,
0.0057815867476165295,
0.00862026121467352,
0.14854177832603455,
-0.03078518994152546,
0.03158912807703018,
-0.01989573985338211,
0.05097142979502678,
0.07625165581703186,
0.035418733954429626,
-0.06465262919664383,
0.06084076315164566,
-0.18116584420204163,
0.0976080372929573,
-0.09363938122987747,
0.029483750462532043,
-0.15418295562267303,
-0.022794360294938087,
-0.008301490917801857,
0.010840047150850296,
0.019788043573498726,
0.13959112763404846,
-0.2082710564136505,
-0.01603749394416809,
0.1510235220193863,
-0.09820152074098587,
-0.0984523743391037,
0.055874355137348175,
-0.05638592317700386,
0.13198819756507874,
0.017671385779976845,
-0.023405225947499275,
0.06300906091928482,
-0.14752383530139923,
-0.02388322725892067,
-0.020488886162638664,
-0.005886280443519354,
0.12591561675071716,
0.09948395937681198,
-0.07408441603183746,
0.03904219716787338,
0.01951315440237522,
-0.0314985029399395,
-0.04448748379945755,
-0.05267114192247391,
-0.12112768739461899,
-0.00005241762119112536,
-0.06754790991544724,
0.03716275468468666,
-0.019693052396178246,
-0.07184379547834396,
-0.021440744400024414,
-0.15619634091854095,
0.00942773837596178,
0.08456156402826309,
0.02162831276655197,
-0.03126232326030731,
-0.08645360171794891,
0.02014193870127201,
-0.014598427340388298,
-0.03984639421105385,
-0.15214477479457855,
-0.022213676944375038,
0.031280796974897385,
-0.14853323996067047,
0.01846787892282009,
-0.061373062431812286,
0.04960339888930321,
0.019453272223472595,
-0.059474680572748184,
-0.01965804398059845,
-0.02564174123108387,
0.020096158608794212,
-0.04337465763092041,
-0.22930921614170074,
-0.01567833311855793,
-0.03681488335132599,
0.14490655064582825,
-0.22415666282176971,
0.03404434770345688,
0.08144089579582214,
0.11991775035858154,
-0.010537219233810902,
-0.05374373123049736,
0.025398574769496918,
-0.06640118360519409,
-0.033992502838373184,
-0.05636591464281082,
-0.012973162345588207,
-0.01433741394430399,
-0.05266156420111656,
-0.0005115034291520715,
-0.10776709765195847,
-0.01702011562883854,
0.09951122850179672,
0.08192305266857147,
-0.16829073429107666,
-0.041250549256801605,
-0.03686511144042015,
-0.07712925970554352,
-0.09763047099113464,
-0.055782102048397064,
0.13581496477127075,
0.043888941407203674,
0.03399880975484848,
-0.08912636339664459,
-0.06487195193767548,
0.012451977469027042,
-0.028515182435512543,
-0.03224638104438782,
0.10991358757019043,
0.09389375895261765,
-0.09349247068166733,
0.10076984018087387,
0.06764553487300873,
0.026246612891554832,
0.09691835194826126,
-0.012002255767583847,
-0.11832378804683685,
-0.03223657235503197,
0.03545214980840683,
0.0028183944523334503,
0.15369358658790588,
-0.09405240416526794,
0.06548050045967102,
0.04353950172662735,
-0.018122779205441475,
0.049917712807655334,
-0.10720637440681458,
0.012759794481098652,
0.009171892888844013,
-0.014753195457160473,
-0.010837302543222904,
-0.03354879468679428,
0.013117716647684574,
0.08774641156196594,
0.03299545869231224,
0.0355222187936306,
0.030268961563706398,
-0.03421920910477638,
-0.12085206806659698,
0.19459819793701172,
-0.09528505802154541,
-0.24163185060024261,
-0.15526904165744781,
0.062302108854055405,
0.047449834644794464,
-0.02749410644173622,
0.021346652880311012,
-0.04742330312728882,
-0.1055731251835823,
-0.08361420035362244,
-0.00039546561310999095,
0.03635384142398834,
-0.07915467768907547,
-0.07845200598239899,
0.051477424800395966,
0.04846074804663658,
-0.12984667718410492,
0.036723148077726364,
0.0610891617834568,
-0.03295305743813515,
0.008582452312111855,
0.0619388185441494,
0.09036744385957718,
0.15640206634998322,
-0.004931429401040077,
-0.024657756090164185,
0.04704143851995468,
0.2719167470932007,
-0.15196353197097778,
0.09720314294099808,
0.1022605299949646,
-0.06671904027462006,
0.08595680445432663,
0.19222666323184967,
0.03607945144176483,
-0.10882322490215302,
0.0451304167509079,
0.030477162450551987,
-0.024386469274759293,
-0.2708514928817749,
-0.05876094102859497,
-0.0018689918797463179,
-0.10161111503839493,
0.06907721608877182,
0.08977340906858444,
0.0945926159620285,
0.05254832282662392,
-0.0645088255405426,
-0.08533436805009842,
0.018494615331292152,
0.08621569722890854,
-0.041601963341236115,
0.0028982528019696474,
0.08623456954956055,
-0.031073473393917084,
0.004328708630055189,
0.11064061522483826,
0.011352477595210075,
0.19512437283992767,
0.061486534774303436,
0.12344178557395935,
0.09381356090307236,
0.09702591598033905,
0.019298134371638298,
0.026385217905044556,
0.02217572182416916,
0.0129170473664999,
-0.0008279408211819828,
-0.07924013584852219,
0.03056258335709572,
0.11385618150234222,
0.06632399559020996,
0.045597054064273834,
0.022528821602463722,
-0.06482691317796707,
0.06569404155015945,
0.1917334347963333,
-0.012017269618809223,
-0.1959526389837265,
-0.053490325808525085,
0.06896208971738815,
-0.0851702019572258,
-0.12060036510229111,
-0.018054677173495293,
0.04837043955922127,
-0.18192659318447113,
0.011718378402292728,
-0.04065157100558281,
0.09811842441558838,
-0.07873722165822983,
-0.036867089569568634,
0.05958390235900879,
0.07496959716081619,
-0.03250272199511528,
0.0864126905798912,
-0.18947194516658783,
0.14660952985286713,
0.00990512315183878,
0.06039281189441681,
-0.08924386650323868,
0.10462214797735214,
-0.001082205562852323,
0.015063677914440632,
0.1563015878200531,
0.006031855940818787,
-0.034790415316820145,
-0.05483856052160263,
-0.09858734905719757,
-0.001799840247258544,
0.09024632722139359,
-0.12027058750391006,
0.05983880162239075,
-0.0010615811916068196,
-0.021510958671569824,
0.005801995284855366,
-0.08695042133331299,
-0.1328638195991516,
-0.1718437522649765,
0.06639543175697327,
-0.13937757909297943,
0.06324155628681183,
-0.1004611998796463,
-0.07015091925859451,
-0.006417619064450264,
0.17005611956119537,
-0.1949271857738495,
-0.06804788112640381,
-0.13728970289230347,
-0.1003413125872612,
0.17403961718082428,
-0.03923213109374046,
0.07828294485807419,
0.0031397438142448664,
0.17337098717689514,
0.030938420444726944,
0.02155054546892643,
0.10224317014217377,
-0.08596967905759811,
-0.1853671371936798,
-0.06955811381340027,
0.14610201120376587,
0.1388292908668518,
0.04728168621659279,
-0.013696000911295414,
0.014029072597622871,
-0.054878417402505875,
-0.1268937736749649,
0.0017182552255690098,
0.10863930732011795,
0.09812178462743759,
0.0054083578288555145,
-0.029638569802045822,
-0.10143581032752991,
-0.0736239030957222,
-0.0716557502746582,
0.017895465716719627,
0.17368124425411224,
-0.07446413487195969,
0.1426355242729187,
0.1283392757177353,
-0.06194353103637695,
-0.19356487691402435,
0.05267338454723358,
0.06517971307039261,
0.009936291724443436,
0.04799584671854973,
-0.1888042390346527,
0.09455383569002151,
0.049109332263469696,
-0.05111798271536827,
0.1233590841293335,
-0.1509721875190735,
-0.15502357482910156,
0.0880524143576622,
0.05617637559771538,
-0.2546672821044922,
-0.13222455978393555,
-0.09379991143941879,
-0.03447761386632919,
-0.12469414621591568,
0.07845084369182587,
0.011642609722912312,
0.014734731055796146,
0.042478375136852264,
0.03314938396215439,
0.006471023894846439,
-0.04740611091256142,
0.21586823463439941,
0.008961997926235199,
0.03756359592080116,
-0.04260288178920746,
-0.11326581239700317,
0.033454377204179764,
-0.036875758320093155,
0.09160994738340378,
0.005449037533253431,
0.017608024179935455,
-0.11865557730197906,
-0.046287935227155685,
-0.061400894075632095,
0.03128238022327423,
-0.09651097655296326,
-0.0988936498761177,
-0.057365987449884415,
0.09669125825166702,
0.07357320934534073,
-0.03756877779960632,
-0.01751025579869747,
-0.0902455672621727,
0.041597187519073486,
0.1916513592004776,
0.20554502308368683,
0.06076790392398834,
-0.0925254374742508,
0.009815567173063755,
-0.021853215992450714,
0.04655712842941284,
-0.23896899819374084,
0.05179910734295845,
0.04969661310315132,
0.020281950011849403,
0.1234702542424202,
-0.03258603811264038,
-0.1564047634601593,
-0.05299060046672821,
0.06947419047355652,
-0.041592467576265335,
-0.16859117150306702,
-0.020782941952347755,
0.043952107429504395,
-0.20221151411533356,
-0.044886987656354904,
-0.0010661543346941471,
-0.020278627052903175,
-0.042954809963703156,
0.008565015159547329,
0.07786791026592255,
-0.01800180785357952,
0.1298576295375824,
0.07810857892036438,
0.09521783888339996,
-0.10330510139465332,
0.08117228746414185,
0.06428221613168716,
-0.0637427493929863,
0.0034494439605623484,
0.06473523378372192,
-0.042289864271879196,
-0.03296075016260147,
0.06823636591434479,
0.07298111915588379,
0.03805408626794815,
-0.05514591932296753,
-0.007618797477334738,
-0.07338498532772064,
0.05754892900586128,
0.1171792671084404,
0.04616347327828407,
0.003900317708030343,
0.044424399733543396,
0.019767964258790016,
-0.08453744649887085,
0.09254894405603409,
0.05350496247410774,
0.026928603649139404,
-0.04146123677492142,
-0.02540326863527298,
0.0253231693059206,
-0.020540574565529823,
-0.014544550329446793,
-0.015915261581540108,
-0.08151138573884964,
-0.017824219539761543,
-0.13018913567066193,
0.038008131086826324,
-0.06696050614118576,
0.02135787531733513,
0.02476874366402626,
-0.056890785694122314,
-0.014818033203482628,
0.012015022337436676,
-0.07976625114679337,
-0.0333869531750679,
-0.0008458365919068456,
0.11724702268838882,
-0.11804770678281784,
0.04485635831952095,
0.0949908047914505,
-0.10253472626209259,
0.07388123124837875,
0.003271200694143772,
0.005108632147312164,
0.024271594360470772,
-0.18614400923252106,
0.08566157519817352,
-0.009183336980640888,
0.00038683670572936535,
0.02315308153629303,
-0.2138887643814087,
-0.004740460775792599,
-0.03862430155277252,
-0.01940225064754486,
0.00015240137872751802,
-0.043935030698776245,
-0.13593019545078278,
0.09100288152694702,
-0.00822089146822691,
-0.10274481028318405,
-0.029177064076066017,
0.02999051846563816,
0.11506883054971695,
-0.03467099368572235,
0.15291598439216614,
-0.007025748025625944,
0.06151536852121353,
-0.1664181351661682,
-0.012966767884790897,
-0.020379414781928062,
0.032014213502407074,
-0.022945664823055267,
-0.014804475009441376,
0.050258975476026535,
-0.021909769624471664,
0.23054292798042297,
-0.03346623107790947,
0.06856810301542282,
0.054374776780605316,
0.031701430678367615,
-0.003240398596972227,
0.09476543962955475,
0.06081933528184891,
0.0009382049320265651,
0.02063041552901268,
0.017618920654058456,
-0.013492322526872158,
-0.04101676866412163,
-0.14976894855499268,
0.06555032730102539,
0.16241790354251862,
0.030957384034991264,
0.0034989239647984505,
0.05261339619755745,
-0.11065027117729187,
-0.07215322554111481,
0.12035957723855972,
-0.01132875680923462,
-0.0332702100276947,
-0.07279792428016663,
0.14635050296783447,
0.13279102742671967,
-0.1987827569246292,
0.07613182812929153,
-0.06301223486661911,
-0.07304418832063675,
-0.11530610173940659,
-0.14102694392204285,
-0.06618500500917435,
-0.0443999245762825,
-0.010827573947608471,
-0.07618795335292816,
0.05545991659164429,
0.10594402253627777,
0.003701878013089299,
-0.02124817483127117,
0.1070900708436966,
-0.00707379961386323,
-0.024889033287763596,
0.03095541149377823,
0.06344648450613022,
0.0314173549413681,
-0.10744474083185196,
0.01893029175698757,
0.003208362963050604,
0.02197791263461113,
0.05179955065250397,
0.013858441263437271,
-0.03768325597047806,
-0.016162846237421036,
-0.028599930927157402,
-0.10084518045186996,
0.036897264420986176,
-0.036095622926950455,
-0.05579696223139763,
0.12587207555770874,
0.019808931276202202,
0.006759841460734606,
-0.018657170236110687,
0.22253714501857758,
-0.072800412774086,
-0.06734568625688553,
-0.16841500997543335,
0.05561809241771698,
-0.07213427126407623,
0.056736964732408524,
0.05134982243180275,
-0.10258441418409348,
0.02363690175116062,
0.1456812620162964,
0.12665311992168427,
-0.01195480301976204,
0.009508932009339333,
0.05191682651638985,
0.0033114778343588114,
-0.03329380229115486,
0.02869967743754387,
0.047302208840847015,
0.08716906607151031,
-0.06701745837926865,
0.08971919119358063,
-0.010545551776885986,
-0.07966030389070511,
-0.0016359201399609447,
0.12094978988170624,
-0.004410291090607643,
0.005478385835886002,
-0.07900834828615189,
0.13781271874904633,
-0.06398323178291321,
-0.23752515017986298,
0.05266335606575012,
-0.07053449004888535,
-0.16833730041980743,
-0.030783798545598984,
0.0024391827173531055,
-0.013698852621018887,
0.01716524176299572,
0.07800401747226715,
-0.05892074108123779,
0.17167769372463226,
0.05293157696723938,
-0.0777592882514,
-0.0797143504023552,
0.07076380401849747,
-0.10944768041372299,
0.27625542879104614,
0.016663027927279472,
0.050119128078222275,
0.10306360572576523,
-0.015759648755192757,
-0.1203935444355011,
0.04929736629128456,
0.1030619814991951,
-0.06945538520812988,
0.07329802215099335,
0.16093887388706207,
-0.0017831953009590507,
0.14507171511650085,
0.057811278849840164,
-0.05124129727482796,
0.04804515093564987,
-0.11119244992733002,
-0.05265291407704353,
-0.1030273512005806,
0.08202602714300156,
-0.07868272066116333,
0.15116943418979645,
0.13212238252162933,
-0.06562936305999756,
-0.017156895250082016,
-0.025923069566488266,
0.08904757350683212,
-0.0019337424309924245,
0.11923245340585709,
0.010843786410987377,
-0.19485466182231903,
0.022077197209000587,
-0.0022802124731242657,
0.10111592710018158,
-0.20171299576759338,
-0.0618155412375927,
0.055813584476709366,
-0.021655842661857605,
-0.06382283568382263,
0.10796709358692169,
0.04598083347082138,
0.04603837430477142,
-0.035964302718639374,
-0.05324787274003029,
-0.00979053508490324,
0.14544498920440674,
-0.100653275847435,
-0.008005415089428425
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Data Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| {"library_name": "peft", "base_model": "meta-llama/Llama-2-13b-chat-hf"} | null | bmehrba/Llama-2-13b-chat-hf-fine-tuned-adapters_Gpt4_t1_Llama13b_Seed101 | [
"peft",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"region:us"
] | 2024-02-08T15:56:00+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
38,
6,
3,
45,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
154,
14,
154,
14
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08950838446617126,
0.17622625827789307,
-0.003707088530063629,
0.032576385885477066,
0.08380123972892761,
0.019701125100255013,
0.05203324928879738,
0.11702486872673035,
-0.05330678075551987,
0.09448089450597763,
0.048484884202480316,
0.10060896724462509,
0.09846198558807373,
0.18868719041347504,
-0.0011855853954330087,
-0.2060726284980774,
0.015578063204884529,
-0.10931064933538437,
0.005876870360225439,
0.12358442693948746,
0.15569306910037994,
-0.09741293638944626,
0.08712729811668396,
-0.01551457867026329,
-0.010067826136946678,
-0.025396287441253662,
-0.07361544668674469,
-0.05290524289011955,
0.04710441827774048,
0.07490185648202896,
0.047730859369039536,
0.003742797765880823,
0.08045824617147446,
-0.2711505889892578,
0.01725192740559578,
0.03912210091948509,
-0.010164672508835793,
0.08416316658258438,
0.08157632499933243,
-0.061213672161102295,
0.10719792544841766,
-0.04486960545182228,
0.12389195710420609,
0.06922121345996857,
-0.06562015414237976,
-0.1487942785024643,
-0.0805540531873703,
0.06815578043460846,
0.16221418976783752,
0.07476766407489777,
-0.04304589703679085,
0.16949640214443207,
-0.13273242115974426,
0.007597264833748341,
0.046794891357421875,
-0.035554688423871994,
-0.08115267008543015,
0.060742560774087906,
0.09725039452314377,
0.07205293327569962,
-0.13358467817306519,
-0.029269445687532425,
0.031876083463430405,
0.026171350851655006,
0.07599646598100662,
0.02472980134189129,
0.14272165298461914,
0.05110684782266617,
-0.13597595691680908,
-0.032095685601234436,
0.1667022556066513,
0.05657454952597618,
-0.05146843194961548,
-0.20977118611335754,
0.010412882082164288,
-0.06257046014070511,
-0.019110077992081642,
-0.0394989438354969,
0.04172099754214287,
-0.026554755866527557,
0.06876977533102036,
0.0052980040200054646,
-0.0955195426940918,
-0.042122215032577515,
0.08467143774032593,
0.03501870483160019,
0.025577984750270844,
-0.03146751970052719,
-0.005369491875171661,
0.13237224519252777,
0.05266503989696503,
-0.11971335113048553,
-0.06415551900863647,
-0.06459555774927139,
-0.05922604724764824,
-0.05847278982400894,
0.025247467681765556,
0.031127413734793663,
0.0707581415772438,
0.20909400284290314,
0.02113768272101879,
0.04728280380368233,
0.06350736320018768,
0.01767423190176487,
0.07364732772111893,
0.08452971279621124,
-0.08042320609092712,
-0.13752959668636322,
-0.026864496991038322,
0.09401044249534607,
-0.004670456051826477,
-0.015377101488411427,
-0.04042273387312889,
0.04590466991066933,
0.03928038105368614,
0.09635873883962631,
0.08342839032411575,
-0.006302335299551487,
-0.08958663791418076,
-0.05172271281480789,
0.21430253982543945,
-0.1486416757106781,
0.022579502314329147,
0.00532573601230979,
-0.046220771968364716,
-0.050389427691698074,
0.013791119679808617,
0.021902183070778847,
-0.01725425384938717,
0.09078584611415863,
-0.07412354648113251,
-0.030390940606594086,
-0.11564502120018005,
-0.00758272223174572,
0.035115793347358704,
0.05083532631397247,
-0.0026497903745621443,
-0.019051065668463707,
-0.06038069352507591,
-0.07015779614448547,
0.08611448109149933,
-0.08802679926156998,
-0.06949871778488159,
-0.022058209404349327,
-0.08482711762189865,
0.008333494886755943,
0.004399609286338091,
0.13455772399902344,
-0.032166268676519394,
0.04013873636722565,
-0.009890900924801826,
0.05181796848773956,
0.06774567812681198,
0.03500198572874069,
-0.053186893463134766,
0.056685443967580795,
-0.19885419309139252,
0.10022944211959839,
-0.09629994630813599,
0.028232630342245102,
-0.15368616580963135,
-0.016224225983023643,
0.024259883910417557,
0.00603050272911787,
0.023533180356025696,
0.13508757948875427,
-0.2269131988286972,
-0.009413540363311768,
0.1492016613483429,
-0.08191759884357452,
-0.11286741495132446,
0.05882270261645317,
-0.06703686714172363,
0.13632111251354218,
0.024114999920129776,
-0.03846221789717674,
0.05126623064279556,
-0.1477012187242508,
-0.034279413521289825,
-0.027603546157479286,
-0.011836200952529907,
0.11866577714681625,
0.09630073606967926,
-0.0608704648911953,
0.048884205520153046,
0.020479585975408554,
-0.032701265066862106,
-0.042141854763031006,
-0.050704531371593475,
-0.12829554080963135,
0.0009587573586031795,
-0.07328714430332184,
0.04790837690234184,
-0.02088468335568905,
-0.06889110058546066,
-0.018932033330202103,
-0.16518932580947876,
0.002006813418120146,
0.09172286838293076,
0.02033841609954834,
-0.03539799153804779,
-0.10069174319505692,
0.0036235731095075607,
-0.011536587961018085,
-0.035604726523160934,
-0.13578550517559052,
-0.02210777997970581,
0.019318837672472,
-0.13882264494895935,
0.030753053724765778,
-0.07345959544181824,
0.051180385053157806,
0.016524922102689743,
-0.05861951783299446,
-0.010977345518767834,
-0.023012345656752586,
0.024373451247811317,
-0.0456857830286026,
-0.24518829584121704,
-0.01426833588629961,
-0.032443173229694366,
0.1618536114692688,
-0.23377619683742523,
0.038241252303123474,
0.06515999883413315,
0.11937034130096436,
-0.02269211784005165,
-0.050194818526506424,
0.02402755618095398,
-0.0810660794377327,
-0.03478178381919861,
-0.05240238085389137,
-0.0170640479773283,
-0.02249637059867382,
-0.06970936059951782,
0.013335862196981907,
-0.10944215208292007,
-0.04154296964406967,
0.10713886469602585,
0.08292265236377716,
-0.15724287927150726,
-0.043278347700834274,
-0.03408950939774513,
-0.08576270937919617,
-0.08529800176620483,
-0.0566803403198719,
0.13487502932548523,
0.05090935528278351,
0.02855822816491127,
-0.08846847712993622,
-0.07940267771482468,
0.00988192018121481,
-0.03207101300358772,
-0.028083765879273415,
0.10094649344682693,
0.07611845433712006,
-0.10813652724027634,
0.08834784477949142,
0.07578150928020477,
0.012136061675846577,
0.11384404450654984,
-0.011400082148611546,
-0.11351825296878815,
-0.04137531667947769,
0.03633233532309532,
0.002555434126406908,
0.1695048063993454,
-0.09464383870363235,
0.06803114712238312,
0.03927377983927727,
-0.022211823612451553,
0.05476415529847145,
-0.10076725482940674,
0.01427049096673727,
0.006726768799126148,
-0.012228100560605526,
-0.011376895941793919,
-0.036163002252578735,
0.020614514127373695,
0.07891662418842316,
0.03816615790128708,
0.036182720214128494,
0.03572281077504158,
-0.04122483730316162,
-0.1245279312133789,
0.19345727562904358,
-0.10554436594247818,
-0.2273423671722412,
-0.1516016721725464,
0.05401213839650154,
0.03572985157370567,
-0.030572842806577682,
0.008941974490880966,
-0.05140937119722366,
-0.0966159775853157,
-0.08070044219493866,
0.005514310672879219,
0.03883929178118706,
-0.07613059133291245,
-0.07262902706861496,
0.05921752378344536,
0.05427297204732895,
-0.13442036509513855,
0.0406947135925293,
0.054035235196352005,
-0.04148136079311371,
0.008404599502682686,
0.06944910436868668,
0.07862463593482971,
0.15086530148983002,
-0.020428497344255447,
-0.020412612706422806,
0.05437345430254936,
0.2643863558769226,
-0.15086820721626282,
0.09670513868331909,
0.09954504668712616,
-0.06504277884960175,
0.07992210984230042,
0.18344183266162872,
0.033216435462236404,
-0.10660552978515625,
0.045308101922273636,
0.031075740233063698,
-0.0188649483025074,
-0.2811678647994995,
-0.06357815116643906,
0.0033266504760831594,
-0.10220301896333694,
0.062428005039691925,
0.0793466567993164,
0.09731262922286987,
0.04918764531612396,
-0.06440604478120804,
-0.07534892857074738,
0.02199655771255493,
0.07507231831550598,
-0.04625728353857994,
0.0006049389485269785,
0.08203481882810593,
-0.0200007613748312,
0.008962401188910007,
0.11015255749225616,
0.013906295411288738,
0.1873634159564972,
0.04269689694046974,
0.11463924497365952,
0.10168035328388214,
0.10507753491401672,
0.000024342234610230662,
0.015555954538285732,
0.02079109288752079,
0.012282595038414001,
-0.002983907237648964,
-0.08613301068544388,
0.02277722768485546,
0.12184786051511765,
0.06945348531007767,
0.04476168751716614,
0.024970298632979393,
-0.050061535090208054,
0.05980529636144638,
0.1768452227115631,
-0.01209972519427538,
-0.1998264193534851,
-0.062326882034540176,
0.06751304864883423,
-0.082801952958107,
-0.11640139669179916,
-0.02261449582874775,
0.050769247114658356,
-0.17440687119960785,
0.015001747757196426,
-0.04254560545086861,
0.09033802151679993,
-0.09127394109964371,
-0.037229955196380615,
0.05321357026696205,
0.07545126974582672,
-0.023492055013775826,
0.09048163145780563,
-0.17921186983585358,
0.13352392613887787,
0.01737614907324314,
0.06370522826910019,
-0.09815072268247604,
0.10393797606229782,
0.015243546105921268,
-0.0071698566898703575,
0.14627893269062042,
0.008973979391157627,
-0.019879506900906563,
-0.058314017951488495,
-0.10938628017902374,
-0.0015536772552877665,
0.08220188319683075,
-0.11720426380634308,
0.06481732428073883,
0.00044200546108186245,
-0.019408708438277245,
0.010529479943215847,
-0.0697939544916153,
-0.14233455061912537,
-0.1691078543663025,
0.06332679092884064,
-0.12960782647132874,
0.05657918378710747,
-0.10196143388748169,
-0.07344398647546768,
-0.006228356156498194,
0.1857890486717224,
-0.19167372584342957,
-0.0651763305068016,
-0.13295814394950867,
-0.08307469636201859,
0.17686748504638672,
-0.038926977664232254,
0.07132517546415329,
0.017756011337041855,
0.17197521030902863,
0.030676020309329033,
0.013996497727930546,
0.10165295004844666,
-0.0863775908946991,
-0.18250107765197754,
-0.06872538477182388,
0.145328551530838,
0.15727265179157257,
0.04947395995259285,
-0.01222315151244402,
0.0006382534629665315,
-0.05825969576835632,
-0.12492486834526062,
0.00552456034347415,
0.14077237248420715,
0.09738009423017502,
0.015011516399681568,
-0.02072962000966072,
-0.12298290431499481,
-0.06933344155550003,
-0.07234511524438858,
0.010791660286486149,
0.1811780333518982,
-0.06657543778419495,
0.1483541578054428,
0.12124106287956238,
-0.0507206916809082,
-0.18955619633197784,
0.04781363531947136,
0.0678601861000061,
0.021055543795228004,
0.06329847872257233,
-0.1708568036556244,
0.10241113603115082,
0.03779063746333122,
-0.056044332683086395,
0.12532320618629456,
-0.13762390613555908,
-0.15448996424674988,
0.08908607810735703,
0.059379611164331436,
-0.23717626929283142,
-0.10756765305995941,
-0.09208329766988754,
-0.04467558488249779,
-0.11974717676639557,
0.07756773382425308,
-0.008080631494522095,
0.01312070433050394,
0.038425788283348083,
0.04747161641716957,
0.010422809049487114,
-0.04883774369955063,
0.2077513337135315,
0.00663892924785614,
0.03319171071052551,
-0.04891526326537132,
-0.10318257659673691,
0.04049978777766228,
-0.04806138575077057,
0.09715691953897476,
-0.014642413705587387,
0.021955221891403198,
-0.1253223717212677,
-0.0439610481262207,
-0.06654173135757446,
0.030696231871843338,
-0.09619533270597458,
-0.09483709931373596,
-0.05548068508505821,
0.10141977667808533,
0.07960876822471619,
-0.03827962279319763,
-0.018101584166288376,
-0.08076406270265579,
0.028281690552830696,
0.192597895860672,
0.20835207402706146,
0.049149978905916214,
-0.06995424628257751,
0.007349140010774136,
-0.012700160034000874,
0.04521884396672249,
-0.2468501627445221,
0.056316666305065155,
0.04637942090630531,
0.019014067947864532,
0.11265500634908676,
-0.035475291311740875,
-0.16250301897525787,
-0.05557123199105263,
0.07098683714866638,
-0.039137084037065506,
-0.15694621205329895,
-0.024994002655148506,
0.05066932737827301,
-0.20187702775001526,
-0.029669208452105522,
0.010474429465830326,
-0.02148980274796486,
-0.04393318295478821,
0.011044103652238846,
0.08090483397245407,
-0.018578581511974335,
0.1367349922657013,
0.07980240881443024,
0.09522033482789993,
-0.10692083835601807,
0.07168128341436386,
0.06122429668903351,
-0.051465462893247604,
0.021644625812768936,
0.06818753480911255,
-0.04446205869317055,
-0.032580625265836716,
0.07838873565196991,
0.058368146419525146,
0.04023381322622299,
-0.0497741736471653,
-0.009552556090056896,
-0.05499427020549774,
0.049196142703294754,
0.10447074472904205,
0.05076836422085762,
0.0006935194251127541,
0.047793444246053696,
0.018387768417596817,
-0.08049451559782028,
0.10598240047693253,
0.05339374020695686,
0.02360537275671959,
-0.0398079976439476,
-0.03602069616317749,
0.018247995525598526,
-0.010786417871713638,
-0.0149832833558321,
-0.016455529257655144,
-0.07099823653697968,
-0.013593231327831745,
-0.13733075559139252,
0.04016523063182831,
-0.08189219981431961,
0.01841694675385952,
0.022008292376995087,
-0.05440347641706467,
-0.007398437242954969,
0.015957478433847427,
-0.07759089022874832,
-0.04222242161631584,
-0.0045568388886749744,
0.12033451348543167,
-0.11743347346782684,
0.041315708309412,
0.0889706164598465,
-0.10073781758546829,
0.08179357647895813,
0.005519764963537455,
0.006593905854970217,
0.027770070359110832,
-0.18307223916053772,
0.07270024716854095,
-0.02148648537695408,
0.003687589429318905,
0.03217103332281113,
-0.22772879898548126,
-0.010953521355986595,
-0.03648538142442703,
-0.016809485852718353,
0.0019160229712724686,
-0.03937701880931854,
-0.13335061073303223,
0.07287079840898514,
-0.01058956515043974,
-0.08660455048084259,
-0.032185930758714676,
0.03226194903254509,
0.1112515926361084,
-0.03534836322069168,
0.15059389173984528,
-0.005941883195191622,
0.05801843851804733,
-0.17130136489868164,
-0.011426819488406181,
-0.019129110500216484,
0.03652174770832062,
-0.018265437334775925,
-0.014729461632668972,
0.053084973245859146,
-0.03412574157118797,
0.2234855443239212,
-0.03480256348848343,
0.06502514332532883,
0.05183198302984238,
0.02280556410551071,
-0.006614799611270428,
0.08636770397424698,
0.06560425460338593,
-0.01096076425164938,
0.02718065120279789,
0.028059065341949463,
-0.012954981066286564,
-0.037562232464551926,
-0.1630524843931198,
0.05572279915213585,
0.1581650972366333,
0.04094236344099045,
0.011616811156272888,
0.06928509473800659,
-0.10752071440219879,
-0.07898375391960144,
0.1387312412261963,
-0.01259393710643053,
-0.032576363533735275,
-0.07013807445764542,
0.13943122327327728,
0.124080128967762,
-0.19758351147174835,
0.07208021730184555,
-0.0731193795800209,
-0.07801702618598938,
-0.10079838335514069,
-0.14738084375858307,
-0.061444323509931564,
-0.052179500460624695,
-0.011450962163507938,
-0.06768535077571869,
0.05396997556090355,
0.10480605065822601,
0.0069710006937384605,
-0.026146549731492996,
0.10475686937570572,
0.0007574855699203908,
-0.027480410411953926,
0.0275881364941597,
0.06416697055101395,
0.01868068240582943,
-0.10241235792636871,
0.016462087631225586,
0.0009010558133013546,
0.028261849656701088,
0.058421481400728226,
0.0037333546206355095,
-0.035359520465135574,
-0.012541528791189194,
-0.022329136729240417,
-0.11025683581829071,
0.038418930023908615,
-0.031967371702194214,
-0.03549599647521973,
0.11972174793481827,
0.021107889711856842,
0.0024782961700111628,
-0.022964047268033028,
0.22632580995559692,
-0.07606904208660126,
-0.0824858620762825,
-0.1684485524892807,
0.048732075840234756,
-0.06246444582939148,
0.03944636881351471,
0.04816613346338272,
-0.1110905185341835,
0.02492443658411503,
0.13681943714618683,
0.13383808732032776,
-0.017702074721455574,
0.0072706313803792,
0.041554342955350876,
-0.001966990763321519,
-0.051138825714588165,
0.022816691547632217,
0.04751669988036156,
0.09492984414100647,
-0.05958498641848564,
0.09289880096912384,
-0.006714127957820892,
-0.08313115686178207,
0.011414550244808197,
0.11385775357484818,
-0.004354037344455719,
0.008586743846535683,
-0.06612556427717209,
0.14033369719982147,
-0.05520116165280342,
-0.2502851188182831,
0.03959165886044502,
-0.0734434500336647,
-0.16861815750598907,
-0.03511347249150276,
0.018955450505018234,
-0.019131824374198914,
0.017461534589529037,
0.07813186943531036,
-0.05068197101354599,
0.17512299120426178,
0.04293905943632126,
-0.08064883947372437,
-0.06616055220365524,
0.07387921214103699,
-0.11062787473201752,
0.28079262375831604,
0.012751048430800438,
0.06857820600271225,
0.10455191880464554,
-0.016430502757430077,
-0.11872978508472443,
0.042664192616939545,
0.10075171291828156,
-0.07164205610752106,
0.08039859682321548,
0.18360178172588348,
0.0013276869431138039,
0.15462037920951843,
0.06878916919231415,
-0.0453730933368206,
0.03654608130455017,
-0.12163300812244415,
-0.05294680967926979,
-0.10768717527389526,
0.08729486167430878,
-0.07798956334590912,
0.15596513450145721,
0.13275524973869324,
-0.07110930234193802,
-0.006204865872859955,
-0.025767024606466293,
0.08593760430812836,
-0.009336618706583977,
0.1176052987575531,
0.00486786337569356,
-0.20527753233909607,
0.022964732721447945,
0.006658138707280159,
0.10234756767749786,
-0.21353045105934143,
-0.06055140495300293,
0.06063069403171539,
-0.027994666248559952,
-0.050338197499513626,
0.11621229350566864,
0.05960828810930252,
0.04527933895587921,
-0.034697841852903366,
-0.03217756003141403,
-0.02518811635673046,
0.13280846178531647,
-0.11107352375984192,
-0.014744595624506474
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Data Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| {"library_name": "peft", "base_model": "meta-llama/Llama-2-13b-chat-hf"} | null | bmehrba/Llama-2-13b-chat-hf-fine-tuned_Gpt4_t1_Llama13b_Seed101 | [
"peft",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-13b-chat-hf",
"region:us"
] | 2024-02-08T15:56:26+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
### Framework versions
- PEFT 0.7.0.dev0
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: True\n- bnb_4bit_compute_dtype: bfloat16",
"### Framework versions\n\n\n- PEFT 0.7.0.dev0"
] | [
38,
6,
3,
45,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
154,
14
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-meta-llama/Llama-2-13b-chat-hf #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.08950838446617126,
0.17622625827789307,
-0.003707088530063629,
0.032576385885477066,
0.08380123972892761,
0.019701125100255013,
0.05203324928879738,
0.11702486872673035,
-0.05330678075551987,
0.09448089450597763,
0.048484884202480316,
0.10060896724462509,
0.09846198558807373,
0.18868719041347504,
-0.0011855853954330087,
-0.2060726284980774,
0.015578063204884529,
-0.10931064933538437,
0.005876870360225439,
0.12358442693948746,
0.15569306910037994,
-0.09741293638944626,
0.08712729811668396,
-0.01551457867026329,
-0.010067826136946678,
-0.025396287441253662,
-0.07361544668674469,
-0.05290524289011955,
0.04710441827774048,
0.07490185648202896,
0.047730859369039536,
0.003742797765880823,
0.08045824617147446,
-0.2711505889892578,
0.01725192740559578,
0.03912210091948509,
-0.010164672508835793,
0.08416316658258438,
0.08157632499933243,
-0.061213672161102295,
0.10719792544841766,
-0.04486960545182228,
0.12389195710420609,
0.06922121345996857,
-0.06562015414237976,
-0.1487942785024643,
-0.0805540531873703,
0.06815578043460846,
0.16221418976783752,
0.07476766407489777,
-0.04304589703679085,
0.16949640214443207,
-0.13273242115974426,
0.007597264833748341,
0.046794891357421875,
-0.035554688423871994,
-0.08115267008543015,
0.060742560774087906,
0.09725039452314377,
0.07205293327569962,
-0.13358467817306519,
-0.029269445687532425,
0.031876083463430405,
0.026171350851655006,
0.07599646598100662,
0.02472980134189129,
0.14272165298461914,
0.05110684782266617,
-0.13597595691680908,
-0.032095685601234436,
0.1667022556066513,
0.05657454952597618,
-0.05146843194961548,
-0.20977118611335754,
0.010412882082164288,
-0.06257046014070511,
-0.019110077992081642,
-0.0394989438354969,
0.04172099754214287,
-0.026554755866527557,
0.06876977533102036,
0.0052980040200054646,
-0.0955195426940918,
-0.042122215032577515,
0.08467143774032593,
0.03501870483160019,
0.025577984750270844,
-0.03146751970052719,
-0.005369491875171661,
0.13237224519252777,
0.05266503989696503,
-0.11971335113048553,
-0.06415551900863647,
-0.06459555774927139,
-0.05922604724764824,
-0.05847278982400894,
0.025247467681765556,
0.031127413734793663,
0.0707581415772438,
0.20909400284290314,
0.02113768272101879,
0.04728280380368233,
0.06350736320018768,
0.01767423190176487,
0.07364732772111893,
0.08452971279621124,
-0.08042320609092712,
-0.13752959668636322,
-0.026864496991038322,
0.09401044249534607,
-0.004670456051826477,
-0.015377101488411427,
-0.04042273387312889,
0.04590466991066933,
0.03928038105368614,
0.09635873883962631,
0.08342839032411575,
-0.006302335299551487,
-0.08958663791418076,
-0.05172271281480789,
0.21430253982543945,
-0.1486416757106781,
0.022579502314329147,
0.00532573601230979,
-0.046220771968364716,
-0.050389427691698074,
0.013791119679808617,
0.021902183070778847,
-0.01725425384938717,
0.09078584611415863,
-0.07412354648113251,
-0.030390940606594086,
-0.11564502120018005,
-0.00758272223174572,
0.035115793347358704,
0.05083532631397247,
-0.0026497903745621443,
-0.019051065668463707,
-0.06038069352507591,
-0.07015779614448547,
0.08611448109149933,
-0.08802679926156998,
-0.06949871778488159,
-0.022058209404349327,
-0.08482711762189865,
0.008333494886755943,
0.004399609286338091,
0.13455772399902344,
-0.032166268676519394,
0.04013873636722565,
-0.009890900924801826,
0.05181796848773956,
0.06774567812681198,
0.03500198572874069,
-0.053186893463134766,
0.056685443967580795,
-0.19885419309139252,
0.10022944211959839,
-0.09629994630813599,
0.028232630342245102,
-0.15368616580963135,
-0.016224225983023643,
0.024259883910417557,
0.00603050272911787,
0.023533180356025696,
0.13508757948875427,
-0.2269131988286972,
-0.009413540363311768,
0.1492016613483429,
-0.08191759884357452,
-0.11286741495132446,
0.05882270261645317,
-0.06703686714172363,
0.13632111251354218,
0.024114999920129776,
-0.03846221789717674,
0.05126623064279556,
-0.1477012187242508,
-0.034279413521289825,
-0.027603546157479286,
-0.011836200952529907,
0.11866577714681625,
0.09630073606967926,
-0.0608704648911953,
0.048884205520153046,
0.020479585975408554,
-0.032701265066862106,
-0.042141854763031006,
-0.050704531371593475,
-0.12829554080963135,
0.0009587573586031795,
-0.07328714430332184,
0.04790837690234184,
-0.02088468335568905,
-0.06889110058546066,
-0.018932033330202103,
-0.16518932580947876,
0.002006813418120146,
0.09172286838293076,
0.02033841609954834,
-0.03539799153804779,
-0.10069174319505692,
0.0036235731095075607,
-0.011536587961018085,
-0.035604726523160934,
-0.13578550517559052,
-0.02210777997970581,
0.019318837672472,
-0.13882264494895935,
0.030753053724765778,
-0.07345959544181824,
0.051180385053157806,
0.016524922102689743,
-0.05861951783299446,
-0.010977345518767834,
-0.023012345656752586,
0.024373451247811317,
-0.0456857830286026,
-0.24518829584121704,
-0.01426833588629961,
-0.032443173229694366,
0.1618536114692688,
-0.23377619683742523,
0.038241252303123474,
0.06515999883413315,
0.11937034130096436,
-0.02269211784005165,
-0.050194818526506424,
0.02402755618095398,
-0.0810660794377327,
-0.03478178381919861,
-0.05240238085389137,
-0.0170640479773283,
-0.02249637059867382,
-0.06970936059951782,
0.013335862196981907,
-0.10944215208292007,
-0.04154296964406967,
0.10713886469602585,
0.08292265236377716,
-0.15724287927150726,
-0.043278347700834274,
-0.03408950939774513,
-0.08576270937919617,
-0.08529800176620483,
-0.0566803403198719,
0.13487502932548523,
0.05090935528278351,
0.02855822816491127,
-0.08846847712993622,
-0.07940267771482468,
0.00988192018121481,
-0.03207101300358772,
-0.028083765879273415,
0.10094649344682693,
0.07611845433712006,
-0.10813652724027634,
0.08834784477949142,
0.07578150928020477,
0.012136061675846577,
0.11384404450654984,
-0.011400082148611546,
-0.11351825296878815,
-0.04137531667947769,
0.03633233532309532,
0.002555434126406908,
0.1695048063993454,
-0.09464383870363235,
0.06803114712238312,
0.03927377983927727,
-0.022211823612451553,
0.05476415529847145,
-0.10076725482940674,
0.01427049096673727,
0.006726768799126148,
-0.012228100560605526,
-0.011376895941793919,
-0.036163002252578735,
0.020614514127373695,
0.07891662418842316,
0.03816615790128708,
0.036182720214128494,
0.03572281077504158,
-0.04122483730316162,
-0.1245279312133789,
0.19345727562904358,
-0.10554436594247818,
-0.2273423671722412,
-0.1516016721725464,
0.05401213839650154,
0.03572985157370567,
-0.030572842806577682,
0.008941974490880966,
-0.05140937119722366,
-0.0966159775853157,
-0.08070044219493866,
0.005514310672879219,
0.03883929178118706,
-0.07613059133291245,
-0.07262902706861496,
0.05921752378344536,
0.05427297204732895,
-0.13442036509513855,
0.0406947135925293,
0.054035235196352005,
-0.04148136079311371,
0.008404599502682686,
0.06944910436868668,
0.07862463593482971,
0.15086530148983002,
-0.020428497344255447,
-0.020412612706422806,
0.05437345430254936,
0.2643863558769226,
-0.15086820721626282,
0.09670513868331909,
0.09954504668712616,
-0.06504277884960175,
0.07992210984230042,
0.18344183266162872,
0.033216435462236404,
-0.10660552978515625,
0.045308101922273636,
0.031075740233063698,
-0.0188649483025074,
-0.2811678647994995,
-0.06357815116643906,
0.0033266504760831594,
-0.10220301896333694,
0.062428005039691925,
0.0793466567993164,
0.09731262922286987,
0.04918764531612396,
-0.06440604478120804,
-0.07534892857074738,
0.02199655771255493,
0.07507231831550598,
-0.04625728353857994,
0.0006049389485269785,
0.08203481882810593,
-0.0200007613748312,
0.008962401188910007,
0.11015255749225616,
0.013906295411288738,
0.1873634159564972,
0.04269689694046974,
0.11463924497365952,
0.10168035328388214,
0.10507753491401672,
0.000024342234610230662,
0.015555954538285732,
0.02079109288752079,
0.012282595038414001,
-0.002983907237648964,
-0.08613301068544388,
0.02277722768485546,
0.12184786051511765,
0.06945348531007767,
0.04476168751716614,
0.024970298632979393,
-0.050061535090208054,
0.05980529636144638,
0.1768452227115631,
-0.01209972519427538,
-0.1998264193534851,
-0.062326882034540176,
0.06751304864883423,
-0.082801952958107,
-0.11640139669179916,
-0.02261449582874775,
0.050769247114658356,
-0.17440687119960785,
0.015001747757196426,
-0.04254560545086861,
0.09033802151679993,
-0.09127394109964371,
-0.037229955196380615,
0.05321357026696205,
0.07545126974582672,
-0.023492055013775826,
0.09048163145780563,
-0.17921186983585358,
0.13352392613887787,
0.01737614907324314,
0.06370522826910019,
-0.09815072268247604,
0.10393797606229782,
0.015243546105921268,
-0.0071698566898703575,
0.14627893269062042,
0.008973979391157627,
-0.019879506900906563,
-0.058314017951488495,
-0.10938628017902374,
-0.0015536772552877665,
0.08220188319683075,
-0.11720426380634308,
0.06481732428073883,
0.00044200546108186245,
-0.019408708438277245,
0.010529479943215847,
-0.0697939544916153,
-0.14233455061912537,
-0.1691078543663025,
0.06332679092884064,
-0.12960782647132874,
0.05657918378710747,
-0.10196143388748169,
-0.07344398647546768,
-0.006228356156498194,
0.1857890486717224,
-0.19167372584342957,
-0.0651763305068016,
-0.13295814394950867,
-0.08307469636201859,
0.17686748504638672,
-0.038926977664232254,
0.07132517546415329,
0.017756011337041855,
0.17197521030902863,
0.030676020309329033,
0.013996497727930546,
0.10165295004844666,
-0.0863775908946991,
-0.18250107765197754,
-0.06872538477182388,
0.145328551530838,
0.15727265179157257,
0.04947395995259285,
-0.01222315151244402,
0.0006382534629665315,
-0.05825969576835632,
-0.12492486834526062,
0.00552456034347415,
0.14077237248420715,
0.09738009423017502,
0.015011516399681568,
-0.02072962000966072,
-0.12298290431499481,
-0.06933344155550003,
-0.07234511524438858,
0.010791660286486149,
0.1811780333518982,
-0.06657543778419495,
0.1483541578054428,
0.12124106287956238,
-0.0507206916809082,
-0.18955619633197784,
0.04781363531947136,
0.0678601861000061,
0.021055543795228004,
0.06329847872257233,
-0.1708568036556244,
0.10241113603115082,
0.03779063746333122,
-0.056044332683086395,
0.12532320618629456,
-0.13762390613555908,
-0.15448996424674988,
0.08908607810735703,
0.059379611164331436,
-0.23717626929283142,
-0.10756765305995941,
-0.09208329766988754,
-0.04467558488249779,
-0.11974717676639557,
0.07756773382425308,
-0.008080631494522095,
0.01312070433050394,
0.038425788283348083,
0.04747161641716957,
0.010422809049487114,
-0.04883774369955063,
0.2077513337135315,
0.00663892924785614,
0.03319171071052551,
-0.04891526326537132,
-0.10318257659673691,
0.04049978777766228,
-0.04806138575077057,
0.09715691953897476,
-0.014642413705587387,
0.021955221891403198,
-0.1253223717212677,
-0.0439610481262207,
-0.06654173135757446,
0.030696231871843338,
-0.09619533270597458,
-0.09483709931373596,
-0.05548068508505821,
0.10141977667808533,
0.07960876822471619,
-0.03827962279319763,
-0.018101584166288376,
-0.08076406270265579,
0.028281690552830696,
0.192597895860672,
0.20835207402706146,
0.049149978905916214,
-0.06995424628257751,
0.007349140010774136,
-0.012700160034000874,
0.04521884396672249,
-0.2468501627445221,
0.056316666305065155,
0.04637942090630531,
0.019014067947864532,
0.11265500634908676,
-0.035475291311740875,
-0.16250301897525787,
-0.05557123199105263,
0.07098683714866638,
-0.039137084037065506,
-0.15694621205329895,
-0.024994002655148506,
0.05066932737827301,
-0.20187702775001526,
-0.029669208452105522,
0.010474429465830326,
-0.02148980274796486,
-0.04393318295478821,
0.011044103652238846,
0.08090483397245407,
-0.018578581511974335,
0.1367349922657013,
0.07980240881443024,
0.09522033482789993,
-0.10692083835601807,
0.07168128341436386,
0.06122429668903351,
-0.051465462893247604,
0.021644625812768936,
0.06818753480911255,
-0.04446205869317055,
-0.032580625265836716,
0.07838873565196991,
0.058368146419525146,
0.04023381322622299,
-0.0497741736471653,
-0.009552556090056896,
-0.05499427020549774,
0.049196142703294754,
0.10447074472904205,
0.05076836422085762,
0.0006935194251127541,
0.047793444246053696,
0.018387768417596817,
-0.08049451559782028,
0.10598240047693253,
0.05339374020695686,
0.02360537275671959,
-0.0398079976439476,
-0.03602069616317749,
0.018247995525598526,
-0.010786417871713638,
-0.0149832833558321,
-0.016455529257655144,
-0.07099823653697968,
-0.013593231327831745,
-0.13733075559139252,
0.04016523063182831,
-0.08189219981431961,
0.01841694675385952,
0.022008292376995087,
-0.05440347641706467,
-0.007398437242954969,
0.015957478433847427,
-0.07759089022874832,
-0.04222242161631584,
-0.0045568388886749744,
0.12033451348543167,
-0.11743347346782684,
0.041315708309412,
0.0889706164598465,
-0.10073781758546829,
0.08179357647895813,
0.005519764963537455,
0.006593905854970217,
0.027770070359110832,
-0.18307223916053772,
0.07270024716854095,
-0.02148648537695408,
0.003687589429318905,
0.03217103332281113,
-0.22772879898548126,
-0.010953521355986595,
-0.03648538142442703,
-0.016809485852718353,
0.0019160229712724686,
-0.03937701880931854,
-0.13335061073303223,
0.07287079840898514,
-0.01058956515043974,
-0.08660455048084259,
-0.032185930758714676,
0.03226194903254509,
0.1112515926361084,
-0.03534836322069168,
0.15059389173984528,
-0.005941883195191622,
0.05801843851804733,
-0.17130136489868164,
-0.011426819488406181,
-0.019129110500216484,
0.03652174770832062,
-0.018265437334775925,
-0.014729461632668972,
0.053084973245859146,
-0.03412574157118797,
0.2234855443239212,
-0.03480256348848343,
0.06502514332532883,
0.05183198302984238,
0.02280556410551071,
-0.006614799611270428,
0.08636770397424698,
0.06560425460338593,
-0.01096076425164938,
0.02718065120279789,
0.028059065341949463,
-0.012954981066286564,
-0.037562232464551926,
-0.1630524843931198,
0.05572279915213585,
0.1581650972366333,
0.04094236344099045,
0.011616811156272888,
0.06928509473800659,
-0.10752071440219879,
-0.07898375391960144,
0.1387312412261963,
-0.01259393710643053,
-0.032576363533735275,
-0.07013807445764542,
0.13943122327327728,
0.124080128967762,
-0.19758351147174835,
0.07208021730184555,
-0.0731193795800209,
-0.07801702618598938,
-0.10079838335514069,
-0.14738084375858307,
-0.061444323509931564,
-0.052179500460624695,
-0.011450962163507938,
-0.06768535077571869,
0.05396997556090355,
0.10480605065822601,
0.0069710006937384605,
-0.026146549731492996,
0.10475686937570572,
0.0007574855699203908,
-0.027480410411953926,
0.0275881364941597,
0.06416697055101395,
0.01868068240582943,
-0.10241235792636871,
0.016462087631225586,
0.0009010558133013546,
0.028261849656701088,
0.058421481400728226,
0.0037333546206355095,
-0.035359520465135574,
-0.012541528791189194,
-0.022329136729240417,
-0.11025683581829071,
0.038418930023908615,
-0.031967371702194214,
-0.03549599647521973,
0.11972174793481827,
0.021107889711856842,
0.0024782961700111628,
-0.022964047268033028,
0.22632580995559692,
-0.07606904208660126,
-0.0824858620762825,
-0.1684485524892807,
0.048732075840234756,
-0.06246444582939148,
0.03944636881351471,
0.04816613346338272,
-0.1110905185341835,
0.02492443658411503,
0.13681943714618683,
0.13383808732032776,
-0.017702074721455574,
0.0072706313803792,
0.041554342955350876,
-0.001966990763321519,
-0.051138825714588165,
0.022816691547632217,
0.04751669988036156,
0.09492984414100647,
-0.05958498641848564,
0.09289880096912384,
-0.006714127957820892,
-0.08313115686178207,
0.011414550244808197,
0.11385775357484818,
-0.004354037344455719,
0.008586743846535683,
-0.06612556427717209,
0.14033369719982147,
-0.05520116165280342,
-0.2502851188182831,
0.03959165886044502,
-0.0734434500336647,
-0.16861815750598907,
-0.03511347249150276,
0.018955450505018234,
-0.019131824374198914,
0.017461534589529037,
0.07813186943531036,
-0.05068197101354599,
0.17512299120426178,
0.04293905943632126,
-0.08064883947372437,
-0.06616055220365524,
0.07387921214103699,
-0.11062787473201752,
0.28079262375831604,
0.012751048430800438,
0.06857820600271225,
0.10455191880464554,
-0.016430502757430077,
-0.11872978508472443,
0.042664192616939545,
0.10075171291828156,
-0.07164205610752106,
0.08039859682321548,
0.18360178172588348,
0.0013276869431138039,
0.15462037920951843,
0.06878916919231415,
-0.0453730933368206,
0.03654608130455017,
-0.12163300812244415,
-0.05294680967926979,
-0.10768717527389526,
0.08729486167430878,
-0.07798956334590912,
0.15596513450145721,
0.13275524973869324,
-0.07110930234193802,
-0.006204865872859955,
-0.025767024606466293,
0.08593760430812836,
-0.009336618706583977,
0.1176052987575531,
0.00486786337569356,
-0.20527753233909607,
0.022964732721447945,
0.006658138707280159,
0.10234756767749786,
-0.21353045105934143,
-0.06055140495300293,
0.06063069403171539,
-0.027994666248559952,
-0.050338197499513626,
0.11621229350566864,
0.05960828810930252,
0.04527933895587921,
-0.034697841852903366,
-0.03217756003141403,
-0.02518811635673046,
0.13280846178531647,
-0.11107352375984192,
-0.014744595624506474
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | khanhnto/kyt-dragon-13b | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T15:57:18+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
56,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06061961501836777,
0.15481999516487122,
-0.004844071343541145,
0.02074851468205452,
0.0983177199959755,
0.007407687604427338,
0.07119518518447876,
0.11185134947299957,
-0.023851769044995308,
0.1167980208992958,
0.031993988901376724,
0.09781743586063385,
0.11217817664146423,
0.16186554729938507,
0.0015333457849919796,
-0.22897611558437347,
0.049678247421979904,
-0.125278040766716,
-0.0294334813952446,
0.11977242678403854,
0.1422213912010193,
-0.10954539477825165,
0.0752737894654274,
-0.038042325526475906,
-0.005828251596540213,
-0.0323176346719265,
-0.06205610930919647,
-0.05266609415411949,
0.05311284959316254,
0.06794639676809311,
0.07308239489793777,
0.01171939354389906,
0.09106900542974472,
-0.2724283039569855,
0.02348201349377632,
0.0805930644273758,
-0.0006441773730330169,
0.07586129754781723,
0.04993962123990059,
-0.08749990910291672,
0.07524524629116058,
-0.060156844556331635,
0.1498761922121048,
0.07955671846866608,
-0.09018243104219437,
-0.19217631220817566,
-0.07921334356069565,
0.09916994720697403,
0.1890910118818283,
0.05953684076666832,
-0.026427440345287323,
0.11642678081989288,
-0.08593545109033585,
0.013638701289892197,
0.06446459144353867,
-0.06054406240582466,
-0.055855002254247665,
0.06904532760381699,
0.08335285633802414,
0.08567540347576141,
-0.12976622581481934,
-0.010767064057290554,
0.015032444149255753,
0.008952446281909943,
0.08948688954114914,
0.017146794125437737,
0.1335189938545227,
0.040557652711868286,
-0.13501930236816406,
-0.043155476450920105,
0.09761431813240051,
0.03665134683251381,
-0.04888195917010307,
-0.2485782504081726,
-0.023432478308677673,
-0.04339504987001419,
-0.03198111802339554,
-0.03649339824914932,
0.043764639645814896,
-0.014506848528981209,
0.07738617807626724,
-0.004502781666815281,
-0.0837155357003212,
-0.04301247000694275,
0.07241875678300858,
0.06128999963402748,
0.02571401372551918,
-0.015821760520339012,
0.0059297760017216206,
0.12327717989683151,
0.11431120336055756,
-0.126715749502182,
-0.052547648549079895,
-0.06306339055299759,
-0.08449548482894897,
-0.044861067086458206,
0.030838407576084137,
0.037995077669620514,
0.045936476439237595,
0.23867325484752655,
0.007765117567032576,
0.053257301449775696,
0.04455438256263733,
0.014407169073820114,
0.06501194834709167,
0.11008983850479126,
-0.05894824117422104,
-0.09719445556402206,
-0.028582042083144188,
0.10156717151403427,
0.007986726239323616,
-0.04139331728219986,
-0.05712985619902611,
0.07059531658887863,
0.018587570637464523,
0.12360043078660965,
0.08000938594341278,
0.003056557849049568,
-0.0755772516131401,
-0.062465377151966095,
0.17764076590538025,
-0.15825673937797546,
0.04532013460993767,
0.03055616281926632,
-0.0341108962893486,
-0.009745313785970211,
0.012105142697691917,
0.025474950671195984,
-0.021481726318597794,
0.09522198140621185,
-0.05601342022418976,
-0.034448131918907166,
-0.11389608681201935,
-0.03694311901926994,
0.030394554138183594,
0.011153047904372215,
-0.02865210548043251,
-0.03502652049064636,
-0.08865131437778473,
-0.06405586749315262,
0.09101516753435135,
-0.07148737460374832,
-0.04784895107150078,
-0.016645915806293488,
-0.07833752781152725,
0.021804187446832657,
0.01691517047584057,
0.09064167737960815,
-0.0222476739436388,
0.03985358029603958,
-0.0550384595990181,
0.061440225690603256,
0.11723454296588898,
0.027987057343125343,
-0.05787884071469307,
0.061519939452409744,
-0.2424532175064087,
0.10252492874860764,
-0.07715212553739548,
0.04971238598227501,
-0.15203025937080383,
-0.02478341944515705,
0.03986154496669769,
0.01284773275256157,
-0.008251311257481575,
0.14196595549583435,
-0.21994100511074066,
-0.030957341194152832,
0.16964265704154968,
-0.10025953501462936,
-0.08109250664710999,
0.060782887041568756,
-0.05354252830147743,
0.11210215091705322,
0.04557164013385773,
-0.02375967986881733,
0.05775221437215805,
-0.14725260436534882,
-0.011030761525034904,
-0.041942402720451355,
-0.0180682260543108,
0.16207332909107208,
0.0703711211681366,
-0.06047816202044487,
0.07456906884908676,
0.01960151270031929,
-0.014246034435927868,
-0.04887177795171738,
-0.02822130173444748,
-0.1047162413597107,
0.01184528972953558,
-0.06102835759520531,
0.018109694123268127,
-0.021768750622868538,
-0.09445013850927353,
-0.029118487611413002,
-0.17402999103069305,
-0.0031633328180760145,
0.08821269869804382,
-0.011630427092313766,
-0.021509924903512,
-0.11245372891426086,
0.009332616813480854,
0.030967719852924347,
0.0002618339203763753,
-0.13677829504013062,
-0.06033218279480934,
0.026970699429512024,
-0.16097871959209442,
0.029791243374347687,
-0.05741601809859276,
0.04530094936490059,
0.04005871340632439,
-0.03433511033654213,
-0.03489551320672035,
0.010874404571950436,
0.010431389324367046,
-0.01894843392074108,
-0.25422003865242004,
-0.01882786676287651,
-0.0234990194439888,
0.1751047968864441,
-0.22956320643424988,
0.042598169296979904,
0.07489731162786484,
0.1460893303155899,
0.007349682506173849,
-0.03550100699067116,
0.015185600146651268,
-0.07262228429317474,
-0.03268764168024063,
-0.06316669285297394,
-0.01207790058106184,
-0.038400664925575256,
-0.05820201337337494,
0.04906858503818512,
-0.1686294972896576,
-0.030321966856718063,
0.10717973858118057,
0.06342670321464539,
-0.1473218947649002,
-0.02780107781291008,
-0.04056945815682411,
-0.04624456167221069,
-0.06676914542913437,
-0.05461418256163597,
0.11812574416399002,
0.056411582976579666,
0.04860803112387657,
-0.07140495628118515,
-0.07455260306596756,
0.008036690764129162,
-0.01956399530172348,
-0.014917809516191483,
0.09334591031074524,
0.07554110884666443,
-0.12264352291822433,
0.09177418053150177,
0.09668384492397308,
0.08576478064060211,
0.10314212739467621,
-0.014663571491837502,
-0.08914592862129211,
-0.040637146681547165,
0.02245822176337242,
0.016187267377972603,
0.15129362046718597,
-0.012961224652826786,
0.055492039769887924,
0.0358695350587368,
-0.014034898020327091,
0.011105312965810299,
-0.09736533463001251,
0.02655916102230549,
0.030835967510938644,
-0.016302183270454407,
0.03745110332965851,
-0.0447014644742012,
0.019208140671253204,
0.09039704501628876,
0.040895868092775345,
0.040978945791721344,
0.010155045427381992,
-0.04354988783597946,
-0.11037563532590866,
0.1787576973438263,
-0.12389461696147919,
-0.24818050861358643,
-0.13812170922756195,
0.010281167924404144,
0.04737642779946327,
-0.010411068797111511,
0.006690691225230694,
-0.06616118550300598,
-0.1175973042845726,
-0.09878289699554443,
0.018617089837789536,
0.045352302491664886,
-0.07590975612401962,
-0.06842505931854248,
0.06414616107940674,
0.03875524550676346,
-0.13939815759658813,
0.024007495492696762,
0.04662325978279114,
-0.08205481618642807,
-0.0029386086389422417,
0.0791812464594841,
0.06965780258178711,
0.17661017179489136,
0.013885351829230785,
-0.023669935762882233,
0.026634456589818,
0.20819635689258575,
-0.1436755359172821,
0.10975687950849533,
0.13545554876327515,
-0.08767466992139816,
0.08120133727788925,
0.1998777538537979,
0.03777998685836792,
-0.10680917650461197,
0.03608465939760208,
0.028374753892421722,
-0.028325283899903297,
-0.2502254545688629,
-0.06958996504545212,
0.0019060121849179268,
-0.05172049254179001,
0.07064855098724365,
0.08791537582874298,
0.09593888372182846,
0.016860228031873703,
-0.09976044297218323,
-0.07697858661413193,
0.046900223940610886,
0.10824491083621979,
-0.00015424020239152014,
-0.015208319760859013,
0.0904119610786438,
-0.03033481352031231,
0.01743943803012371,
0.09215071052312851,
0.0030607767403125763,
0.17535938322544098,
0.051709048449993134,
0.17189906537532806,
0.07866133749485016,
0.06444311141967773,
0.02004685252904892,
0.007725914940237999,
0.021817529574036598,
0.017227526754140854,
-0.0030957073904573917,
-0.08709781616926193,
-0.0034981227945536375,
0.1202581599354744,
0.049845851957798004,
0.029173865914344788,
0.012042860500514507,
-0.030704669654369354,
0.08337877690792084,
0.1770893782377243,
0.0029054484330117702,
-0.1893385946750641,
-0.07169844210147858,
0.07795937359333038,
-0.08648337423801422,
-0.10729733109474182,
-0.029470939189195633,
0.041069481521844864,
-0.1729043871164322,
0.016882894560694695,
-0.019335895776748657,
0.10788324475288391,
-0.13190391659736633,
-0.01772487722337246,
0.05657728388905525,
0.06932812184095383,
-0.009677323512732983,
0.06694949418306351,
-0.16090403497219086,
0.11770165711641312,
0.01751571334898472,
0.06636732816696167,
-0.09608277678489685,
0.09618937969207764,
-0.007830657996237278,
0.0041499207727611065,
0.1410749852657318,
0.010120149701833725,
-0.05952107161283493,
-0.09608154743909836,
-0.10546442121267319,
-0.009841260500252247,
0.1306990385055542,
-0.14852415025234222,
0.08813067525625229,
-0.02661319263279438,
-0.044553373008966446,
0.003614129964262247,
-0.12497276812791824,
-0.13103094696998596,
-0.18366187810897827,
0.05707118660211563,
-0.12947207689285278,
0.04045100137591362,
-0.10902881622314453,
-0.045833900570869446,
-0.02098964899778366,
0.20040063560009003,
-0.23137451708316803,
-0.06714103370904922,
-0.1551055610179901,
-0.08061286807060242,
0.14446212351322174,
-0.046455029398202896,
0.08550118654966354,
0.0008278203313238919,
0.19068008661270142,
0.021319707855582237,
-0.017237508669495583,
0.1072206199169159,
-0.10052918642759323,
-0.2010865956544876,
-0.09273224323987961,
0.15895552933216095,
0.13766798377037048,
0.03809428587555885,
-0.004381525795906782,
0.03171157464385033,
-0.02098114788532257,
-0.12076930701732635,
0.020226983353495598,
0.17317426204681396,
0.08982043713331223,
0.025265544652938843,
-0.02972041629254818,
-0.11267432570457458,
-0.07061342149972916,
-0.03774050623178482,
0.024755435064435005,
0.18072067201137543,
-0.07222156971693039,
0.18405316770076752,
0.13775517046451569,
-0.05534014105796814,
-0.19904261827468872,
0.021996473893523216,
0.04293542355298996,
0.0070380112156271935,
0.0323902890086174,
-0.20307663083076477,
0.09384101629257202,
0.0008334947633557022,
-0.05131231248378754,
0.1379684954881668,
-0.1823476254940033,
-0.151598259806633,
0.06042521819472313,
0.043563615530729294,
-0.19374065101146698,
-0.12374074012041092,
-0.08848230540752411,
-0.04693066328763962,
-0.15487661957740784,
0.10312657803297043,
0.0020827590487897396,
0.008401188999414444,
0.03778626397252083,
0.02252252586185932,
0.012139533646404743,
-0.04198719933629036,
0.1914343535900116,
-0.025891713798046112,
0.03347287327051163,
-0.0790715217590332,
-0.060851071029901505,
0.062408581376075745,
-0.058187782764434814,
0.0755455270409584,
-0.025226406753063202,
0.015947066247463226,
-0.10598332434892654,
-0.048235729336738586,
-0.02852320298552513,
0.019321219995617867,
-0.09431382268667221,
-0.09348297864198685,
-0.04829427972435951,
0.09367614984512329,
0.09042316675186157,
-0.03652578964829445,
-0.03649144619703293,
-0.078715980052948,
0.038977332413196564,
0.17627815902233124,
0.18159319460391998,
0.04659178853034973,
-0.07959239184856415,
-0.001915142871439457,
-0.014336181804537773,
0.04684065282344818,
-0.22077152132987976,
0.060553863644599915,
0.04557652771472931,
0.016117896884679794,
0.11537692695856094,
-0.0208132341504097,
-0.16198977828025818,
-0.06710557639598846,
0.061360616236925125,
-0.06944561004638672,
-0.17825035750865936,
0.0039279889315366745,
0.07344977557659149,
-0.16578389704227448,
-0.037031736224889755,
0.04200848564505577,
-0.01189455483108759,
-0.0403641052544117,
0.012352054007351398,
0.08063354343175888,
0.007078902795910835,
0.07699975371360779,
0.055281639099121094,
0.09124495089054108,
-0.10227900743484497,
0.07410510629415512,
0.08149529248476028,
-0.08644098788499832,
0.030720343813300133,
0.09573426842689514,
-0.06469762325286865,
-0.0346054881811142,
0.04237886518239975,
0.08354541659355164,
0.024281201884150505,
-0.04682289808988571,
0.0023111123591661453,
-0.09734189510345459,
0.05927345156669617,
0.11483542621135712,
0.03496333956718445,
0.011234734207391739,
0.03813567012548447,
0.04486291855573654,
-0.08093374222517014,
0.11926916986703873,
0.023795632645487785,
0.020354853942990303,
-0.04112942889332771,
-0.040553025901317596,
0.035851649940013885,
-0.026020776480436325,
-0.011440055444836617,
-0.035174157470464706,
-0.0722682997584343,
-0.014069457538425922,
-0.16000694036483765,
-0.0076758842915296555,
-0.03660871088504791,
0.005114538595080376,
0.022510098293423653,
-0.03652830421924591,
0.00792311318218708,
0.012217256240546703,
-0.06868947297334671,
-0.05553458258509636,
-0.023233558982610703,
0.09422210603952408,
-0.16494666039943695,
0.0220257006585598,
0.0823851153254509,
-0.12121747434139252,
0.09289738535881042,
0.016782134771347046,
0.00412249518558383,
0.026962365955114365,
-0.1545863002538681,
0.04763968288898468,
-0.020152103155851364,
0.013473534025251865,
0.04222847521305084,
-0.21637047827243805,
-0.004404853098094463,
-0.04015503451228142,
-0.05566934496164322,
-0.008993052877485752,
-0.0319182425737381,
-0.11338426172733307,
0.09645436704158783,
0.011025024577975273,
-0.08443772792816162,
-0.02965564839541912,
0.03353232145309448,
0.07690354436635971,
-0.027447547763586044,
0.1498211771249771,
-0.004663881380110979,
0.07559948414564133,
-0.17581342160701752,
-0.02282017655670643,
-0.011197620071470737,
0.022367527708411217,
-0.021871577948331833,
-0.01622559316456318,
0.04623444378376007,
-0.02704801969230175,
0.19120801985263824,
-0.024701936170458794,
0.049393873661756516,
0.06364397704601288,
0.009232889860868454,
-0.013832193799316883,
0.11151392012834549,
0.05708572641015053,
0.024334950372576714,
0.022262847051024437,
0.003451440716162324,
-0.04008655622601509,
-0.009981024079024792,
-0.18596695363521576,
0.06803664565086365,
0.14585918188095093,
0.09060460329055786,
-0.012669353745877743,
0.0707244873046875,
-0.10161512345075607,
-0.12005364894866943,
0.10127941519021988,
-0.06415384262800217,
-0.010188822634518147,
-0.06542414426803589,
0.14027701318264008,
0.14953285455703735,
-0.1886233240365982,
0.06583356112241745,
-0.06602055579423904,
-0.0566304549574852,
-0.11457879096269608,
-0.1930263340473175,
-0.057075321674346924,
-0.050602465867996216,
-0.018466074019670486,
-0.05384097993373871,
0.06939727067947388,
0.05750798434019089,
0.01126816775649786,
0.00868057832121849,
0.08568526059389114,
-0.009656033478677273,
0.00248199631460011,
0.030120067298412323,
0.06713981181383133,
0.016768986359238625,
-0.0321255661547184,
0.0179112758487463,
-0.00597198773175478,
0.034156378358602524,
0.059282708913087845,
0.03608176112174988,
-0.028436895459890366,
0.015559280291199684,
-0.034912437200546265,
-0.11309733241796494,
0.042801856994628906,
-0.029640642926096916,
-0.0749855786561966,
0.1347348988056183,
0.026981467381119728,
0.005015076603740454,
-0.023140020668506622,
0.2503887414932251,
-0.07436972856521606,
-0.09334370493888855,
-0.14373961091041565,
0.11701542884111404,
-0.04212593287229538,
0.0635172426700592,
0.03596310690045357,
-0.10810714215040207,
0.017985546961426735,
0.1320217251777649,
0.15442703664302826,
-0.04732590913772583,
0.019251897931098938,
0.028577854856848717,
0.00439635943621397,
-0.04075566306710243,
0.05177190154790878,
0.07100846618413925,
0.14500564336776733,
-0.05157303810119629,
0.08530787378549576,
0.002609728369861841,
-0.1021018698811531,
-0.041973695158958435,
0.11415864527225494,
-0.014296893030405045,
0.017620453611016273,
-0.057136841118335724,
0.124222531914711,
-0.05874236673116684,
-0.23697422444820404,
0.06316976249217987,
-0.0765061303973198,
-0.1432730257511139,
-0.024886758998036385,
0.071670763194561,
-0.016632623970508575,
0.02605951391160488,
0.07167234271764755,
-0.0754380151629448,
0.18880942463874817,
0.03957989811897278,
-0.05233397334814072,
-0.05954399332404137,
0.0744764655828476,
-0.11850855499505997,
0.27879106998443604,
0.010482731275260448,
0.051307905465364456,
0.1042102724313736,
-0.02021743729710579,
-0.13270841538906097,
0.023401619866490364,
0.09579801559448242,
-0.08917027711868286,
0.04087764397263527,
0.21448291838169098,
-0.00629545608535409,
0.11935057491064072,
0.07611140608787537,
-0.07468950748443604,
0.047562725841999054,
-0.11468592286109924,
-0.07639975845813751,
-0.08699081838130951,
0.09244474768638611,
-0.06785612553358078,
0.14258281886577606,
0.12599852681159973,
-0.05530165135860443,
0.011584274470806122,
-0.028389399871230125,
0.045467376708984375,
0.005578654818236828,
0.100032277405262,
0.011115525849163532,
-0.18496567010879517,
0.024811718612909317,
0.016259413212537766,
0.10884406417608261,
-0.18112654983997345,
-0.09105053544044495,
0.046958595514297485,
0.0005061255069449544,
-0.06443515419960022,
0.12483241409063339,
0.057313691824674606,
0.04654949903488159,
-0.0451689288020134,
-0.026830285787582397,
-0.006042256020009518,
0.14264579117298126,
-0.10707559436559677,
-0.005129707511514425
] |
null | null | stable-baselines3 |
# **A2C** Agent playing **PandaReachDense-v3**
This is a trained model of a **A2C** agent playing **PandaReachDense-v3**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["PandaReachDense-v3", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "A2C", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "PandaReachDense-v3", "type": "PandaReachDense-v3"}, "metrics": [{"type": "mean_reward", "value": "-0.28 +/- 0.10", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | bmschopp/a2c-PandaReachDense-v3 | [
"stable-baselines3",
"PandaReachDense-v3",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-08T16:04:52+00:00 | [] | [] | TAGS
#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# A2C Agent playing PandaReachDense-v3
This is a trained model of a A2C agent playing PandaReachDense-v3
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
41,
45,
17
] | [
"passage: TAGS\n#stable-baselines3 #PandaReachDense-v3 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# A2C Agent playing PandaReachDense-v3\nThis is a trained model of a A2C agent playing PandaReachDense-v3\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.028780510649085045,
0.06549051403999329,
-0.004174588713794947,
0.028733979910612106,
0.12748076021671295,
-0.010029550641775131,
0.16130082309246063,
0.07903143763542175,
0.052706290036439896,
-0.055043965578079224,
0.09157051891088486,
-0.079488605260849,
0.04699381813406944,
0.3393711447715759,
0.029525093734264374,
-0.186785027384758,
0.08573613315820694,
0.015584449283778667,
0.018966808915138245,
0.09867662936449051,
0.03466832637786865,
-0.08736564218997955,
0.04568251967430115,
0.03800429776310921,
-0.07686931639909744,
-0.04319252818822861,
-0.03975098207592964,
-0.06744661927223206,
0.10361767560243607,
-0.044310007244348526,
0.1670169234275818,
-0.03489987552165985,
0.10219604521989822,
-0.12577489018440247,
0.031373992562294006,
-0.04813149571418762,
-0.05141052231192589,
0.002818689215928316,
-0.011371237225830555,
0.05937984213232994,
0.04167760908603668,
0.05197896435856819,
0.07366002351045609,
0.04871916025876999,
-0.08704962581396103,
-0.11396265029907227,
-0.006845315918326378,
0.07931416481733322,
0.17974808812141418,
0.04054044932126999,
-0.02474738284945488,
0.09696658700704575,
-0.11350683122873306,
0.01657135598361492,
-0.019304286688566208,
-0.4018571078777313,
0.006876560393720865,
0.15550047159194946,
0.04677277058362961,
0.010903568007051945,
-0.0061170910485088825,
-0.004642391111701727,
0.02805398777127266,
-0.037410516291856766,
0.08670840412378311,
-0.09000635892152786,
0.06153826415538788,
-0.019131680950522423,
-0.04113767296075821,
-0.01751464419066906,
0.2419518232345581,
0.01633240468800068,
-0.08024721592664719,
-0.07922019064426422,
0.009968155063688755,
-0.028026137501001358,
-0.0877801775932312,
-0.06134319305419922,
0.07644549012184143,
0.057131536304950714,
0.10696670413017273,
-0.030399860814213753,
-0.058683689683675766,
-0.04541248828172684,
0.08352918922901154,
-0.03953780233860016,
-0.017566127702593803,
-0.01754307933151722,
-0.06739802658557892,
-0.003707833355292678,
0.015629740431904793,
-0.06615205854177475,
-0.015486059710383415,
-0.044966671615839005,
-0.1556774228811264,
-0.009128551930189133,
-0.0599384643137455,
0.03310214728116989,
0.10073909163475037,
0.13065455853939056,
0.06838785856962204,
0.09685135632753372,
-0.08001106232404709,
0.0389438234269619,
0.06625691801309586,
0.09461154788732529,
-0.044509198516607285,
-0.011874453164637089,
0.14630302786827087,
0.10327376425266266,
0.09657767415046692,
-0.09182082861661911,
-0.12403369694948196,
0.04173071309924126,
0.10965418070554733,
0.03382069617509842,
0.0046537998132407665,
0.04452834278345108,
-0.14144757390022278,
0.023916395381093025,
0.0006972529226914048,
-0.045244041830301285,
-0.03088594414293766,
0.06111180782318115,
-0.04433412477374077,
0.02348744124174118,
-0.012718633748590946,
0.10830001533031464,
0.10152670741081238,
-0.023899899795651436,
-0.052799396216869354,
-0.04201658070087433,
-0.0440504252910614,
-0.05507666990160942,
0.04012975096702576,
0.01289378758519888,
0.04624854028224945,
-0.1184653639793396,
-0.13997629284858704,
0.051258668303489685,
0.019622454419732094,
-0.026321161538362503,
-0.13472233712673187,
-0.09338399767875671,
-0.03747362270951271,
-0.011210841126739979,
0.0030350966844707727,
-0.19588395953178406,
-0.02434816211462021,
-0.03428230062127113,
0.13725687563419342,
0.10810749977827072,
-0.06433141976594925,
-0.06369391083717346,
-0.12834231555461884,
0.06795675307512283,
-0.23485252261161804,
0.038750845938920975,
-0.09932064265012741,
0.12411006540060043,
0.007471752353012562,
0.023616313934326172,
0.1410844624042511,
0.02330038882791996,
0.004575210623443127,
0.1702503114938736,
-0.18833371996879578,
-0.046672217547893524,
0.17527204751968384,
-0.0857074186205864,
-0.17703735828399658,
0.05021136254072189,
-0.02124672941863537,
-0.013779462315142155,
0.06350992619991302,
0.09937554597854614,
-0.01727774553000927,
-0.17061583697795868,
0.02558896690607071,
-0.0014508399181067944,
-0.05959303304553032,
0.021542999893426895,
0.12072649598121643,
0.08040176331996918,
-0.027203790843486786,
-0.0016989230643957853,
-0.15452547371387482,
0.09701786935329437,
-0.023543400689959526,
-0.08447092026472092,
0.022736359387636185,
-0.10411997884511948,
0.10016260296106339,
-0.015677137300372124,
0.10591494292020798,
-0.02265925332903862,
-0.018805475905537605,
-0.032891299575567245,
0.10408006608486176,
-0.0068649593740701675,
0.039593957364559174,
-0.17728297412395477,
0.1326225996017456,
0.02176543138921261,
0.046730607748031616,
-0.10109715908765793,
-0.10202061384916306,
0.06674831360578537,
0.15375585854053497,
0.05606463924050331,
0.03833417221903801,
0.07328703999519348,
0.03443831577897072,
-0.0030986627098172903,
-0.1205538883805275,
-0.12789975106716156,
0.019881807267665863,
0.06068658083677292,
-0.08039596676826477,
-0.05172275751829147,
-0.10460081696510315,
0.21138279139995575,
-0.10705634206533432,
0.012047823518514633,
-0.09333895146846771,
0.010153836570680141,
0.08388294279575348,
0.01348812971264124,
0.08132237941026688,
0.02585482969880104,
-0.04426883906126022,
0.009419471956789494,
0.0882885605096817,
0.044275086373090744,
-0.1379590630531311,
0.03784618154168129,
0.024114131927490234,
0.23272188007831573,
0.15174852311611176,
-0.016499420627951622,
-0.055556558072566986,
0.006534850224852562,
0.03740030899643898,
0.03533044084906578,
0.034956689924001694,
0.06951800733804703,
0.1090264692902565,
0.07713755965232849,
0.1276414394378662,
-0.05066131055355072,
0.17763042449951172,
-0.006530070677399635,
-0.14888496696949005,
0.02993084490299225,
-0.07033783197402954,
0.0941668227314949,
-0.06030277907848358,
0.048379335552453995,
0.05410725995898247,
0.0304675605148077,
0.08504439890384674,
-0.00693494314327836,
0.022639812901616096,
-0.04341154545545578,
0.04943868890404701,
0.06790532171726227,
0.06545940041542053,
0.06452376395463943,
-0.007423467002809048,
0.015456308610737324,
-0.05288444459438324,
-0.0518295019865036,
-0.10519610345363617,
-0.12370408326387405,
0.037892695516347885,
-0.015912096947431564,
-0.04463989660143852,
-0.01629551686346531,
-0.07266248762607574,
0.050321705639362335,
0.05250744894146919,
-0.07199236750602722,
0.028561361134052277,
-0.007090074475854635,
-0.09633425623178482,
0.1130511462688446,
-0.14269201457500458,
-0.31355980038642883,
-0.02000165916979313,
-0.13154496252536774,
-0.02077566273510456,
0.15819574892520905,
-0.057956792414188385,
-0.1681092083454132,
0.03305667266249657,
-0.02401961199939251,
-0.09238096326589584,
0.04225420579314232,
-0.018061356619000435,
0.10221174359321594,
0.0857708528637886,
0.043082691729068756,
0.00862243864685297,
-0.01184127852320671,
-0.03903079405426979,
-0.08788500726222992,
0.07608162611722946,
-0.06721128523349762,
0.1173204705119133,
0.13519366085529327,
0.04123268276453018,
-0.015909500420093536,
-0.02043113484978676,
0.06215733662247658,
0.012027861550450325,
-0.036599598824977875,
0.13453175127506256,
-0.03608042374253273,
-0.00864011887460947,
0.04470202699303627,
0.008029532618820667,
-0.10533943772315979,
0.09432658553123474,
-0.05022074654698372,
-0.06974482536315918,
-0.017500806599855423,
-0.08790571242570877,
-0.09950723499059677,
0.18995612859725952,
0.0490412712097168,
0.007856572046875954,
-0.05151839926838875,
0.036120012402534485,
0.07772433012723923,
0.044773608446121216,
0.007161281071603298,
0.03985898196697235,
-0.005716364365071058,
-0.013170693069696426,
0.05278664082288742,
-0.023887991905212402,
0.009960537776350975,
-0.007844919338822365,
0.13077811896800995,
-0.015673788264393806,
0.10317149013280869,
0.0030158995650708675,
0.008619097992777824,
0.08018261194229126,
0.12394148856401443,
0.08064290136098862,
0.019240466877818108,
-0.11554506421089172,
-0.04732639715075493,
-0.030522609129548073,
-0.18181301653385162,
0.11669926345348358,
0.10738886147737503,
0.05268440023064613,
-0.05564067140221596,
0.22832486033439636,
0.0012100599706172943,
0.10802210867404938,
0.03496129810810089,
-0.17664514482021332,
0.024751557037234306,
0.03574612736701965,
0.050895314663648605,
0.007034227252006531,
0.062039270997047424,
-0.09453237801790237,
-0.1839483082294464,
0.03968557342886925,
0.018860090523958206,
0.05523261800408363,
-0.018427258357405663,
0.018512532114982605,
-0.12044285237789154,
-0.05746040865778923,
0.02161633037030697,
0.02076297253370285,
-0.3029120862483978,
0.06816349923610687,
-0.04133946821093559,
0.07392577081918716,
0.009542034938931465,
0.01343793235719204,
0.06604447960853577,
0.01652485318481922,
0.1375029981136322,
-0.017935138195753098,
0.1707022786140442,
-0.1572514772415161,
-0.16084668040275574,
0.025680551305413246,
-0.059293005615472794,
0.07245437800884247,
0.082563117146492,
0.017692390829324722,
0.0069250138476490974,
-0.00047057756455615163,
0.20794180035591125,
-0.13032017648220062,
-0.0346711240708828,
-0.035274047404527664,
0.019543148577213287,
0.022580156102776527,
-0.03844551369547844,
-0.021310672163963318,
0.06112392246723175,
0.1489492505788803,
0.07546767592430115,
-0.02780069410800934,
-0.04611911624670029,
-0.03938353434205055,
-0.09507237374782562,
-0.044778671115636826,
0.10472412407398224,
-0.07841785997152328,
0.10144548118114471,
-0.07513871043920517,
-0.04432075098156929,
0.11707907915115356,
-0.09250949323177338,
-0.053160861134529114,
-0.07627046853303909,
0.05462219938635826,
0.008296831510961056,
0.13374868035316467,
0.03642493113875389,
0.02114485390484333,
0.10089845955371857,
-0.05001259222626686,
0.08662480860948563,
0.03777577355504036,
-0.03541218861937523,
0.03517242521047592,
-0.05375073477625847,
-0.04829130321741104,
-0.010828596539795399,
0.03814345970749855,
0.24244728684425354,
0.302570104598999,
-0.012830551713705063,
0.1897524893283844,
0.09193363785743713,
0.029696941375732422,
-0.16292639076709747,
-0.1200476586818695,
0.05548451840877533,
0.059938978403806686,
0.06154406815767288,
-0.2788083851337433,
0.057189684361219406,
-0.053967077285051346,
-0.08999616652727127,
-0.06829255819320679,
-0.08560561388731003,
-0.07613074034452438,
0.088682159781456,
0.08794322609901428,
0.09100460261106491,
-0.12551987171173096,
0.015924450010061264,
-0.012671655975282192,
-0.1664767563343048,
0.12128932029008865,
-0.039350032806396484,
0.07007917016744614,
-0.025050386786460876,
-0.06438229978084564,
0.025165842846035957,
-0.02775278501212597,
0.04424511641263962,
-0.1206880658864975,
0.0005293674184940755,
-0.04527926817536354,
-0.03749620169401169,
0.1088484600186348,
0.020565982908010483,
-0.0028168195858597755,
-0.09558401256799698,
-0.011945599690079689,
-0.3103867173194885,
0.01988539844751358,
0.02114551141858101,
-0.039148375391960144,
-0.0012507046340033412,
-0.08678091317415237,
-0.042053963989019394,
0.10508828610181808,
0.03930897265672684,
0.08641290664672852,
0.15335260331630707,
-0.005581455305218697,
-0.021082017570734024,
0.17506572604179382,
0.05701295658946037,
-0.014002309180796146,
0.10069113969802856,
-0.06732672452926636,
-0.06576105207204819,
0.04418903961777687,
-0.1016126498579979,
-0.005435575265437365,
0.005642053205519915,
-0.007821558974683285,
0.07107745110988617,
0.09962856024503708,
-0.03340476378798485,
0.18194207549095154,
0.09798844903707504,
-0.15048468112945557,
0.0030947427731007338,
0.052597809582948685,
-0.032650984823703766,
0.04424609988927841,
-0.04443032294511795,
0.05541829764842987,
-0.07521786540746689,
-0.03790169581770897,
0.02031708136200905,
-0.01010141521692276,
-0.07618512213230133,
0.00011962707503698766,
0.03176301345229149,
0.029956085607409477,
-0.08340912312269211,
0.14036758244037628,
0.016359949484467506,
0.0652431845664978,
0.11902019381523132,
0.019259776920080185,
-0.10460162162780762,
-0.014167122542858124,
-0.02339506521821022,
0.2028627097606659,
-0.007937151938676834,
-0.018536100164055824,
-0.11391238868236542,
-0.12847240269184113,
0.018047582358121872,
-0.10348039865493774,
0.10282431542873383,
-0.052032727748155594,
-0.06570395082235336,
-0.03704213351011276,
-0.05561172217130661,
0.031932998448610306,
0.017090078443288803,
-0.015642894431948662,
-0.16111870110034943,
-0.04170334339141846,
0.06846143305301666,
0.039452772587537766,
-0.06145704537630081,
-0.06289087235927582,
-0.16302458941936493,
0.03506235405802727,
-0.1278870701789856,
0.0010145133128389716,
-0.047339316457509995,
-0.05002537742257118,
-0.05195476487278938,
0.01521157007664442,
-0.0177876316010952,
0.008817745372653008,
-0.05148332938551903,
0.03292781487107277,
0.011250603944063187,
0.0014076961670070887,
-0.06952075660228729,
-0.04419080913066864,
0.032172493636608124,
-0.04430563375353813,
0.0661356970667839,
0.04131564497947693,
-0.005653871223330498,
0.021474739536643028,
-0.07005896419286728,
-0.10248169302940369,
0.10313672572374344,
-0.014939527027308941,
0.050572704523801804,
-0.0603681318461895,
-0.012018447741866112,
0.007195405196398497,
-0.07569561898708344,
-0.007751014549285173,
0.24328774213790894,
-0.010914106853306293,
-0.05394120141863823,
-0.07426224648952484,
-0.036970075219869614,
-0.09100507944822311,
-0.0004900419735349715,
0.1948854625225067,
0.05477539822459221,
0.14600017666816711,
-0.0532439760863781,
0.08785777539014816,
-0.06481330841779709,
-0.01534446980804205,
-0.08259234577417374,
0.030320849269628525,
-0.157977893948555,
-0.08130980283021927,
-0.028043894097208977,
-0.03728124126791954,
0.13441862165927887,
-0.19242097437381744,
0.0032852457370609045,
-0.010904400609433651,
-0.04910553991794586,
0.11381126195192337,
0.0557032972574234,
0.24474471807479858,
0.1050342544913292,
-0.035265225917100906,
0.10503548383712769,
0.12215624749660492,
0.0929517149925232,
-0.03347417712211609,
0.058777112513780594,
-0.05078745633363724,
-0.0868106484413147,
0.09736774861812592,
0.012061800807714462,
0.036776214838027954,
-0.08157306164503098,
0.022900743409991264,
-0.10047483444213867,
0.002025678288191557,
0.02005080319941044,
0.2473200410604477,
0.1967000812292099,
-0.09632564336061478,
-0.012216159142553806,
-0.05708231031894684,
-0.032561756670475006,
-0.04091155156493187,
-0.002459051087498665,
-0.07821618020534515,
-0.21873407065868378,
0.051539067178964615,
-0.0930585265159607,
-0.07632365822792053,
-0.06189138814806938,
-0.04064059257507324,
-0.02870149537920952,
0.046939339488744736,
0.03212931379675865,
0.04136762022972107,
0.05070297420024872,
-0.0371626541018486,
-0.09345480799674988,
0.06879863888025284,
-0.11172787100076675,
-0.042014576494693756,
-0.03408866748213768,
0.014045859687030315,
0.032319605350494385,
-0.07429610192775726,
0.07487598061561584,
-0.012149554677307606,
-0.07710553705692291,
0.036456044763326645,
-0.03482281416654587,
0.02153356932103634,
0.07482071220874786,
0.04184282198548317,
-0.09644174575805664,
0.015602846629917622,
0.18867559731006622,
0.020273970440030098,
0.008802177384495735,
-0.14742465317249298,
0.2000039666891098,
-0.02619965374469757,
0.07266447693109512,
-0.03337041288614273,
-0.015141828916966915,
-0.10115411877632141,
0.19129611551761627,
0.11998134851455688,
-0.24376079440116882,
0.024953339248895645,
-0.12912821769714355,
0.022151969373226166,
-0.13376696407794952,
0.20840151607990265,
0.05465596541762352,
0.10847201198339462,
-0.06020665541291237,
-0.02479162998497486,
-0.1493310034275055,
-0.09408020973205566,
-0.08478302508592606,
-0.0414455346763134,
0.10249399393796921,
0.0031611735466867685,
-0.05072701349854469,
-0.00887248944491148,
-0.1566619724035263,
0.10201162099838257,
-0.048264030367136,
-0.11855816096067429,
-0.0679796114563942,
-0.059141192585229874,
-0.06102965027093887,
0.11088541150093079,
0.11637356877326965,
-0.01684124954044819,
0.024554423987865448,
-0.07280154526233673,
-0.012559473514556885,
0.011003518477082253,
0.005383014678955078,
0.0626269057393074,
-0.04783647879958153,
0.1594477891921997,
-0.021524829789996147,
0.0008918871753849089,
0.04285505786538124,
0.05263057351112366,
-0.07584847509860992,
0.06380704790353775,
0.02512199431657791,
0.028178859502077103,
-0.006920731160789728,
0.059795111417770386,
-0.0196672473102808,
0.08964395523071289,
0.08038042485713959,
-0.007235884666442871,
0.09868589043617249,
-0.03191833570599556,
0.006547331809997559,
-0.057698819786310196,
0.06932510435581207,
-0.12982366979122162,
0.05436630919575691,
0.043436627835035324,
-0.10945180803537369,
0.03841061517596245,
0.02560393325984478,
0.11603125184774399,
0.058632634580135345,
-0.040632184594869614,
-0.10494323819875717,
-0.13799439370632172,
0.023235952481627464,
0.058803655207157135,
-0.06312531977891922,
-0.13800419867038727,
-0.052970461547374725,
-0.2062724232673645,
0.04198472201824188,
-0.07393307238817215,
0.06842854619026184,
0.045238204300403595,
0.01849091611802578,
-0.05578908324241638,
-0.06200101599097252,
0.01771395653486252,
0.13669656217098236,
-0.06059794872999191,
-0.13932769000530243
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# flan-t5-small-finetuned-xsum
This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: nan
- Rouge1: 1.2294
- Rouge2: 0.0439
- Rougel: 1.149
- Rougelsum: 1.1538
- Gen Len: 14.526
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
| 0.0 | 1.0 | 625 | nan | 1.2294 | 0.0439 | 1.149 | 1.1538 | 14.526 |
| 0.0 | 2.0 | 1250 | nan | 1.2294 | 0.0439 | 1.149 | 1.1538 | 14.526 |
| 0.0 | 3.0 | 1875 | nan | 1.2294 | 0.0439 | 1.149 | 1.1538 | 14.526 |
| 0.0 | 4.0 | 2500 | nan | 1.2294 | 0.0439 | 1.149 | 1.1538 | 14.526 |
| 0.0 | 5.0 | 3125 | nan | 1.2294 | 0.0439 | 1.149 | 1.1538 | 14.526 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["rouge"], "base_model": "google/flan-t5-small", "model-index": [{"name": "flan-t5-small-finetuned-xsum", "results": []}]} | text2text-generation | Americo/flan-t5-small-finetuned-xsum | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"generated_from_trainer",
"base_model:google/flan-t5-small",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T16:05:02+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google/flan-t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| flan-t5-small-finetuned-xsum
============================
This model is a fine-tuned version of google/flan-t5-small on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: nan
* Rouge1: 1.2294
* Rouge2: 0.0439
* Rougel: 1.149
* Rougelsum: 1.1538
* Gen Len: 14.526
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google/flan-t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Tokenizers 0.15.1"
] | [
81,
113,
4,
27
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #t5 #text2text-generation #generated_from_trainer #base_model-google/flan-t5-small #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Tokenizers 0.15.1"
] | [
-0.08758394420146942,
0.0742587149143219,
-0.003380618756636977,
0.08743523061275482,
0.09674795717000961,
-0.021990040317177773,
0.20061835646629333,
0.12105880677700043,
-0.08934551477432251,
0.06936907768249512,
0.14749236404895782,
0.1010870486497879,
0.02998383343219757,
0.1759670078754425,
-0.07591711729764938,
-0.2202523648738861,
0.05400790646672249,
0.038801465183496475,
-0.012622351758182049,
0.12594135105609894,
0.08730345219373703,
-0.11960800737142563,
0.09170892834663391,
0.03420399874448776,
-0.1695907562971115,
-0.007818309590220451,
0.024260887876152992,
-0.08056420832872391,
0.11992181092500687,
0.0392422117292881,
0.09930126368999481,
0.06690029799938202,
0.028277015313506126,
-0.16455458104610443,
0.018459336832165718,
0.05075855180621147,
-0.01320156641304493,
0.09054690599441528,
0.05440696328878403,
-0.009293928742408752,
0.06923176348209381,
-0.06450091302394867,
0.06307502090930939,
0.022570570930838585,
-0.13071787357330322,
-0.22464486956596375,
-0.0975765734910965,
0.05728015676140785,
0.06553668528795242,
0.07509463280439377,
-0.012177453376352787,
0.18722839653491974,
-0.016796110197901726,
0.10398908704519272,
0.24415697157382965,
-0.3408496081829071,
-0.06385964900255203,
-0.012460000813007355,
0.06001909077167511,
0.08111030608415604,
-0.07373817265033722,
-0.01570480689406395,
0.04115554317831993,
0.022915730252861977,
0.1547688990831375,
-0.017197556793689728,
0.008336536586284637,
-0.03817281872034073,
-0.12820030748844147,
-0.06350498646497726,
0.15967518091201782,
0.05524175241589546,
-0.06105205789208412,
-0.07752695679664612,
-0.08075267821550369,
-0.15309156477451324,
-0.0516069233417511,
-0.018503403291106224,
0.04707830771803856,
-0.01757381483912468,
-0.08292756974697113,
-0.04404320567846298,
-0.09188172221183777,
-0.04918770119547844,
-0.03628401830792427,
0.16882702708244324,
0.043132808059453964,
-0.003055593464523554,
-0.050688523799180984,
0.07081969827413559,
-0.03481702879071236,
-0.15954342484474182,
-0.0018422594293951988,
0.012717016041278839,
0.03484492003917694,
-0.04717771336436272,
-0.04574289172887802,
-0.14472727477550507,
0.026559915393590927,
0.17508602142333984,
-0.09555261582136154,
0.0635150596499443,
-0.03680040314793587,
0.04169158264994621,
-0.10675638914108276,
0.17366524040699005,
-0.01768231764435768,
0.03222567215561867,
0.056266698986291885,
0.10237640142440796,
0.08605175465345383,
-0.02905694954097271,
-0.12240368127822876,
0.03388694301247597,
0.13510942459106445,
0.025135071948170662,
-0.03800434619188309,
0.0753030851483345,
-0.05470382049679756,
-0.004162714816629887,
0.05997603014111519,
-0.09479610621929169,
0.02647222951054573,
-0.00388250220566988,
-0.028262678533792496,
-0.05008430406451225,
0.019618095830082893,
0.023437784984707832,
-0.02317812293767929,
0.07313636690378189,
-0.08577537536621094,
0.008578822016716003,
-0.0651804730296135,
-0.12677912414073944,
0.028090639039874077,
-0.0755094438791275,
0.009779096581041813,
-0.12497732043266296,
-0.1223570853471756,
-0.00934958178550005,
0.04431839659810066,
-0.02985428459942341,
-0.05821938067674637,
-0.050508733838796616,
-0.09009657055139542,
0.03560023754835129,
-0.011630148626863956,
0.04057722166180611,
-0.06546377390623093,
0.0896487832069397,
0.060527026653289795,
0.06509684771299362,
-0.059368763118982315,
0.02785637229681015,
-0.10158504545688629,
0.0608665831387043,
-0.22652596235275269,
0.037647780030965805,
-0.03501865267753601,
0.10469953715801239,
-0.09120379388332367,
-0.08192885667085648,
-0.014087572693824768,
-0.018514035269618034,
0.09742163866758347,
0.09912294149398804,
-0.1301315724849701,
-0.07036197930574417,
0.21706944704055786,
-0.11027780920267105,
-0.1672755777835846,
0.1508532613515854,
-0.037580978125333786,
0.036088503897190094,
0.05852862447500229,
0.21293878555297852,
0.058093633502721786,
-0.09549982100725174,
0.005130146164447069,
-0.06209122762084007,
0.048952050507068634,
-0.036867205053567886,
0.06897439062595367,
-0.0012509439839050174,
0.02456817775964737,
0.004524642135947943,
-0.004988022614270449,
0.04263043403625488,
-0.06354216486215591,
-0.07376162707805634,
-0.06469110399484634,
-0.07373976707458496,
0.02836006134748459,
0.025596842169761658,
0.07514191418886185,
-0.13697440922260284,
-0.1092129498720169,
0.041589442640542984,
0.04698671028017998,
-0.08613879978656769,
0.04216255247592926,
-0.10993992537260056,
0.13833434879779816,
-0.07260096818208694,
0.004350572824478149,
-0.15624114871025085,
-0.031065121293067932,
0.04636527597904205,
-0.01887342520058155,
0.021686671301722527,
-0.06751753389835358,
0.07734932005405426,
0.0711619034409523,
-0.04571595415472984,
-0.04532500356435776,
-0.019493650645017624,
0.002327136229723692,
-0.10702046006917953,
-0.1802261918783188,
-0.010146667249500751,
-0.041880641132593155,
0.11131403595209122,
-0.18448713421821594,
0.05406107008457184,
0.07884618639945984,
0.1110665425658226,
0.053735047578811646,
-0.01316563505679369,
-0.002887573093175888,
0.06050228700041771,
-0.046024322509765625,
-0.07140602916479111,
0.0633440837264061,
0.045935966074466705,
-0.10617941617965698,
0.03266735002398491,
-0.18324515223503113,
0.20006497204303741,
0.14099043607711792,
-0.018844496458768845,
-0.06376934796571732,
-0.020971162244677544,
-0.03634975478053093,
-0.02373296208679676,
-0.003743326524272561,
0.004507911391556263,
0.16310206055641174,
0.01266040001064539,
0.16509689390659332,
-0.11610160768032074,
-0.048308610916137695,
0.047218553721904755,
-0.03782748430967331,
-0.01211271807551384,
0.09156053513288498,
0.01979208178818226,
-0.14664538204669952,
0.13201214373111725,
0.17017915844917297,
-0.060171645134687424,
0.12461207807064056,
-0.0658261701464653,
-0.05211896821856499,
-0.02440156601369381,
0.040163397789001465,
0.045062121003866196,
0.08121275156736374,
-0.10299770534038544,
-0.018198812380433083,
0.011803222820162773,
0.0070221517235040665,
0.003648325800895691,
-0.2012449949979782,
0.001722131622955203,
0.04996814206242561,
-0.05230419337749481,
-0.026610583066940308,
-0.005775414872914553,
-0.002006686758249998,
0.08972273021936417,
0.016522135585546494,
-0.06206660345196724,
0.043590426445007324,
0.016588086262345314,
-0.07629659026861191,
0.1809549182653427,
-0.07250837981700897,
-0.16910697519779205,
-0.13747645914554596,
-0.054494451731443405,
-0.059311944991350174,
0.017247453331947327,
0.08563455939292908,
-0.05739708989858627,
-0.04852316528558731,
-0.12904126942157745,
-0.0386645682156086,
0.028787579387426376,
0.02839479222893715,
0.03371232748031616,
-0.009141458198428154,
0.08906671404838562,
-0.09332522749900818,
-0.024166718125343323,
0.00035894993925467134,
0.00441413838416338,
0.06576208025217056,
0.0010666358284652233,
0.11575019359588623,
0.10849872976541519,
-0.03668687120079994,
0.01620735414326191,
-0.04222162440419197,
0.2220354527235031,
-0.06912990659475327,
-0.01671881414949894,
0.14327894151210785,
-0.023164916783571243,
0.08276211470365524,
0.13164463639259338,
0.038953159004449844,
-0.10404268652200699,
0.02184908464550972,
-0.0035019037313759327,
-0.034207213670015335,
-0.2054617553949356,
0.0048980494029819965,
-0.04506700858473778,
-0.006940571125596762,
0.11169087886810303,
0.04027749225497246,
0.04540795460343361,
0.06020098179578781,
-0.0007111908635124564,
0.0804162472486496,
0.020393038168549538,
0.11436984688043594,
0.14957456290721893,
0.06389854848384857,
0.13936160504817963,
-0.057231221348047256,
-0.0534336194396019,
0.033889807760715485,
-0.003079763613641262,
0.1915411800146103,
0.021391339600086212,
0.2250758409500122,
0.03659472241997719,
0.13763652741909027,
0.04348823055624962,
0.06525180488824844,
-0.011733795516192913,
-0.022532209753990173,
-0.01053240243345499,
-0.054438333958387375,
-0.03591286763548851,
0.03068428486585617,
-0.07468458265066147,
0.037340644747018814,
-0.10399804264307022,
0.031866010278463364,
0.05603884533047676,
0.3147098422050476,
0.018518520519137383,
-0.36270326375961304,
-0.10601293295621872,
0.009110074490308762,
-0.04792587459087372,
-0.04045089706778526,
0.015344349667429924,
0.11784423142671585,
-0.05151709169149399,
0.06517790257930756,
-0.08645027130842209,
0.09681703895330429,
-0.04138263687491417,
0.04180925711989403,
0.02643543854355812,
0.08027835190296173,
-0.013110360130667686,
0.03284177929162979,
-0.282226026058197,
0.2654060125350952,
0.02462063357234001,
0.07147304713726044,
-0.04392314702272415,
0.012110970914363861,
0.022533245384693146,
0.0758388340473175,
0.057826608419418335,
-0.021383462473750114,
-0.09508989006280899,
-0.14355704188346863,
-0.11123698949813843,
0.021910926327109337,
0.09658311307430267,
-0.008767640218138695,
0.11887475848197937,
-0.03594231978058815,
-0.0046785809099674225,
0.05189547687768936,
-0.02128763683140278,
-0.07327709347009659,
-0.09714828431606293,
0.010432837530970573,
0.05004117265343666,
0.020857198163866997,
-0.09950566291809082,
-0.09956282377243042,
-0.08277580142021179,
0.17542795836925507,
0.006080911960452795,
-0.05135617405176163,
-0.11316893249750137,
0.019908210262656212,
0.04485279694199562,
-0.08138908445835114,
0.05090935155749321,
-0.020713889971375465,
0.13071861863136292,
-0.0078749880194664,
-0.060631174594163895,
0.13254696130752563,
-0.04466286674141884,
-0.17969149351119995,
-0.03996322676539421,
0.12494326382875443,
-0.02952086366713047,
0.0396595373749733,
0.0007028322434052825,
0.03253733366727829,
-0.02448870800435543,
-0.06646482646465302,
0.02635197341442108,
-0.03497399389743805,
0.061539992690086365,
-0.05963662266731262,
-0.010735969990491867,
0.015298586338758469,
-0.07094255834817886,
-0.027990104630589485,
0.17035505175590515,
0.303959459066391,
-0.07539678364992142,
0.028003064915537834,
0.042295776307582855,
-0.033724963665008545,
-0.1578126698732376,
0.007558472920209169,
0.03162616863846779,
0.004246735945343971,
-0.0005362214869819582,
-0.1385193020105362,
0.009904706850647926,
0.07929665595293045,
-0.023966394364833832,
0.08445575088262558,
-0.266026109457016,
-0.14659912884235382,
0.08849207311868668,
0.157660573720932,
0.10954000800848007,
-0.16986070573329926,
-0.06371532380580902,
-0.03521738946437836,
-0.12764309346675873,
0.10441616922616959,
-0.12040863186120987,
0.1005869209766388,
-0.013282068073749542,
0.05518631264567375,
0.010410434566438198,
-0.056386977434158325,
0.11344576627016068,
-0.062394317239522934,
0.09528323262929916,
-0.07195933908224106,
0.016069721430540085,
0.09976611286401749,
-0.08100418001413345,
0.03820133954286575,
-0.12482180446386337,
0.0415770560503006,
-0.045201826840639114,
-0.0318928137421608,
-0.04051603376865387,
0.015353405848145485,
-0.03272831439971924,
-0.022576218470931053,
-0.04009462893009186,
-0.006113505456596613,
0.049844153225421906,
-0.007929245010018349,
0.19873957335948944,
0.0017617231933400035,
0.14740677177906036,
0.16533353924751282,
0.12429193407297134,
-0.11987725645303726,
-0.003727732226252556,
0.008901741355657578,
-0.04549475759267807,
0.05734129250049591,
-0.14751698076725006,
0.05130832642316818,
0.11123689264059067,
-0.0016587626887485385,
0.12439928948879242,
0.059377189725637436,
-0.05105544999241829,
0.03269210830330849,
0.06667915731668472,
-0.18042533099651337,
-0.15491898357868195,
-0.016744952648878098,
0.040044475346803665,
-0.11348528414964676,
0.04765138402581215,
0.13771113753318787,
-0.06801737844944,
-0.014154513366520405,
0.004706894047558308,
-0.003302160417661071,
-0.003942941315472126,
0.13167963922023773,
0.029942456632852554,
0.05481795594096184,
-0.08154238015413284,
0.06610740721225739,
0.058599360287189484,
-0.10851985961198807,
0.03807954862713814,
0.07156412303447723,
-0.106997549533844,
-0.03644989803433418,
0.03904891386628151,
0.15822143852710724,
-0.023773280903697014,
-0.06092539429664612,
-0.17358209192752838,
-0.1317732036113739,
0.06747817993164062,
0.2051730901002884,
0.06023505702614784,
0.015092577785253525,
-0.003188398666679859,
0.0027187217492610216,
-0.09939773380756378,
0.1033819168806076,
0.027219412848353386,
0.09427373111248016,
-0.14924056828022003,
0.10819917917251587,
-0.00240961997769773,
0.007801827508956194,
-0.016955286264419556,
0.03418970853090286,
-0.11995875090360641,
-0.00870829913765192,
-0.14088906347751617,
0.02588232234120369,
-0.03394904360175133,
0.0033050677739083767,
-0.009765339083969593,
-0.042349692434072495,
-0.07145494967699051,
0.020535802468657494,
-0.10372104495763779,
-0.0346842035651207,
0.02407321333885193,
0.03644189611077309,
-0.11325909942388535,
-0.016635997220873833,
0.03050106018781662,
-0.09435644745826721,
0.08597642183303833,
0.03150539472699165,
-0.010012648068368435,
0.019558358937501907,
-0.08738090097904205,
0.025355936959385872,
0.0651998445391655,
0.0008692097035236657,
0.03220917657017708,
-0.12032719701528549,
-0.012009602971374989,
0.021672481670975685,
0.015573590993881226,
0.019422076642513275,
0.12086927145719528,
-0.11637824773788452,
-0.009924151003360748,
0.005905839614570141,
-0.056760404258966446,
-0.058158162981271744,
0.03479829058051109,
0.08246270567178726,
-0.004529770463705063,
0.20133085548877716,
-0.09905996918678284,
0.0036467700265347958,
-0.19518110156059265,
0.010674377903342247,
0.010041381232440472,
-0.14398522675037384,
-0.099847711622715,
-0.04841051623225212,
0.04776856303215027,
-0.06372907757759094,
0.11774354428052902,
-0.023810742422938347,
0.02042515017092228,
0.05110717564821243,
-0.03527403622865677,
0.0024910017382353544,
0.03179740533232689,
0.19213487207889557,
0.021328553557395935,
-0.044430673122406006,
0.04613086208701134,
0.0044735209085047245,
0.09576160460710526,
0.07006146758794785,
0.16989411413669586,
0.1439114511013031,
-0.007800066377967596,
0.1110219731926918,
0.04487518221139908,
0.006337742321193218,
-0.1544336974620819,
0.05974052846431732,
-0.042345836758613586,
0.12124548107385635,
-0.011327613145112991,
0.20263056457042694,
0.17079435288906097,
-0.13174743950366974,
0.025033915415406227,
-0.03881866857409477,
-0.0638078898191452,
-0.09976448118686676,
-0.10126223415136337,
-0.11197888851165771,
-0.15689696371555328,
-0.0038167734164744616,
-0.10139629244804382,
0.017203740775585175,
0.02647690661251545,
0.005946759134531021,
-0.015265772119164467,
0.16276760399341583,
0.03451156243681908,
0.009979899041354656,
0.06932733207941055,
-0.012195942923426628,
-0.05363057926297188,
-0.04746096953749657,
-0.10980293154716492,
0.023067748174071312,
-0.016160860657691956,
0.0347093865275383,
0.001121061504818499,
0.0012476701522246003,
0.03558702394366264,
-0.025446107611060143,
-0.1134156584739685,
0.018035003915429115,
0.0362619087100029,
0.045046474784612656,
0.016489462926983833,
0.0113893523812294,
-0.012685243040323257,
-0.00127783278003335,
0.190326526761055,
-0.05924747884273529,
-0.05880066752433777,
-0.10963262617588043,
0.197675883769989,
0.002531651873141527,
-0.018675852566957474,
0.016303738579154015,
-0.07603862136602402,
0.026291463524103165,
0.1923551857471466,
0.1479606181383133,
-0.03553399071097374,
0.011625818908214569,
-0.03176902234554291,
-0.009402130730450153,
-0.024145059287548065,
0.08696801960468292,
0.11226969212293625,
-0.022539885714650154,
-0.057045187801122665,
-0.037605538964271545,
-0.05083286017179489,
0.00003339137037983164,
-0.05676230788230896,
0.07492769509553909,
0.013874371536076069,
0.0068093957379460335,
-0.04690631106495857,
0.048201195895671844,
-0.029210420325398445,
-0.033615563064813614,
0.009692505933344364,
-0.19626595079898834,
-0.134061798453331,
-0.008196093142032623,
0.06892772018909454,
-0.026097804307937622,
0.052376918494701385,
-0.006240627262741327,
-0.007389961276203394,
0.045282039791345596,
-0.023786690086126328,
-0.059402309358119965,
-0.045799776911735535,
0.06655064225196838,
-0.15577790141105652,
0.19884377717971802,
-0.02557012438774109,
0.02493707276880741,
0.13805608451366425,
0.043243829160928726,
-0.09956677258014679,
0.08442926406860352,
0.04201338812708855,
-0.05544883385300636,
0.026939302682876587,
0.10026486217975616,
-0.02323850989341736,
0.08678488433361053,
0.05375521257519722,
-0.09384417533874512,
0.009349585510790348,
-0.08275832235813141,
-0.03757903724908829,
-0.03667406365275383,
-0.03963494300842285,
-0.05901803821325302,
0.11682560294866562,
0.16917724907398224,
-0.04647452011704445,
-0.002621809486299753,
-0.042164482176303864,
0.022498412057757378,
0.08397982269525528,
0.0053193229250609875,
-0.03717183321714401,
-0.26232093572616577,
0.01408667117357254,
0.102743960916996,
-0.0039129299111664295,
-0.30722904205322266,
-0.09344260394573212,
-0.02973974123597145,
-0.02025637775659561,
-0.11140746623277664,
0.08628658205270767,
0.12652891874313354,
0.05072397738695145,
-0.06397295743227005,
-0.06175034120678902,
-0.07427813857793808,
0.17207810282707214,
-0.11914034187793732,
-0.06988254189491272
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | fill-mask | Fm505/dummy-model | [
"transformers",
"safetensors",
"bert",
"fill-mask",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:09:11+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #bert #fill-mask #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #bert #fill-mask #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
46,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #bert #fill-mask #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.0695682242512703,
0.16317413747310638,
-0.0039503732696175575,
0.015582166612148285,
0.11206504702568054,
0.009388403967022896,
0.07999511063098907,
0.10805515944957733,
-0.022488119080662727,
0.12826035916805267,
0.04157862439751625,
0.09647717326879501,
0.11301155388355255,
0.19367553293704987,
0.004903987515717745,
-0.20981737971305847,
0.06215637922286987,
-0.11348576843738556,
0.012589743360877037,
0.12234765291213989,
0.14233623445034027,
-0.10432587563991547,
0.07009522616863251,
-0.03240275755524635,
-0.015827082097530365,
-0.030783042311668396,
-0.05974305048584938,
-0.06185486167669296,
0.06711262464523315,
0.06498897075653076,
0.06968989223241806,
0.024118594825267792,
0.07975957542657852,
-0.2886873483657837,
0.01945551112294197,
0.07693088054656982,
0.004277593456208706,
0.06179933622479439,
0.07438474148511887,
-0.06548646092414856,
0.12410084903240204,
-0.04917580634355545,
0.15509648621082306,
0.07449730485677719,
-0.09788993746042252,
-0.1928139179944992,
-0.07736650109291077,
0.07917608320713043,
0.15369024872779846,
0.05949998274445534,
-0.0366918221116066,
0.14726826548576355,
-0.08467528223991394,
0.014476358890533447,
0.06632605195045471,
-0.07833176851272583,
-0.05415208265185356,
0.041515614837408066,
0.07901278138160706,
0.0834973081946373,
-0.1278284192085266,
-0.00972729828208685,
0.03875104710459709,
0.021097054705023766,
0.10629333555698395,
0.023433465510606766,
0.12429918348789215,
0.02415166236460209,
-0.1392766535282135,
-0.062283553183078766,
0.12270510196685791,
0.030049407854676247,
-0.05623685196042061,
-0.23220579326152802,
-0.009600076824426651,
-0.02659357711672783,
-0.027964934706687927,
-0.03921128436923027,
0.03545486927032471,
-0.029904281720519066,
0.07641863077878952,
0.011728300713002682,
-0.07156947255134583,
-0.05115693062543869,
0.08266618847846985,
0.06655919551849365,
0.023309092968702316,
-0.022748295217752457,
0.022040950134396553,
0.1145513653755188,
0.08442392945289612,
-0.12301572412252426,
-0.06391360610723495,
-0.0636562705039978,
-0.09367027133703232,
-0.045953843742609024,
0.035427894443273544,
0.0689372792840004,
0.052030809223651886,
0.1975124478340149,
0.011451225727796555,
0.053010325878858566,
0.03033238649368286,
0.015536453574895859,
0.061831213533878326,
0.0769730731844902,
-0.050268951803445816,
-0.1342480480670929,
-0.043217532336711884,
0.11456252634525299,
0.009491410106420517,
-0.03323882073163986,
-0.03446515277028084,
0.06311985105276108,
0.05336147919297218,
0.11711134761571884,
0.07178932428359985,
0.018411245197057724,
-0.06926928460597992,
-0.03773912414908409,
0.1873171627521515,
-0.15400058031082153,
0.019539861008524895,
0.01581503264605999,
-0.05302942171692848,
-0.04518861696124077,
0.015980584546923637,
0.008176358416676521,
-0.025257200002670288,
0.10878665745258331,
-0.06910517811775208,
-0.039905957877635956,
-0.10782934725284576,
-0.05966084450483322,
0.030464332550764084,
-0.01236342079937458,
-0.031870488077402115,
-0.04436861723661423,
-0.11301329731941223,
-0.0747642070055008,
0.06969720125198364,
-0.06395140290260315,
-0.06758769601583481,
-0.030247729271650314,
-0.05435582250356674,
0.014192626811563969,
0.003177133621647954,
0.1288391500711441,
-0.03037290647625923,
0.04630573093891144,
-0.042741283774375916,
0.07184016704559326,
0.12946291267871857,
0.033582016825675964,
-0.07494695484638214,
0.06655236333608627,
-0.21820411086082458,
0.10199115425348282,
-0.0967235192656517,
0.03035767190158367,
-0.16142524778842926,
-0.028911394998431206,
0.01647994853556156,
0.032515235245227814,
-0.010486545972526073,
0.13834860920906067,
-0.19198662042617798,
-0.03841867297887802,
0.1795300990343094,
-0.13333173096179962,
-0.08988800644874573,
0.06576739251613617,
-0.05458313226699829,
0.12602677941322327,
0.04838111251592636,
-0.03003455139696598,
0.058201808482408524,
-0.14024828374385834,
-0.02112412638962269,
-0.054867181926965714,
-0.004416458308696747,
0.14790405333042145,
0.06459064781665802,
-0.05876573175191879,
0.0235891230404377,
0.021585118025541306,
-0.0243860874325037,
-0.042450159788131714,
-0.035232093185186386,
-0.09779409319162369,
0.007044002879410982,
-0.0766984298825264,
0.01927781291306019,
-0.016798585653305054,
-0.08684395998716354,
-0.03863377124071121,
-0.16315165162086487,
-0.010619483888149261,
0.09515801072120667,
0.01122341025620699,
-0.03048161417245865,
-0.0940304845571518,
0.010681629180908203,
0.011915594339370728,
-0.013544750399887562,
-0.15514391660690308,
-0.05313099920749664,
0.02975001558661461,
-0.1710706353187561,
0.03037792071700096,
-0.05140242353081703,
0.036749161779880524,
0.04263495281338692,
-0.04429660737514496,
-0.02103106863796711,
0.013221964240074158,
0.01574954204261303,
-0.020732590928673744,
-0.23971620202064514,
-0.01838381588459015,
-0.0490109957754612,
0.180399551987648,
-0.23546038568019867,
0.04174007847905159,
0.05565045401453972,
0.12047853320837021,
0.004876249469816685,
-0.04898517578840256,
0.035101015120744705,
-0.0464148111641407,
-0.043804459273815155,
-0.06390430778265,
-0.0032493185717612505,
-0.030191048979759216,
-0.048357024788856506,
0.04138917848467827,
-0.19590826332569122,
-0.027627017349004745,
0.11042598634958267,
0.07558931410312653,
-0.169208362698555,
-0.06871634721755981,
-0.030856039375066757,
-0.06003820523619652,
-0.09412489086389542,
-0.047238849103450775,
0.10385532677173615,
0.04374808073043823,
0.05180130526423454,
-0.06982806324958801,
-0.04795414209365845,
0.014942421577870846,
-0.008059881627559662,
-0.03664791211485863,
0.08609483391046524,
0.09363002330064774,
-0.11547045409679413,
0.09921012818813324,
0.06918627768754959,
0.05994226783514023,
0.10326741635799408,
0.0071405707858502865,
-0.0977831706404686,
-0.015172232873737812,
0.022144995629787445,
0.016030732542276382,
0.14287981390953064,
-0.08233530074357986,
0.031467825174331665,
0.045298222452402115,
-0.036048755049705505,
0.011318118311464787,
-0.10247772186994553,
0.02277444861829281,
0.03092500939965248,
-0.008242631331086159,
0.01892855390906334,
-0.050711531192064285,
0.013711366802453995,
0.10258015990257263,
0.034890126436948776,
0.024561753496527672,
0.013505036942660809,
-0.03989248350262642,
-0.12325675040483475,
0.17909187078475952,
-0.09560354053974152,
-0.2589256465435028,
-0.1297009140253067,
0.0016485642408952117,
0.047613367438316345,
-0.014485974796116352,
0.017094451934099197,
-0.054489344358444214,
-0.1085980087518692,
-0.10300207883119583,
0.01934879459440708,
0.053916629403829575,
-0.08475877344608307,
-0.062334731221199036,
0.05043650045990944,
0.04047852009534836,
-0.12173960357904434,
0.02034258283674717,
0.043894823640584946,
-0.0658656656742096,
0.007203231565654278,
0.05971246957778931,
0.08562132716178894,
0.18097670376300812,
0.015700936317443848,
-0.01629323698580265,
0.012648820877075195,
0.22811394929885864,
-0.14797993004322052,
0.09497489780187607,
0.1383752077817917,
-0.05721057578921318,
0.08420403301715851,
0.21159671247005463,
0.03094991110265255,
-0.09150044620037079,
0.04039512202143669,
0.03325594961643219,
-0.04082412272691727,
-0.23298248648643494,
-0.08047252148389816,
-0.0032772731501609087,
-0.08068809658288956,
0.10162109136581421,
0.09027010202407837,
0.1132882758975029,
0.05186156928539276,
-0.10623718798160553,
-0.07151743024587631,
0.04082353785634041,
0.11879902333021164,
-0.02256317250430584,
0.0035680357832461596,
0.09804364293813705,
-0.030343100428581238,
0.014163176529109478,
0.09006884694099426,
0.006501318886876106,
0.18543729186058044,
0.04132460802793503,
0.12941181659698486,
0.08460376411676407,
0.06855804473161697,
0.02408376708626747,
0.018935227766633034,
0.028948647901415825,
0.02635204792022705,
-0.019921185448765755,
-0.09093207120895386,
-0.014699726365506649,
0.14122939109802246,
0.041546013206243515,
0.030846212059259415,
0.011316456831991673,
-0.03433994576334953,
0.06460443139076233,
0.16659797728061676,
0.01839955896139145,
-0.23006980121135712,
-0.06272536516189575,
0.07289306819438934,
-0.07035499066114426,
-0.11501950770616531,
-0.011477392166852951,
0.02908293344080448,
-0.17825914919376373,
0.0479411743581295,
-0.022636432200670242,
0.10284267365932465,
-0.09900184720754623,
-0.023873277008533478,
0.04412706196308136,
0.0705353394150734,
-0.03580007329583168,
0.07728645950555801,
-0.20810243487358093,
0.15388257801532745,
0.007997404783964157,
0.06405343860387802,
-0.10958539694547653,
0.08351796120405197,
0.023163659498095512,
0.0035931505262851715,
0.1632837951183319,
-0.0070585873909294605,
-0.09036380797624588,
-0.07494185119867325,
-0.07656881213188171,
-0.01101732812821865,
0.09567800909280777,
-0.10157354176044464,
0.08578468859195709,
-0.0066426945850253105,
-0.034807685762643814,
-0.0036719576455652714,
-0.10101240128278732,
-0.13781799376010895,
-0.18216682970523834,
0.05830308049917221,
-0.108928382396698,
0.02895812690258026,
-0.11352406442165375,
-0.06376964598894119,
-0.02617688849568367,
0.19453708827495575,
-0.1856314241886139,
-0.08167315274477005,
-0.14844723045825958,
-0.07225494831800461,
0.12610159814357758,
-0.045480526983737946,
0.07762189209461212,
-0.001880245516076684,
0.20876388251781464,
-0.0052230339497327805,
0.00003858540367218666,
0.08722095936536789,
-0.0969884842634201,
-0.20284906029701233,
-0.09463591128587723,
0.13733142614364624,
0.12461096793413162,
0.04729393869638443,
-0.004690513946115971,
0.02620808035135269,
-0.006243519019335508,
-0.10859202593564987,
0.02910296991467476,
0.14064766466617584,
0.08289998024702072,
0.038931865245103836,
-0.02449015900492668,
-0.14890262484550476,
-0.10413921624422073,
-0.0505983941257,
0.016270531341433525,
0.178590789437294,
-0.07276389747858047,
0.15994256734848022,
0.15645509958267212,
-0.061302632093429565,
-0.2092013657093048,
0.031304605305194855,
0.03814060986042023,
-0.00735123734921217,
0.032533854246139526,
-0.20631249248981476,
0.07322324812412262,
0.013627852313220501,
-0.06360501050949097,
0.13831163942813873,
-0.16617220640182495,
-0.14868974685668945,
0.09640426188707352,
0.06936991959810257,
-0.21415942907333374,
-0.13356217741966248,
-0.09905771166086197,
-0.05417881906032562,
-0.10609032213687897,
0.07912828028202057,
0.01456306129693985,
0.0063131884671747684,
0.039144620299339294,
0.026788141578435898,
0.018952853977680206,
-0.055193133652210236,
0.19719533622264862,
-0.005659637041389942,
0.03867407143115997,
-0.079157754778862,
-0.08333847671747208,
0.030260656028985977,
-0.06024036556482315,
0.07474341988563538,
-0.023070191964507103,
0.0043883658945560455,
-0.10337463766336441,
-0.06357163935899734,
-0.05533557012677193,
0.04131639003753662,
-0.08671651035547256,
-0.09687067568302155,
-0.054821133613586426,
0.1052565649151802,
0.08825810253620148,
-0.03520786389708519,
-0.06399983167648315,
-0.09821294248104095,
0.07732317596673965,
0.22283773124217987,
0.18066252768039703,
0.07087353616952896,
-0.07483634352684021,
0.0010324176400899887,
-0.022433919832110405,
0.054983608424663544,
-0.1989203691482544,
0.044562239199876785,
0.04126801714301109,
0.0311158187687397,
0.13412833213806152,
-0.023162340745329857,
-0.16252556443214417,
-0.04666705057024956,
0.06344481557607651,
-0.062314748764038086,
-0.16319546103477478,
-0.00016717356629669666,
0.0916345939040184,
-0.16376787424087524,
-0.0648001953959465,
0.02249295450747013,
-0.03433338552713394,
-0.024365736171603203,
-0.0000157451031554956,
0.0820402204990387,
0.023767979815602303,
0.11757155507802963,
0.0667911022901535,
0.11306948214769363,
-0.10177480429410934,
0.07651326805353165,
0.08987756073474884,
-0.10732097923755646,
0.03177256882190704,
0.0679590180516243,
-0.05994531139731407,
-0.03372235968708992,
0.02310117892920971,
0.08390586078166962,
0.033371008932590485,
-0.07280343025922775,
-0.0027108066715300083,
-0.1101594865322113,
0.06614562124013901,
0.1410103440284729,
0.03333422541618347,
0.0003071172977797687,
0.04553673043847084,
0.026930835098028183,
-0.0945608988404274,
0.11191196739673615,
0.038921058177948,
0.03493048995733261,
-0.04539057984948158,
0.004739639814943075,
0.04245656356215477,
-0.010762917809188366,
-0.017839528620243073,
-0.03884020075201988,
-0.06404419243335724,
-0.010227293707430363,
-0.15227684378623962,
0.028882548213005066,
-0.07475166767835617,
0.005362969357520342,
0.017854763194918633,
-0.037447962909936905,
-0.0017063911072909832,
0.007931754924356937,
-0.07987163960933685,
-0.040919411927461624,
-0.006045978050678968,
0.10301655530929565,
-0.1571531742811203,
0.0073095266707241535,
0.08932953327894211,
-0.12466500699520111,
0.0767330452799797,
-0.003056043293327093,
-0.009896434843540192,
0.019439974799752235,
-0.14304398000240326,
0.05748452618718147,
-0.009228454902768135,
0.00885720830410719,
0.026310641318559647,
-0.2028958648443222,
0.005276820156723261,
-0.046354740858078,
-0.0541728176176548,
-0.00875329039990902,
-0.04280844330787659,
-0.1152573749423027,
0.10480808466672897,
0.01848764158785343,
-0.08730915933847427,
-0.016358980908989906,
0.04895014688372612,
0.1039031445980072,
-0.05420120432972908,
0.13693569600582123,
-0.025150449946522713,
0.06211975961923599,
-0.17961733043193817,
-0.015974286943674088,
-0.019317906349897385,
0.015642758458852768,
-0.026861993595957756,
-0.007615713868290186,
0.05647684261202812,
-0.015129351057112217,
0.22813443839550018,
-0.020140036940574646,
0.030034644529223442,
0.06510136276483536,
0.006892532110214233,
-0.012034238316118717,
0.08359529078006744,
0.04697233811020851,
0.022989440709352493,
0.0157330222427845,
0.01879745163023472,
-0.04617619141936302,
-0.0199876856058836,
-0.12531277537345886,
0.08793473988771439,
0.16457070410251617,
0.09647742658853531,
-0.004848872311413288,
0.05518500506877899,
-0.11906541883945465,
-0.08156134188175201,
0.09614262729883194,
-0.0331968292593956,
0.0012238750932738185,
-0.05372428148984909,
0.1372125893831253,
0.15581628680229187,
-0.17664477229118347,
0.06854353100061417,
-0.06903261691331863,
-0.05629098787903786,
-0.11328249424695969,
-0.1729356348514557,
-0.06526507437229156,
-0.031787171959877014,
-0.004383164457976818,
-0.057878151535987854,
0.0651032105088234,
0.10321546345949173,
0.008594287559390068,
0.0038617050740867853,
0.08901966363191605,
-0.037351686507463455,
0.0009274887852370739,
0.043105632066726685,
0.053729452192783356,
0.019777609035372734,
-0.06573960185050964,
0.010367924347519875,
0.0030578807927668095,
0.03969988226890564,
0.05550984665751457,
0.03013552725315094,
-0.021250499412417412,
0.015762019902467728,
-0.012240002863109112,
-0.10046600550413132,
0.036009300500154495,
-0.031009560450911522,
-0.0476452112197876,
0.14803250133991241,
0.01959061250090599,
0.00536824157461524,
-0.022027885541319847,
0.22717911005020142,
-0.0692228451371193,
-0.08300955593585968,
-0.14375057816505432,
0.1483703851699829,
-0.04850876331329346,
0.04777670279145241,
0.048611290752887726,
-0.1006619781255722,
0.03077290952205658,
0.14936116337776184,
0.14990805089473724,
-0.019973209127783775,
0.00984974205493927,
0.014924170449376106,
0.004702849313616753,
-0.027548549696803093,
0.05295790731906891,
0.04582742601633072,
0.12960462272167206,
-0.06669346988201141,
0.09011545777320862,
-0.011044532991945744,
-0.08640255779027939,
-0.023667417466640472,
0.12734681367874146,
0.004126773681491613,
0.023254431784152985,
-0.07918302714824677,
0.12160252779722214,
-0.0662197694182396,
-0.2606266736984253,
0.06323975324630737,
-0.06686537712812424,
-0.15521176159381866,
-0.018760496750473976,
0.028445512056350708,
0.005920670460909605,
0.027520710602402687,
0.06006603315472603,
-0.06003495678305626,
0.14988677203655243,
0.03705393522977829,
-0.07608085125684738,
-0.07932386547327042,
0.0791633352637291,
-0.07992571592330933,
0.29982754588127136,
0.007550139911472797,
0.04989086836576462,
0.09792694449424744,
-0.0319637730717659,
-0.13174396753311157,
0.05019649490714073,
0.09466034173965454,
-0.06565423309803009,
0.0679093673825264,
0.198545441031456,
-0.008660713210701942,
0.11422579735517502,
0.07441084831953049,
-0.08171522617340088,
0.058534253388643265,
-0.07837630808353424,
-0.08840835094451904,
-0.09223968535661697,
0.09403426945209503,
-0.06329182535409927,
0.15364859998226166,
0.13498902320861816,
-0.04389459639787674,
-0.0027436274103820324,
-0.031526047736406326,
0.05292224511504173,
-0.0004144559206906706,
0.11318539083003998,
0.01770014688372612,
-0.19339020550251007,
0.033284399658441544,
-0.018599338829517365,
0.096232570707798,
-0.2523307204246521,
-0.07822466641664505,
0.03999762609601021,
-0.016596345230937004,
-0.048957161605358124,
0.11946418881416321,
0.050940077751874924,
0.05071447789669037,
-0.05539802834391594,
-0.058477628976106644,
0.0015120115131139755,
0.15989357233047485,
-0.11395517736673355,
-0.002205098979175091
] |
null | null | transformers |
# NeuralTrix-7B-v1
NeuralTrix-7B-v1 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [mlabonne/OmniBeagle-7B](https://huggingface.co/mlabonne/OmniBeagle-7B)
* [flemmingmiguel/MBX-7B-v3](https://huggingface.co/flemmingmiguel/MBX-7B-v3)
* [AiMavenAi/AiMaven-Prometheus](https://huggingface.co/AiMavenAi/AiMaven-Prometheus)
## 🧩 Configuration
```yaml
models:
- model: mistralai/Mistral-7B-v0.1
# no parameters necessary for base model
- model: mlabonne/OmniBeagle-7B
parameters:
density: 0.65
weight: 0.4
- model: flemmingmiguel/MBX-7B-v3
parameters:
density: 0.6
weight: 0.35
- model: AiMavenAi/AiMaven-Prometheus
parameters:
density: 0.6
weight: 0.35
merge_method: dare_ties
base_model: mistralai/Mistral-7B-v0.1
parameters:
int8_mask: true
dtype: float16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "CultriX/NeuralTrix-7B-v1"
messages = [{"role": "user", "content": "What is a large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "mlabonne/OmniBeagle-7B", "flemmingmiguel/MBX-7B-v3", "AiMavenAi/AiMaven-Prometheus"], "base_model": ["mlabonne/OmniBeagle-7B", "flemmingmiguel/MBX-7B-v3", "AiMavenAi/AiMaven-Prometheus"]} | text-generation | CultriX/NeuralTrix-7B-v1 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"mlabonne/OmniBeagle-7B",
"flemmingmiguel/MBX-7B-v3",
"AiMavenAi/AiMaven-Prometheus",
"base_model:mlabonne/OmniBeagle-7B",
"base_model:flemmingmiguel/MBX-7B-v3",
"base_model:AiMavenAi/AiMaven-Prometheus",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T16:09:29+00:00 | [] | [] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #base_model-mlabonne/OmniBeagle-7B #base_model-flemmingmiguel/MBX-7B-v3 #base_model-AiMavenAi/AiMaven-Prometheus #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# NeuralTrix-7B-v1
NeuralTrix-7B-v1 is a merge of the following models using LazyMergekit:
* mlabonne/OmniBeagle-7B
* flemmingmiguel/MBX-7B-v3
* AiMavenAi/AiMaven-Prometheus
## Configuration
## Usage
| [
"# NeuralTrix-7B-v1\n\nNeuralTrix-7B-v1 is a merge of the following models using LazyMergekit:\n* mlabonne/OmniBeagle-7B\n* flemmingmiguel/MBX-7B-v3\n* AiMavenAi/AiMaven-Prometheus",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #base_model-mlabonne/OmniBeagle-7B #base_model-flemmingmiguel/MBX-7B-v3 #base_model-AiMavenAi/AiMaven-Prometheus #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# NeuralTrix-7B-v1\n\nNeuralTrix-7B-v1 is a merge of the following models using LazyMergekit:\n* mlabonne/OmniBeagle-7B\n* flemmingmiguel/MBX-7B-v3\n* AiMavenAi/AiMaven-Prometheus",
"## Configuration",
"## Usage"
] | [
156,
72,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #mlabonne/OmniBeagle-7B #flemmingmiguel/MBX-7B-v3 #AiMavenAi/AiMaven-Prometheus #base_model-mlabonne/OmniBeagle-7B #base_model-flemmingmiguel/MBX-7B-v3 #base_model-AiMavenAi/AiMaven-Prometheus #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# NeuralTrix-7B-v1\n\nNeuralTrix-7B-v1 is a merge of the following models using LazyMergekit:\n* mlabonne/OmniBeagle-7B\n* flemmingmiguel/MBX-7B-v3\n* AiMavenAi/AiMaven-Prometheus## Configuration## Usage"
] | [
-0.059001754969358444,
0.08325124531984329,
-0.007921796292066574,
0.024844132363796234,
0.0448136143386364,
0.033553462475538254,
0.16382953524589539,
0.10190489888191223,
0.040268152952194214,
0.08261700719594955,
0.0705629512667656,
0.1419818103313446,
0.05329621583223343,
0.1186363622546196,
-0.02658309042453766,
-0.18704283237457275,
0.07040562480688095,
-0.004859836306422949,
-0.032237447798252106,
0.06463626772165298,
0.11072766035795212,
-0.04569612815976143,
0.1008160337805748,
0.016483504325151443,
-0.09464192390441895,
-0.0045705027878284454,
-0.014700359664857388,
-0.0031572713050991297,
0.08686505258083344,
0.09294097870588303,
0.026483768597245216,
0.02294938825070858,
0.006064539775252342,
-0.1838098019361496,
0.018865596503019333,
0.009898054413497448,
-0.023402711376547813,
0.05447951331734657,
0.10380975902080536,
-0.03177003934979439,
0.13696880638599396,
-0.04400208964943886,
0.03382370248436928,
0.060558851808309555,
-0.10589990764856339,
-0.02196517214179039,
-0.06714779883623123,
0.03587101399898529,
0.06779850274324417,
0.04587318003177643,
-0.02830035239458084,
0.15356752276420593,
0.028371084481477737,
0.10699018090963364,
0.17417845129966736,
-0.3232269287109375,
-0.036063894629478455,
0.11377842724323273,
0.07358206063508987,
-0.052019041031599045,
-0.020351199433207512,
0.05758606269955635,
0.011717337183654308,
0.02279072441160679,
0.07213007658720016,
-0.05627790093421936,
0.13468964397907257,
-0.08952512592077255,
-0.11682054400444031,
0.009400691837072372,
0.12192521244287491,
0.03065662458539009,
-0.00482142623513937,
-0.10716070979833603,
-0.06906116753816605,
0.09923715889453888,
-0.07070212811231613,
-0.02180558256804943,
0.031919971108436584,
-0.03042883239686489,
0.10070474445819855,
-0.06764017045497894,
-0.039735935628414154,
0.009435155428946018,
-0.06005299091339111,
0.1118338331580162,
0.019152430817484856,
-0.017737342044711113,
0.011110711842775345,
0.02713705785572529,
-0.14213693141937256,
-0.12267027795314789,
-0.0032846727408468723,
-0.049217723309993744,
-0.043216828256845474,
-0.0315176323056221,
-0.04573982208967209,
-0.10829729586839676,
0.10528425127267838,
0.1773538589477539,
-0.028177225962281227,
0.06388077884912491,
0.0031208533328026533,
0.029116224497556686,
0.019567402079701424,
0.0040541947819292545,
-0.09040951728820801,
-0.13418257236480713,
0.04187442734837532,
0.1099165827035904,
0.043873682618141174,
0.021045850589871407,
-0.05755462497472763,
-0.01959271915256977,
0.008730115368962288,
0.011839537881314754,
0.08010032773017883,
0.06912052631378174,
-0.08896219730377197,
-0.05777158960700035,
0.09534676373004913,
-0.10411503911018372,
0.023866869509220123,
-0.004546014126390219,
-0.028200402855873108,
0.0264957994222641,
0.03704483434557915,
0.046232596039772034,
-0.03167857974767685,
0.0268828384578228,
-0.06333738565444946,
-0.02604147046804428,
-0.03171895071864128,
-0.06311438977718353,
0.02053983323276043,
-0.02969609573483467,
-0.03249778226017952,
-0.10091884434223175,
-0.1632944792509079,
-0.03504682704806328,
0.0588766448199749,
-0.07241520285606384,
-0.02593885362148285,
-0.06033413112163544,
0.0010793132241815329,
-0.005394310690462589,
0.0240174550563097,
-0.015738118439912796,
0.008178750053048134,
-0.00751187140122056,
0.03015642613172531,
0.07018153369426727,
-0.048723701387643814,
0.010512975044548512,
-0.035451579838991165,
0.08145486563444138,
-0.1731569468975067,
0.08739718049764633,
-0.0503636971116066,
0.08172210305929184,
-0.14166027307510376,
-0.0054216836579144,
-0.014868193306028843,
0.015485204756259918,
0.06181129068136215,
0.146945521235466,
-0.11985752731561661,
-0.08175697177648544,
0.12664194405078888,
-0.10295150429010391,
-0.16635972261428833,
0.12112592905759811,
0.03316476196050644,
0.004633261822164059,
0.051820073276758194,
0.1924603432416916,
0.046792589128017426,
-0.0707608312368393,
-0.05808737128973007,
-0.02867705002427101,
0.033712029457092285,
0.03644975274801254,
0.0697520300745964,
0.002280763816088438,
-0.005469275638461113,
0.030863892287015915,
0.014811943285167217,
0.04881048947572708,
-0.019153442233800888,
-0.06328842788934708,
-0.03616595268249512,
-0.07914270460605621,
0.042131442576646805,
-0.01996397227048874,
0.00003432197991060093,
-0.039262741804122925,
-0.04154646396636963,
-0.023942207917571068,
0.10032882541418076,
-0.03188911825418472,
0.013752390630543232,
-0.09526670724153519,
0.11642204225063324,
-0.0351715050637722,
0.04270031303167343,
-0.13076306879520416,
-0.11227260529994965,
0.037644531577825546,
-0.08313432335853577,
0.03884343430399895,
-0.07069004327058792,
0.07200628519058228,
0.05471267178654671,
-0.052064765244722366,
-0.05123285949230194,
0.10541573166847229,
-0.007608497515320778,
-0.025254225358366966,
-0.18961647152900696,
-0.034901540726423264,
-0.07233429700136185,
0.1972523033618927,
-0.14627550542354584,
0.05198492109775543,
-0.008612626232206821,
0.19928832352161407,
0.0074850087985396385,
-0.01780828833580017,
0.04366077855229378,
-0.014527597464621067,
-0.03404605761170387,
0.00262823817320168,
0.07937219738960266,
-0.019942261278629303,
-0.12131287902593613,
0.07455597817897797,
-0.10019483417272568,
0.12981733679771423,
0.08287670463323593,
0.01765131764113903,
-0.0275997593998909,
-0.08181224763393402,
-0.012407831847667694,
-0.04169507324695587,
0.07054119557142258,
-0.06183944270014763,
0.08118519932031631,
0.04716304689645767,
0.12006577849388123,
-0.07672751694917679,
-0.04520081356167793,
0.016980787739157677,
-0.046528395265340805,
-0.08758623152971268,
0.07455316931009293,
-0.03546661511063576,
-0.16450150310993195,
0.0730128288269043,
0.15436340868473053,
0.013727151788771152,
0.08523587137460709,
0.006634091027081013,
0.007619316224008799,
-0.08582980930805206,
0.012288968078792095,
0.041047897189855576,
-0.031124727800488472,
-0.08636369556188583,
0.024274058640003204,
0.056564778089523315,
-0.010711206123232841,
0.04096231237053871,
-0.027197614312171936,
0.031042061746120453,
0.007518362253904343,
-0.002381847007200122,
0.1501314640045166,
0.07162819057703018,
-0.0036032439675182104,
0.07437687367200851,
-0.0034937968011945486,
-0.04581837356090546,
-0.02937544323503971,
-0.02326800487935543,
-0.09108790010213852,
0.16212809085845947,
-0.13448107242584229,
-0.22820425033569336,
-0.1162143126130104,
-0.11677911132574081,
-0.11929512768983841,
-0.00921351183205843,
0.01964559406042099,
-0.030259953811764717,
-0.054761890321969986,
-0.11063653230667114,
0.046133529394865036,
0.04774117097258568,
-0.01993647962808609,
0.02723478339612484,
0.0170596931129694,
0.054934658110141754,
-0.0910160094499588,
-0.02480236440896988,
0.029206758365035057,
-0.046707943081855774,
0.039964091032743454,
-0.04171450063586235,
0.08286648243665695,
0.0564151331782341,
0.044818636029958725,
-0.00363663281314075,
-0.004595637321472168,
0.29033273458480835,
-0.06870701909065247,
0.08958042412996292,
0.1553361564874649,
-0.0040833973325788975,
0.055097661912441254,
0.18384328484535217,
0.024241775274276733,
-0.06648077815771103,
-0.017791975289583206,
-0.011301766149699688,
0.015825916081666946,
-0.20200961828231812,
-0.12789320945739746,
-0.04361174255609512,
0.013318359851837158,
0.07138794660568237,
0.04748792201280594,
0.06864041090011597,
0.06556589901447296,
-0.05825497955083847,
0.0020526954904198647,
0.05633709952235222,
0.08838057518005371,
0.20188014209270477,
-0.0023653414100408554,
0.10107365995645523,
-0.017843348905444145,
-0.028727050870656967,
0.046926651149988174,
0.03708630055189133,
0.0795530453324318,
0.050300780683755875,
0.13572901487350464,
0.051770880818367004,
0.01945662312209606,
0.0013553487369790673,
0.04267744719982147,
-0.048404403030872345,
-0.029164176434278488,
-0.014854750595986843,
-0.09645277261734009,
-0.0034897709265351295,
0.04524145647883415,
0.005623234435915947,
0.031028296798467636,
-0.024905819445848465,
0.02867226116359234,
0.10122176259756088,
0.11926164478063583,
0.09935382008552551,
-0.24950848519802094,
-0.019037356600165367,
0.046194035559892654,
-0.016991570591926575,
-0.05151640623807907,
-0.007413757964968681,
0.0071443780325353146,
-0.08470664918422699,
0.1540992110967636,
-0.035582199692726135,
0.07378891855478287,
-0.05228468403220177,
0.024730520322918892,
-0.004765741061419249,
0.09834637492895126,
0.001015215995721519,
0.03971440717577934,
-0.2220603972673416,
0.17777906358242035,
0.056267283856868744,
0.009151840582489967,
0.029086999595165253,
0.05549520254135132,
0.030799072235822678,
0.1609574854373932,
0.11660437285900116,
-0.004530925303697586,
0.03938332945108414,
-0.034851375967264175,
-0.08746878802776337,
-0.01719284988939762,
0.05221523717045784,
-0.036322738975286484,
0.08173228800296783,
-0.027433648705482483,
-0.08112894743680954,
-0.02738041989505291,
0.04241147264838219,
-0.18120957911014557,
-0.1543685644865036,
0.09534536302089691,
0.08221127092838287,
0.03426946699619293,
-0.08625343441963196,
-0.04146117344498634,
-0.1282719075679779,
0.19416417181491852,
-0.06836149841547012,
-0.054400280117988586,
-0.10131097584962845,
0.03123556450009346,
0.12363254278898239,
-0.061381518840789795,
0.06466558575630188,
-0.06828487664461136,
0.09825849533081055,
-0.09558693319559097,
-0.08290713280439377,
0.044835545122623444,
-0.11989272385835648,
-0.077021524310112,
-0.045987024903297424,
0.15473191440105438,
-0.043135374784469604,
0.027282996103167534,
0.03783373534679413,
0.047938305884599686,
-0.001955841202288866,
-0.03848890960216522,
0.01853829436004162,
0.09552941471338272,
0.0079806474968791,
0.058223433792591095,
-0.09608284384012222,
-0.16293998062610626,
-0.06928964704275131,
-0.021121418103575706,
0.13683927059173584,
0.32028982043266296,
-0.003671043086796999,
0.038297638297080994,
0.18937994539737701,
-0.07965492457151413,
-0.16112980246543884,
-0.08114180713891983,
0.05625132471323013,
-0.009510956704616547,
0.029760541394352913,
-0.1404469609260559,
0.07289744168519974,
0.08861259371042252,
0.005517824552953243,
0.008017187938094139,
-0.3054029643535614,
-0.11170195788145065,
0.03379688411951065,
0.05245375633239746,
0.03670698031783104,
-0.12633968889713287,
-0.0887121707201004,
-0.06801740825176239,
-0.13964064419269562,
0.015763353556394577,
0.05241571366786957,
0.07351648807525635,
-0.019989969208836555,
-0.014513911679387093,
0.031241796910762787,
-0.029324837028980255,
0.12697374820709229,
-0.0018549843225628138,
0.024989064782857895,
-0.06719764322042465,
-0.03355104848742485,
0.08753135055303574,
-0.043567776679992676,
0.0863674208521843,
-0.11591406166553497,
0.012561745941638947,
-0.06905791908502579,
-0.012246563099324703,
-0.06933283805847168,
0.06618189066648483,
-0.061866603791713715,
-0.02711832895874977,
-0.016717538237571716,
0.07553696632385254,
0.056054309010505676,
0.0442032665014267,
0.05642559751868248,
-0.0400627925992012,
0.06407888978719711,
0.2217288315296173,
0.11275508254766464,
-0.051283784210681915,
-0.042560119181871414,
-0.01726149395108223,
-0.026873944327235222,
0.007230780087411404,
-0.03357832133769989,
-0.0024366870056837797,
0.1034824475646019,
-0.032064057886600494,
0.10623238235712051,
0.004016768652945757,
-0.08382435888051987,
-0.056007832288742065,
0.09442516416311264,
-0.13375705480575562,
-0.16134744882583618,
-0.02137739025056362,
0.135798841714859,
-0.0433795228600502,
0.03184393420815468,
0.2052423357963562,
-0.006202720571309328,
-0.026777660474181175,
0.025456096976995468,
0.011096854694187641,
-0.037082333117723465,
0.1299593150615692,
-0.010295010171830654,
0.04636668413877487,
-0.08916615694761276,
0.02747473493218422,
0.06096233054995537,
-0.09801500290632248,
-0.0018911283696070313,
0.15191660821437836,
-0.08048386871814728,
-0.09603728353977203,
-0.05447497218847275,
0.1514333337545395,
-0.01753644086420536,
0.002998960204422474,
-0.07260959595441818,
-0.0895141139626503,
0.03653153032064438,
0.1508188396692276,
0.05437562242150307,
-0.003121785121038556,
0.03204876184463501,
-0.06257577240467072,
-0.027922818437218666,
0.12421747297048569,
0.02288185805082321,
0.09090317785739899,
-0.10892058163881302,
-0.02104068174958229,
-0.050687097012996674,
-0.0032096654176712036,
-0.0239880234003067,
-0.003799101337790489,
-0.11882542818784714,
-0.05536872148513794,
-0.06856923550367355,
-0.02872883528470993,
-0.11580760031938553,
-0.023104410618543625,
-0.0148091996088624,
-0.011386520229279995,
-0.012285086326301098,
-0.012348552234470844,
-0.03239202871918678,
-0.0530671663582325,
-0.017330875620245934,
0.08903282135725021,
-0.1017034575343132,
-0.03399532660841942,
0.01307227835059166,
-0.07747264206409454,
0.08326515555381775,
-0.012225812301039696,
0.009673372842371464,
-0.04253552109003067,
-0.09642799943685532,
-0.047144655138254166,
0.027713391929864883,
0.02432321198284626,
0.03317389264702797,
-0.1395789533853531,
-0.0035958499647676945,
-0.018652960658073425,
-0.03012697584927082,
0.0016761461738497019,
0.10906535387039185,
-0.08176553249359131,
0.04307286813855171,
-0.003887276630848646,
-0.080922432243824,
-0.055826470255851746,
-0.00016033581050578505,
0.035269130021333694,
-0.020275332033634186,
0.1266368180513382,
-0.06398121267557144,
0.04195407032966614,
-0.16182608902454376,
0.016024945303797722,
0.00005489433169714175,
-0.13100075721740723,
-0.004495127126574516,
-0.04357920214533806,
0.030336959287524223,
-0.019998544827103615,
0.0648876428604126,
0.007458403240889311,
-0.09922052919864655,
0.04964210093021393,
-0.028320780023932457,
-0.03563299402594566,
0.04389374703168869,
0.05503404513001442,
0.06954964995384216,
-0.009788933210074902,
-0.03995988145470619,
0.03859995678067207,
0.03156277909874916,
0.07953867316246033,
0.025100354105234146,
0.10419688373804092,
0.02946186065673828,
0.08319027721881866,
0.1036076694726944,
-0.016155816614627838,
-0.047010019421577454,
0.040054239332675934,
-0.04480792582035065,
0.09596127271652222,
-0.03049146570265293,
0.134913831949234,
0.13001395761966705,
-0.13722094893455505,
0.06422954797744751,
0.003115203930065036,
-0.03696707263588905,
-0.064303457736969,
-0.14214883744716644,
-0.10085468739271164,
-0.05445185303688049,
-0.017875531688332558,
-0.1281179040670395,
0.0020592377986758947,
-0.029638076201081276,
0.034088123589754105,
0.005129644647240639,
0.14416128396987915,
-0.015205197967588902,
-0.00752267288044095,
0.0748148187994957,
-0.0009280004887841642,
-0.04373423382639885,
0.00616227462887764,
-0.004751477390527725,
-0.02901453711092472,
0.007363146170973778,
-0.027238693088293076,
0.03736912086606026,
0.008351418189704418,
0.04174897447228432,
-0.027849815785884857,
-0.12093910574913025,
0.00455124955624342,
-0.0032549602910876274,
0.000023906895876280032,
0.03489476069808006,
0.011903801932930946,
-0.047830965369939804,
-0.023588210344314575,
0.11316654831171036,
-0.019527493044734,
-0.07777620851993561,
-0.09303058683872223,
0.14712116122245789,
-0.025236276909708977,
0.04253339394927025,
-0.023136382922530174,
-0.03902706503868103,
-0.011823761276900768,
0.2122587412595749,
0.23245151340961456,
-0.05775346979498863,
-0.008755363523960114,
0.033549290150403976,
0.006954081356525421,
0.04132149741053581,
0.03738899901509285,
0.0563274584710598,
0.15100854635238647,
-0.05445811524987221,
0.07289785146713257,
-0.035741548985242844,
-0.06764572113752365,
-0.09092894196510315,
0.02599378488957882,
0.040914539247751236,
0.024682529270648956,
0.028337305411696434,
0.08725295215845108,
-0.023363787680864334,
-0.00830257497727871,
-0.029805025085806847,
-0.15428373217582703,
-0.12376934289932251,
-0.08187124133110046,
0.04993731528520584,
0.03162764757871628,
0.09609954059123993,
-0.02002827823162079,
-0.06132788956165314,
0.08989985287189484,
-0.03863148018717766,
-0.05747615545988083,
-0.08327547460794449,
0.03589770197868347,
-0.12358395010232925,
0.1384902000427246,
-0.014335447922348976,
0.034005437046289444,
0.11321718245744705,
-0.061330657452344894,
-0.1231926828622818,
0.03116900660097599,
0.03488465026021004,
-0.014565711840987206,
0.056591834872961044,
0.1294688880443573,
0.00449182465672493,
0.11763232946395874,
0.034287795424461365,
-0.11550827324390411,
0.043358683586120605,
0.050991255789995193,
-0.04488883912563324,
-0.0298650823533535,
0.12446750700473785,
-0.06161074340343475,
0.10992616415023804,
0.16085773706436157,
-0.036328140646219254,
-0.020791655406355858,
-0.0336017981171608,
0.033953696489334106,
0.08609063923358917,
0.09996232390403748,
-0.04962247237563133,
-0.2142663598060608,
0.018817709758877754,
-0.030001511797308922,
0.022696135565638542,
-0.23578563332557678,
-0.10790016502141953,
-0.10445978492498398,
-0.0070149158127605915,
-0.10552076995372772,
0.1000761166214943,
0.1469956934452057,
-0.003039542119950056,
-0.009777533821761608,
-0.12035378813743591,
-0.0025247568264603615,
0.07504251599311829,
-0.07605046033859253,
-0.08301708847284317
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-large-lora-1.57M-squad-model2
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 65
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-large", "model-index": [{"name": "roberta-large-lora-1.57M-squad-model2", "results": []}]} | question-answering | varun-v-rao/roberta-large-lora-1.57M-squad-model2 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-large",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:15:55+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us
|
# roberta-large-lora-1.57M-squad-model2
This model is a fine-tuned version of roberta-large on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 65
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-large-lora-1.57M-squad-model2\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 65\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n",
"# roberta-large-lora-1.57M-squad-model2\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 65\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
70,
41,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n# roberta-large-lora-1.57M-squad-model2\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 65\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.08628564327955246,
0.1204436644911766,
-0.0023692853283137083,
0.09558846056461334,
0.14734505116939545,
0.008069266565144062,
0.1204904243350029,
0.1206308975815773,
-0.09404516220092773,
0.06422611325979233,
0.07532741129398346,
0.045794278383255005,
0.02471904642879963,
0.1048365905880928,
-0.028502536937594414,
-0.21740582585334778,
-0.008179706521332264,
-0.019202716648578644,
-0.09648295491933823,
0.10011418908834457,
0.10703703761100769,
-0.10460344702005386,
0.08124324679374695,
-0.01697753742337227,
-0.152287095785141,
0.053049128502607346,
-0.017025858163833618,
-0.03740499168634415,
0.09971154481172562,
0.03438189625740051,
0.09712555259466171,
0.005171172320842743,
0.14086705446243286,
-0.2177458107471466,
0.008614426478743553,
0.06937361508607864,
0.02843828685581684,
0.07506591826677322,
0.02605622075498104,
-0.01686936616897583,
0.08884910494089127,
-0.16610446572303772,
0.09807361662387848,
0.02878008782863617,
-0.07621103525161743,
-0.17076057195663452,
-0.09304960072040558,
0.05157500132918358,
0.0834818035364151,
0.10039785504341125,
0.00431022560223937,
0.15796980261802673,
-0.08254880458116531,
0.08630942553281784,
0.23546288907527924,
-0.26728832721710205,
-0.06600649654865265,
0.0818355455994606,
0.0775112733244896,
0.10201912373304367,
-0.12521643936634064,
-0.0026651350781321526,
0.05606807768344879,
0.008299501612782478,
0.08482852578163147,
-0.0389813594520092,
-0.09567089378833771,
0.019331060349941254,
-0.1330844908952713,
-0.0017747444799169898,
0.18086695671081543,
0.05686492845416069,
-0.04084758087992668,
-0.07302583754062653,
-0.03904533013701439,
-0.045646000653505325,
-0.025053635239601135,
-0.03728706017136574,
0.050216104835271835,
-0.06566792726516724,
-0.09612470120191574,
-0.07124452292919159,
-0.0873640775680542,
-0.08970349282026291,
0.004917382262647152,
0.10450824350118637,
0.04370546713471413,
0.01104950811713934,
-0.04342520982027054,
0.080828957259655,
-0.03390241041779518,
-0.10295587033033371,
-0.034547559916973114,
0.006125433370471001,
-0.08365099132061005,
-0.05301358550786972,
-0.01865522563457489,
-0.03281192481517792,
0.034577880054712296,
0.1747005730867386,
-0.054525259882211685,
0.05009398236870766,
0.01778160035610199,
0.0055147758685052395,
-0.025216439738869667,
0.14325012266635895,
-0.06705670058727264,
-0.06165864318609238,
0.021351303905248642,
0.08076512068510056,
0.021858813241124153,
0.0060857078060507774,
-0.09056182205677032,
-0.02743583545088768,
0.0845087468624115,
0.08548992872238159,
-0.03282121941447258,
0.030704600736498833,
0.005028203595429659,
-0.008254782296717167,
0.021313723176717758,
-0.13047148287296295,
0.03810952231287956,
-0.01534206047654152,
-0.06460344046354294,
-0.05759194865822792,
0.035517461597919464,
-0.007482076529413462,
0.009540945291519165,
0.05521745607256889,
-0.08639198541641235,
-0.009609484113752842,
-0.07968802750110626,
-0.09037646651268005,
0.01512970495969057,
-0.06722571700811386,
0.014261821284890175,
-0.08467002213001251,
-0.1948431432247162,
-0.045230183750391006,
0.009817246347665787,
-0.06153174117207527,
-0.01724667102098465,
-0.03847089409828186,
-0.07066893577575684,
-0.0154018085449934,
0.0020755198784172535,
0.13216792047023773,
-0.037654031068086624,
0.08739257603883743,
0.025357302278280258,
0.033243902027606964,
0.012095768004655838,
0.03184404969215393,
-0.0999893769621849,
0.028885815292596817,
-0.13973020017147064,
0.04464549571275711,
-0.10116036981344223,
0.05882330983877182,
-0.1267354041337967,
-0.09513991326093674,
-0.0014033537590876222,
-0.007937293499708176,
0.04976160451769829,
0.10930713266134262,
-0.14571623504161835,
-0.018176259472966194,
0.16407287120819092,
-0.07808385044336319,
-0.09255508333444595,
0.09965518116950989,
-0.0548575334250927,
0.05760267376899719,
0.07816065102815628,
0.16857166588306427,
0.11872806400060654,
-0.13134071230888367,
-0.011362428776919842,
0.001547320862300694,
0.04917725548148155,
-0.015982722863554955,
0.06424513459205627,
-0.0020533581264317036,
0.024171823635697365,
0.010054998099803925,
-0.08062592893838882,
0.008168447762727737,
-0.08945667743682861,
-0.07784991711378098,
-0.04721808061003685,
-0.1123838871717453,
0.012846022844314575,
0.04534491151571274,
0.028968360275030136,
-0.08089137077331543,
-0.09995850920677185,
0.10518744587898254,
0.13587278127670288,
-0.05489474907517433,
0.0037966654635965824,
-0.06699032336473465,
0.044552143663167953,
-0.06901673972606659,
-0.032376132905483246,
-0.17991411685943604,
-0.12440185248851776,
0.016811273992061615,
-0.04179093614220619,
0.05489795282483101,
0.029186950996518135,
0.0775265321135521,
0.0573885440826416,
-0.06275820732116699,
-0.009084810502827168,
-0.059154294431209564,
0.0031519215553998947,
-0.08866076916456223,
-0.19229666888713837,
-0.04241817817091942,
-0.03661002591252327,
0.13452215492725372,
-0.2609156668186188,
0.021421389654278755,
-0.03288421034812927,
0.10827454179525375,
0.03586791455745697,
-0.03850634768605232,
0.004007161594927311,
0.032698553055524826,
-0.008987599983811378,
-0.0928855612874031,
0.023144932463765144,
-0.016108093783259392,
-0.0760422870516777,
-0.05632559582591057,
-0.14231114089488983,
0.07242802530527115,
0.06790266931056976,
0.07087607681751251,
-0.10864503681659698,
0.004207220859825611,
-0.05471847951412201,
-0.03403957188129425,
-0.07616724073886871,
-0.014472732320427895,
0.11766798049211502,
-0.002327104564756155,
0.11832383275032043,
-0.06345651298761368,
-0.06572453677654266,
0.000774191168602556,
-0.02013517916202545,
0.005006179213523865,
0.10697207599878311,
0.08155211061239243,
-0.13978736102581024,
0.09557298570871353,
0.07565902918577194,
-0.06491623818874359,
0.14507199823856354,
-0.05086009204387665,
-0.08800332993268967,
-0.036580078303813934,
0.02640967071056366,
0.0027674620505422354,
0.13616400957107544,
-0.06093982979655266,
-0.005914622452110052,
0.022825762629508972,
-0.005561958998441696,
0.016560418531298637,
-0.1577383279800415,
-0.03282354772090912,
0.023792795836925507,
-0.045300908386707306,
0.009786166250705719,
-0.01133724395185709,
0.02075323462486267,
0.09248937666416168,
0.012435946613550186,
-0.02285531349480152,
0.005859224125742912,
-0.008122319355607033,
-0.07727523893117905,
0.19500933587551117,
-0.07096441835165024,
-0.11905001103878021,
-0.133577361702919,
0.026100309565663338,
-0.03885801509022713,
-0.028998348861932755,
0.025209937244653702,
-0.08746979385614395,
-0.054143063724040985,
-0.10092951357364655,
0.017692014575004578,
-0.02932792715728283,
-0.019504046067595482,
-0.0004094771284144372,
0.040570493787527084,
0.08007887005805969,
-0.14346365630626678,
0.02543918415904045,
-0.03357135131955147,
-0.12878787517547607,
-0.014750349335372448,
0.04671040177345276,
0.1457597017288208,
0.09587214142084122,
-0.02053108997642994,
0.007720873691141605,
-0.03485005721449852,
0.20747894048690796,
-0.07284671068191528,
0.011761640198528767,
0.11755518615245819,
0.026174111291766167,
0.0406528078019619,
0.12957032024860382,
0.03459561988711357,
-0.09409500658512115,
0.03805044665932655,
0.08541804552078247,
-0.0336339958012104,
-0.2547321617603302,
-0.029606478288769722,
-0.032478004693984985,
-0.05973105505108833,
0.07001025229692459,
0.06613128632307053,
0.027085674926638603,
0.048495739698410034,
-0.005016063805669546,
0.03432360291481018,
-0.010134581476449966,
0.08884400129318237,
0.09276097267866135,
0.029184985905885696,
0.10089026391506195,
-0.06033625453710556,
-0.05368136987090111,
0.05525064468383789,
0.017185000702738762,
0.31117570400238037,
-0.015418533235788345,
0.1187945231795311,
0.053139109164476395,
0.12224596738815308,
-0.02569279447197914,
0.02584873139858246,
0.006784969940781593,
-0.011715014465153217,
0.005437551066279411,
-0.058860793709754944,
0.012581756338477135,
0.03250904381275177,
-0.03562438488006592,
0.0470428504049778,
-0.06395050138235092,
0.0358785018324852,
0.022513046860694885,
0.2131364345550537,
0.030552001670002937,
-0.278706818819046,
-0.07649333775043488,
0.025389574468135834,
-0.02568056620657444,
-0.03985818848013878,
0.0187771487981081,
0.13807320594787598,
-0.10699480772018433,
0.033285629004240036,
-0.05025589466094971,
0.08174948394298553,
-0.03502641245722771,
-0.011490631848573685,
0.01625729911029339,
0.12253322452306747,
-0.017233388498425484,
0.10002556443214417,
-0.18430301547050476,
0.2375001609325409,
0.02542950212955475,
0.1110132709145546,
-0.03781487047672272,
0.012567335739731789,
0.006035097409039736,
0.08240178227424622,
0.1338871270418167,
-0.015791377052664757,
-0.03704230114817619,
-0.1808379739522934,
-0.0647100955247879,
0.048944588750600815,
0.11161120980978012,
-0.029913250356912613,
0.10356999933719635,
-0.05788836628198624,
-0.009012625552713871,
0.05487504228949547,
-0.09960034489631653,
-0.15288636088371277,
-0.11975035816431046,
-0.010770940221846104,
0.0001130820601247251,
-0.07484035193920135,
-0.0864885151386261,
-0.08244546502828598,
-0.029182609170675278,
0.16516242921352386,
0.011379572562873363,
-0.038901716470718384,
-0.12858687341213226,
0.09548361599445343,
0.1276014745235443,
-0.07789232581853867,
0.008976925164461136,
0.018921179696917534,
0.10679952800273895,
0.03576701134443283,
-0.06556897610425949,
0.056447360664606094,
-0.05669335648417473,
-0.15359075367450714,
-0.044558335095644,
0.12628959119319916,
0.052838221192359924,
0.04517762362957001,
-0.00010759438009699807,
0.027545861899852753,
0.013166461139917374,
-0.09512288868427277,
0.000947140681091696,
0.07287195324897766,
0.04636459797620773,
0.05693695694208145,
-0.08689854294061661,
0.019446030259132385,
-0.04068950191140175,
-0.024018704891204834,
0.13261419534683228,
0.22063343226909637,
-0.09244083613157272,
0.06857031583786011,
0.08327952027320862,
-0.08061592280864716,
-0.19148191809654236,
0.07036609202623367,
0.04912351816892624,
0.005585249979048967,
0.07818780094385147,
-0.1549956351518631,
0.1218319907784462,
0.1119038388133049,
-0.02091323956847191,
0.057569149881601334,
-0.311900794506073,
-0.12823925912380219,
0.07878651469945908,
0.12997935712337494,
0.016946643590927124,
-0.16424372792243958,
-0.04277946427464485,
-0.023085813969373703,
-0.13677731156349182,
0.0707615539431572,
-0.11455263942480087,
0.10842257738113403,
-0.0031410465016961098,
0.09702416509389877,
0.020908938720822334,
-0.03992956504225731,
0.13181784749031067,
0.020269447937607765,
0.10083811730146408,
-0.05647148936986923,
-0.02374320477247238,
0.13599373400211334,
-0.06185602396726608,
0.07084420323371887,
-0.05616825819015503,
0.07404473423957825,
-0.11431941390037537,
-0.036450229585170746,
-0.06887350231409073,
0.06612275540828705,
-0.0545981340110302,
-0.058437786996364594,
-0.07535599172115326,
0.06938420981168747,
0.02810845710337162,
-0.02835172414779663,
0.07958059757947922,
0.00036176363937556744,
0.09279005229473114,
0.06598438322544098,
0.09985137730836868,
0.03622930869460106,
-0.08383592963218689,
0.008720995858311653,
-0.02782576158642769,
0.05070452764630318,
-0.1463647186756134,
0.026436975225806236,
0.10968166589736938,
0.06168399006128311,
0.13449466228485107,
0.023368969559669495,
-0.05864369869232178,
0.010251905769109726,
0.037140436470508575,
-0.08584010601043701,
-0.20449599623680115,
0.007590591907501221,
-0.06430233269929886,
-0.15424907207489014,
0.04015284776687622,
0.10613296180963516,
-0.06830007582902908,
-0.015886595472693443,
-0.027062352746725082,
0.03405797481536865,
-0.02135474979877472,
0.17616453766822815,
0.08082907646894455,
0.059457022696733475,
-0.07650136202573776,
0.09423478692770004,
0.08171876519918442,
-0.021179843693971634,
0.04110727086663246,
0.07631796598434448,
-0.06699623167514801,
-0.03107684850692749,
0.05972200259566307,
0.20086388289928436,
-0.06621517241001129,
-0.033638179302215576,
-0.10843079537153244,
-0.07634881883859634,
0.028855180367827415,
0.1420256346464157,
0.042801160365343094,
-0.024324212223291397,
-0.002642946783453226,
0.03401348739862442,
-0.15945173799991608,
0.1164713054895401,
0.040665481239557266,
0.07556235790252686,
-0.14655163884162903,
0.07116496562957764,
0.0031710457988083363,
0.031678326427936554,
-0.015307513996958733,
0.03677014261484146,
-0.10562793910503387,
-0.013667472638189793,
-0.13623853027820587,
-0.006115380208939314,
-0.01467962097376585,
0.014851736836135387,
-0.016804812476038933,
-0.06586910784244537,
-0.057733118534088135,
0.05215315893292427,
-0.06181732937693596,
-0.052300162613391876,
0.022328274324536324,
0.063271664083004,
-0.1671844720840454,
-0.022309213876724243,
0.027922846376895905,
-0.07615529745817184,
0.059400297701358795,
0.029147375375032425,
0.03969063237309456,
0.032181113958358765,
-0.13274480402469635,
0.003363207448273897,
0.02193048596382141,
0.0197153203189373,
0.0633566677570343,
-0.09676756709814072,
-0.012436822056770325,
-0.027007313445210457,
0.07145307958126068,
0.020450789481401443,
0.013347121886909008,
-0.12082602083683014,
0.01421282347291708,
-0.09381162375211716,
-0.05300632864236832,
-0.052907902747392654,
0.041176993399858475,
0.08646489679813385,
0.03676755353808403,
0.1471269130706787,
-0.08461105078458786,
0.04484022781252861,
-0.21776309609413147,
-0.03176277503371239,
-0.005977439694106579,
-0.03764388710260391,
-0.07530570030212402,
-0.03452518209815025,
0.06813279539346695,
-0.055988624691963196,
0.12301158905029297,
0.0053331851959228516,
0.08242719620466232,
0.04734718054533005,
-0.034435778856277466,
0.022943703457713127,
0.004111729562282562,
0.17265716195106506,
0.04342947527766228,
-0.020250119268894196,
0.07448393851518631,
-0.011803281493484974,
0.06735293567180634,
0.015250126831233501,
0.16720779240131378,
0.18908344209194183,
-0.034681711345911026,
0.037330273538827896,
0.08836846798658371,
-0.09534414112567902,
-0.09905857592821121,
0.10883402079343796,
-0.0013112765736877918,
0.06597405672073364,
-0.048399459570646286,
0.15790772438049316,
0.16098779439926147,
-0.1581116020679474,
0.03710072487592697,
-0.058867182582616806,
-0.10658805072307587,
-0.10750880837440491,
-0.03451156243681908,
-0.1055501401424408,
-0.10196203738451004,
0.03306879848241806,
-0.1211085319519043,
0.020870478823781013,
0.07784496247768402,
0.0064174416474998,
0.011440382339060307,
0.17301784455776215,
-0.007506903260946274,
0.032083552330732346,
0.029125701636075974,
0.004434030037373304,
-0.010616286657750607,
-0.044037651270627975,
-0.027264412492513657,
0.07052180916070938,
-0.0030634235590696335,
0.0555180087685585,
-0.03497421368956566,
0.001299917115829885,
0.04275951161980629,
-0.029393289238214493,
-0.07301438599824905,
0.007225836627185345,
0.04820752516388893,
0.04287531599402428,
0.039606790989637375,
0.05767088755965233,
-0.00969341304153204,
-0.03552822396159172,
0.2908284664154053,
-0.06269705295562744,
-0.09619857370853424,
-0.11023996025323868,
0.16541509330272675,
0.027230968698859215,
-0.0040830946527421474,
0.04029650613665581,
-0.1337006390094757,
0.02604595199227333,
0.18277636170387268,
0.16861224174499512,
-0.031151147559285164,
-0.007285759784281254,
-0.022322027012705803,
-0.013993766158819199,
-0.05027586594223976,
0.08938068151473999,
0.09818515926599503,
0.051511093974113464,
-0.05603218451142311,
-0.034845542162656784,
-0.0073250760324299335,
-0.02254842035472393,
-0.07591630518436432,
0.07226401567459106,
0.02940886840224266,
0.018396813422441483,
-0.03701125457882881,
0.09004826098680496,
0.009344356134533882,
-0.22804296016693115,
0.05698535963892937,
-0.15362155437469482,
-0.16623938083648682,
-0.02675016038119793,
0.09421078860759735,
-0.022508060559630394,
0.041376322507858276,
-0.03216582164168358,
-0.007022892590612173,
0.10312585532665253,
-0.007647327613085508,
-0.08912505954504013,
-0.10984431952238083,
0.08607695251703262,
-0.045799240469932556,
0.23542766273021698,
-0.012989255599677563,
0.047079406678676605,
0.117290198802948,
-0.022279130294919014,
-0.14537666738033295,
0.06155485287308693,
0.07297508418560028,
-0.06574172526597977,
0.03885211795568466,
0.15590490400791168,
-0.04824643209576607,
0.11937230080366135,
0.04061003029346466,
-0.11017198860645294,
-0.00683259405195713,
-0.027136368677020073,
-0.02768111601471901,
-0.10488526523113251,
0.0029295117128640413,
-0.07831145823001862,
0.15172939002513885,
0.18068450689315796,
-0.039511267095804214,
0.019531650468707085,
-0.07756359130144119,
0.037589360028505325,
0.08806968480348587,
0.08072283118963242,
-0.010021112859249115,
-0.18887341022491455,
0.014369729906320572,
0.010172384791076183,
0.0300008375197649,
-0.29729029536247253,
-0.07916664332151413,
0.034181006252765656,
-0.03975924104452133,
-0.04712151736021042,
0.11016102135181427,
0.11339009553194046,
0.04519147425889969,
-0.03575790300965309,
-0.16293010115623474,
-0.04959971085190773,
0.1497361958026886,
-0.16506119072437286,
-0.05099382624030113
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
<details><summary>See axolotl config</summary>
axolotl version: `0.4.0`
```yaml
base_model: croissantllm/CroissantLLMBase
model_type: LlamaForCausalLM
tokenizer_type: LlamaTokenizerFast
is_llama_derived_model: true
load_in_8bit: false
load_in_4bit: false
strict: false
datasets:
- path: manu/mmlu_auxiliary_train_formatted_extra
split: train
type: completion
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./out
sequence_len: 2048
sample_packing: false
pad_to_sequence_len: false
adapter:
lora_model_dir:
lora_r:
lora_alpha:
lora_dropout:
lora_target_linear:
lora_fan_in_fan_out:
wandb_project:
wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
gradient_accumulation_steps: 2
micro_batch_size: 24
num_epochs: 1
optimizer: adamw_bnb_8bit
lr_scheduler: cosine
learning_rate: 0.0002
train_on_inputs: false
group_by_length: false
bf16: auto
fp16:
tf32: false
gradient_checkpointing: true
early_stopping_patience:
resume_from_checkpoint:
local_rank:
logging_steps: 1
xformers_attention:
flash_attention: true
flash_attn_cross_entropy: false
flash_attn_rms_norm: true
flash_attn_fuse_qkv: false
flash_attn_fuse_mlp: true
warmup_steps: 50
evals_per_epoch: 4
eval_table_size:
saves_per_epoch: 1
debug:
deepspeed: #deepspeed_configs/zero2.json # multi-gpu only
weight_decay: 0.1
fsdp:
fsdp_config:
special_tokens:
```
</details><br>
# out
This model is a fine-tuned version of [croissantllm/CroissantLLMBase](https://huggingface.co/croissantllm/CroissantLLMBase) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.4229
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 24
- eval_batch_size: 24
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 48
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 50
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 3.2302 | 0.0 | 1 | 3.1796 |
| 2.3772 | 0.25 | 134 | 2.3697 |
| 1.8538 | 0.5 | 268 | 1.8599 |
| 1.5117 | 0.75 | 402 | 1.4229 |
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.1.2+cu121
- Datasets 2.16.1
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "base_model": "croissantllm/CroissantLLMBase", "model-index": [{"name": "out", "results": []}]} | text-generation | manu/croissant_mmlu | [
"transformers",
"pytorch",
"tensorboard",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"base_model:croissantllm/CroissantLLMBase",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T16:18:23+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #safetensors #llama #text-generation #generated_from_trainer #base_model-croissantllm/CroissantLLMBase #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| <img src="URL alt="Built with Axolotl" width="200" height="32"/>
See axolotl config
axolotl version: '0.4.0'
out
===
This model is a fine-tuned version of croissantllm/CroissantLLMBase on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 1.4229
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0002
* train\_batch\_size: 24
* eval\_batch\_size: 24
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 48
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: cosine
* lr\_scheduler\_warmup\_steps: 50
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.38.0.dev0
* Pytorch 2.1.2+cu121
* Datasets 2.16.1
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 24\n* eval\\_batch\\_size: 24\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 50\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #pytorch #tensorboard #safetensors #llama #text-generation #generated_from_trainer #base_model-croissantllm/CroissantLLMBase #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 24\n* eval\\_batch\\_size: 24\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 50\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
83,
144,
4,
38
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #safetensors #llama #text-generation #generated_from_trainer #base_model-croissantllm/CroissantLLMBase #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 24\n* eval\\_batch\\_size: 24\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 48\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* lr\\_scheduler\\_warmup\\_steps: 50\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.0"
] | [
-0.11413402110338211,
0.1473250687122345,
-0.003845338011160493,
0.07173480838537216,
0.10013110935688019,
0.021188462153077126,
0.12177955359220505,
0.1244693472981453,
-0.057829599827528,
0.13629718124866486,
0.13471752405166626,
0.12358919531106949,
0.04972682148218155,
0.13707655668258667,
-0.01194140873849392,
-0.2823347747325897,
0.004557095933705568,
-0.004531492479145527,
-0.13133540749549866,
0.12804502248764038,
0.06446482986211777,
-0.1183379516005516,
0.07332832366228104,
-0.005733155179768801,
-0.11276327073574066,
-0.035295743495225906,
-0.04472370818257332,
-0.04552456736564636,
0.10857260972261429,
0.016129478812217712,
0.09450458735227585,
0.041121706366539,
0.07782646268606186,
-0.2364056557416916,
0.013402090407907963,
0.05675079673528671,
0.009300313889980316,
0.08032859861850739,
0.09506899118423462,
-0.020362086594104767,
0.15458807349205017,
-0.10704642534255981,
0.06335164606571198,
0.03503446653485298,
-0.09462732076644897,
-0.19958515465259552,
-0.08170219510793686,
0.0681028813123703,
0.11454489082098007,
0.07458332180976868,
-0.01396262738853693,
0.069998599588871,
-0.07670309394598007,
0.08172030001878738,
0.2437221109867096,
-0.266507089138031,
-0.08261413127183914,
0.036947257816791534,
0.06788266450166702,
0.05637790262699127,
-0.1205296516418457,
-0.011727944947779179,
0.027918044477701187,
0.010343587957322598,
0.10888837277889252,
0.0006171578424982727,
0.13638901710510254,
0.0011885777348652482,
-0.15383602678775787,
-0.052507393062114716,
0.09226881712675095,
0.07612654566764832,
-0.025021888315677643,
-0.10139520466327667,
-0.05620379373431206,
-0.21212251484394073,
-0.04602326452732086,
-0.010322934947907925,
0.03099086508154869,
-0.04313087835907936,
-0.08814284205436707,
0.0016119761858135462,
-0.07209666073322296,
-0.07202626764774323,
0.03799869865179062,
0.14242173731327057,
0.05081666633486748,
-0.04393615201115608,
0.023229414597153664,
0.11828385293483734,
0.094305619597435,
-0.15165379643440247,
-0.0010845557553693652,
0.0263545960187912,
-0.061889562755823135,
-0.018592119216918945,
-0.005233473144471645,
0.011895301751792431,
0.033796217292547226,
0.1604929119348526,
-0.04260295629501343,
0.058347515761852264,
0.05476102977991104,
0.029500078409910202,
-0.09015484154224396,
0.13223524391651154,
-0.06995944678783417,
-0.07968270778656006,
-0.03592279180884361,
0.124559186398983,
0.029602142050862312,
-0.018348125740885735,
-0.08778001368045807,
0.0335257388651371,
0.11079062521457672,
0.04968477785587311,
0.005941139534115791,
0.00904445257037878,
-0.06425552070140839,
-0.025343142449855804,
0.05594230815768242,
-0.08788878470659256,
0.046586986631155014,
0.0345228873193264,
-0.0578373558819294,
0.004443905781954527,
-0.007977362722158432,
0.007613393012434244,
-0.007634537760168314,
0.1488339602947235,
-0.11024781316518784,
-0.028464363887906075,
-0.07042865455150604,
-0.07956285029649734,
0.03730987384915352,
-0.0929049476981163,
-0.006193786393851042,
-0.06432704627513885,
-0.11905556172132492,
-0.05761685222387314,
0.057316940277814865,
-0.05149618536233902,
-0.0735214576125145,
-0.07006092369556427,
-0.09059695899486542,
0.04783474653959274,
0.006498781498521566,
0.10808050632476807,
-0.04973125457763672,
0.11450403928756714,
-0.014799331314861774,
0.0576934814453125,
0.07968191057443619,
0.05075380206108093,
-0.06561450660228729,
0.06912994384765625,
-0.15578784048557281,
0.059291474521160126,
-0.07127799093723297,
0.057071585208177567,
-0.13233761489391327,
-0.11290009319782257,
-0.03263986483216286,
-0.01618996262550354,
0.065611332654953,
0.13843189179897308,
-0.1479237973690033,
-0.09999450296163559,
0.1925673633813858,
-0.06666715443134308,
-0.10687489807605743,
0.13304992020130157,
-0.01860722340643406,
-0.045544642955064774,
0.02599024772644043,
0.1377238631248474,
0.1051645278930664,
-0.0805075541138649,
-0.030831150710582733,
-0.03159293532371521,
0.1050289124250412,
0.0002132902736775577,
0.11072304844856262,
-0.030914250761270523,
0.020378492772579193,
-0.011151131242513657,
-0.05184808745980263,
0.0471370555460453,
-0.09102366119623184,
-0.08857795596122742,
-0.01293141394853592,
-0.07875090837478638,
0.04577050730586052,
0.05526253581047058,
0.029043743386864662,
-0.07290252298116684,
-0.13917484879493713,
-0.0133454455062747,
0.09729312360286713,
-0.08624110370874405,
0.020675087347626686,
-0.03503727540373802,
0.10415946692228317,
-0.0362144410610199,
-0.008791755884885788,
-0.13976287841796875,
-0.0655592530965805,
0.041387107223272324,
-0.04614328220486641,
0.0076436917297542095,
-0.019898511469364166,
0.06726846098899841,
0.09586885571479797,
-0.038055405020713806,
-0.0764540359377861,
-0.053617119789123535,
-0.019407950341701508,
-0.08671583980321884,
-0.23953987658023834,
-0.05795201659202576,
-0.032641276717185974,
0.1658637672662735,
-0.2213837206363678,
0.03167091682553291,
0.018892183899879456,
0.12952426075935364,
0.028766369447112083,
-0.06206469610333443,
-0.004032591823488474,
0.05642487108707428,
-0.041013363748788834,
-0.07512480765581131,
0.04815889522433281,
-0.013525551185011864,
-0.11871219426393509,
-0.026831982657313347,
-0.1439690887928009,
0.1217556744813919,
0.09840425103902817,
0.02394551783800125,
-0.09283363819122314,
-0.08463522791862488,
-0.07811673730611801,
-0.04436729475855827,
-0.015438271686434746,
0.013799643144011497,
0.11531529575586319,
0.03378259390592575,
0.11445261538028717,
-0.08257579058408737,
-0.05758961662650108,
0.04325103014707565,
0.01341975387185812,
-0.03364498168230057,
0.1594698131084442,
0.11036714166402817,
-0.04854907840490341,
0.1258331835269928,
0.13653405010700226,
-0.04151146113872528,
0.13214215636253357,
-0.04247097298502922,
-0.0944061279296875,
-0.043189454823732376,
0.04887404292821884,
0.032011184841394424,
0.0993562713265419,
-0.1131923496723175,
0.0035615928936749697,
0.007260181941092014,
0.007769015617668629,
0.01327156275510788,
-0.18750008940696716,
-0.026124117895960808,
0.04288516193628311,
-0.05392148345708847,
0.02628297172486782,
-0.0031540857162326574,
-0.019811982288956642,
0.0945073664188385,
0.028083566576242447,
-0.05998009815812111,
-0.008402712643146515,
-0.003964001778513193,
-0.07768913358449936,
0.21589729189872742,
-0.08456578850746155,
-0.13123807311058044,
-0.1343163698911667,
0.001985071925446391,
-0.039699457585811615,
0.0005197187419980764,
0.040757764130830765,
-0.08592560142278671,
-0.01749425381422043,
-0.07231458276510239,
0.0012632953003048897,
-0.04398677870631218,
0.04695015400648117,
-0.0077981483191251755,
-0.004041971173137426,
0.018298152834177017,
-0.09336082637310028,
0.011722573079168797,
-0.01511841919273138,
-0.04398014396429062,
0.021517451852560043,
0.04200148209929466,
0.12438629567623138,
0.15806719660758972,
0.030860088765621185,
0.013046813197433949,
-0.026195473968982697,
0.1736280471086502,
-0.09920989722013474,
0.0028050870168954134,
0.09885944426059723,
0.027099521830677986,
0.07714279741048813,
0.12980522215366364,
0.060718558728694916,
-0.07105984538793564,
0.017288586124777794,
0.050738707184791565,
-0.028324270620942116,
-0.20969723165035248,
-0.03739612549543381,
-0.05011763051152229,
0.01890501007437706,
0.13118836283683777,
0.05263245478272438,
0.021700767800211906,
0.05805215239524841,
-0.02604518085718155,
0.02436203882098198,
-0.018789634108543396,
0.08076190948486328,
0.04144591838121414,
0.042342860251665115,
0.10247304290533066,
-0.012659337371587753,
-0.04619525372982025,
0.033864133059978485,
0.001119640888646245,
0.21687737107276917,
-0.0073473178781569,
0.16904276609420776,
0.05014899745583534,
0.15318192541599274,
-0.004219541791826487,
0.05996837839484215,
0.013642429374158382,
-0.017019787803292274,
0.005598261021077633,
-0.06655487418174744,
-0.03003338910639286,
0.05483480542898178,
0.050781916826963425,
0.044833190739154816,
-0.11765643954277039,
0.0511348620057106,
0.0348542258143425,
0.27048352360725403,
0.05881645902991295,
-0.30635949969291687,
-0.07705936580896378,
0.01660657487809658,
-0.03574972227215767,
-0.022037452086806297,
0.0311554204672575,
0.11298409104347229,
-0.08541557937860489,
0.04839033633470535,
-0.08566290885210037,
0.07837719470262527,
-0.07410374283790588,
-0.004819820635020733,
0.062393590807914734,
0.10071870684623718,
-0.00925632007420063,
0.05855073779821396,
-0.2244594842195511,
0.27176278829574585,
-0.003840549848973751,
0.04951687157154083,
-0.051344115287065506,
0.028114505112171173,
0.025204043835401535,
0.015172882936894894,
0.1267091929912567,
-0.007504863664507866,
-0.058914680033922195,
-0.16158725321292877,
-0.12519387900829315,
0.0077065094374120235,
0.12655407190322876,
-0.11263659596443176,
0.11857307702302933,
-0.02021031267940998,
-0.035845108330249786,
0.03844612464308739,
-0.06969156861305237,
-0.07465070486068726,
-0.1161305233836174,
0.03474220260977745,
-0.023164590820670128,
0.022284671664237976,
-0.0890716090798378,
-0.11594422161579132,
-0.08130403608083725,
0.17515598237514496,
-0.1390151083469391,
-0.028065521270036697,
-0.1368962973356247,
0.06707420200109482,
0.13844352960586548,
-0.09913387149572372,
0.047605812549591064,
-0.018189219757914543,
0.10753461718559265,
-0.001211414928548038,
-0.021401531994342804,
0.09829208254814148,
-0.07928799837827682,
-0.21158766746520996,
-0.06150161102414131,
0.15924085676670074,
0.03595638647675514,
0.061047643423080444,
-0.0315508209168911,
0.026582172140479088,
-0.022099027410149574,
-0.10635679215192795,
0.06029164046049118,
0.01758347451686859,
0.05763597786426544,
0.02755575068295002,
-0.04534177854657173,
0.04353486746549606,
-0.04986842721700668,
-0.03628161549568176,
0.12692174315452576,
0.3123571574687958,
-0.09747359156608582,
0.038289669901132584,
0.06529348343610764,
-0.04572533816099167,
-0.16302378475666046,
-0.013008811511099339,
0.0848807841539383,
0.020278146490454674,
-0.0071145398542284966,
-0.2048732340335846,
0.06497333943843842,
0.10167767852544785,
-0.01688559725880623,
0.11811714619398117,
-0.3584245443344116,
-0.13021652400493622,
0.07104083150625229,
0.12078024446964264,
-0.0021578974556177855,
-0.1820363998413086,
-0.07624747604131699,
0.013537924736738205,
-0.08218097686767578,
0.08274227380752563,
-0.009062036871910095,
0.11430144309997559,
-0.025803713127970695,
0.01441275980323553,
0.017365597188472748,
-0.07427453994750977,
0.13738632202148438,
0.013810889795422554,
0.051358263939619064,
-0.015564302913844585,
-0.005677629727870226,
0.009886096231639385,
-0.06241857632994652,
0.007225266192108393,
-0.1087249144911766,
0.0315869078040123,
-0.10607826709747314,
-0.02309533953666687,
-0.07786167412996292,
0.04169714078307152,
-0.06899327784776688,
-0.050350371748209,
-0.03797294571995735,
0.043679360300302505,
0.07934080809354782,
0.0022094298619776964,
0.1290576457977295,
0.004971324931830168,
0.15103405714035034,
0.09838838130235672,
0.07523517310619354,
-0.001326842582784593,
-0.09695124626159668,
-0.02460024319589138,
-0.014114737510681152,
0.03896651789546013,
-0.11347857862710953,
-0.0015843097353354096,
0.15144383907318115,
0.051132284104824066,
0.13038372993469238,
0.06456631422042847,
-0.05578942596912384,
-0.004583206959068775,
0.07691065967082977,
-0.12788286805152893,
-0.15955720841884613,
-0.024038398638367653,
-0.020482808351516724,
-0.1476888209581375,
0.017400527372956276,
0.09608981758356094,
-0.04785079509019852,
-0.008609112352132797,
0.0022785519249737263,
0.031195031479001045,
0.00029270511004142463,
0.18763671815395355,
0.04951786622405052,
0.08406724035739899,
-0.08304472267627716,
0.06419941782951355,
0.05519687011837959,
-0.15922211110591888,
0.01577403023838997,
0.11076252162456512,
-0.07091466337442398,
-0.019099518656730652,
0.04377160593867302,
0.09556485712528229,
-0.04257043078541756,
-0.011903144419193268,
-0.12485133856534958,
-0.10777250677347183,
0.08974805474281311,
0.10473720729351044,
0.04997485131025314,
0.033055517822504044,
-0.030103081837296486,
0.03920343145728111,
-0.1150742620229721,
0.13000796735286713,
0.07275791466236115,
0.08970407396554947,
-0.14661696553230286,
0.12675385177135468,
-0.0005338586634024978,
0.014149419032037258,
-0.004416462499648333,
0.024144494906067848,
-0.1141093373298645,
-0.02271529845893383,
-0.05448894575238228,
-0.028084779158234596,
-0.08318053185939789,
-0.014482463710010052,
-0.013681110925972462,
-0.04287688806653023,
-0.04320729151368141,
0.0016162041574716568,
-0.08745032548904419,
-0.06389538198709488,
-0.015190714038908482,
0.06873508542776108,
-0.1345461905002594,
-0.009237093850970268,
0.03948003798723221,
-0.1116793155670166,
0.08669998496770859,
0.016407214105129242,
0.04891011863946915,
0.0038645321037620306,
-0.09021848440170288,
0.053380683064460754,
0.0262764785438776,
-0.012576046399772167,
0.02611825056374073,
-0.1401364803314209,
-0.0034743223804980516,
-0.05011119320988655,
0.008027957752346992,
0.00000789365640230244,
0.023999320343136787,
-0.14092418551445007,
0.0080027487128973,
-0.03864584490656853,
-0.0708332359790802,
-0.06682442128658295,
0.04916369542479515,
0.06405699253082275,
-0.01032745186239481,
0.15891970694065094,
-0.06450703740119934,
0.06096680089831352,
-0.20907753705978394,
-0.0053688231855630875,
-0.0009118278976529837,
-0.06852845847606659,
-0.052138738334178925,
-0.030652323737740517,
0.06443479657173157,
-0.04894746094942093,
0.06427665054798126,
-0.039004310965538025,
0.03771696239709854,
0.028162159025669098,
-0.06367359310388565,
0.05691974610090256,
0.039194244891405106,
0.17442657053470612,
0.04981507733464241,
-0.034918662160634995,
0.047524288296699524,
0.016858741641044617,
0.07532252371311188,
0.06588555872440338,
0.18811708688735962,
0.13086090981960297,
0.011980772018432617,
0.08239663392305374,
0.04216183349490166,
-0.10685011744499207,
-0.19662582874298096,
0.09344648569822311,
-0.041411709040403366,
0.1111857146024704,
-0.01591343805193901,
0.2021518498659134,
0.1214587390422821,
-0.1950114369392395,
0.022747809067368507,
0.005439746659249067,
-0.08411528170108795,
-0.10418011993169785,
-0.07825914025306702,
-0.07691922783851624,
-0.13959990441799164,
0.00718242721632123,
-0.10703999549150467,
0.023425884544849396,
0.03389455005526543,
0.028701109811663628,
0.03111286833882332,
0.15322791039943695,
0.04999912530183792,
0.026300327852368355,
0.0750337615609169,
0.048428408801555634,
-0.012909572571516037,
-0.020326824858784676,
-0.09254161268472672,
-0.008708516135811806,
-0.06156104803085327,
0.0483052097260952,
-0.05695527046918869,
-0.06834868341684341,
0.07950258255004883,
0.019438037648797035,
-0.09423012286424637,
0.018334394320845604,
-0.0031148300040513277,
0.050111379474401474,
0.07835250347852707,
0.0052269622683525085,
-0.01756596565246582,
-0.029189560562372208,
0.2004067301750183,
-0.08632058650255203,
-0.007238220889121294,
-0.09424485266208649,
0.22822794318199158,
0.011022130027413368,
-0.01565631665289402,
0.02892448753118515,
-0.07322220504283905,
0.009801848791539669,
0.1660582274198532,
0.15632274746894836,
-0.029342837631702423,
-0.013777975924313068,
0.043975017964839935,
-0.00515739805996418,
-0.011332862079143524,
0.07286296784877777,
0.10984615981578827,
0.11079249531030655,
-0.0717208981513977,
-0.038699690252542496,
-0.04728333652019501,
-0.04946333169937134,
-0.05686233192682266,
0.052596479654312134,
0.04412991181015968,
-0.004512628074735403,
-0.027992280200123787,
0.09237490594387054,
-0.06116838380694389,
-0.10343294590711594,
0.08254192769527435,
-0.17960207164287567,
-0.17121268808841705,
-0.036231640726327896,
0.07318632304668427,
0.001709005911834538,
0.06595408916473389,
0.007546524051576853,
-0.05083104595541954,
0.09685644507408142,
0.001764815067872405,
-0.0910799652338028,
-0.060960348695516586,
0.04832148179411888,
-0.037643857300281525,
0.2165004312992096,
-0.03742758929729462,
-0.0019352673552930355,
0.13606809079647064,
0.04690035432577133,
-0.09924673289060593,
0.012008502148091793,
0.10153459757566452,
-0.09757636487483978,
0.0499567873775959,
0.153281569480896,
-0.030072640627622604,
0.0944785326719284,
0.04882102087140083,
-0.09684649109840393,
0.009785041213035583,
-0.08464595675468445,
-0.042616311460733414,
-0.055362820625305176,
0.016664570197463036,
-0.036118339747190475,
0.14744223654270172,
0.20910948514938354,
-0.04712003841996193,
-0.017676904797554016,
-0.04274049028754234,
0.016139425337314606,
0.05301129072904587,
0.11997370421886444,
-0.012763031758368015,
-0.2424439936876297,
0.01812627539038658,
0.009063669480383396,
0.020289914682507515,
-0.2611302435398102,
-0.09453698992729187,
0.017391188070178032,
-0.03379002586007118,
-0.10431544482707977,
0.10295532643795013,
0.04828507453203201,
0.03855697065591812,
-0.04953194037079811,
-0.08309672772884369,
-0.06171393394470215,
0.16256041824817657,
-0.17836524546146393,
-0.06609833240509033
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | asifhaider/opt-6.7b-lora | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:23:33+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Data Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.1 | {"library_name": "peft", "base_model": "mistralai/Mistral-7B-Instruct-v0.1"} | null | razla/Mistral7b-4 | [
"peft",
"arxiv:1910.09700",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"region:us"
] | 2024-02-08T16:24:06+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-Instruct-v0.1 #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.1 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
"TAGS\n#peft #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-Instruct-v0.1 #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.1"
] | [
37,
6,
3,
45,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-Instruct-v0.1 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.1"
] | [
-0.09280140697956085,
0.17553231120109558,
-0.004129678476601839,
0.03642214834690094,
0.09151369333267212,
0.024356184527277946,
0.05630626156926155,
0.11064136028289795,
-0.0512177050113678,
0.10234527289867401,
0.04918515309691429,
0.09312760829925537,
0.09987960010766983,
0.18844445049762726,
0.0044515617191791534,
-0.2298787385225296,
0.022205254063010216,
-0.10839727520942688,
0.0054216827265918255,
0.12652306258678436,
0.15383900701999664,
-0.09363751113414764,
0.08660339564085007,
-0.020208241418004036,
-0.022401196882128716,
-0.024437034502625465,
-0.07442793995141983,
-0.0547286681830883,
0.04495560750365257,
0.07894964516162872,
0.05899076163768768,
-0.0021642688661813736,
0.08817755430936813,
-0.2684805989265442,
0.016117561608552933,
0.04567277431488037,
-0.006486745551228523,
0.0870618000626564,
0.10494299232959747,
-0.05409630760550499,
0.1079624742269516,
-0.039122387766838074,
0.12606076896190643,
0.06517945975065231,
-0.0748048648238182,
-0.15703070163726807,
-0.08680015802383423,
0.08485505729913712,
0.16156896948814392,
0.07370838522911072,
-0.039233360439538956,
0.1563233584165573,
-0.1273135095834732,
0.01226890366524458,
0.06147238612174988,
-0.04897502809762955,
-0.09079809486865997,
0.05956657603383064,
0.09386191517114639,
0.06967685371637344,
-0.13985790312290192,
-0.03749403357505798,
0.029868973419070244,
0.030951591208577156,
0.0656435489654541,
0.020778879523277283,
0.14568881690502167,
0.042210306972265244,
-0.14004461467266083,
-0.03354855626821518,
0.16809672117233276,
0.052482280880212784,
-0.05460592731833458,
-0.20523302257061005,
0.0064969174563884735,
-0.07145493477582932,
-0.017837362363934517,
-0.04521559551358223,
0.037540797144174576,
-0.019887777045369148,
0.0734168142080307,
0.009323974139988422,
-0.0943276584148407,
-0.043193716555833817,
0.07530920207500458,
0.036856651306152344,
0.025279687717556953,
-0.030881814658641815,
-0.002523052506148815,
0.13105356693267822,
0.06183216720819473,
-0.11815487593412399,
-0.06711763888597488,
-0.06731227785348892,
-0.05688854679465294,
-0.06774375587701797,
0.029381798580288887,
0.0341888926923275,
0.05726776644587517,
0.21743467450141907,
-0.0017759406473487616,
0.0531357042491436,
0.0718580037355423,
0.014807856641709805,
0.06343251466751099,
0.09150100499391556,
-0.08367465436458588,
-0.13622336089611053,
-0.03753688186407089,
0.08673257380723953,
-0.010459864512085915,
-0.012879528105258942,
-0.036834221333265305,
0.04203631728887558,
0.04098881781101227,
0.0975450947880745,
0.07343113422393799,
-0.009003736078739166,
-0.093410924077034,
-0.05015890300273895,
0.2226303219795227,
-0.14671891927719116,
0.035942379385232925,
0.011975139379501343,
-0.04285179078578949,
-0.02785581536591053,
0.013376380316913128,
0.016976170241832733,
-0.023072922602295876,
0.10212952643632889,
-0.07462310791015625,
-0.02124623768031597,
-0.11470510065555573,
-0.01788192242383957,
0.03439907729625702,
0.0529630072414875,
-0.0066626849584281445,
-0.023058347404003143,
-0.06662661582231522,
-0.07004140317440033,
0.08101058006286621,
-0.09316902607679367,
-0.06837931275367737,
-0.014659283682703972,
-0.08941911906003952,
0.006737235933542252,
0.009094251319766045,
0.13268963992595673,
-0.028196316212415695,
0.03652012348175049,
-0.01697647199034691,
0.050805818289518356,
0.07148025929927826,
0.03420931473374367,
-0.05234179273247719,
0.05657242611050606,
-0.19244246184825897,
0.10032042860984802,
-0.08967317640781403,
0.027554666623473167,
-0.16106866300106049,
-0.030086223036050797,
0.009738510474562645,
0.0009234542376361787,
0.027862004935741425,
0.13860416412353516,
-0.22148101031780243,
-0.012080254964530468,
0.14048239588737488,
-0.08888334035873413,
-0.09990011900663376,
0.056695498526096344,
-0.0518028549849987,
0.1250401884317398,
0.023333925753831863,
-0.03633715957403183,
0.05794636532664299,
-0.1412118375301361,
-0.025946129113435745,
-0.02926640398800373,
-0.00694421399384737,
0.12583860754966736,
0.0990971028804779,
-0.06228255853056908,
0.04755837470293045,
0.016736596822738647,
-0.027001217007637024,
-0.041733913123607635,
-0.055725812911987305,
-0.12173914909362793,
0.004647215362638235,
-0.06409628689289093,
0.047816202044487,
-0.016353726387023926,
-0.06972769647836685,
-0.01848401501774788,
-0.1566595882177353,
0.021310046315193176,
0.08648834377527237,
0.017312515527009964,
-0.03392726555466652,
-0.0932558998465538,
0.011130395345389843,
-0.016422078013420105,
-0.036545976996421814,
-0.14865054190158844,
-0.027592573314905167,
0.023116426542401314,
-0.13729165494441986,
0.015186065807938576,
-0.06979985535144806,
0.057803720235824585,
0.022895408794283867,
-0.05880001187324524,
-0.013972318731248379,
-0.026389461010694504,
0.01941268891096115,
-0.04985291510820389,
-0.23562972247600555,
-0.02462669648230076,
-0.03286711499094963,
0.1528421938419342,
-0.23668429255485535,
0.03411379083991051,
0.0673738643527031,
0.1197752058506012,
-0.01760704629123211,
-0.056770507246255875,
0.021966520696878433,
-0.0697036013007164,
-0.031892526894807816,
-0.059108931571245193,
-0.014704322442412376,
-0.019338427111506462,
-0.05527006834745407,
0.0009777687955647707,
-0.1056194081902504,
-0.02718171663582325,
0.09734366089105606,
0.08863259106874466,
-0.16836224496364594,
-0.04584725946187973,
-0.035072118043899536,
-0.08037076890468597,
-0.09111011028289795,
-0.05259617790579796,
0.13759195804595947,
0.05323825031518936,
0.031488124281167984,
-0.08571740239858627,
-0.07425107061862946,
0.010409596376121044,
-0.028131747618317604,
-0.037951305508613586,
0.10798171907663345,
0.07276655733585358,
-0.10890458524227142,
0.09681094437837601,
0.07587725669145584,
0.018951309844851494,
0.09812012314796448,
-0.010683861561119556,
-0.11388456076383591,
-0.04029306769371033,
0.038622524589300156,
0.0036920371931046247,
0.1539752036333084,
-0.0787297934293747,
0.07122835516929626,
0.039123959839344025,
-0.016568060964345932,
0.046934619545936584,
-0.10205834358930588,
0.013179399073123932,
0.01345333643257618,
-0.014760328456759453,
-0.01473155990242958,
-0.037206292152404785,
0.019513985142111778,
0.08078134059906006,
0.03634136542677879,
0.03785719722509384,
0.029095569625496864,
-0.034919146448373795,
-0.12574078142642975,
0.19285397231578827,
-0.108273446559906,
-0.22202986478805542,
-0.15460915863513947,
0.063909150660038,
0.037947848439216614,
-0.03173588961362839,
0.009550401009619236,
-0.04621097072958946,
-0.09866507351398468,
-0.08420389890670776,
0.005803098436444998,
0.04042823240160942,
-0.07612355798482895,
-0.06007229909300804,
0.04722877964377403,
0.05174887180328369,
-0.13321377336978912,
0.045457128435373306,
0.06056974083185196,
-0.046728238463401794,
0.008427056483924389,
0.06173811852931976,
0.07651295512914658,
0.14833810925483704,
-0.010570581071078777,
-0.021728701889514923,
0.04776778444647789,
0.2792262136936188,
-0.14995747804641724,
0.10044047236442566,
0.10502827912569046,
-0.06418829411268234,
0.07816198468208313,
0.18219122290611267,
0.034221868962049484,
-0.10822553187608719,
0.04635465517640114,
0.02743881195783615,
-0.016123786568641663,
-0.27496376633644104,
-0.05808459594845772,
-0.0001963985851034522,
-0.10156700760126114,
0.07027044892311096,
0.08469092100858688,
0.09762902557849884,
0.047377459704875946,
-0.06813536584377289,
-0.07971060276031494,
0.028800632804632187,
0.08003538846969604,
-0.04540814459323883,
0.0038365612272173166,
0.08131581544876099,
-0.03000347875058651,
0.012995688244700432,
0.11099456995725632,
0.0104020144790411,
0.19995735585689545,
0.05192889645695686,
0.11287040263414383,
0.08933474123477936,
0.10060347616672516,
0.008031192235648632,
0.0303626861423254,
0.017929013818502426,
0.010704973712563515,
0.000892018259037286,
-0.08279723674058914,
0.019597651436924934,
0.11011762171983719,
0.06930366158485413,
0.057437170296907425,
0.02457294799387455,
-0.06569267064332962,
0.05918578431010246,
0.19545742869377136,
-0.012235857546329498,
-0.20744788646697998,
-0.06899867206811905,
0.06353529542684555,
-0.08495686948299408,
-0.11704366654157639,
-0.026448743417859077,
0.06354955583810806,
-0.17582543194293976,
0.023424971848726273,
-0.04433430731296539,
0.09506151080131531,
-0.08094653487205505,
-0.04054001718759537,
0.07163584977388382,
0.08226625621318817,
-0.022900981828570366,
0.0825568214058876,
-0.17911314964294434,
0.134952113032341,
0.01124594546854496,
0.06835493445396423,
-0.09108690917491913,
0.1060989499092102,
0.0076294238679111,
0.010437210090458393,
0.1418825089931488,
0.010359355248510838,
-0.01666693575680256,
-0.07191362231969833,
-0.10390663146972656,
-0.008360704407095909,
0.08567162603139877,
-0.11327662318944931,
0.05995211750268936,
-0.00601981719955802,
-0.02192329242825508,
0.00525386817753315,
-0.07246677577495575,
-0.14264167845249176,
-0.16625890135765076,
0.06217791885137558,
-0.13703610002994537,
0.06331935524940491,
-0.0995209813117981,
-0.076270192861557,
-0.016536366194486618,
0.1616908609867096,
-0.20347173511981964,
-0.06328960508108139,
-0.1403825730085373,
-0.08620019257068634,
0.18163928389549255,
-0.04111087694764137,
0.07168667763471603,
0.015510933473706245,
0.17265568673610687,
0.03624330833554268,
0.016094062477350235,
0.09735419601202011,
-0.08624356985092163,
-0.18919724225997925,
-0.060787394642829895,
0.1391448974609375,
0.15334343910217285,
0.049866605550050735,
-0.010254969820380211,
0.010891235433518887,
-0.0520683191716671,
-0.12840910255908966,
-0.0030074205715209246,
0.1400311291217804,
0.0930049940943718,
0.006655999459326267,
-0.026147056370973587,
-0.10597392916679382,
-0.06996311247348785,
-0.06644007563591003,
0.013902216218411922,
0.17106148600578308,
-0.07025249302387238,
0.13722220063209534,
0.11461041867733002,
-0.0567951425909996,
-0.18885108828544617,
0.05145948752760887,
0.08010780811309814,
0.021004775539040565,
0.057196542620658875,
-0.17626060545444489,
0.09668416529893875,
0.059279654175043106,
-0.04636157676577568,
0.12424521148204803,
-0.1619100719690323,
-0.15067945420742035,
0.07453273236751556,
0.06056422367691994,
-0.25804877281188965,
-0.11728063970804214,
-0.08798902481794357,
-0.03798328712582588,
-0.12678970396518707,
0.06945738941431046,
0.013073763810098171,
0.013933585956692696,
0.041606705635786057,
0.03890233859419823,
0.005724054761230946,
-0.043294087052345276,
0.2153140902519226,
0.0030695884488523006,
0.03852320834994316,
-0.046578798443078995,
-0.09692947566509247,
0.03276998549699783,
-0.0416768379509449,
0.09456208348274231,
0.0014304842334240675,
0.017584851011633873,
-0.11626597493886948,
-0.04216182231903076,
-0.06378909945487976,
0.028017211705446243,
-0.09675832092761993,
-0.08909492939710617,
-0.05019475892186165,
0.10160789638757706,
0.06938590109348297,
-0.04267105832695961,
-0.010447661392390728,
-0.0793536975979805,
0.044714294373989105,
0.1796419769525528,
0.20086117088794708,
0.05427732691168785,
-0.08484169840812683,
0.008440880104899406,
-0.015389878302812576,
0.045206986367702484,
-0.23343156278133392,
0.05068710818886757,
0.0515897311270237,
0.018641550093889236,
0.11812181025743484,
-0.0314321331679821,
-0.15550829470157623,
-0.05418111011385918,
0.06598928570747375,
-0.03180372342467308,
-0.17265617847442627,
-0.018539337441325188,
0.0566074438393116,
-0.20496104657649994,
-0.03535214066505432,
0.008391361683607101,
-0.02777557633817196,
-0.04602481424808502,
0.011125941760838032,
0.07819054275751114,
-0.021893084049224854,
0.14624153077602386,
0.07418452948331833,
0.0917057991027832,
-0.10350585728883743,
0.07308387756347656,
0.06054608151316643,
-0.05309276282787323,
0.010431909002363682,
0.07692719995975494,
-0.04342715069651604,
-0.035810526460409164,
0.0755297914147377,
0.06031150370836258,
0.041875410825014114,
-0.04347313567996025,
0.0006184701924212277,
-0.06781867146492004,
0.056906990706920624,
0.1073373332619667,
0.04937898740172386,
0.004621774423867464,
0.04077240079641342,
0.019863922148942947,
-0.08327043801546097,
0.09639332443475723,
0.05704358592629433,
0.027141287922859192,
-0.048162344843149185,
-0.025877393782138824,
0.016309576109051704,
-0.01954711228609085,
-0.014441832900047302,
-0.018663445487618446,
-0.07643482834100723,
-0.016430044546723366,
-0.1265610158443451,
0.03233778849244118,
-0.08421196788549423,
0.021872960031032562,
0.0220671147108078,
-0.058295030146837234,
-0.008038690313696861,
0.017150482162833214,
-0.07364658266305923,
-0.04431304335594177,
-0.005954784341156483,
0.1177598163485527,
-0.11498303711414337,
0.04244931787252426,
0.08591064810752869,
-0.09997779875993729,
0.07655265182256699,
0.008442788384854794,
0.005498629994690418,
0.023045795038342476,
-0.1828068047761917,
0.07369755953550339,
-0.011758057400584221,
0.0006842000875622034,
0.029342051595449448,
-0.23023691773414612,
-0.005714702419936657,
-0.03707250580191612,
-0.01843038573861122,
0.003834768198430538,
-0.03663549944758415,
-0.13203682005405426,
0.07417566329240799,
-0.020970625802874565,
-0.0936848372220993,
-0.0319860465824604,
0.033740729093551636,
0.1225227415561676,
-0.03200415521860123,
0.15821939706802368,
-0.00906917080283165,
0.0649491399526596,
-0.1711292862892151,
-0.014952152967453003,
-0.02135513536632061,
0.033199556171894073,
-0.02435232140123844,
-0.012292984873056412,
0.05490924045443535,
-0.023990483954548836,
0.20033374428749084,
-0.04184139519929886,
0.06262174993753433,
0.054103951901197433,
0.024146607145667076,
-0.020402248948812485,
0.08973324298858643,
0.06319645047187805,
0.0002773959713522345,
0.022030185908079147,
0.02817329578101635,
-0.008174042217433453,
-0.040237702429294586,
-0.14966976642608643,
0.06057923659682274,
0.16396024823188782,
0.027885043993592262,
0.012112509459257126,
0.05497776344418526,
-0.11124580353498459,
-0.0713535025715828,
0.12290359288454056,
-0.014367940835654736,
-0.03286096453666687,
-0.07215135544538498,
0.1299908459186554,
0.12879472970962524,
-0.20197851955890656,
0.07172134518623352,
-0.0681663304567337,
-0.07644392549991608,
-0.10871770977973938,
-0.14241477847099304,
-0.060334958136081696,
-0.048038870096206665,
-0.016741672530770302,
-0.07400241494178772,
0.05250579118728638,
0.10200504213571548,
0.01264016143977642,
-0.019112732261419296,
0.10301186144351959,
-0.00682412413880229,
-0.029484031721949577,
0.02866247296333313,
0.06557506322860718,
0.028308803215622902,
-0.10213936865329742,
0.017442524433135986,
0.005516501609236002,
0.02330516278743744,
0.06420427560806274,
0.009063515812158585,
-0.035803649574518204,
-0.011254461482167244,
-0.026391366496682167,
-0.11106916517019272,
0.04074976593255997,
-0.029713837429881096,
-0.051178961992263794,
0.12278436869382858,
0.02286147139966488,
0.005061932373791933,
-0.023791847750544548,
0.2253320962190628,
-0.07532037049531937,
-0.07834893465042114,
-0.17119061946868896,
0.061915479600429535,
-0.05587479844689369,
0.047718651592731476,
0.04622887447476387,
-0.10432832688093185,
0.02296709269285202,
0.125065878033638,
0.12809623777866364,
-0.0148500707000494,
0.007627879735082388,
0.048726651817560196,
-0.00010584416304482147,
-0.042940136045217514,
0.02861046977341175,
0.05274439975619316,
0.09011797606945038,
-0.05989954248070717,
0.10586516559123993,
-0.009326326660811901,
-0.0785234346985817,
-0.0000036407379866432166,
0.10629300773143768,
-0.009711070917546749,
0.006864482071250677,
-0.07048671692609787,
0.14522288739681244,
-0.05517352744936943,
-0.22764207422733307,
0.05436788871884346,
-0.0693909302353859,
-0.1733580380678177,
-0.028718266636133194,
0.022029919549822807,
-0.014364778064191341,
0.019665976986289024,
0.08285193145275116,
-0.04734303429722786,
0.18775229156017303,
0.046516988426446915,
-0.07150484621524811,
-0.08031817525625229,
0.07235243916511536,
-0.10452817380428314,
0.27277448773384094,
0.013079485855996609,
0.05795780196785927,
0.10425643622875214,
-0.014952000230550766,
-0.13144458830356598,
0.0393635630607605,
0.09976226836442947,
-0.07083691656589508,
0.07705918699502945,
0.17356643080711365,
-0.0015672051813453436,
0.1515946090221405,
0.0642058327794075,
-0.05329511687159538,
0.03494327887892723,
-0.11915883421897888,
-0.04787831008434296,
-0.1036909818649292,
0.09083699434995651,
-0.07426229119300842,
0.1541290581226349,
0.13766998052597046,
-0.06963549554347992,
-0.010891376994550228,
-0.023986872285604477,
0.08351199328899384,
-0.0028523458167910576,
0.127402663230896,
0.004235271830111742,
-0.20352177321910858,
0.016253268346190453,
-0.026195025071501732,
0.10275648534297943,
-0.18882392346858978,
-0.06528139114379883,
0.056331388652324677,
-0.032463327050209045,
-0.05764108523726463,
0.11344319581985474,
0.04501274973154068,
0.04604769125580788,
-0.03247056156396866,
-0.04172878712415695,
-0.01871451921761036,
0.1412779688835144,
-0.10212189704179764,
-0.010496157221496105
] |
null | null | transformers | A Moe model built on top of Qwen1.5-7B-Chat, Qwen1.5-7B and Crystalcareai/CrystalQwen-1.5-7B, then qlora via mlx
```
pip install mlx-moe
```
```
python -m mlx_moe.generate --model mzbac/qwen-1.5-2x3-sft-hf-4bit-mlx --prompt "how backpropagation works?" --eos-token "<|im_end|>"
``` | {} | text-generation | mzbac/qwen-1.5-2x3-sft-hf-4bit-mlx | [
"transformers",
"safetensors",
"qwen2moe",
"text-generation",
"conversational",
"custom_code",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:24:55+00:00 | [] | [] | TAGS
#transformers #safetensors #qwen2moe #text-generation #conversational #custom_code #autotrain_compatible #endpoints_compatible #region-us
| A Moe model built on top of Qwen1.5-7B-Chat, Qwen1.5-7B and Crystalcareai/CrystalQwen-1.5-7B, then qlora via mlx
| [] | [
"TAGS\n#transformers #safetensors #qwen2moe #text-generation #conversational #custom_code #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
50
] | [
"passage: TAGS\n#transformers #safetensors #qwen2moe #text-generation #conversational #custom_code #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
-0.043063003569841385,
0.05287229269742966,
-0.0063142189756035805,
-0.004617184866219759,
0.12316770851612091,
-0.03785243630409241,
0.09963001310825348,
0.10694342106580734,
0.0048498790711164474,
-0.01170374546200037,
0.13090260326862335,
0.16591177880764008,
-0.015226278454065323,
0.11746543645858765,
-0.12875567376613617,
-0.14235937595367432,
0.10136227309703827,
0.001570118241943419,
-0.007506511174142361,
0.09000573307275772,
0.09645487368106842,
-0.06170446798205376,
0.09076471626758575,
-0.04574745148420334,
-0.12279772013425827,
0.03696261718869209,
0.047050971537828445,
-0.11368441581726074,
0.1115524098277092,
0.06762789189815521,
0.11302667111158371,
0.0868280753493309,
-0.05802116170525551,
-0.22486655414104462,
0.040278978645801544,
-0.010086286813020706,
-0.08455274999141693,
-0.008813346736133099,
0.03472984582185745,
-0.03853096067905426,
-0.060806598514318466,
0.05428686365485191,
-0.023463476449251175,
0.10750874876976013,
-0.1356325000524521,
0.040948107838630676,
-0.010076211765408516,
-0.0466783307492733,
0.11384612321853638,
0.08621593564748764,
-0.025027751922607422,
0.17129947245121002,
-0.07562392204999924,
0.1358332335948944,
0.06353824585676193,
-0.386687308549881,
0.0054303728975355625,
0.09609819203615189,
0.10662779957056046,
0.0627303496003151,
-0.05246191844344139,
0.044244907796382904,
0.07002291828393936,
-0.007456681225448847,
-0.03723248839378357,
-0.08601528406143188,
-0.11777934432029724,
0.021960260346531868,
-0.06457681208848953,
-0.00918578915297985,
0.17064912617206573,
-0.015230468474328518,
0.042867328971624374,
-0.04965118691325188,
-0.1255093216896057,
-0.034909967333078384,
-0.015123403631150723,
-0.021605361253023148,
-0.05632920190691948,
0.10112716257572174,
-0.04682152718305588,
-0.008871825411915779,
-0.12797290086746216,
0.011479238979518414,
-0.1873456984758377,
0.15423603355884552,
-0.007029378786683083,
0.031028661876916885,
-0.23006001114845276,
0.01749378815293312,
0.0055840411223471165,
-0.13496078550815582,
-0.021897708997130394,
-0.1076003834605217,
0.03247881680727005,
0.010125837288796902,
-0.025294091552495956,
-0.05658666789531708,
0.19359658658504486,
0.1268244832754135,
0.0025249035097658634,
0.08232367783784866,
-0.05333118885755539,
0.03873242437839508,
0.0030353334732353687,
0.08640710264444351,
0.0794377475976944,
-0.13770398497581482,
0.0911199152469635,
-0.08129744976758957,
0.01826079934835434,
-0.06596466153860092,
-0.1459478735923767,
-0.026897205039858818,
0.05514821037650108,
0.1373569220304489,
0.03607110679149628,
0.10230162739753723,
-0.01834416575729847,
0.0469185896217823,
0.10443504899740219,
-0.0673375353217125,
-0.028024373576045036,
0.04012839123606682,
0.07028166949748993,
0.010248994454741478,
-0.021108830347657204,
0.07567258179187775,
-0.06475847959518433,
-0.04427893087267876,
-0.03056197613477707,
-0.0684320330619812,
-0.005034947767853737,
-0.0626710057258606,
0.019733143970370293,
0.0015726089477539062,
0.0525742843747139,
-0.2039468139410019,
-0.07948847860097885,
-0.0033654156140983105,
-0.03400719538331032,
0.018305286765098572,
-0.017894553020596504,
-0.09284596890211105,
-0.05320270359516144,
0.027256928384304047,
-0.04955189675092697,
-0.11018794775009155,
-0.06531751900911331,
0.07554019242525101,
0.04759802669286728,
0.06493491679430008,
-0.15620380640029907,
0.02250588871538639,
-0.10862290114164352,
0.022629044950008392,
-0.019794227555394173,
0.058584731072187424,
-0.01069143507629633,
0.15984629094600677,
0.025640705600380898,
0.024096878245472908,
-0.06221695616841316,
0.05357763171195984,
-0.02933191880583763,
0.22602307796478271,
-0.094344362616539,
-0.049723073840141296,
0.32118308544158936,
-0.10356265306472778,
-0.23585017025470734,
0.1155068501830101,
0.008947841823101044,
0.02470068633556366,
0.09141705930233002,
0.20288369059562683,
0.09079279750585556,
-0.04606740549206734,
0.033999234437942505,
0.1148567646741867,
-0.10705956071615219,
-0.07102622836828232,
-0.011130603030323982,
0.0037911240942776203,
-0.08822821080684662,
0.05089690163731575,
0.10704384744167328,
0.07345765829086304,
-0.03882751613855362,
-0.06198600307106972,
-0.04917820170521736,
-0.061678823083639145,
0.04506054148077965,
-0.03413965925574303,
0.03995732590556145,
-0.1062026396393776,
0.015645233914256096,
-0.10335098952054977,
-0.001861759927123785,
-0.025829864665865898,
-0.0009297278593294322,
-0.14521126449108124,
0.06390691548585892,
0.006223296746611595,
0.05405907332897186,
-0.10612750053405762,
-0.12809059023857117,
-0.004876904655247927,
0.09118686616420746,
-0.036012474447488785,
0.06512226164340973,
0.07948587834835052,
0.00893654115498066,
0.014992625452578068,
-0.07543452829122543,
0.18499498069286346,
0.03513623774051666,
-0.0777682289481163,
-0.039462942630052567,
0.0991055965423584,
-0.08653666079044342,
0.03324253857135773,
-0.09153788536787033,
0.02733391895890236,
0.07437063008546829,
0.10071936994791031,
0.03824013099074364,
0.05359040945768356,
-0.016299128532409668,
0.020878054201602936,
-0.07975935190916061,
0.014566008932888508,
0.06843708455562592,
0.032161589711904526,
-0.10656880587339401,
0.18428704142570496,
-0.25940224528312683,
0.2818271815776825,
0.16637714207172394,
-0.22599977254867554,
0.05456111952662468,
-0.0610545352101326,
-0.01405227743089199,
0.008392874151468277,
0.058667898178100586,
-0.04714805632829666,
0.06433086842298508,
-0.0029321869369596243,
0.16617849469184875,
-0.05516858026385307,
-0.03727037459611893,
0.008906188420951366,
-0.06549006700515747,
-0.03826947882771492,
0.05697764456272125,
0.015775183215737343,
-0.17816437780857086,
0.14793051779270172,
0.24363571405410767,
0.07479053735733032,
0.19275358319282532,
-0.009370177052915096,
0.04780842363834381,
0.047618839889764786,
0.06435687094926834,
-0.02460542507469654,
-0.043901849538087845,
-0.20795297622680664,
-0.03872480243444443,
0.03482815623283386,
0.037709858268499374,
0.09108589589595795,
-0.08809815347194672,
-0.05103784054517746,
0.008117950521409512,
-0.015538964420557022,
-0.026386534795165062,
0.05807211622595787,
0.0287484060972929,
0.11070121824741364,
-0.012329546734690666,
-0.04815464839339256,
0.09635122865438461,
-0.05253596603870392,
-0.11605452746152878,
0.17912913858890533,
-0.1511935144662857,
-0.2800846993923187,
-0.14271678030490875,
-0.1284559667110443,
-0.058975949883461,
0.04499815031886101,
0.14789623022079468,
-0.1455530971288681,
-0.0352913923561573,
-0.026804998517036438,
0.015952054411172867,
-0.007631248794496059,
0.02969791553914547,
-0.004542654845863581,
0.06911126524209976,
-0.0316094234585762,
-0.10382644087076187,
-0.04273027554154396,
0.02328007109463215,
-0.08368434756994247,
0.13578854501247406,
-0.13378922641277313,
0.12596212327480316,
0.14695805311203003,
0.0067641232162714005,
0.03969926759600639,
-0.031151819974184036,
0.25106921792030334,
-0.06510676443576813,
0.0003979597822763026,
0.18152697384357452,
-0.07478295266628265,
0.06491614878177643,
0.18674276769161224,
-0.029553543776273727,
-0.1069251149892807,
0.062232937663793564,
-0.042487215250730515,
-0.07109047472476959,
-0.20247116684913635,
-0.11402525752782822,
-0.07637018710374832,
0.12041693180799484,
-0.05397060140967369,
0.05899381637573242,
0.09441513568162918,
0.06590332090854645,
-0.014272715896368027,
-0.0691797286272049,
0.03810529038310051,
0.083393894135952,
0.1586124747991562,
0.0003772262134589255,
0.12199496477842331,
-0.07464583963155746,
-0.115252286195755,
0.06570960581302643,
0.10420744866132736,
0.0255412794649601,
0.10317918658256531,
0.06140907108783722,
0.024691922590136528,
0.166356161236763,
0.15276098251342773,
0.0984635055065155,
0.03759699687361717,
-0.06516645848751068,
0.013747141696512699,
-0.01951591670513153,
-0.08093000948429108,
0.030140183866024017,
0.000312562333419919,
-0.11332875490188599,
-0.05829630419611931,
0.05203130468726158,
0.10647264122962952,
0.12954171001911163,
0.05943172052502632,
-0.1399608999490738,
-0.03534097224473953,
0.12026194483041763,
-0.03218381106853485,
-0.08380278199911118,
0.14295831322669983,
0.1127745732665062,
-0.09383426606655121,
0.06070268526673317,
-0.012633015401661396,
0.11255140602588654,
-0.010951950214803219,
0.10942234843969345,
-0.11500140279531479,
-0.10885048657655716,
0.015362207777798176,
0.10002965480089188,
-0.31364408135414124,
0.17993299663066864,
0.011664754711091518,
0.014253566972911358,
-0.06267908960580826,
-0.023194124922156334,
0.0017748861573636532,
0.13938361406326294,
0.11649585515260696,
-0.012513118796050549,
-0.06932920962572098,
-0.1450270563364029,
-0.0001888665574369952,
0.07648278027772903,
0.15617822110652924,
0.040805038064718246,
0.03515132889151573,
-0.05314848944544792,
-0.017563287168741226,
-0.010818450711667538,
-0.06097680702805519,
-0.044177159667015076,
-0.15126611292362213,
0.01779887266457081,
0.0974588394165039,
0.15568876266479492,
-0.029348568990826607,
0.05413476750254631,
-0.1354747861623764,
0.16720545291900635,
-0.16985541582107544,
-0.05832938477396965,
-0.08325286954641342,
-0.15540914237499237,
-0.01833241991698742,
-0.08701562881469727,
0.05886229872703552,
-0.069981150329113,
0.07694602012634277,
-0.08453212678432465,
-0.16954363882541656,
0.1138739064335823,
-0.14340469241142273,
-0.07224996387958527,
-0.0439942330121994,
0.1541743129491806,
-0.06306201964616776,
-0.0250607468187809,
0.08064141124486923,
-0.004674367140978575,
-0.08194591104984283,
-0.11426772177219391,
-0.008265141397714615,
-0.06688673794269562,
-0.004174412228167057,
-0.005543473642319441,
-0.04262849688529968,
-0.12053560465574265,
-0.042791400104761124,
-0.03364025428891182,
0.2938517928123474,
0.25642696022987366,
-0.0603390596807003,
0.0996706634759903,
0.20303189754486084,
-0.033874351531267166,
-0.3408950865268707,
-0.13414904475212097,
-0.14875556528568268,
-0.029512910172343254,
-0.021702779456973076,
-0.03018416464328766,
0.08708039671182632,
-0.017136763781309128,
-0.009876161813735962,
0.03818678855895996,
-0.18749265372753143,
-0.07455087453126907,
0.15061557292938232,
0.02021222934126854,
0.36018091440200806,
-0.1612364649772644,
-0.10562185198068619,
-0.06597989052534103,
-0.23051777482032776,
0.10494038462638855,
-0.16400092840194702,
0.04432554543018341,
0.04821665957570076,
0.07282466441392899,
0.03780359774827957,
-0.041427131742239,
0.09573282301425934,
-0.027500655502080917,
0.010254718363285065,
-0.10389354825019836,
-0.08701618760824203,
0.008087586611509323,
-0.026410674676299095,
-0.02037918195128441,
-0.07255994528532028,
0.05960261449217796,
-0.0855325236916542,
-0.008485737256705761,
-0.06266212463378906,
0.03710117191076279,
0.005346227437257767,
-0.05905802920460701,
0.01847703941166401,
-0.09277082979679108,
0.05859776958823204,
0.02919743023812771,
0.2333907186985016,
-0.0920783132314682,
0.17854948341846466,
0.2163303941488266,
0.19503366947174072,
-0.18174488842487335,
0.14139340817928314,
-0.025421857833862305,
-0.08493242412805557,
0.06674351543188095,
-0.04944855719804764,
0.1097462847828865,
0.08909899741411209,
-0.05785746872425079,
0.06473016738891602,
0.08360537886619568,
0.03874976933002472,
0.056853942573070526,
0.13648927211761475,
-0.2277878373861313,
-0.0773196816444397,
-0.053923726081848145,
0.07315419614315033,
0.059674207121133804,
0.12676529586315155,
0.16543267667293549,
0.04631292074918747,
-0.018884535878896713,
-0.04500520974397659,
0.03494015336036682,
-0.054271385073661804,
0.05060233920812607,
0.053809043020009995,
0.043258555233478546,
-0.12309203296899796,
0.06006801500916481,
-0.04585874453186989,
-0.16868191957473755,
0.021852411329746246,
0.1292654275894165,
-0.15588174760341644,
-0.1235262006521225,
-0.02937212772667408,
0.18502123653888702,
-0.12436992675065994,
-0.08697862923145294,
-0.060848504304885864,
-0.13019658625125885,
0.010040368884801865,
0.2607008218765259,
0.04100089147686958,
0.10408692061901093,
0.033268146216869354,
-0.011687038466334343,
-0.03447972610592842,
0.010886633768677711,
-0.02200428955256939,
0.014282766729593277,
-0.11614646017551422,
0.025743918493390083,
-0.06315503269433975,
0.15196934342384338,
-0.09564312547445297,
-0.06159433349967003,
-0.17122869193553925,
0.02274947613477707,
-0.18162676692008972,
-0.025001654401421547,
-0.10913007706403732,
-0.014754015021026134,
-0.013701584190130234,
-0.0514698289334774,
-0.06306591629981995,
-0.036239754408597946,
-0.08324627578258514,
0.029060864821076393,
-0.028476672247052193,
0.014801506884396076,
-0.0828404501080513,
0.015794478356838226,
0.07395409047603607,
-0.034348394721746445,
0.11348778754472733,
0.06855247169733047,
-0.1100957840681076,
0.09502297639846802,
-0.16457988321781158,
-0.08406361937522888,
0.09989892691373825,
0.0019419932505115867,
0.05380544066429138,
0.1367984563112259,
-0.013153301551938057,
0.09496305882930756,
0.008460761047899723,
0.046204742044210434,
0.14409704506397247,
-0.09302747994661331,
0.07167544960975647,
-0.030554022639989853,
-0.11977113783359528,
-0.04291042685508728,
-0.08512654155492783,
0.07359012961387634,
-0.056793730705976486,
0.16197215020656586,
-0.09905941039323807,
0.0785503014922142,
-0.04526858776807785,
0.011827931739389896,
0.04848310351371765,
-0.14087370038032532,
-0.09602519869804382,
-0.052064940333366394,
0.029485248029232025,
-0.02799629420042038,
0.19837912917137146,
-0.08857370167970657,
0.06502816081047058,
0.07066000252962112,
-0.030333828181028366,
-0.02549811638891697,
0.014687647111713886,
0.21839502453804016,
0.10194796323776245,
-0.0487581267952919,
-0.1078709065914154,
0.03382601961493492,
0.01180790551006794,
-0.08221820741891861,
0.06253744661808014,
0.05931324139237404,
-0.021497592329978943,
0.1279948204755783,
0.017460724338889122,
0.044749531894922256,
-0.08316391706466675,
-0.18265409767627716,
-0.10431192815303802,
0.017063871026039124,
-0.021148357540369034,
0.02481912262737751,
0.2201521098613739,
-0.0030107127968221903,
-0.019902050495147705,
-0.06982078403234482,
-0.035051606595516205,
-0.17421619594097137,
-0.04750453680753708,
-0.127236470580101,
-0.11521444469690323,
0.012032270431518555,
-0.11968265473842621,
0.022968871518969536,
0.015410910360515118,
0.05209808424115181,
-0.05533929169178009,
0.14666207134723663,
0.02111164666712284,
-0.06573200970888138,
0.029411183670163155,
-0.030946720391511917,
0.07954572141170502,
0.060687679797410965,
-0.017575273290276527,
-0.07240605354309082,
0.006457648705691099,
-0.0308812465518713,
0.08654051274061203,
-0.04098736494779587,
0.03991345688700676,
-0.15730531513690948,
-0.0998629480600357,
-0.035404469817876816,
0.0930785983800888,
-0.07072439044713974,
0.1084585040807724,
0.02872467413544655,
-0.004297163337469101,
0.06907650828361511,
0.21287782490253448,
-0.0783853605389595,
-0.13454389572143555,
-0.0864691287279129,
0.12374312430620193,
0.034548647701740265,
0.17962665855884552,
-0.058995358645915985,
-0.03362147510051727,
-0.0696672797203064,
0.3700457811355591,
0.2349603921175003,
-0.09110670536756516,
0.05697229504585266,
-0.030094800516963005,
0.0495280884206295,
0.07249809056520462,
0.11956935375928879,
0.10037709027528763,
0.28574395179748535,
-0.062465354800224304,
-0.047864094376564026,
-0.014840890653431416,
-0.02058103308081627,
-0.11099456250667572,
0.10385292023420334,
-0.022493751719594002,
-0.017338190227746964,
-0.04366178438067436,
0.09521190077066422,
-0.18865278363227844,
0.15666964650154114,
-0.12688709795475006,
-0.16749274730682373,
-0.030833695083856583,
-0.005510621704161167,
0.16010063886642456,
0.013879344798624516,
0.07542198896408081,
0.01462320052087307,
-0.07917548716068268,
0.05759221687912941,
-0.010355202481150627,
-0.1501915603876114,
0.027959682047367096,
0.013230116106569767,
-0.08339410275220871,
0.04617690667510033,
0.008174864575266838,
0.0039430600591003895,
0.09904026240110397,
0.04688134044408798,
-0.028324075043201447,
0.14786142110824585,
0.04565662518143654,
-0.0533314049243927,
-0.007598178926855326,
0.00786502193659544,
0.01102074608206749,
0.0195435993373394,
0.06436983495950699,
-0.1924717277288437,
0.06522340327501297,
0.02570509910583496,
-0.04661737009882927,
-0.026546239852905273,
0.058767419308423996,
-0.027245985344052315,
0.06423929333686829,
0.019021667540073395,
-0.031387489289045334,
0.021021580323576927,
0.00948591809719801,
0.018379895016551018,
-0.02351708710193634,
-0.06556422263383865,
-0.08474491536617279,
-0.1766437143087387,
-0.06406129896640778,
0.09896894544363022,
0.008794368244707584,
-0.18544642627239227,
0.007064018398523331,
-0.09771694242954254,
0.10349821299314499,
-0.1527138352394104,
0.11495500057935715,
0.1555929183959961,
0.012873918749392033,
-0.019642017781734467,
-0.08482743799686432,
0.04736330360174179,
0.12406177818775177,
-0.06796565651893616,
-0.15078523755073547
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | technocrat3128/sentiment_analysis_FB_roberta_fine_tune_hashtag_removed | [
"transformers",
"safetensors",
"roberta",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:24:57+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #roberta #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #roberta #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
47,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #roberta #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07097413390874863,
0.1667894870042801,
-0.003666079370304942,
0.021719951182603836,
0.1169515997171402,
0.008284984156489372,
0.07978618890047073,
0.10868363827466965,
-0.0287666916847229,
0.12472489476203918,
0.03958536684513092,
0.100002720952034,
0.11150532960891724,
0.19375544786453247,
0.0031385887414216995,
-0.2044016718864441,
0.0626462921500206,
-0.11493270099163055,
0.012976650148630142,
0.12250365316867828,
0.1423846185207367,
-0.10432620346546173,
0.07270357757806778,
-0.038304854184389114,
-0.02555735409259796,
-0.02992432750761509,
-0.06288715451955795,
-0.06023360788822174,
0.06700628250837326,
0.060681991279125214,
0.0691668838262558,
0.022982222959399223,
0.08022861182689667,
-0.2922787070274353,
0.0204449649900198,
0.07901842147111893,
0.0025627717841416597,
0.06292086839675903,
0.07761354744434357,
-0.07406430691480637,
0.11906187236309052,
-0.055618006736040115,
0.15586547553539276,
0.07508774101734161,
-0.09551873803138733,
-0.1806546300649643,
-0.08346612006425858,
0.09468797594308853,
0.16155311465263367,
0.05840203911066055,
-0.03498850762844086,
0.14096277952194214,
-0.08077409118413925,
0.016175461933016777,
0.06836850196123123,
-0.07082965970039368,
-0.05214150249958038,
0.05027921497821808,
0.07374554872512817,
0.09283687174320221,
-0.1307315081357956,
-0.010299519635736942,
0.04289179667830467,
0.018450511619448662,
0.10514726489782333,
0.023654451593756676,
0.11117657274007797,
0.028080156072974205,
-0.13854995369911194,
-0.0629403293132782,
0.12449787557125092,
0.032600224018096924,
-0.060118548572063446,
-0.23809053003787994,
-0.004739876836538315,
-0.03291091322898865,
-0.021559255197644234,
-0.04033812880516052,
0.04089386388659477,
-0.029875123873353004,
0.08085108548402786,
0.006228691898286343,
-0.07215629518032074,
-0.052459850907325745,
0.08968427777290344,
0.05656275153160095,
0.02484809421002865,
-0.02587738446891308,
0.02272983267903328,
0.1178721934556961,
0.09954223781824112,
-0.11430655419826508,
-0.06540504842996597,
-0.06454093754291534,
-0.08725714683532715,
-0.04798560217022896,
0.03668604791164398,
0.07030875980854034,
0.046011876314878464,
0.19716040790081024,
0.00269759027287364,
0.05310693010687828,
0.02948983944952488,
0.016264021396636963,
0.06394703686237335,
0.06814983487129211,
-0.05028722807765007,
-0.12879575788974762,
-0.03764699772000313,
0.11916788667440414,
0.006199078168720007,
-0.032489337027072906,
-0.0359247624874115,
0.05915093049407005,
0.05401566997170448,
0.11897829174995422,
0.06329395622015,
0.01798311620950699,
-0.06940232217311859,
-0.041600339114665985,
0.1802349090576172,
-0.15566512942314148,
0.020810957998037338,
0.015188573859632015,
-0.055285364389419556,
-0.042803313583135605,
0.023135531693696976,
0.009154262952506542,
-0.02835596352815628,
0.10190211236476898,
-0.06733403354883194,
-0.03976709768176079,
-0.10797561705112457,
-0.05405012145638466,
0.03397495672106743,
-0.023470325395464897,
-0.0295037180185318,
-0.04243861138820648,
-0.12201286107301712,
-0.07699050009250641,
0.06754051148891449,
-0.06153370812535286,
-0.06873670220375061,
-0.039611347019672394,
-0.0595494844019413,
0.012631349265575409,
0.002569467294961214,
0.12651456892490387,
-0.029271753504872322,
0.04821896180510521,
-0.05001894757151604,
0.06899536401033401,
0.13243289291858673,
0.03377855569124222,
-0.07108951359987259,
0.06568732112646103,
-0.21172066032886505,
0.10609432309865952,
-0.09376481175422668,
0.03139003366231918,
-0.16234688460826874,
-0.024603933095932007,
0.027305902913212776,
0.03785030171275139,
-0.01080681849271059,
0.14199338853359222,
-0.18069814145565033,
-0.0368097685277462,
0.18344907462596893,
-0.13205336034297943,
-0.09475254267454147,
0.06138899549841881,
-0.06121760979294777,
0.13521988689899445,
0.05350736901164055,
-0.023734936490654945,
0.05919842794537544,
-0.1313716471195221,
-0.025711048394441605,
-0.05855761095881462,
-0.008730711415410042,
0.14457696676254272,
0.06305380910634995,
-0.05395244061946869,
0.026081737130880356,
0.018570488318800926,
-0.019653121009469032,
-0.04788879305124283,
-0.03383433073759079,
-0.09769339114427567,
0.005564571358263493,
-0.08092508465051651,
0.01647305116057396,
-0.018450245261192322,
-0.08526520431041718,
-0.040688324719667435,
-0.1556665003299713,
0.002278752624988556,
0.09955150634050369,
0.00428511667996645,
-0.028858626261353493,
-0.09412826597690582,
0.0010344249894842505,
0.014487308450043201,
-0.011919399723410606,
-0.15230298042297363,
-0.04881799593567848,
0.026287198066711426,
-0.1660703718662262,
0.03276912495493889,
-0.05137382447719574,
0.03549332916736603,
0.044475916773080826,
-0.0461643822491169,
-0.024930957704782486,
0.014140721410512924,
0.017139434814453125,
-0.02374175749719143,
-0.24581307172775269,
-0.013825789093971252,
-0.04980151727795601,
0.18321004509925842,
-0.24926310777664185,
0.04708116501569748,
0.061669353395700455,
0.11851225793361664,
0.005235658492892981,
-0.047542329877614975,
0.04137692227959633,
-0.049956902861595154,
-0.04146651551127434,
-0.06634348630905151,
-0.003162689507007599,
-0.03229794278740883,
-0.04517984762787819,
0.041715674102306366,
-0.190790057182312,
-0.026901599019765854,
0.11227498203516006,
0.0757983848452568,
-0.172615647315979,
-0.07716702669858932,
-0.032237354665994644,
-0.06075936183333397,
-0.08794162422418594,
-0.04783210903406143,
0.1008552610874176,
0.040512196719646454,
0.05559884011745453,
-0.07266750186681747,
-0.05319931358098984,
0.01357761025428772,
-0.011024849489331245,
-0.03397562727332115,
0.09108693152666092,
0.08268309384584427,
-0.1214602142572403,
0.1049051284790039,
0.06639663875102997,
0.06267683953046799,
0.10698967427015305,
0.00657825730741024,
-0.09603405743837357,
-0.011494014412164688,
0.02610582485795021,
0.01399005576968193,
0.1426926553249359,
-0.07498639076948166,
0.030146000906825066,
0.042116980999708176,
-0.030949845910072327,
0.012327825650572777,
-0.1036502942442894,
0.0180202666670084,
0.03208072856068611,
-0.00865737535059452,
0.015741337090730667,
-0.056021902710199356,
0.013855318538844585,
0.10478801280260086,
0.03378775715827942,
0.028230587020516396,
0.01726776920258999,
-0.039543844759464264,
-0.12787796556949615,
0.1788790076971054,
-0.09654324501752853,
-0.25226548314094543,
-0.13508103787899017,
0.00398462638258934,
0.046175748109817505,
-0.011871901340782642,
0.01716979779303074,
-0.05890684202313423,
-0.10515041649341583,
-0.10667064040899277,
0.0201003048568964,
0.05383666977286339,
-0.0892900750041008,
-0.06228140741586685,
0.054920271039009094,
0.037761468440294266,
-0.12314439564943314,
0.023868469521403313,
0.045075058937072754,
-0.07041552662849426,
0.006107859779149294,
0.0533861480653286,
0.08385360985994339,
0.17951355874538422,
0.008954732678830624,
-0.01666683331131935,
0.009695366024971008,
0.2122192084789276,
-0.14731436967849731,
0.09077969938516617,
0.13881319761276245,
-0.05979595333337784,
0.08313421159982681,
0.2037704885005951,
0.028123976662755013,
-0.09402397274971008,
0.04063381254673004,
0.03390387445688248,
-0.04006393998861313,
-0.24267235398292542,
-0.07322345674037933,
0.003262192476540804,
-0.06732312589883804,
0.10439975559711456,
0.09025674313306808,
0.11683525890111923,
0.05397806316614151,
-0.10713037848472595,
-0.06759513914585114,
0.048913486301898956,
0.11977367848157883,
-0.031416673213243484,
0.004043765366077423,
0.0945456475019455,
-0.03565856069326401,
0.023265741765499115,
0.09114573150873184,
0.016503766179084778,
0.18390727043151855,
0.03800623491406441,
0.1316893994808197,
0.08565124869346619,
0.06437213718891144,
0.019635362550616264,
0.020543145015835762,
0.02446223795413971,
0.029573822394013405,
-0.020725078880786896,
-0.08386499434709549,
-0.01673775352537632,
0.14245417714118958,
0.024977976456284523,
0.0360303595662117,
0.0035881029907613993,
-0.03668458014726639,
0.06648463755846024,
0.16771891713142395,
0.012426458299160004,
-0.23092281818389893,
-0.06617839634418488,
0.0744035392999649,
-0.06997847557067871,
-0.11378128826618195,
-0.011021395213901997,
0.03138574957847595,
-0.1820969581604004,
0.04344824329018593,
-0.02262875810265541,
0.10113118588924408,
-0.11110582947731018,
-0.025930341333150864,
0.038102298974990845,
0.06534357368946075,
-0.03762030974030495,
0.07928402721881866,
-0.20931516587734222,
0.14360107481479645,
0.006824537646025419,
0.0662476122379303,
-0.1070002019405365,
0.08016478270292282,
0.02166800945997238,
0.003962947987020016,
0.1639549285173416,
-0.006579355336725712,
-0.08173416554927826,
-0.08530484884977341,
-0.07792343199253082,
-0.01293096225708723,
0.09758622944355011,
-0.1083054319024086,
0.08659143000841141,
-0.007826665416359901,
-0.031724635511636734,
-0.002251977799460292,
-0.11942502856254578,
-0.13174815475940704,
-0.1834975630044937,
0.04998116195201874,
-0.11704818159341812,
0.03951563686132431,
-0.11107321828603745,
-0.061925847083330154,
-0.03264854475855827,
0.19538627564907074,
-0.1968487948179245,
-0.08019812405109406,
-0.14597393572330475,
-0.0734013020992279,
0.11603168398141861,
-0.04134324938058853,
0.07953549921512604,
0.003630653489381075,
0.2027774453163147,
-0.0049596261233091354,
0.0008760678465478122,
0.0867764800786972,
-0.09453176707029343,
-0.2061263471841812,
-0.095054991543293,
0.13693954050540924,
0.12683691084384918,
0.045141350477933884,
-0.001808035303838551,
0.022449098527431488,
-0.003543051891028881,
-0.10818912833929062,
0.030619105324149132,
0.1492711752653122,
0.09657935798168182,
0.0439409464597702,
-0.025595664978027344,
-0.14485116302967072,
-0.10286740958690643,
-0.05655559152364731,
0.017665131017565727,
0.183364138007164,
-0.06921800225973129,
0.16631482541561127,
0.15526814758777618,
-0.06590348482131958,
-0.20979635417461395,
0.03614744171500206,
0.03232418745756149,
-0.008261965587735176,
0.03450188785791397,
-0.20602507889270782,
0.07341582328081131,
0.01584669202566147,
-0.057912733405828476,
0.13327911496162415,
-0.17143049836158752,
-0.1488124579191208,
0.08935882896184921,
0.0777517780661583,
-0.211406409740448,
-0.13250647485256195,
-0.09561139345169067,
-0.05425233766436577,
-0.1049741804599762,
0.08951318264007568,
-0.00034212550963275135,
0.00718674948439002,
0.035794805735349655,
0.02204621024429798,
0.017388639971613884,
-0.05294845998287201,
0.19216622412204742,
-0.0027261620853096247,
0.04696741700172424,
-0.08073770254850388,
-0.08108212053775787,
0.03557400405406952,
-0.06683739274740219,
0.0830080509185791,
-0.023084791377186775,
0.004679036792367697,
-0.11573878675699234,
-0.06346318125724792,
-0.04972510412335396,
0.03276039659976959,
-0.08775602281093597,
-0.09446880221366882,
-0.05614441633224487,
0.10359876602888107,
0.08980697393417358,
-0.03496086969971657,
-0.061577584594488144,
-0.09710686653852463,
0.07230715453624725,
0.22115971148014069,
0.18863879144191742,
0.07647473365068436,
-0.0675395280122757,
-0.0029117604717612267,
-0.023906316608190536,
0.054080698639154434,
-0.2153206318616867,
0.044365670531988144,
0.03903084248304367,
0.03029310703277588,
0.13431638479232788,
-0.023942356929183006,
-0.16170847415924072,
-0.044563427567481995,
0.06042470782995224,
-0.0675726905465126,
-0.15454024076461792,
0.003008861094713211,
0.08846548199653625,
-0.16023053228855133,
-0.053557612001895905,
0.02649034932255745,
-0.0350460447371006,
-0.027545088902115822,
0.0010348923970013857,
0.08224344998598099,
0.02413026988506317,
0.11039666831493378,
0.07115603238344193,
0.11174404621124268,
-0.10106326639652252,
0.08730421215295792,
0.0898137167096138,
-0.10840801149606705,
0.03910919278860092,
0.06986082345247269,
-0.06326653808355331,
-0.0329873263835907,
0.031108932569622993,
0.08609528094530106,
0.02910410612821579,
-0.07589155435562134,
-0.0003110404359176755,
-0.10957848280668259,
0.06651712954044342,
0.14283788204193115,
0.03319181129336357,
0.007173929363489151,
0.04601660370826721,
0.029003625735640526,
-0.10390203446149826,
0.11265801638364792,
0.041551440954208374,
0.0359308086335659,
-0.04995269328355789,
0.003519897349178791,
0.04222669079899788,
-0.011379980482161045,
-0.015903322026133537,
-0.039394035935401917,
-0.06654489040374756,
-0.009303842671215534,
-0.1557016372680664,
0.02798696979880333,
-0.0701727494597435,
0.008773589506745338,
0.016523631289601326,
-0.02984452061355114,
0.00421373825520277,
0.010279021225869656,
-0.07592474669218063,
-0.03921252489089966,
-0.00597892701625824,
0.10799336433410645,
-0.16164712607860565,
0.008396364748477936,
0.0864807590842247,
-0.12359115481376648,
0.0785754844546318,
-0.004872123710811138,
-0.008228191174566746,
0.016140099614858627,
-0.14468397200107574,
0.06475566327571869,
-0.009952590800821781,
0.002042346866801381,
0.02179797925055027,
-0.20463575422763824,
0.004303002264350653,
-0.04947707802057266,
-0.05808398127555847,
-0.0054858666844666,
-0.0371098592877388,
-0.11110534518957138,
0.10392162203788757,
0.016078852117061615,
-0.08070345222949982,
-0.01963779143989086,
0.054263342171907425,
0.11082571744918823,
-0.05670001730322838,
0.1372474879026413,
-0.0212235189974308,
0.05716216191649437,
-0.17756865918636322,
-0.01764446683228016,
-0.017806867137551308,
0.013603313826024532,
-0.03553657978773117,
-0.008313823491334915,
0.05301200598478317,
-0.016943611204624176,
0.2223619669675827,
-0.021118339151144028,
0.024224206805229187,
0.06494368612766266,
0.003768373280763626,
-0.014263768680393696,
0.09235488623380661,
0.04585632309317589,
0.016036443412303925,
0.01857670024037361,
0.017341768369078636,
-0.043770886957645416,
-0.012837734073400497,
-0.12971961498260498,
0.0822075679898262,
0.16555102169513702,
0.08159321546554565,
-0.0025153302121907473,
0.04910283908247948,
-0.11745497584342957,
-0.08390354365110397,
0.09864985197782516,
-0.03327968344092369,
-0.010939104482531548,
-0.05545666068792343,
0.1393168866634369,
0.15790824592113495,
-0.18001005053520203,
0.06630969792604446,
-0.06947531551122665,
-0.05813371390104294,
-0.10700751096010208,
-0.17336682975292206,
-0.06371892243623734,
-0.032686252146959305,
-0.007239010650664568,
-0.06101612001657486,
0.06587019562721252,
0.10927599668502808,
0.011966768652200699,
0.0038074126932770014,
0.08745987713336945,
-0.03549729287624359,
0.0032165138982236385,
0.04436643421649933,
0.053415216505527496,
0.018384460359811783,
-0.06577059626579285,
0.0062953270971775055,
0.005856741219758987,
0.03797626122832298,
0.05783457309007645,
0.030920613557100296,
-0.010121258907020092,
0.010502477176487446,
-0.017619112506508827,
-0.09953264892101288,
0.040120288729667664,
-0.02619595266878605,
-0.04532664641737938,
0.14967231452465057,
0.019792752340435982,
-0.0015246463008224964,
-0.020656753331422806,
0.23401693999767303,
-0.06644817441701889,
-0.07621476799249649,
-0.139224112033844,
0.14507612586021423,
-0.04275493323802948,
0.049066413193941116,
0.04825753718614578,
-0.10474846512079239,
0.03731349855661392,
0.1451096534729004,
0.14759132266044617,
-0.028853775933384895,
0.008417275734245777,
0.010314252227544785,
0.004880689084529877,
-0.023987766355276108,
0.05473437160253525,
0.04762427508831024,
0.1143009290099144,
-0.06655429303646088,
0.0966365858912468,
-0.004376910626888275,
-0.08667857199907303,
-0.01979442685842514,
0.1345251053571701,
0.00039378736983053386,
0.02452908828854561,
-0.07970208674669266,
0.11903424561023712,
-0.06584613770246506,
-0.254661500453949,
0.06372799724340439,
-0.06844745576381683,
-0.15212500095367432,
-0.020589474588632584,
0.0192123856395483,
-0.00045748791308142245,
0.021751869469881058,
0.06367123872041702,
-0.060426320880651474,
0.1489165723323822,
0.03729896992444992,
-0.07695354521274567,
-0.07863835990428925,
0.07556982338428497,
-0.08442215621471405,
0.3028976023197174,
0.005024306941777468,
0.04913705214858055,
0.09561952203512192,
-0.03656364977359772,
-0.13419130444526672,
0.03908977657556534,
0.0922386571764946,
-0.052813757210969925,
0.06462136656045914,
0.20481634140014648,
-0.011546511203050613,
0.11757413297891617,
0.07377773523330688,
-0.08157313615083694,
0.05285948887467384,
-0.08638615161180496,
-0.0914662778377533,
-0.09036537259817123,
0.09207529574632645,
-0.060405224561691284,
0.1555742472410202,
0.13106994330883026,
-0.046590786427259445,
0.0021229966077953577,
-0.02774938941001892,
0.05171699821949005,
-0.0023280514869838953,
0.1081797182559967,
0.027423536404967308,
-0.1910448670387268,
0.03156386315822601,
-0.01286277174949646,
0.09995201975107193,
-0.2582627832889557,
-0.0796453133225441,
0.039762452244758606,
-0.01255226694047451,
-0.05679939687252045,
0.12189354747533798,
0.05551491677761078,
0.05127199366688728,
-0.055598754435777664,
-0.0537382997572422,
-0.00422793859615922,
0.16415901482105255,
-0.1055506095290184,
-0.00439038872718811
] |
null | null | transformers | # merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* [hyeogi/SOLAR-10.7B-dpo-v1](https://huggingface.co/hyeogi/SOLAR-10.7B-dpo-v1)
* [LDCC/LDCC-SOLAR-10.7B](https://huggingface.co/LDCC/LDCC-SOLAR-10.7B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
slices:
- sources:
- model: LDCC/LDCC-SOLAR-10.7B
layer_range: [0, 48]
- model: hyeogi/SOLAR-10.7B-dpo-v1
layer_range: [0, 48]
merge_method: slerp
tokenizer_source: base
base_model: LDCC/LDCC-SOLAR-10.7B
embed_slerp: true
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5
dtype: bfloat16
```
## Datasets
Finetuned using LoRA with [kyujinpy/OpenOrca-KO](https://huggingface.co/datasets/kyujinpy/OpenOrca-KO) | {"language": ["ko"], "license": "apache-2.0", "tags": ["mergekit", "merge", "LDCC/LDCC-SOLAR-10.7B", "hyeogi/SOLAR-10.7B-dpo-v1"], "base_model": ["LDCC/LDCC-SOLAR-10.7B", "hyeogi/SOLAR-10.7B-dpo-v1"]} | text-generation | jumtul/LDCC-Hyeogi.05 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"LDCC/LDCC-SOLAR-10.7B",
"hyeogi/SOLAR-10.7B-dpo-v1",
"ko",
"base_model:LDCC/LDCC-SOLAR-10.7B",
"base_model:hyeogi/SOLAR-10.7B-dpo-v1",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T16:25:02+00:00 | [] | [
"ko"
] | TAGS
#transformers #safetensors #llama #text-generation #mergekit #merge #LDCC/LDCC-SOLAR-10.7B #hyeogi/SOLAR-10.7B-dpo-v1 #ko #base_model-LDCC/LDCC-SOLAR-10.7B #base_model-hyeogi/SOLAR-10.7B-dpo-v1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # merge
This is a merge of pre-trained language models created using mergekit.
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* hyeogi/SOLAR-10.7B-dpo-v1
* LDCC/LDCC-SOLAR-10.7B
### Configuration
The following YAML configuration was used to produce this model:
## Datasets
Finetuned using LoRA with kyujinpy/OpenOrca-KO | [
"# merge\nThis is a merge of pre-trained language models created using mergekit.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the SLERP merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* hyeogi/SOLAR-10.7B-dpo-v1\n* LDCC/LDCC-SOLAR-10.7B",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Datasets\n\nFinetuned using LoRA with kyujinpy/OpenOrca-KO"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #LDCC/LDCC-SOLAR-10.7B #hyeogi/SOLAR-10.7B-dpo-v1 #ko #base_model-LDCC/LDCC-SOLAR-10.7B #base_model-hyeogi/SOLAR-10.7B-dpo-v1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# merge\nThis is a merge of pre-trained language models created using mergekit.",
"## Merge Details",
"### Merge Method\n\nThis model was merged using the SLERP merge method.",
"### Models Merged\n\nThe following models were included in the merge:\n* hyeogi/SOLAR-10.7B-dpo-v1\n* LDCC/LDCC-SOLAR-10.7B",
"### Configuration\n\nThe following YAML configuration was used to produce this model:",
"## Datasets\n\nFinetuned using LoRA with kyujinpy/OpenOrca-KO"
] | [
132,
18,
4,
18,
46,
17,
21
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #mergekit #merge #LDCC/LDCC-SOLAR-10.7B #hyeogi/SOLAR-10.7B-dpo-v1 #ko #base_model-LDCC/LDCC-SOLAR-10.7B #base_model-hyeogi/SOLAR-10.7B-dpo-v1 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# merge\nThis is a merge of pre-trained language models created using mergekit.## Merge Details### Merge Method\n\nThis model was merged using the SLERP merge method.### Models Merged\n\nThe following models were included in the merge:\n* hyeogi/SOLAR-10.7B-dpo-v1\n* LDCC/LDCC-SOLAR-10.7B### Configuration\n\nThe following YAML configuration was used to produce this model:## Datasets\n\nFinetuned using LoRA with kyujinpy/OpenOrca-KO"
] | [
-0.05027922987937927,
-0.02658991329371929,
-0.0035859323106706142,
-0.0636521503329277,
0.11659843474626541,
0.03568644821643829,
0.18308036029338837,
0.05488472059369087,
0.015831466764211655,
0.0925271138548851,
-0.027008797973394394,
0.12269053608179092,
0.09501028060913086,
0.14729955792427063,
0.009523372165858746,
-0.1681404709815979,
0.07742597162723541,
-0.031967926770448685,
-0.11170454323291779,
0.05337309092283249,
0.08826649934053421,
-0.09362874180078506,
0.07815533876419067,
0.009125033393502235,
-0.13348661363124847,
0.024273952469229698,
-0.06529183685779572,
0.012913764454424381,
0.11238520592451096,
0.0915316715836525,
0.05763661861419678,
0.07259514927864075,
-0.005170167423784733,
-0.22448177635669708,
0.022889262065291405,
-0.0031064781360328197,
-0.039539188146591187,
0.04566672444343567,
0.06633227318525314,
0.029079020023345947,
0.15914086997509003,
-0.17081153392791748,
-0.03473644331097603,
0.039976175874471664,
-0.04216562211513519,
-0.17332515120506287,
-0.1125662624835968,
0.10242921113967896,
0.13772398233413696,
-0.0034868335351347923,
-0.017638280987739563,
0.026760434731841087,
0.04173848405480385,
0.04438014328479767,
0.10504286736249924,
-0.20554320514202118,
0.038887981325387955,
0.18687157332897186,
0.01847650110721588,
-0.12147217243909836,
0.0014257206348702312,
0.05306779593229294,
0.048831500113010406,
-0.034718625247478485,
0.042547594755887985,
-0.04131777212023735,
0.08150511980056763,
-0.014705312438309193,
-0.12001010030508041,
-0.046333931386470795,
0.22158671915531158,
-0.003935357090085745,
-0.007655973080545664,
-0.12661336362361908,
-0.1136811226606369,
0.06758522987365723,
-0.0077984570525586605,
0.026599999517202377,
0.014686452224850655,
-0.01722482219338417,
0.06569410115480423,
-0.06503686308860779,
-0.09186338633298874,
-0.01967989280819893,
-0.079732745885849,
0.26669982075691223,
0.0932069942355156,
0.010283180512487888,
-0.020443426445126534,
0.13019759953022003,
-0.04949698597192764,
-0.12877996265888214,
-0.009310523979365826,
0.005752275697886944,
-0.04303349554538727,
0.03748733922839165,
-0.08024099469184875,
-0.07942613214254379,
0.12379960715770721,
0.2055646777153015,
-0.07922469824552536,
-0.007758974563330412,
0.13069529831409454,
0.04563794657588005,
0.027635538950562477,
0.031991198658943176,
-0.13356496393680573,
-0.03989579528570175,
0.04230746254324913,
0.004300383385270834,
0.18309882283210754,
0.008174983784556389,
-0.10211006551980972,
-0.03415565937757492,
0.022231820970773697,
0.03538661077618599,
0.10545673221349716,
0.05197840556502342,
-0.052706029266119,
-0.05944322794675827,
0.17240339517593384,
-0.09680373966693878,
-0.014454654417932034,
0.018213843926787376,
-0.0025360179133713245,
-0.08567055314779282,
0.08355765789747238,
0.04695617035031319,
0.010023844428360462,
0.10280055552721024,
-0.00924642849713564,
-0.00037432584213092923,
-0.062136828899383545,
-0.05101826786994934,
-0.011402367614209652,
0.058514658361673355,
0.014769378118216991,
-0.08168993145227432,
-0.21316304802894592,
-0.01800399459898472,
-0.00816649105399847,
-0.06405819207429886,
-0.044942911714315414,
-0.003969334065914154,
-0.009249486960470676,
-0.005920407362282276,
-0.023226339370012283,
0.009123205207288265,
-0.005977391730993986,
0.00012709524889942259,
0.005903144832700491,
0.03270970657467842,
-0.10611594468355179,
0.017129432410001755,
-0.07095058262348175,
0.11875368654727936,
-0.10657915472984314,
0.10342895984649658,
-0.034043408930301666,
0.044626183807849884,
-0.0920703187584877,
-0.04521580785512924,
0.028331944718956947,
0.00629679998382926,
0.08041492849588394,
0.172336146235466,
-0.13140013813972473,
-0.05238882824778557,
0.09060981869697571,
-0.13398034870624542,
-0.08613479882478714,
0.025378504768013954,
0.03369487076997757,
0.035499460995197296,
0.006979570258408785,
0.15893559157848358,
0.09769149124622345,
-0.0590774230659008,
-0.12196283787488937,
-0.03073340840637684,
0.061443157494068146,
-0.018535619601607323,
0.0636526495218277,
-0.033439382910728455,
-0.06076841801404953,
0.03342369571328163,
-0.10003567487001419,
0.11253383010625839,
-0.04314494505524635,
-0.05135568231344223,
-0.06446480751037598,
-0.03197280690073967,
0.03691685199737549,
-0.0011755781015381217,
-0.013754405081272125,
-0.0054747858084738255,
-0.01441330835223198,
0.1287248283624649,
0.16575275361537933,
-0.09814805537462234,
0.04180904105305672,
-0.015235685743391514,
0.14841553568840027,
-0.07412048429250717,
0.012558799237012863,
-0.038110390305519104,
-0.09499210864305496,
0.016460182145237923,
-0.05464520677924156,
-0.04114200547337532,
0.020178480073809624,
0.022920861840248108,
0.04932660609483719,
-0.06394658982753754,
-0.09876542538404465,
0.021496998146176338,
0.02885477989912033,
0.007976728491485119,
-0.17105074226856232,
-0.1244925707578659,
-0.02008877694606781,
0.28737562894821167,
-0.09717968106269836,
0.0787649154663086,
0.0605851486325264,
0.22872082889080048,
-0.07553565502166748,
0.017218371853232384,
0.10560812056064606,
0.02085343562066555,
-0.03946083039045334,
-0.02490290068089962,
0.05347243696451187,
0.024684349074959755,
-0.20540110766887665,
0.1017175167798996,
-0.18871407210826874,
-0.15505777299404144,
0.0871502161026001,
0.02556312270462513,
0.01945536583662033,
-0.08858752995729446,
-0.028301578015089035,
-0.07536309957504272,
0.025297975167632103,
-0.07147081941366196,
-0.03878794610500336,
-0.020160675048828125,
0.09179017692804337,
-0.023766320198774338,
-0.03605762869119644,
-0.024692369624972343,
-0.020869791507720947,
-0.04910184442996979,
0.08167488873004913,
-0.00790298543870449,
-0.041506461799144745,
0.09891687333583832,
0.1104189082980156,
0.06083930283784866,
0.09918299317359924,
-0.024062475189566612,
-0.00890832208096981,
-0.060820482671260834,
0.052059292793273926,
0.00947149470448494,
0.07398579269647598,
0.021048707887530327,
0.06237475574016571,
0.0701143890619278,
-0.05764983594417572,
0.0632910504937172,
-0.138034388422966,
0.021280329674482346,
0.03416208177804947,
-0.01782204769551754,
0.14836205542087555,
0.10763028264045715,
-0.005454279016703367,
0.000057326666137669235,
-0.00860581174492836,
0.03703857958316803,
0.0219378974288702,
-0.023565534502267838,
-0.08528696745634079,
0.2025526911020279,
-0.12755681574344635,
-0.18838365375995636,
-0.18099676072597504,
-0.03485843911767006,
-0.11745169013738632,
-0.03095485456287861,
0.05248378589749336,
-0.05149722471833229,
-0.08180808275938034,
-0.06812699884176254,
0.17533040046691895,
0.032592177391052246,
0.012616022489964962,
-0.036067597568035126,
-0.0456634983420372,
0.03213513270020485,
-0.10870784521102905,
0.004588128998875618,
0.01600288413465023,
-0.0570836178958416,
0.07429937273263931,
0.04071702063083649,
0.12069042026996613,
0.0800243392586708,
0.03524848446249962,
-0.028566943481564522,
0.012984652072191238,
0.22407269477844238,
-0.06246078386902809,
0.08602399379014969,
0.207940474152565,
0.01888977736234665,
0.04793020710349083,
0.21890594065189362,
0.042686715722084045,
-0.034345321357250214,
0.0009178604232147336,
-0.03398304432630539,
-0.07031504064798355,
-0.17911067605018616,
-0.18401066958904266,
-0.07195619493722916,
-0.062465470284223557,
-0.004806128330528736,
0.0790194422006607,
0.09691982716321945,
0.07001032680273056,
-0.13433924317359924,
-0.025622373446822166,
-0.011584043502807617,
0.0738399475812912,
0.22122494876384735,
0.03154471889138222,
0.07298869639635086,
-0.05567668005824089,
0.025506243109703064,
0.041105542331933975,
0.029697857797145844,
0.10538201034069061,
-0.012131760828197002,
0.09475234895944595,
0.11689603328704834,
0.11996761709451675,
-0.004624626599252224,
0.03409852460026741,
-0.050568755716085434,
0.014504334889352322,
0.029627596959471703,
-0.11559761315584183,
0.038632504642009735,
0.06375252455472946,
-0.07552526146173477,
0.07602780312299728,
-0.03892218694090843,
-0.0534033328294754,
0.044919226318597794,
0.10067939013242722,
0.13734279572963715,
-0.17491267621517181,
-0.07252849638462067,
0.09192539751529694,
0.0750996470451355,
-0.02527298405766487,
-0.06696627289056778,
0.060554198920726776,
-0.07303168624639511,
0.2289285510778427,
-0.0072178286500275135,
0.055722229182720184,
-0.011819996871054173,
-0.008786697871983051,
0.006976948119699955,
0.14599475264549255,
0.02457229234278202,
0.06753960996866226,
-0.16664882004261017,
0.11023248732089996,
0.0456041619181633,
-0.00031483141356147826,
0.0005948007456026971,
0.05249444395303726,
-0.018786156550049782,
0.17700187861919403,
0.027331331744790077,
0.04100587218999863,
0.0332062691450119,
-0.02241516299545765,
-0.12086948752403259,
0.008739842101931572,
-0.014157680794596672,
-0.09417227655649185,
0.09287464618682861,
-0.014064067043364048,
-0.02597666345536709,
-0.02078760787844658,
0.10181792080402374,
-0.10389432311058044,
-0.13504523038864136,
0.10717012733221054,
0.09541227668523788,
0.11789995431900024,
-0.11224585771560669,
-0.030563808977603912,
-0.11851033568382263,
0.16968205571174622,
-0.07954378426074982,
-0.1440315544605255,
-0.07245930284261703,
0.015058123506605625,
0.17055974900722504,
-0.10155165195465088,
0.06822678446769714,
-0.023367825895547867,
0.017202353104948997,
-0.0319632962346077,
-0.15845124423503876,
0.020207956433296204,
-0.0907202735543251,
-0.1405634582042694,
0.058308135718107224,
0.14710786938667297,
-0.03043115884065628,
0.023691507056355476,
-0.009711922146379948,
0.05264478549361229,
-0.10633747279644012,
-0.08332665264606476,
0.009593702852725983,
0.20342309772968292,
0.06945422291755676,
0.1354082077741623,
-0.008309715427458286,
-0.13297419250011444,
-0.00800605770200491,
-0.10490155965089798,
0.07963911443948746,
0.2823222577571869,
-0.061041612178087234,
0.09410139918327332,
0.08665556460618973,
-0.06560558080673218,
-0.19055157899856567,
-0.11461025476455688,
-0.052725955843925476,
0.06133626773953438,
-0.01290922611951828,
-0.11520638316869736,
0.023380491882562637,
0.07638516277074814,
-0.018312156200408936,
-0.04682563245296478,
-0.3153823912143707,
-0.18367953598499298,
0.04793655127286911,
0.055686626583337784,
0.10211994498968124,
-0.1608767807483673,
-0.12125138193368912,
-0.056687723845243454,
-0.2066689431667328,
0.05193908140063286,
-0.1356372833251953,
0.05200899392366409,
-0.05870803818106651,
0.02455217018723488,
0.037236541509628296,
-0.030522413551807404,
0.2274216115474701,
-0.024294601753354073,
0.002399920020252466,
-0.06778142601251602,
0.021924976259469986,
0.11228279769420624,
-0.04263249784708023,
0.07526207715272903,
-0.09438157826662064,
0.03676629066467285,
-0.1203000470995903,
0.010996771045029163,
-0.09205199033021927,
0.07540968060493469,
-0.09292803704738617,
-0.06030182167887688,
-0.10222230851650238,
0.012185635045170784,
0.0703987255692482,
0.0036102994345128536,
0.1423756182193756,
-0.003262478159740567,
0.09296087920665741,
0.16846102476119995,
0.06389978528022766,
0.043483465909957886,
-0.013264365494251251,
0.0005328914267010987,
-0.0301645677536726,
0.0369107685983181,
-0.09280768036842346,
-0.008271992206573486,
0.0733913704752922,
0.02237733080983162,
0.050300851464271545,
0.017563659697771072,
-0.0906895101070404,
0.0021771437022835016,
0.09734339267015457,
-0.15242503583431244,
-0.3558131456375122,
-0.045635148882865906,
0.06422369927167892,
-0.02670723758637905,
0.13513703644275665,
0.12310801446437836,
-0.056907087564468384,
-0.052559174597263336,
0.03141818568110466,
0.03974742814898491,
-0.05341176688671112,
0.11427522450685501,
0.03847796469926834,
0.07593289762735367,
-0.12133954465389252,
0.04162644222378731,
0.0628761351108551,
-0.007143449503928423,
-0.03495915234088898,
0.04624766856431961,
-0.1048223227262497,
-0.04272258281707764,
-0.0465436726808548,
0.012359442189335823,
-0.1368020474910736,
-0.11148254573345184,
-0.09919694066047668,
-0.11412345618009567,
-0.012440824881196022,
-0.018211914226412773,
0.05339503288269043,
0.010465609841048717,
0.06187653914093971,
-0.05836264044046402,
-0.05406707897782326,
0.06136160343885422,
-0.006903297733515501,
0.10341153293848038,
-0.12729014456272125,
0.08427749574184418,
-0.026103008538484573,
0.10219202190637589,
-0.031105780974030495,
-0.008806516416370869,
-0.07302939891815186,
-0.020875344052910805,
-0.10252396017313004,
-0.05616709217429161,
-0.16752518713474274,
-0.05238991603255272,
-0.020436767488718033,
-0.0405440591275692,
0.020175915211439133,
0.05747469142079353,
-0.01760307513177395,
-0.046985212713479996,
-0.07257000356912613,
0.02425629086792469,
-0.0740843415260315,
-0.029141878709197044,
0.02026662975549698,
-0.10683678835630417,
0.09503921121358871,
0.011270356364548206,
0.003231093054637313,
-0.03766046464443207,
-0.05660489574074745,
-0.049488816410303116,
0.04464855417609215,
-0.0022468571551144123,
0.0039983224123716354,
-0.13621382415294647,
-0.0856107771396637,
-0.012730843387544155,
-0.09455669671297073,
-0.051413629204034805,
0.07961035519838333,
-0.0670606791973114,
0.01517449226230383,
-0.0963636115193367,
0.04218415915966034,
-0.05443314462900162,
0.058704081922769547,
-0.0052982522174716,
0.06557594984769821,
0.0591881163418293,
-0.052836693823337555,
-0.00842653401196003,
-0.13793152570724487,
-0.0347304493188858,
-0.04449659213423729,
-0.0925467237830162,
-0.07015135884284973,
-0.06403644382953644,
0.029698388651013374,
0.026637639850378036,
0.11538432538509369,
-0.07867585867643356,
-0.0890679806470871,
0.0049660345539450645,
0.09825287014245987,
0.05740273743867874,
0.03600747510790825,
0.25485092401504517,
-0.0045409356243908405,
0.05973801016807556,
0.011683141812682152,
0.035819657146930695,
-0.003000773023813963,
0.016259465366601944,
0.04267038777470589,
0.08117002248764038,
0.038032952696084976,
0.028487207368016243,
0.059648558497428894,
-0.13427042961120605,
-0.06402336806058884,
-0.0453445129096508,
-0.07494273781776428,
0.06929539889097214,
-0.02776751108467579,
0.14913341403007507,
0.1149873211979866,
-0.19321925938129425,
0.07273988425731659,
0.07744872570037842,
-0.039274200797080994,
-0.07426352798938751,
-0.05991000682115555,
-0.07552583515644073,
-0.1509978473186493,
-0.02478589490056038,
-0.07927580177783966,
-0.03423468396067619,
0.08890650421380997,
-0.0383150689303875,
0.020266218110919,
0.14893345534801483,
0.1426004022359848,
-0.047044895589351654,
-0.05471647530794144,
-0.0052354466170072556,
0.011321628466248512,
-0.04065322130918503,
-0.04078754037618637,
0.041304707527160645,
-0.0005893093766644597,
0.007479999214410782,
0.05402877926826477,
0.11047637462615967,
0.09612792730331421,
-0.0007912222063168883,
-0.08054731041193008,
0.034161195158958435,
0.019568804651498795,
0.0922766774892807,
-0.02220872789621353,
0.02038215659558773,
-0.0036586683709174395,
-0.020877163857221603,
0.1203627660870552,
-0.04471822828054428,
-0.09717058390378952,
-0.011007887311279774,
0.0839940533041954,
-0.05480518192052841,
-0.0018809445900842547,
0.09157132357358932,
-0.13254289329051971,
-0.01982564851641655,
0.09754186868667603,
0.2099451720714569,
0.08219394832849503,
0.025135114789009094,
-0.002698321361094713,
0.0031992814037948847,
0.013793460093438625,
0.051722850650548935,
0.03809960559010506,
0.1392531543970108,
-0.06951861083507538,
0.08562204986810684,
-0.03511853888630867,
-0.06427083164453506,
-0.12446114420890808,
0.04264162480831146,
-0.0901595950126648,
-0.09213753789663315,
0.06432390958070755,
0.0555882453918457,
-0.0046454970724880695,
-0.11741550266742706,
0.1158791109919548,
-0.15302513539791107,
-0.09528058022260666,
-0.10418909043073654,
0.15543581545352936,
-0.005657147150486708,
0.005397479049861431,
-0.08234003931283951,
0.046755559742450714,
0.14510220289230347,
0.030599776655435562,
-0.07532230019569397,
-0.029138335958123207,
0.07596323639154434,
-0.08326093852519989,
0.028038589283823967,
-0.02380274422466755,
0.06112868711352348,
0.0972389280796051,
-0.018538648262619972,
-0.13984514772891998,
0.028025835752487183,
0.07574652880430222,
-0.07360862195491791,
0.06316443532705307,
0.06058104708790779,
0.006457298528403044,
0.016783975064754486,
-0.0015608022222295403,
-0.14778415858745575,
0.027886513620615005,
0.1517375260591507,
-0.023564999923110008,
-0.057117823511362076,
0.02184818685054779,
-0.07989461719989777,
0.09361688792705536,
0.08156471699476242,
-0.08479895442724228,
0.0040720487013459206,
-0.04116572439670563,
0.016453001648187637,
0.04065176844596863,
0.17803271114826202,
0.028524817898869514,
-0.13220003247261047,
0.041259922087192535,
-0.004422958008944988,
0.08904135972261429,
-0.16913741827011108,
-0.05637454614043236,
-0.05662380903959274,
-0.049980856478214264,
-0.06918742507696152,
0.134177103638649,
0.12307236343622208,
-0.008005647920072079,
-0.032346226274967194,
-0.11233597248792648,
-0.059514693915843964,
0.08994699269533157,
-0.11322475969791412,
-0.060030337423086166
] |
null | null | diffusers |
<!-- This model card has been generated automatically according to the information the training script had access to. You
should probably proofread and complete it, then remove this comment. -->
# SDXL LoRA DreamBooth - sdubail/alanmarmot_LoRA_DS_extended
<Gallery />
## Model description
These are sdubail/alanmarmot_LoRA_DS_extended LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using [DreamBooth](https://dreambooth.github.io/).
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of alanmarmot to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
[Download](sdubail/alanmarmot_LoRA_DS_extended/tree/main) them in the Files & versions tab.
## Intended uses & limitations
#### How to use
```python
# TODO: add an example code snippet for running this diffusion pipeline
```
#### Limitations and bias
[TODO: provide examples of latent issues and potential remediations]
## Training details
[TODO: describe the data used to train the model] | {"license": "openrail++", "library_name": "diffusers", "tags": ["text-to-image", "stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a photo of alanmarmot", "widget": []} | text-to-image | sdubail/alanmarmot_LoRA_DS_extended | [
"diffusers",
"tensorboard",
"text-to-image",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"has_space",
"region:us"
] | 2024-02-08T16:30:17+00:00 | [] | [] | TAGS
#diffusers #tensorboard #text-to-image #stable-diffusion-xl #stable-diffusion-xl-diffusers #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us
|
# SDXL LoRA DreamBooth - sdubail/alanmarmot_LoRA_DS_extended
<Gallery />
## Model description
These are sdubail/alanmarmot_LoRA_DS_extended LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
The weights were trained using DreamBooth.
LoRA for the text encoder was enabled: False.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
## Trigger words
You should use a photo of alanmarmot to trigger the image generation.
## Download model
Weights for this model are available in Safetensors format.
Download them in the Files & versions tab.
## Intended uses & limitations
#### How to use
#### Limitations and bias
[TODO: provide examples of latent issues and potential remediations]
## Training details
[TODO: describe the data used to train the model] | [
"# SDXL LoRA DreamBooth - sdubail/alanmarmot_LoRA_DS_extended\n\n<Gallery />",
"## Model description\n\nThese are sdubail/alanmarmot_LoRA_DS_extended LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of alanmarmot to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Intended uses & limitations",
"#### How to use",
"#### Limitations and bias\n\n[TODO: provide examples of latent issues and potential remediations]",
"## Training details\n\n[TODO: describe the data used to train the model]"
] | [
"TAGS\n#diffusers #tensorboard #text-to-image #stable-diffusion-xl #stable-diffusion-xl-diffusers #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n",
"# SDXL LoRA DreamBooth - sdubail/alanmarmot_LoRA_DS_extended\n\n<Gallery />",
"## Model description\n\nThese are sdubail/alanmarmot_LoRA_DS_extended LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.",
"## Trigger words\n\nYou should use a photo of alanmarmot to trigger the image generation.",
"## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.",
"## Intended uses & limitations",
"#### How to use",
"#### Limitations and bias\n\n[TODO: provide examples of latent issues and potential remediations]",
"## Training details\n\n[TODO: describe the data used to train the model]"
] | [
86,
30,
95,
19,
28,
9,
5,
24,
16
] | [
"passage: TAGS\n#diffusers #tensorboard #text-to-image #stable-diffusion-xl #stable-diffusion-xl-diffusers #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #has_space #region-us \n# SDXL LoRA DreamBooth - sdubail/alanmarmot_LoRA_DS_extended\n\n<Gallery />## Model description\n\nThese are sdubail/alanmarmot_LoRA_DS_extended LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.\n\nThe weights were trained using DreamBooth.\n\nLoRA for the text encoder was enabled: False.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix.## Trigger words\n\nYou should use a photo of alanmarmot to trigger the image generation.## Download model\n\nWeights for this model are available in Safetensors format.\n\nDownload them in the Files & versions tab.## Intended uses & limitations#### How to use#### Limitations and bias\n\n[TODO: provide examples of latent issues and potential remediations]## Training details\n\n[TODO: describe the data used to train the model]"
] | [
-0.041575729846954346,
0.12671130895614624,
-0.0026700219605118036,
0.029888411983847618,
0.1053159162402153,
-0.028425442054867744,
0.14836184680461884,
0.1131192296743393,
0.03322695940732956,
0.08624915033578873,
0.013305668719112873,
0.03887088969349861,
0.06878984719514847,
0.09885214269161224,
-0.03474953770637512,
-0.24167929589748383,
0.016868632286787033,
-0.025172367691993713,
-0.043003279715776443,
0.06107481196522713,
0.08436356484889984,
-0.0919937863945961,
0.08374808728694916,
0.012661022134125233,
-0.12303509563207626,
0.07733884453773499,
-0.0175432451069355,
-0.02266327664256096,
0.0381232313811779,
0.05979416146874428,
0.0655544176697731,
0.04666056111454964,
0.0944090485572815,
-0.23527617752552032,
-0.002462727716192603,
0.07371442764997482,
0.00289954268373549,
0.036023788154125214,
0.10271933674812317,
-0.04277787357568741,
0.1094142273068428,
-0.1625860631465912,
0.07803387194871902,
0.0654950737953186,
-0.05040600895881653,
-0.11081158369779587,
-0.04738534986972809,
-0.014778150245547295,
0.14913280308246613,
0.10378574579954147,
-0.022007925435900688,
0.06164299324154854,
0.0009847579058259726,
0.0420159250497818,
0.27075251936912537,
-0.1410328447818756,
-0.04936264455318451,
0.13066554069519043,
0.06856963038444519,
0.01982046291232109,
-0.07867568731307983,
0.049255773425102234,
0.047428712248802185,
-0.012939952313899994,
0.11895780265331268,
-0.05941148102283478,
0.06259423494338989,
-0.04416937381029129,
-0.12859483063220978,
-0.031876418739557266,
0.15781910717487335,
-0.00381400715559721,
-0.025876738131046295,
-0.1936858892440796,
-0.045904140919446945,
0.17041809856891632,
-0.01923047937452793,
-0.07324907928705215,
0.02021188661456108,
-0.029400700703263283,
-0.009066532365977764,
-0.10858182609081268,
-0.08116514980792999,
-0.025671163573861122,
0.08330302685499191,
0.11260022222995758,
0.012057861313223839,
-0.015455338172614574,
0.020297596231102943,
0.10711885988712311,
0.00908364076167345,
-0.12565398216247559,
0.024694552645087242,
-0.056455694139003754,
-0.08940884470939636,
0.008236472494900227,
-0.0004987685824744403,
-0.12960267066955566,
0.015194443985819817,
0.06748990714550018,
0.047567449510097504,
0.02594054490327835,
-0.02290547825396061,
0.024415329098701477,
0.023502221331000328,
0.10567550361156464,
-0.01258304063230753,
-0.08487070351839066,
0.03626002371311188,
0.06621846556663513,
-0.03384368121623993,
-0.04202331230044365,
-0.07237689197063446,
-0.03466229885816574,
0.030194256454706192,
0.06880069524049759,
0.04249751195311546,
-0.01772628352046013,
-0.07821507006883621,
-0.07261516153812408,
0.12064523249864578,
-0.12684334814548492,
0.01808694377541542,
-0.06628595292568207,
-0.06033513695001602,
0.02780173346400261,
0.09239707887172699,
0.04001893848180771,
-0.0493972972035408,
0.07502777129411697,
-0.04177958518266678,
-0.011879139579832554,
-0.11780966073274612,
-0.11719353497028351,
-0.019598189741373062,
-0.0693957731127739,
-0.021803660318255424,
-0.06484829634428024,
-0.19427354633808136,
-0.052798137068748474,
0.034421034157276154,
-0.06511697918176651,
-0.030888797715306282,
-0.028487378731369972,
-0.03586363047361374,
-0.021925263106822968,
0.0259721577167511,
0.0262468159198761,
0.018994955345988274,
0.033188119530677795,
0.002411940833553672,
0.06030287593603134,
0.09340817481279373,
0.02603534422814846,
-0.07219236344099045,
0.08115161955356598,
-0.15301164984703064,
0.1868850141763687,
-0.09716155380010605,
0.031071322038769722,
-0.1265564262866974,
-0.06530306488275528,
0.0009400181588716805,
-0.002346336841583252,
0.012714879587292671,
0.12534312903881073,
-0.2525005340576172,
-0.044248927384614944,
0.12747927010059357,
-0.13894972205162048,
-0.023601477965712547,
0.054001472890377045,
-0.07671689242124557,
0.08624912798404694,
0.07666004449129105,
0.1374892294406891,
0.08225589245557785,
-0.15902943909168243,
-0.00486307917162776,
-0.06273560971021652,
0.029361629858613014,
0.08803845196962357,
0.03153203800320625,
0.019092276692390442,
0.07171323895454407,
0.021765874698758125,
-0.0561978816986084,
0.021236276254057884,
-0.013222961686551571,
-0.04420147463679314,
-0.009642252698540688,
-0.05772482976317406,
0.0125796003267169,
0.019426818937063217,
-0.045082591474056244,
0.012225855141878128,
-0.07707687467336655,
0.1585962176322937,
0.05095108598470688,
-0.06715093553066254,
-0.00019040160987060517,
-0.0403025783598423,
-0.0260276198387146,
-0.08684980869293213,
-0.010961909778416157,
-0.15470647811889648,
-0.15323369204998016,
0.0328904427587986,
-0.01764826290309429,
0.08199141174554825,
0.08597271144390106,
0.04780696704983711,
0.02622383087873459,
-0.07089611887931824,
0.0066391173750162125,
-0.0025233591441065073,
-0.010038807988166809,
-0.039397142827510834,
-0.12111128866672516,
-0.045602839440107346,
-0.03474409505724907,
0.08500687777996063,
-0.2047051638364792,
0.03552648797631264,
0.12233232706785202,
0.12806248664855957,
0.05450093001127243,
-0.05472857132554054,
0.05806960165500641,
-0.013016955927014351,
-0.017640264704823494,
-0.09987328201532364,
-0.01097290962934494,
-0.014133719727396965,
-0.07784024626016617,
0.09112728387117386,
-0.18959732353687286,
-0.026546576991677284,
0.09675423800945282,
0.045452333986759186,
-0.06431420892477036,
-0.10906758904457092,
-0.05337842181324959,
-0.03351517394185066,
-0.08371903002262115,
-0.04432639479637146,
0.11471330374479294,
0.017997268587350845,
0.06823447346687317,
-0.07129105925559998,
-0.07450353354215622,
-0.023516321554780006,
0.008258120156824589,
-0.05048335716128349,
0.05323140695691109,
0.014512376859784126,
-0.10059085488319397,
0.05162280797958374,
-0.015403133817017078,
-0.0056509035639464855,
0.09339860081672668,
0.05596967041492462,
-0.08664204925298691,
-0.00428664218634367,
0.02442454732954502,
0.07154674828052521,
0.10591332614421844,
0.08580414205789566,
0.029677215963602066,
0.05605335906147957,
-0.050898388028144836,
0.011978653259575367,
-0.14367833733558655,
0.006492677144706249,
0.05370884761214256,
-0.04276832565665245,
0.11218401789665222,
0.009657369926571846,
-0.007990995422005653,
0.0588671900331974,
0.008185271173715591,
0.07142193615436554,
-0.012843295000493526,
-0.054564181715250015,
-0.09642482548952103,
0.1002829372882843,
-0.0691802054643631,
-0.23567739129066467,
-0.09505568444728851,
0.03374270722270012,
-0.04607352614402771,
0.009276269003748894,
0.015030428767204285,
-0.0469026118516922,
-0.10986030846834183,
-0.07596953958272934,
0.02608504332602024,
-0.04381941631436348,
-0.0484200082719326,
-0.022948889061808586,
0.042984895408153534,
0.05755334347486496,
-0.10712599009275436,
-0.0007609589956700802,
0.02981952764093876,
-0.04416358098387718,
0.008252197876572609,
0.001340540242381394,
0.12436987459659576,
0.09122701734304428,
-0.004060468636453152,
0.020181963220238686,
-0.03877825289964676,
0.22906246781349182,
-0.0878513753414154,
0.10687272250652313,
0.19593515992164612,
-0.01130355428904295,
0.10510573536157608,
0.129354789853096,
0.009490507654845715,
-0.08606018871068954,
0.06586159020662308,
0.10929038375616074,
-0.08600547164678574,
-0.1731523871421814,
-0.07624290138483047,
-0.0737086609005928,
-0.09787857532501221,
0.12370932847261429,
0.06887388974428177,
0.14630410075187683,
0.10962669551372528,
-0.06723111867904663,
0.04646635055541992,
0.09549058228731155,
0.15257379412651062,
0.12159547209739685,
0.03308563306927681,
0.08063213527202606,
-0.03423720598220825,
-0.04920710623264313,
0.07308173179626465,
0.053378112614154816,
0.262194961309433,
-0.07427285611629486,
0.04490788280963898,
0.04539329186081886,
0.02906389907002449,
0.011405549943447113,
0.03469927981495857,
-0.03325717896223068,
0.015532229095697403,
-0.03462203964591026,
-0.11792661249637604,
0.0030336547642946243,
0.14204904437065125,
0.015248186886310577,
-0.03022761270403862,
0.04858122766017914,
-0.006210798863321543,
0.017205681651830673,
0.11931756138801575,
0.0009683335083536804,
-0.22751282155513763,
0.007116040680557489,
0.07674368470907211,
0.009381720796227455,
-0.036088570952415466,
-0.012368001975119114,
0.1286299228668213,
-0.1228744387626648,
0.0921512171626091,
-0.08233606815338135,
0.11746537685394287,
-0.07248683273792267,
-0.0564066581428051,
0.037215638905763626,
0.13006842136383057,
-0.012473632581532001,
0.09980545938014984,
-0.22931459546089172,
0.08514226973056793,
0.012716243974864483,
0.03209322690963745,
-0.08580200374126434,
0.07600059360265732,
0.008276969194412231,
-0.053306810557842255,
0.15209953486919403,
-0.007139585446566343,
-0.0218303631991148,
-0.05803123489022255,
-0.047554295510053635,
-0.034110065549612045,
0.0785464197397232,
-0.08508477360010147,
0.08277673274278641,
-0.01370360516011715,
-0.029882146045565605,
0.014338023960590363,
-0.04149208962917328,
-0.1766831874847412,
-0.21424618363380432,
0.06719862669706345,
0.0009519229643046856,
-0.046402573585510254,
-0.06705770641565323,
-0.0378212109208107,
-0.018349332734942436,
0.2160237431526184,
-0.010727866552770138,
-0.13971619307994843,
-0.17024879157543182,
-0.04238484427332878,
0.19702193140983582,
-0.023879844695329666,
0.05394371598958969,
0.039888594299554825,
0.22864118218421936,
-0.10324189066886902,
-0.0927799642086029,
0.04843582957983017,
-0.08068867772817612,
-0.1289159059524536,
-0.04272221028804779,
0.1199573278427124,
0.0705394372344017,
0.030907033011317253,
0.020930927246809006,
0.044959016144275665,
-0.0014654956758022308,
-0.09345978498458862,
0.014135293662548065,
0.1079605296254158,
0.036669887602329254,
0.013462006114423275,
-0.018307628110051155,
-0.027996057644486427,
-0.11696504801511765,
0.013751056976616383,
0.0141133489087224,
0.23990894854068756,
-0.0628102719783783,
0.08860079199075699,
-0.01957383193075657,
-0.07658132910728455,
-0.13675585389137268,
0.07123997062444687,
0.06307591497898102,
-0.0012500964803621173,
0.07482392340898514,
-0.17753784358501434,
0.049633726477622986,
0.025788182392716408,
-0.023542845621705055,
0.10422543436288834,
-0.2899624705314636,
-0.16152678430080414,
-0.012759683653712273,
0.10669677704572678,
-0.07721766084432602,
-0.13838009536266327,
-0.050734054297208786,
-0.10888707637786865,
-0.0756719559431076,
0.10931137204170227,
0.02213553711771965,
0.012747854925692081,
0.0399554967880249,
0.07215278595685959,
0.048229459673166275,
-0.04395841062068939,
0.1405028998851776,
0.014558729715645313,
0.06364388018846512,
-0.08337961137294769,
0.008866677060723305,
0.10638268291950226,
-0.07416248321533203,
0.05514711141586304,
-0.054185111075639725,
0.015418070368468761,
-0.13577784597873688,
-0.06994128227233887,
-0.015705473721027374,
0.052876826375722885,
-0.03609064221382141,
-0.105767160654068,
-0.06699348241090775,
0.07509790360927582,
0.13801684975624084,
-0.011543451808393002,
-0.10809390246868134,
-0.08748288452625275,
-0.0020517322700470686,
0.15056782960891724,
0.1237398087978363,
0.09994708001613617,
-0.08981519192457199,
-0.0010555139742791653,
0.01775166019797325,
0.04422084242105484,
-0.08852164447307587,
0.02225513383746147,
0.0974355936050415,
0.04355300962924957,
0.10681276023387909,
0.008089764043688774,
-0.10299287736415863,
0.015772094950079918,
0.017512725666165352,
-0.06567224860191345,
-0.0991932675242424,
-0.043426621705293655,
0.09697309136390686,
-0.06842022389173508,
-0.09223605692386627,
0.08610313385725021,
-0.07253948599100113,
-0.009004777297377586,
0.01901366375386715,
0.0401747040450573,
0.024462472647428513,
0.09540541470050812,
0.03136441111564636,
0.05833704397082329,
-0.07482028007507324,
0.07421623915433884,
0.09128642082214355,
-0.07716052234172821,
0.02332012914121151,
0.07984206080436707,
-0.04350893199443817,
0.020152322947978973,
-0.05032409355044365,
0.08193385601043701,
-0.08983160555362701,
0.00003196034595021047,
-0.04420601204037666,
-0.09657356888055801,
0.0024458388797938824,
0.07621830701828003,
0.0333423987030983,
-0.020294828340411186,
0.01010621152818203,
0.00726365577429533,
-0.15894420444965363,
0.09573221951723099,
0.10461325943470001,
0.06777900457382202,
-0.1640843003988266,
0.010559244081377983,
0.02258080616593361,
-0.019102027639746666,
-0.05528933182358742,
0.007236822508275509,
-0.06532256305217743,
0.0034773910883814096,
-0.06520465761423111,
0.07635486125946045,
-0.10236489027738571,
-0.007448885124176741,
-0.05189480260014534,
-0.04694138467311859,
-0.0006601494969800115,
0.053956884890794754,
-0.05936918035149574,
-0.054799895733594894,
-0.015322474762797356,
0.017027199268341064,
-0.1302398443222046,
-0.07493529468774796,
0.06918013095855713,
-0.11173895001411438,
0.03501186519861221,
-0.03116784058511257,
-0.05578118935227394,
-0.04829949513077736,
-0.08431507647037506,
-0.008714384399354458,
0.10193895548582077,
0.014220701530575752,
0.017153440043330193,
-0.06705648452043533,
0.010960069485008717,
-0.005314345005899668,
-0.012355102226138115,
-0.03400095924735069,
-0.046725984662771225,
-0.12183831632137299,
0.017410917207598686,
-0.0425313264131546,
-0.06703253835439682,
-0.03184482082724571,
0.08967362344264984,
0.1537126749753952,
0.03500944375991821,
0.08277315646409988,
-0.07352223992347717,
0.10186563432216644,
-0.173864483833313,
-0.02892729453742504,
0.01498468779027462,
0.001438590930774808,
-0.00042005241266451776,
-0.05268649384379387,
0.043133027851581573,
-0.01218133233487606,
0.19913145899772644,
0.03464174270629883,
0.0523383729159832,
0.03528808429837227,
-0.03902628645300865,
0.07359196990728378,
0.03354403376579285,
0.13522478938102722,
0.012339502573013306,
0.02602771855890751,
0.03714480251073837,
0.011296545155346394,
0.03793533891439438,
-0.03678205981850624,
0.0811995193362236,
0.12693877518177032,
0.01665046066045761,
0.02326544187963009,
0.037258315831422806,
-0.051884837448596954,
-0.10719464719295502,
0.12098793685436249,
-0.007545620668679476,
0.03451629355549812,
-0.08147425949573517,
0.08740412443876266,
0.16101191937923431,
-0.17307887971401215,
0.0944451093673706,
0.07177351415157318,
-0.06273314356803894,
-0.06159254163503647,
-0.15806476771831512,
-0.07078041136264801,
-0.06959410011768341,
0.023746205493807793,
-0.09926341474056244,
0.05613432079553604,
0.04534466937184334,
-0.009791051037609577,
0.021200982853770256,
0.13752298057079315,
-0.060772571712732315,
-0.04909970611333847,
0.030498959124088287,
-0.003151373006403446,
-0.006064796354621649,
0.04425705596804619,
0.006266724783927202,
0.07134928554296494,
0.08028408139944077,
0.026105981320142746,
0.01970604807138443,
0.04722832888364792,
0.011889023706316948,
-0.00666885357350111,
-0.050649385899305344,
-0.021622752770781517,
0.014397074468433857,
0.039204470813274384,
0.1322885900735855,
0.02675756625831127,
-0.041245318949222565,
-0.05911964923143387,
0.15943148732185364,
-0.09020928293466568,
-0.11611206084489822,
-0.07654321938753128,
0.06862696260213852,
-0.01327955350279808,
0.01585613563656807,
-0.014892080798745155,
-0.11047648638486862,
0.04950128123164177,
0.11646566540002823,
0.14336474239826202,
-0.025445692241191864,
-0.0013269652845337987,
-0.04551912099123001,
0.008151654154062271,
-0.04309920594096184,
0.020937273278832436,
0.01855935901403427,
0.21431496739387512,
-0.08023001998662949,
-0.015305067412555218,
-0.025618478655815125,
-0.06348942965269089,
-0.07938898354768753,
0.03661319985985756,
0.019590437412261963,
-0.005632605403661728,
-0.031561464071273804,
0.09840288013219833,
-0.03268444910645485,
-0.1530025750398636,
0.0882243663072586,
-0.10592948645353317,
-0.07271679490804672,
-0.014467031694948673,
-0.027462543919682503,
-0.028482506051659584,
0.02505515143275261,
-0.028847824782133102,
-0.004322601482272148,
0.0851297378540039,
0.03148403763771057,
-0.09117753058671951,
0.025348858907818794,
0.042747315019369125,
-0.026966700330376625,
0.1859070509672165,
-0.03622918576002121,
0.05232327803969383,
0.045949406921863556,
-0.03190920129418373,
-0.11008918285369873,
0.024829795584082603,
0.04782477766275406,
-0.11595745384693146,
0.017450952902436256,
0.20219996571540833,
-0.043060995638370514,
0.10782557725906372,
0.043877970427274704,
-0.10869258642196655,
0.032895904034376144,
-0.01728195697069168,
0.006904382258653641,
-0.09963096678256989,
0.030134735628962517,
-0.12394167482852936,
0.11446041613817215,
0.11881257593631744,
-0.01466637197881937,
0.024876076728105545,
-0.059175170958042145,
0.024673759937286377,
0.025008704513311386,
0.16461972892284393,
-0.0023889518342912197,
-0.0521259531378746,
-0.028339801356196404,
0.008221444673836231,
0.01776350662112236,
-0.20941658318042755,
-0.08838362246751785,
-0.039758384227752686,
-0.04923166707158089,
-0.022727901116013527,
0.10261531919240952,
0.09280848503112793,
0.01629650592803955,
-0.024215739220380783,
-0.25462278723716736,
0.021648088470101357,
0.11299387365579605,
-0.11285550147294998,
0.03297533094882965
] |
null | null | transformers |
# Model Trained Using AutoTrain
- Problem type: Image Classification
## Validation Metricsg
loss: 1.5556857585906982
f1_macro: 0.2
f1_micro: 0.2631578947368421
f1_weighted: 0.21052631578947367
precision_macro: 0.27272727272727276
precision_micro: 0.2631578947368421
precision_weighted: 0.2535885167464115
recall_macro: 0.22666666666666666
recall_micro: 0.2631578947368421
recall_weighted: 0.2631578947368421
accuracy: 0.2631578947368421
| {"tags": ["autotrain", "image-classification"], "datasets": ["autotrain-vhhui-btea8/autotrain-data"], "widget": [{"src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg", "example_title": "Tiger"}, {"src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg", "example_title": "Teapot"}, {"src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg", "example_title": "Palace"}]} | image-classification | basavaakash002/autotrain-vhhui-btea8 | [
"transformers",
"safetensors",
"vit",
"image-classification",
"autotrain",
"dataset:autotrain-vhhui-btea8/autotrain-data",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:30:23+00:00 | [] | [] | TAGS
#transformers #safetensors #vit #image-classification #autotrain #dataset-autotrain-vhhui-btea8/autotrain-data #autotrain_compatible #endpoints_compatible #region-us
|
# Model Trained Using AutoTrain
- Problem type: Image Classification
## Validation Metricsg
loss: 1.5556857585906982
f1_macro: 0.2
f1_micro: 0.2631578947368421
f1_weighted: 0.21052631578947367
precision_macro: 0.27272727272727276
precision_micro: 0.2631578947368421
precision_weighted: 0.2535885167464115
recall_macro: 0.22666666666666666
recall_micro: 0.2631578947368421
recall_weighted: 0.2631578947368421
accuracy: 0.2631578947368421
| [
"# Model Trained Using AutoTrain\n\n- Problem type: Image Classification",
"## Validation Metricsg\nloss: 1.5556857585906982\n\nf1_macro: 0.2\n\nf1_micro: 0.2631578947368421\n\nf1_weighted: 0.21052631578947367\n\nprecision_macro: 0.27272727272727276\n\nprecision_micro: 0.2631578947368421\n\nprecision_weighted: 0.2535885167464115\n\nrecall_macro: 0.22666666666666666\n\nrecall_micro: 0.2631578947368421\n\nrecall_weighted: 0.2631578947368421\n\naccuracy: 0.2631578947368421"
] | [
"TAGS\n#transformers #safetensors #vit #image-classification #autotrain #dataset-autotrain-vhhui-btea8/autotrain-data #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Trained Using AutoTrain\n\n- Problem type: Image Classification",
"## Validation Metricsg\nloss: 1.5556857585906982\n\nf1_macro: 0.2\n\nf1_micro: 0.2631578947368421\n\nf1_weighted: 0.21052631578947367\n\nprecision_macro: 0.27272727272727276\n\nprecision_micro: 0.2631578947368421\n\nprecision_weighted: 0.2535885167464115\n\nrecall_macro: 0.22666666666666666\n\nrecall_micro: 0.2631578947368421\n\nrecall_weighted: 0.2631578947368421\n\naccuracy: 0.2631578947368421"
] | [
61,
16,
142
] | [
"passage: TAGS\n#transformers #safetensors #vit #image-classification #autotrain #dataset-autotrain-vhhui-btea8/autotrain-data #autotrain_compatible #endpoints_compatible #region-us \n# Model Trained Using AutoTrain\n\n- Problem type: Image Classification## Validation Metricsg\nloss: 1.5556857585906982\n\nf1_macro: 0.2\n\nf1_micro: 0.2631578947368421\n\nf1_weighted: 0.21052631578947367\n\nprecision_macro: 0.27272727272727276\n\nprecision_micro: 0.2631578947368421\n\nprecision_weighted: 0.2535885167464115\n\nrecall_macro: 0.22666666666666666\n\nrecall_micro: 0.2631578947368421\n\nrecall_weighted: 0.2631578947368421\n\naccuracy: 0.2631578947368421"
] | [
-0.13183455169200897,
0.10540620982646942,
-0.0018977689323946834,
0.11414158344268799,
0.12355975806713104,
0.0507659874856472,
0.12208354473114014,
0.1124124526977539,
-0.028122631832957268,
0.11870939284563065,
0.14185430109500885,
0.09799717366695404,
0.06449797004461288,
0.13183297216892242,
-0.1067986711859703,
-0.16187801957130432,
0.03694083169102669,
0.030491160228848457,
0.07947291433811188,
0.07837817817926407,
0.07561638206243515,
-0.12008347362279892,
0.10741257667541504,
-0.028192030265927315,
-0.233554407954216,
0.007795991841703653,
0.1434398591518402,
-0.10083971172571182,
0.10549210011959076,
0.09590780735015869,
0.09447557479143143,
0.010239778086543083,
0.12060295045375824,
-0.06769376248121262,
-0.020201051607728004,
0.03133726865053177,
-0.009560743346810341,
0.06780017912387848,
0.1589621752500534,
0.04281548038125038,
-0.06548736244440079,
-0.059006545692682266,
0.0481550358235836,
0.06616290658712387,
-0.08175766468048096,
-0.13664257526397705,
-0.11422808468341827,
0.09355472773313522,
0.1058981865644455,
0.07846789062023163,
-0.0024675524327903986,
0.1884063482284546,
-0.02338450588285923,
0.09169445186853409,
0.12460482865571976,
-0.23465895652770996,
-0.06893455982208252,
0.15508981049060822,
-0.03205527737736702,
-0.06557504087686539,
-0.07433172315359116,
0.07330796867609024,
0.09733022004365921,
-0.008445686660706997,
0.035552043467760086,
-0.030139341950416565,
-0.026171471923589706,
-0.0018707035342231393,
-0.0846589058637619,
-0.02359597571194172,
0.1412181407213211,
0.039770543575286865,
-0.07753956317901611,
-0.0950966477394104,
-0.043707702308893204,
-0.09106462448835373,
-0.08120337873697281,
-0.03651406988501549,
0.003579378593713045,
-0.05666746571660042,
-0.06327632814645767,
0.09764064848423004,
-0.05800168961286545,
-0.1076732873916626,
-0.12237607687711716,
0.0821751281619072,
-0.007523950655013323,
0.033225882798433304,
0.004855967126786709,
0.024935569614171982,
-0.039383623749017715,
-0.08355437219142914,
0.00024593036505393684,
-0.004254440777003765,
-0.08148623257875443,
-0.05619056895375252,
0.020009363070130348,
-0.03765038773417473,
0.0036852580960839987,
0.12762293219566345,
0.002404967090114951,
0.062815822660923,
-0.01745506562292576,
0.007048938423395157,
-0.07776203006505966,
0.15564963221549988,
-0.10759025812149048,
-0.051513075828552246,
0.009267202578485012,
0.02580724097788334,
0.029098954051733017,
-0.005220727529376745,
-0.06477894634008408,
-0.10392500460147858,
0.15913251042366028,
0.035654038190841675,
0.0006894958205521107,
0.037068385630846024,
-0.0675138309597969,
-0.031112397089600563,
0.023569924756884575,
-0.053624652326107025,
0.02521991729736328,
-0.038428060710430145,
-0.09251443296670914,
-0.016809379681944847,
0.04521015286445618,
-0.003680907189846039,
-0.025038989260792732,
0.11909697204828262,
-0.13365483283996582,
0.00582443131133914,
-0.05369051918387413,
-0.12333530932664871,
0.03385504335165024,
-0.02827051468193531,
-0.018900346010923386,
-0.12451276928186417,
-0.18933773040771484,
-0.022075803950428963,
-0.04488489404320717,
-0.09729993343353271,
0.020780576393008232,
0.01907975785434246,
-0.05018467828631401,
0.08383512496948242,
0.018435252830386162,
0.061666447669267654,
-0.004785570781677961,
0.0449238196015358,
0.07699904590845108,
0.1121218204498291,
-0.0335974283516407,
0.006932368967682123,
-0.0636875256896019,
0.013725870288908482,
-0.18185590207576752,
-0.0060004922561347485,
-0.05139412358403206,
0.05751426890492439,
-0.10302001237869263,
-0.020946841686964035,
0.06985010951757431,
0.0021639023907482624,
0.04898637533187866,
0.17151395976543427,
-0.12493177503347397,
-0.02848822996020317,
0.10176123678684235,
-0.09050415456295013,
-0.11235600709915161,
0.12200037389993668,
0.017697470262646675,
0.0009308140724897385,
0.033705465495586395,
0.07300516217947006,
0.043296851217746735,
-0.09328822791576385,
-0.06972399353981018,
-0.04261105880141258,
-0.04618707671761513,
-0.0537458173930645,
0.07152358442544937,
-0.0029675965197384357,
-0.10509920120239258,
0.009858968667685986,
0.028718825429677963,
0.013970175758004189,
-0.10211300849914551,
-0.05884004011750221,
-0.05152865871787071,
-0.04362223669886589,
0.0563121996819973,
0.03074284829199314,
0.0419059656560421,
-0.07928973436355591,
-0.0506238229572773,
0.05805950611829758,
0.08681412786245346,
0.00820882897824049,
-0.051607511937618256,
-0.14646956324577332,
0.0486624650657177,
-0.14948216080665588,
-0.02066454477608204,
-0.1437079906463623,
-0.1070108637213707,
-0.00027838427922688425,
-0.035302843898534775,
-0.028066402301192284,
-0.09076935052871704,
0.08701463043689728,
0.052468441426754,
-0.06327090412378311,
-0.0459638349711895,
0.07364390045404434,
-0.00448215426877141,
-0.11567460745573044,
-0.11968199908733368,
-0.018237318843603134,
0.01766391657292843,
0.15117205679416656,
-0.18172453343868256,
-0.016348181292414665,
0.04553908109664917,
0.08556675165891647,
0.05655704811215401,
-0.007517912425100803,
-0.02624654211103916,
0.029566297307610512,
-0.025583554059267044,
-0.06332505494356155,
0.05243270471692085,
-0.050968509167432785,
-0.04687594622373581,
0.02274824120104313,
-0.23146174848079681,
0.25873881578445435,
0.12910345196723938,
-0.016833845525979996,
-0.09928613156080246,
-0.015408328734338284,
0.01269589550793171,
-0.05485760048031807,
-0.07249213755130768,
-0.01520515140146017,
-0.08546556532382965,
-0.015105404891073704,
0.11408088356256485,
-0.048862896859645844,
-0.017753437161445618,
0.0696362778544426,
-0.017620710656046867,
-0.08047525584697723,
0.11249608546495438,
0.02365340106189251,
-0.16083721816539764,
0.13196533918380737,
0.06723843514919281,
-0.06181919574737549,
0.11202914267778397,
0.0300085861235857,
-0.04163552075624466,
-0.06699509918689728,
-0.04354492574930191,
0.06782639771699905,
0.13149331510066986,
0.02640158124268055,
0.029796747490763664,
0.06260991841554642,
-0.01837647333741188,
-0.034831032156944275,
-0.09923914819955826,
-0.010754670016467571,
0.007941246964037418,
-0.014051996171474457,
0.04258327931165695,
0.00765966298058629,
0.044679850339889526,
0.1584940254688263,
0.013675146736204624,
-0.0508791022002697,
0.04618990048766136,
-0.003324755234643817,
-0.08289778232574463,
0.2096116989850998,
-0.0811273455619812,
-0.20604196190834045,
-0.18612369894981384,
-0.04839063808321953,
-0.08507265150547028,
0.0005246470682322979,
-0.016483696177601814,
-0.10459208488464355,
-0.11442212760448456,
-0.07140450179576874,
0.00680919736623764,
0.03395022824406624,
0.003388361306861043,
0.051115136593580246,
-0.003835705341771245,
0.08390385657548904,
-0.05832217261195183,
-0.033328037708997726,
-0.0036493423394858837,
-0.045062772929668427,
0.10595766454935074,
0.011833040975034237,
0.12987327575683594,
0.12463828176259995,
-0.026038404554128647,
0.023907998576760292,
-0.031243611127138138,
0.16492798924446106,
-0.045273397117853165,
0.015171848237514496,
0.20171096920967102,
0.048183947801589966,
0.010052318684756756,
0.13179896771907806,
0.024843383580446243,
-0.0908624604344368,
0.03172488138079643,
0.05871612951159477,
-0.02398928813636303,
-0.13511668145656586,
-0.12669402360916138,
0.012531140819191933,
0.08081843703985214,
0.1058802455663681,
0.02064361609518528,
0.10716499388217926,
0.10768266022205353,
-0.0014599626883864403,
0.09477914124727249,
-0.04787229374051094,
0.10083334892988205,
0.11798251420259476,
0.04387499764561653,
0.1855088621377945,
-0.05911507084965706,
-0.01048020925372839,
0.07837902009487152,
-0.05059459060430527,
0.10024945437908173,
0.052426405251026154,
0.07774993032217026,
-0.02502945065498352,
0.11910723894834518,
0.07566533982753754,
0.09694234281778336,
0.059042952954769135,
-0.07552832365036011,
0.03523648530244827,
-0.08884670585393906,
-0.0450877882540226,
0.030483024194836617,
-0.035587750375270844,
0.07137420028448105,
-0.12132526934146881,
0.09683798253536224,
0.035689521580934525,
0.16224656999111176,
0.08293598890304565,
-0.43898606300354004,
-0.09984131902456284,
0.03473087027668953,
0.03209066763520241,
-0.1202661544084549,
-0.03311197832226753,
0.0496540404856205,
-0.1025143712759018,
0.0347905196249485,
-0.07016917318105698,
0.08489102125167847,
-0.03489508107304573,
-0.0018439769046381116,
-0.0047113229520618916,
0.033633697777986526,
-0.02873220667243004,
0.019403085112571716,
-0.2088318169116974,
0.2098681628704071,
0.0657019093632698,
0.040633320808410645,
-0.08132219314575195,
0.002857463899999857,
0.04884856194257736,
0.0766824334859848,
0.16549532115459442,
-0.01071593351662159,
-0.11795111745595932,
-0.3331942856311798,
-0.0476037859916687,
-0.03157668560743332,
0.06081121787428856,
0.033372990787029266,
0.08359689265489578,
-0.07596254348754883,
-0.02837175503373146,
0.037019215524196625,
-0.022815529257059097,
-0.14586959779262543,
-0.020761480554938316,
0.032457973808050156,
0.06951190531253815,
0.00920012779533863,
-0.07741586118936539,
-0.08359027653932571,
-0.04454610124230385,
0.10783790796995163,
-0.13708263635635376,
-0.022327091544866562,
-0.17774680256843567,
0.09727375954389572,
0.08346342295408249,
-0.09178708493709564,
0.08036915212869644,
-0.027381908148527145,
0.12496354430913925,
0.012841305695474148,
-0.11703991889953613,
0.13444459438323975,
-0.11849717050790787,
-0.13041937351226807,
-0.03484434261918068,
0.03103288635611534,
-0.03489290922880173,
0.037578437477350235,
0.037965450435876846,
0.04464627057313919,
0.0423203781247139,
-0.0717715248465538,
0.08131073415279388,
-0.002393461298197508,
0.13504986464977264,
0.05872597545385361,
-0.004100334364920855,
-0.15734276175498962,
-0.03247418627142906,
-0.018879622220993042,
0.08169176429510117,
0.3095724582672119,
-0.07851070910692215,
-0.048938948661088943,
0.08167446404695511,
-0.016352687031030655,
-0.2904200255870819,
0.06807678937911987,
-0.03683140128850937,
0.03217099606990814,
-0.0036115467082709074,
-0.018752630800008774,
0.1230909451842308,
0.1662214696407318,
-0.04716360196471214,
0.013176369480788708,
-0.2495499849319458,
-0.13722455501556396,
0.18078629672527313,
0.13992013037204742,
0.10827992111444473,
-0.14518572390079498,
-0.057919807732105255,
-0.1286325305700302,
-0.16176794469356537,
0.04908263310790062,
0.002683135448023677,
0.06537370383739471,
-0.07270986586809158,
-0.017525387927889824,
0.04604288563132286,
-0.08515642583370209,
0.12855607271194458,
-0.05946546047925949,
0.11310815811157227,
-0.09985889494419098,
-0.05522917956113815,
0.032339271157979965,
-0.06469972431659698,
0.10612444579601288,
0.051395874470472336,
0.07237154245376587,
-0.1047617718577385,
-0.019585473462939262,
-0.032832227647304535,
0.06574124842882156,
-0.05193650722503662,
-0.07257286459207535,
-0.035425540059804916,
0.07571840286254883,
-0.09084497392177582,
-0.0575687475502491,
0.025696318596601486,
-0.018728751689195633,
0.07065151631832123,
0.1953071504831314,
0.08063586801290512,
-0.06528706103563309,
-0.011225016787648201,
0.01723366789519787,
-0.06164566054940224,
0.0985344722867012,
-0.03440649062395096,
0.07182985544204712,
0.10936342179775238,
0.0133908586576581,
0.12782683968544006,
0.06415046751499176,
-0.007039071526378393,
-0.021296072751283646,
0.07767294347286224,
-0.13392114639282227,
-0.036332495510578156,
-0.016110483556985855,
0.025173576548695564,
-0.019506286829710007,
-0.0016022803029045463,
0.14891114830970764,
-0.04181590676307678,
0.01856139302253723,
-0.010026403702795506,
0.014843440614640713,
0.02909270115196705,
0.24877823889255524,
0.010649271309375763,
0.0633431002497673,
-0.08018501847982407,
0.0940123200416565,
0.05897006392478943,
-0.11109522730112076,
-0.0009760022512637079,
0.04471389576792717,
-0.07842295616865158,
-0.03682896867394447,
0.05540195852518082,
0.16445347666740417,
-0.1818131059408188,
-0.04204495623707771,
-0.10309799760580063,
-0.14948096871376038,
0.04923862963914871,
0.24738730490207672,
0.0765259638428688,
0.005375317297875881,
0.04098951816558838,
-0.05620858073234558,
-0.12083613127470016,
0.08038034290075302,
0.05699237436056137,
0.06117577105760574,
-0.1766631305217743,
0.139816552400589,
-0.053127944469451904,
0.008483102545142174,
-0.04998287186026573,
0.04474318400025368,
-0.18871602416038513,
-0.005655843764543533,
-0.11346395313739777,
0.05011314898729324,
-0.051121775060892105,
-0.00525114219635725,
-0.030354153364896774,
-0.006518542766571045,
-0.043069105595350266,
0.02272777445614338,
-0.03216046839952469,
-0.04815513268113136,
0.04555762559175491,
0.05137929692864418,
-0.09611151367425919,
-0.05676230043172836,
0.03462473675608635,
-0.02106456458568573,
0.06404618918895721,
0.0158794317394495,
0.0890076532959938,
-0.0044344221241772175,
-0.07359082996845245,
-0.017928291112184525,
0.08769620954990387,
-0.041871484369039536,
0.054706934839487076,
-0.11874353140592575,
0.022122712805867195,
-0.00020615810353774577,
0.01381335686892271,
0.053916994482278824,
0.10166124254465103,
-0.11324725300073624,
0.0017622357700020075,
-0.0916932076215744,
-0.039301835000514984,
-0.12195062637329102,
0.043632328510284424,
0.15319935977458954,
-0.0337236113846302,
0.09412897378206253,
-0.10378764569759369,
-0.007176809944212437,
-0.20880119502544403,
0.00766115915030241,
-0.04477480798959732,
-0.1067800372838974,
-0.0854259729385376,
0.020869852975010872,
0.08013791590929031,
-0.02150229923427105,
0.029101761057972908,
-0.0003955483844038099,
-0.008487606421113014,
0.07200564444065094,
0.08672524988651276,
-0.030748067423701286,
0.014835173264145851,
0.18429745733737946,
0.07373128086328506,
-0.023315897211432457,
0.05097556114196777,
0.10034555196762085,
0.10094519704580307,
0.04170289263129234,
0.07883273810148239,
0.1332196742296219,
-0.09461263567209244,
0.12161438167095184,
0.06259432435035706,
-0.08829246461391449,
-0.0935894027352333,
0.07145798951387405,
-0.16559112071990967,
0.06723428517580032,
-0.050115372985601425,
-0.009046905674040318,
0.18586766719818115,
-0.10229498893022537,
-0.005100556183606386,
-0.08153348416090012,
-0.062285881489515305,
-0.1338142305612564,
-0.09814905375242233,
-0.14146341383457184,
-0.061160534620285034,
0.012804378755390644,
-0.07868786156177521,
0.03489634767174721,
0.1617690622806549,
0.030345233157277107,
-0.007267663720995188,
0.18590465188026428,
-0.16347487270832062,
0.01381315104663372,
0.037180591374635696,
0.001383344642817974,
-0.05366434529423714,
-0.02917834185063839,
-0.043149858713150024,
0.03198675066232681,
0.019355226308107376,
0.0208910945802927,
-0.005772547330707312,
-0.013889730907976627,
0.04154794663190842,
0.03230652958154678,
-0.1223493367433548,
-0.015546189621090889,
0.02740613743662834,
0.012666089460253716,
0.054252661764621735,
-0.021515648812055588,
0.06638097018003464,
-0.02936539612710476,
0.1980808824300766,
-0.08274446427822113,
-0.009670223109424114,
-0.07015778869390488,
0.1890718787908554,
-0.012734211049973965,
0.0883806124329567,
-0.0008463835692964494,
-0.05160723254084587,
0.06009732559323311,
0.14934946596622467,
0.09062901884317398,
-0.04079368710517883,
-0.038510896265506744,
-0.00997097883373499,
-0.01801079884171486,
-0.07314198464155197,
0.13503386080265045,
0.024993866682052612,
0.07634735852479935,
-0.054770562797784805,
0.06555705517530441,
-0.009625540114939213,
-0.057793013751506805,
-0.02012898214161396,
0.05917052552103996,
0.02062227949500084,
0.058293599635362625,
-0.041313450783491135,
0.06606639921665192,
0.024078737944364548,
0.08348724991083145,
0.14906810224056244,
-0.12081222981214523,
-0.14101670682430267,
0.035589415580034256,
0.0523940734565258,
-0.05836111679673195,
0.07126979529857635,
-0.06749076396226883,
0.012534043751657009,
-0.04699559882283211,
-0.009565910324454308,
-0.1409883350133896,
-0.1323915272951126,
-0.011118692345917225,
0.0912381261587143,
0.2920004427433014,
0.0011966257588937879,
0.061400406062603,
0.13613727688789368,
-0.05213532596826553,
-0.18468478322029114,
0.08548185974359512,
-0.016923891380429268,
-0.12804611027240753,
0.10286321491003036,
0.17235442996025085,
-0.02047640271484852,
0.18433202803134918,
0.012241019867360592,
-0.2258330136537552,
0.005236930213868618,
-0.016600973904132843,
-0.017113272100687027,
-0.05260747671127319,
0.03898373246192932,
-0.0601433590054512,
0.13975407183170319,
0.12396782636642456,
-0.027422403916716576,
-0.019162800163030624,
-0.04611778259277344,
0.06355956196784973,
0.031209083274006844,
0.06780038774013519,
0.009661147370934486,
-0.14219774305820465,
0.09724344313144684,
-0.11828631907701492,
-0.013502919115126133,
-0.29033321142196655,
-0.07502917945384979,
-0.04922955483198166,
-0.06340866535902023,
-0.08932983875274658,
0.10000861436128616,
0.12768836319446564,
0.04149710759520531,
-0.041628025472164154,
-0.22436483204364777,
0.025852596387267113,
0.13430507481098175,
-0.07139253616333008,
-0.11997535079717636
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# test_AsymmetricLoss_25K_bs64_P4_N1
This model is a fine-tuned version of [allenai/scibert_scivocab_uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6203
- Accuracy: 0.7448
- Precision: 0.0101
- Recall: 0.2592
- F1: 0.0194
- Hamming: 0.2552
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 40
- eval_batch_size: 40
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Hamming |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 0.6901 | 0.0 | 5 | 0.6457 | 0.6626 | 0.0099 | 0.3394 | 0.0192 | 0.3374 |
| 0.6344 | 0.0 | 10 | 0.6203 | 0.7448 | 0.0101 | 0.2592 | 0.0194 | 0.2552 |
### Framework versions
- Transformers 4.35.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.7.1
- Tokenizers 0.14.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy", "precision", "recall", "f1"], "base_model": "allenai/scibert_scivocab_uncased", "model-index": [{"name": "test_AsymmetricLoss_25K_bs64_P4_N1", "results": []}]} | text-classification | bdpc/test_AsymmetricLoss_25K_bs64_P4_N1 | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:allenai/scibert_scivocab_uncased",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:32:27+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us
| test\_AsymmetricLoss\_25K\_bs64\_P4\_N1
=======================================
This model is a fine-tuned version of allenai/scibert\_scivocab\_uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6203
* Accuracy: 0.7448
* Precision: 0.0101
* Recall: 0.2592
* F1: 0.0194
* Hamming: 0.2552
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 40
* eval\_batch\_size: 40
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 10
### Training results
### Framework versions
* Transformers 4.35.0.dev0
* Pytorch 2.0.1+cu118
* Datasets 2.7.1
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 40\n* eval\\_batch\\_size: 40\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 40\n* eval\\_batch\\_size: 40\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
61,
115,
4,
36
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 40\n* eval\\_batch\\_size: 40\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10### Training results### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
-0.11941082775592804,
0.05689161270856857,
-0.0024710167199373245,
0.12643443048000336,
0.17027325928211212,
0.0345444530248642,
0.1227070763707161,
0.12827399373054504,
-0.06712691485881805,
0.033240336924791336,
0.11953604966402054,
0.14740608632564545,
0.021579165011644363,
0.1281055510044098,
-0.051371194422245026,
-0.2940415143966675,
-0.004633181728422642,
0.017219306901097298,
-0.06507152318954468,
0.13465914130210876,
0.08588126301765442,
-0.1353035867214203,
0.0919848158955574,
-0.014839192852377892,
-0.16075195372104645,
-0.004124085884541273,
0.015248768962919712,
-0.047049906104803085,
0.15233498811721802,
0.014021219685673714,
0.12384684383869171,
0.02246846631169319,
0.11766361445188522,
-0.1779567301273346,
0.009681671857833862,
0.04676750302314758,
0.024497507140040398,
0.09262643754482269,
0.057925425469875336,
-0.005231502465903759,
0.11612623184919357,
-0.08387722074985504,
0.06262906640768051,
0.022417256608605385,
-0.13601265847682953,
-0.18222875893115997,
-0.07619091123342514,
0.0326886884868145,
0.07395829260349274,
0.08984249085187912,
-0.008051051758229733,
0.10209646075963974,
-0.093891441822052,
0.11129812896251678,
0.22811077535152435,
-0.2635839879512787,
-0.07715153694152832,
0.0062596723437309265,
0.00930891465395689,
0.09182406216859818,
-0.11135493218898773,
-0.012772854417562485,
0.0323949009180069,
0.03682177513837814,
0.12333262711763382,
-0.027334261685609818,
-0.12498282641172409,
0.02575238049030304,
-0.1400059312582016,
-0.02680302783846855,
0.09956303983926773,
0.022871151566505432,
-0.042318496853113174,
-0.03585714101791382,
-0.07302577048540115,
-0.16949069499969482,
-0.0393105074763298,
-0.013815590180456638,
0.03150599077343941,
-0.04578305035829544,
-0.07483170181512833,
-0.029444620013237,
-0.08769520372152328,
-0.0989767462015152,
-0.03359294682741165,
0.17496801912784576,
0.03958776593208313,
0.01137162372469902,
-0.006126420106738806,
0.13257160782814026,
0.01456708088517189,
-0.15358154475688934,
0.011612565256655216,
0.01073485892266035,
-0.04618196189403534,
-0.047736793756484985,
-0.05500136315822601,
-0.020882342010736465,
-0.02011677622795105,
0.1314883679151535,
-0.022660188376903534,
0.04753918573260307,
0.036841344088315964,
0.019608721137046814,
-0.10114539414644241,
0.20029771327972412,
-0.055815812200307846,
-0.02341451495885849,
-0.014462986961007118,
0.09352404624223709,
0.0038181766867637634,
-0.033196549862623215,
-0.10503692924976349,
-0.01127643696963787,
0.10295660048723221,
0.02512827143073082,
-0.08098543435335159,
0.06994519382715225,
-0.0442170687019825,
-0.03680218383669853,
0.01224500685930252,
-0.09943114966154099,
0.036005303263664246,
0.020423641428351402,
-0.10004527121782303,
-0.011710675433278084,
0.026532161980867386,
0.007409719284623861,
-0.014904431998729706,
0.1325729638338089,
-0.0962897539138794,
0.037460871040821075,
-0.09163089841604233,
-0.12124553322792053,
-0.0081277284771204,
-0.09890499711036682,
0.0035750640090554953,
-0.0789196789264679,
-0.18824484944343567,
-0.02723303623497486,
0.042748529464006424,
-0.0484558641910553,
-0.032766759395599365,
-0.06847461313009262,
-0.0638132095336914,
0.024265918880701065,
-0.005678042769432068,
0.12003441154956818,
-0.05680233985185623,
0.12103956192731857,
0.03955712914466858,
0.0687522441148758,
-0.020830105990171432,
0.05860484018921852,
-0.10721904039382935,
0.009518629871308804,
-0.18861989676952362,
0.06542586535215378,
-0.055847980082035065,
0.04832913726568222,
-0.08050663769245148,
-0.10648266971111298,
0.022618446499109268,
0.0008140904828906059,
0.08971016854047775,
0.14277146756649017,
-0.1803087294101715,
-0.07825534045696259,
0.1449977457523346,
-0.0692165419459343,
-0.09332185983657837,
0.12101507931947708,
-0.06636898964643478,
0.0404399074614048,
0.06498192250728607,
0.15949009358882904,
0.08415135741233826,
-0.07419700175523758,
0.02308039925992489,
-0.00907408818602562,
0.08309328556060791,
0.00025823793839663267,
0.07009844481945038,
0.012607919983565807,
0.0004221163981128484,
0.024573741480708122,
-0.06655017286539078,
0.043409865349531174,
-0.10868864506483078,
-0.09572090208530426,
-0.03206147998571396,
-0.10295705497264862,
0.09153115749359131,
0.07322042435407639,
0.07614745199680328,
-0.08826982975006104,
-0.08965049684047699,
0.054531678557395935,
0.09892402589321136,
-0.06331802159547806,
0.019741399213671684,
-0.06115872785449028,
0.056163761764764786,
-0.018993159756064415,
-0.022395169362425804,
-0.18869908154010773,
-0.043161917477846146,
0.018316196277737617,
0.047765351831912994,
0.02319898083806038,
0.001203516498208046,
0.08814572542905807,
0.08726538717746735,
-0.062226444482803345,
-0.0365716889500618,
-0.046432968229055405,
-0.007808587048202753,
-0.1394963562488556,
-0.2052653282880783,
-0.06703058630228043,
-0.018573986366391182,
0.1396852284669876,
-0.2171928584575653,
0.037465352565050125,
-0.008897167630493641,
0.08425871282815933,
0.024791227653622627,
-0.024280564859509468,
-0.034077439457178116,
0.07526900619268417,
-0.034903742372989655,
-0.055090632289648056,
0.0670602023601532,
-0.013024544343352318,
-0.08498640358448029,
-0.05271292105317116,
-0.1203758642077446,
0.13664975762367249,
0.10524983704090118,
-0.09785842150449753,
-0.09588475525379181,
-0.00042442663107067347,
-0.05249154567718506,
-0.03043815866112709,
-0.050303876399993896,
0.024082617834210396,
0.1709488183259964,
0.00887475535273552,
0.14942754805088043,
-0.05497026816010475,
-0.03879962116479874,
0.017505839467048645,
-0.011985646560788155,
0.044180892407894135,
0.13263346254825592,
0.09654363244771957,
-0.08034911751747131,
0.12682129442691803,
0.12050750106573105,
-0.08391153067350388,
0.14659741520881653,
-0.024326195940375328,
-0.08145109564065933,
-0.013740350492298603,
-0.02581750974059105,
-0.0002363555831834674,
0.10515327751636505,
-0.1216050535440445,
-0.021122457459568977,
0.011681010015308857,
0.013074880465865135,
-0.004846036434173584,
-0.20881278812885284,
-0.0366685688495636,
0.04456609860062599,
-0.05043037608265877,
-0.03437589108943939,
-0.011092466302216053,
0.007840475998818874,
0.11810828745365143,
0.010891380719840527,
-0.09057675302028656,
0.013287216424942017,
-0.0032869097776710987,
-0.06825215369462967,
0.20857064425945282,
-0.06972363591194153,
-0.11458795517683029,
-0.12076853960752487,
-0.05864904075860977,
-0.05884627252817154,
0.02805626019835472,
0.03582477197051048,
-0.09472296386957169,
-0.006726567167788744,
-0.05719101056456566,
0.03207826241850853,
0.007422778755426407,
0.03952089324593544,
-0.02646268531680107,
-0.020159916952252388,
0.05684160441160202,
-0.09083171933889389,
-0.005380300339311361,
-0.05940210819244385,
-0.08008269220590591,
0.033961039036512375,
0.03287670761346817,
0.11700290441513062,
0.15422740578651428,
-0.018841644749045372,
0.01430301833897829,
-0.024687238037586212,
0.23196370899677277,
-0.07571941614151001,
-0.007187602110207081,
0.14114345610141754,
-0.02880832739174366,
0.05052047222852707,
0.12935848534107208,
0.06680785864591599,
-0.07974467426538467,
0.008131050504744053,
0.04277414083480835,
-0.033677540719509125,
-0.2110627144575119,
-0.03450912982225418,
-0.05298471078276634,
0.005995623767375946,
0.10154087841510773,
0.012234868481755257,
0.03257445991039276,
0.07184985280036926,
0.014652218669652939,
0.04571181908249855,
-0.028516732156276703,
0.06015104800462723,
0.0928637906908989,
0.043075405061244965,
0.1351725310087204,
-0.024904213845729828,
-0.06950903683900833,
0.038976039737463,
-0.02205088548362255,
0.19929058849811554,
-0.0232085008174181,
0.11824875324964523,
0.02222968079149723,
0.15645602345466614,
-0.012928477488458157,
0.0851994976401329,
0.00813725870102644,
-0.038219548761844635,
-0.009534357115626335,
-0.0434056855738163,
-0.06025325879454613,
0.024733634665608406,
-0.05706885829567909,
0.06612584739923477,
-0.1562170833349228,
0.011451841332018375,
0.05596570298075676,
0.27542731165885925,
0.06483660638332367,
-0.32468995451927185,
-0.11161945015192032,
0.005418392363935709,
-0.0347122959792614,
-0.029549939557909966,
0.009482262656092644,
0.1243046224117279,
-0.1073383167386055,
0.025322649627923965,
-0.07278795540332794,
0.09035878628492355,
-0.06243490055203438,
0.05531174689531326,
0.07820191234350204,
0.09349574148654938,
-0.010707205161452293,
0.07210732251405716,
-0.27364400029182434,
0.2782500684261322,
0.013020721264183521,
0.06469059735536575,
-0.0711253434419632,
-0.009829212911427021,
0.053608331829309464,
0.08993028849363327,
0.06341760605573654,
-0.015769748017191887,
-0.03299962729215622,
-0.2364710122346878,
-0.0687323734164238,
0.01994018629193306,
0.11703118681907654,
-0.07204641401767731,
0.11341281235218048,
-0.04362628608942032,
-0.0018420206615701318,
0.056374527513980865,
-0.03873105347156525,
-0.05155652016401291,
-0.07939059287309647,
0.004147980362176895,
0.001652750768698752,
-0.00426686555147171,
-0.05166452378034592,
-0.12035423517227173,
-0.0734204426407814,
0.15181703865528107,
-0.03581453114748001,
-0.03502956032752991,
-0.1346813589334488,
0.08224271982908249,
0.0966954156756401,
-0.09119637310504913,
0.024425305426120758,
0.014696158468723297,
0.0754246860742569,
0.035809118300676346,
-0.05702006071805954,
0.11207885295152664,
-0.0671042874455452,
-0.2035500556230545,
-0.06299130618572235,
0.11304900050163269,
0.048077814280986786,
0.07022537291049957,
-0.02079686149954796,
0.020786024630069733,
-0.023686600849032402,
-0.08414462208747864,
0.02895437180995941,
-0.013073107227683067,
0.04547605663537979,
0.039673976600170135,
-0.06286906450986862,
0.04863113909959793,
-0.07041546702384949,
-0.005239865742623806,
0.16169345378875732,
0.2463109791278839,
-0.0989377573132515,
0.02836993709206581,
0.03454482927918434,
-0.06436572223901749,
-0.20104436576366425,
0.037525638937950134,
0.07090476900339127,
0.021690910682082176,
0.04903193935751915,
-0.20309439301490784,
0.08994486927986145,
0.09188999980688095,
-0.01075468398630619,
0.0898975282907486,
-0.3080082833766937,
-0.12531070411205292,
0.1143195852637291,
0.12457635998725891,
0.11185476928949356,
-0.14193686842918396,
-0.02192842587828636,
-0.007527681067585945,
-0.08839596062898636,
0.08642870932817459,
-0.0765356793999672,
0.12811842560768127,
-0.026245569810271263,
0.0908486545085907,
0.03430899232625961,
-0.04449109360575676,
0.11854904890060425,
0.016400443390011787,
0.09506392478942871,
-0.05001704767346382,
-0.051701150834560394,
0.020240020006895065,
-0.050003282725811005,
0.008656074292957783,
-0.07730698585510254,
0.04309589043259621,
-0.11217888444662094,
-0.01756550744175911,
-0.09370698779821396,
0.02412417158484459,
-0.03759457543492317,
-0.0701608806848526,
-0.014783675782382488,
0.03771252930164337,
0.044160161167383194,
-0.0036420433316379786,
0.12435220927000046,
-0.017596255987882614,
0.15743495523929596,
0.10123665630817413,
0.08845678716897964,
-0.01815642975270748,
-0.033200837671756744,
-0.004137715324759483,
-0.004919292405247688,
0.059770435094833374,
-0.12345252931118011,
0.03433575481176376,
0.1461084932088852,
0.03646758571267128,
0.13137713074684143,
0.08424169570207596,
-0.008558755740523338,
0.007576341275125742,
0.05163588747382164,
-0.1768711507320404,
-0.06006920337677002,
-0.020819179713726044,
-0.0635937750339508,
-0.13244237005710602,
0.05507943034172058,
0.11136427521705627,
-0.06426382809877396,
-0.014217616058886051,
-0.01164198387414217,
0.010686593130230904,
-0.0422467403113842,
0.21675266325473785,
0.06146189942955971,
0.061332400888204575,
-0.10426873713731766,
0.06286017596721649,
0.049955449998378754,
-0.05537387728691101,
0.0034978301264345646,
0.10051389783620834,
-0.08901343494653702,
-0.03690258413553238,
0.05758373439311981,
0.15156544744968414,
-0.06722260266542435,
-0.016543397679924965,
-0.15841253101825714,
-0.11436798423528671,
0.06784309446811676,
0.17373153567314148,
0.1012011244893074,
0.02012472227215767,
-0.07093483209609985,
0.0276210717856884,
-0.11946321278810501,
0.09682004153728485,
0.05734072998166084,
0.07279308885335922,
-0.14781832695007324,
0.19293612241744995,
-0.020270878449082375,
0.053248655050992966,
-0.02808334492146969,
0.015606974251568317,
-0.11160936951637268,
0.013042286038398743,
-0.12391256541013718,
-0.04402898624539375,
-0.009007132612168789,
0.0018418030813336372,
-0.01091555692255497,
-0.07351028174161911,
-0.04429762065410614,
0.0007218311657197773,
-0.11689268052577972,
-0.025169724598526955,
0.02289270982146263,
0.036684468388557434,
-0.1254495233297348,
-0.03941281512379646,
0.034817226231098175,
-0.08433393388986588,
0.08067414164543152,
0.06069564074277878,
0.01994679868221283,
0.05169682577252388,
-0.13104334473609924,
0.010100162588059902,
0.03205852583050728,
0.0062888567335903645,
0.06151974946260452,
-0.06480182707309723,
-0.003064954187721014,
-0.03933309391140938,
0.062351044267416,
0.020387794822454453,
0.104458749294281,
-0.12107010930776596,
0.029471557587385178,
-0.005121536087244749,
-0.07548190653324127,
-0.06528885662555695,
0.05888805538415909,
0.06980879604816437,
0.02881932072341442,
0.16804076731204987,
-0.08921903371810913,
0.058024175465106964,
-0.21902844309806824,
-0.0011340922210365534,
-0.008171962574124336,
-0.1193128451704979,
-0.12852510809898376,
-0.08211040496826172,
0.08542174845933914,
-0.051245298236608505,
0.08564119040966034,
0.018790360540151596,
0.1012205258011818,
0.0255010724067688,
-0.02900291420519352,
0.030534274876117706,
0.03000527434051037,
0.17774829268455505,
0.047429803758859634,
-0.04888748377561569,
0.0708225816488266,
0.05078502371907234,
0.08916313946247101,
0.14640186727046967,
0.2191113531589508,
0.1241406574845314,
-0.013185635209083557,
0.08806779235601425,
0.043789442628622055,
-0.0700586587190628,
-0.17052699625492096,
-0.0022810245864093304,
-0.04063553735613823,
0.08486941456794739,
-0.04465412721037865,
0.21721401810646057,
0.053341031074523926,
-0.16820265352725983,
0.04762478545308113,
-0.0483129657804966,
-0.1017690896987915,
-0.12305083870887756,
-0.021088246256113052,
-0.07332833856344223,
-0.1408700793981552,
-0.002929703099653125,
-0.10740658640861511,
0.03730329871177673,
0.10275425761938095,
0.009882263839244843,
-0.013154194690287113,
0.17029641568660736,
0.01783895678818226,
0.0384795144200325,
0.07409556210041046,
0.010711281560361385,
-0.008429507724940777,
-0.08308129757642746,
-0.08936379849910736,
-0.005937411915510893,
-0.015932176262140274,
0.02887357957661152,
-0.07782825827598572,
-0.06474451720714569,
0.02858194336295128,
-0.007592073641717434,
-0.11373495310544968,
0.016776282340288162,
0.011319180019199848,
0.07081134617328644,
0.058975700289011,
0.009538812562823296,
0.010830020532011986,
-0.02298208698630333,
0.24963462352752686,
-0.09587863832712173,
-0.04916398599743843,
-0.11041835695505142,
0.28982603549957275,
0.03650680184364319,
0.0005231188260950148,
0.023580636829137802,
-0.07616019994020462,
-0.005048795603215694,
0.23171764612197876,
0.18280664086341858,
-0.11613310873508453,
-0.006419157609343529,
0.0030950854998081923,
-0.006066089496016502,
-0.005403527989983559,
0.1017594262957573,
0.09589381515979767,
0.04005729779601097,
-0.09853312373161316,
-0.02876577526330948,
-0.026272786781191826,
-0.0347575917840004,
-0.0317276306450367,
0.08160807192325592,
0.03342248871922493,
0.01781618595123291,
-0.058720432221889496,
0.0502341091632843,
-0.07154948264360428,
-0.1068183109164238,
0.06408531218767166,
-0.21247076988220215,
-0.17123322188854218,
-0.032001785933971405,
0.06596580892801285,
0.024079708382487297,
0.07594675570726395,
-0.02429892309010029,
-0.011666842736303806,
0.08678372949361801,
-0.012133315205574036,
-0.06070742756128311,
-0.10849189013242722,
0.11147274821996689,
-0.09675052016973495,
0.2028314769268036,
-0.04950588941574097,
0.04335247352719307,
0.11512768268585205,
0.049350377172231674,
-0.06858568638563156,
0.04235092177987099,
0.060175977647304535,
-0.09834157675504684,
0.021339306607842445,
0.13644739985466003,
-0.045377474278211594,
0.07564523071050644,
0.03614232689142227,
-0.15671567618846893,
0.021805290132761,
-0.06909941136837006,
-0.07450833171606064,
-0.02752526104450226,
-0.044688157737255096,
-0.03293673321604729,
0.13862518966197968,
0.25002068281173706,
-0.03095857799053192,
0.011795494705438614,
-0.07558390498161316,
0.005595171824097633,
0.059709079563617706,
0.051001425832509995,
-0.06742645055055618,
-0.2326880246400833,
0.0024419219698756933,
0.06404910236597061,
-0.012952064163982868,
-0.2339019775390625,
-0.09254217892885208,
0.00920872576534748,
-0.052385762333869934,
-0.08577456325292587,
0.09313642233610153,
0.06112116202712059,
0.043386124074459076,
-0.04213502258062363,
-0.06246321275830269,
-0.06773554533720016,
0.16411122679710388,
-0.15736673772335052,
-0.08144693076610565
] |
null | null | transformers | <p>
This is the organization grouping all the models and datasets used in the <a href="https://huggingface.co/course/chapter1" class="underline">Hugging Face course</a>.
</p> | {} | token-classification | ramin153/bert-per-ai-token-classification | [
"transformers",
"safetensors",
"bert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:35:04+00:00 | [] | [] | TAGS
#transformers #safetensors #bert #token-classification #autotrain_compatible #endpoints_compatible #region-us
| <p>
This is the organization grouping all the models and datasets used in the <a href="URL class="underline">Hugging Face course</a>.
</p> | [] | [
"TAGS\n#transformers #safetensors #bert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
38
] | [
"passage: TAGS\n#transformers #safetensors #bert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] | [
-0.04248703643679619,
0.05589418113231659,
-0.007779605221003294,
0.005975014995783567,
0.14228811860084534,
-0.003972362261265516,
0.10242500901222229,
0.058972541242837906,
0.04169780761003494,
0.002145511796697974,
0.14366257190704346,
0.21196110546588898,
-0.05068425461649895,
0.16102389991283417,
-0.1237838864326477,
-0.2217583954334259,
0.12375142425298691,
0.03026631288230419,
-0.03727564960718155,
0.0902625024318695,
0.09037412703037262,
-0.0967470332980156,
0.07488518953323364,
-0.05090109631419182,
-0.1376856565475464,
0.049563270062208176,
0.0839284285902977,
-0.14270757138729095,
0.10078749060630798,
0.04070557653903961,
0.22726549208164215,
0.04865041375160217,
-0.032048214226961136,
-0.15798932313919067,
0.01838853396475315,
0.030550576746463776,
-0.05543762072920799,
0.040641892701387405,
0.07014014571905136,
-0.08013498038053513,
-0.10443559288978577,
0.04803421348333359,
0.054942306131124496,
0.06037287414073944,
-0.13030794262886047,
-0.12795723974704742,
-0.02580907754600048,
-0.00901036150753498,
0.10975174605846405,
0.030131513252854347,
0.020070532336831093,
0.20710545778274536,
-0.12010206282138824,
0.10003583878278732,
0.08605201542377472,
-0.30016574263572693,
0.0023791054263710976,
0.1332450956106186,
0.0421329066157341,
0.00451376847922802,
-0.05640370026230812,
0.05039897933602333,
0.06639780104160309,
-0.02077716588973999,
0.057250965386629105,
-0.07618380337953568,
-0.0688806027173996,
0.0364333912730217,
-0.08905501663684845,
0.0062831491231918335,
0.18765956163406372,
-0.021330108866095543,
0.02292724698781967,
-0.031017953529953957,
-0.10920095443725586,
-0.050521235913038254,
-0.018926039338111877,
-0.04628022015094757,
-0.01644023507833481,
0.06491198390722275,
0.005576660856604576,
0.07540241628885269,
-0.10437337309122086,
0.005423572845757008,
-0.19631293416023254,
0.25547268986701965,
0.004369556903839111,
0.056324079632759094,
-0.16151459515094757,
0.027957787737250328,
-0.020141342654824257,
-0.08862339705228806,
0.027558565139770508,
-0.11224106699228287,
-0.022180864587426186,
-0.07589983940124512,
-0.01834099553525448,
-0.012282801792025566,
0.11337893456220627,
0.19191771745681763,
0.0675460696220398,
0.06007479503750801,
-0.0584198534488678,
0.053691864013671875,
0.02209564670920372,
0.10508986562490463,
0.025074759498238564,
-0.05211947485804558,
0.05784158408641815,
-0.11048170179128647,
0.003547482658177614,
-0.05090351775288582,
-0.10990578681230545,
0.005735517013818026,
0.09707682579755783,
0.11265046149492264,
-0.024116521701216698,
0.07747364044189453,
-0.07780924439430237,
0.014841865748167038,
0.12138054519891739,
-0.07021868973970413,
0.02989245019853115,
0.00856633111834526,
0.05652039498090744,
0.018842920660972595,
-0.044785089790821075,
0.025152644142508507,
0.014738974161446095,
0.13217131793498993,
-0.07575051486492157,
-0.041517022997140884,
-0.036889541894197464,
-0.09067166596651077,
0.03520917147397995,
-0.09280415624380112,
0.06292741745710373,
-0.20217235386371613,
-0.0957656279206276,
0.026789585128426552,
0.03218908607959747,
0.0117246238514781,
0.02809700183570385,
-0.011423136107623577,
-0.013438625261187553,
0.01772036962211132,
-0.07518564164638519,
-0.1284305602312088,
-0.07292903959751129,
0.04700606316328049,
-0.008660794235765934,
0.04552876204252243,
-0.08779054880142212,
0.024526922032237053,
-0.12770192325115204,
0.03937569260597229,
-0.1407272070646286,
-0.017806483432650566,
-0.09429708868265152,
0.21487095952033997,
-0.026979006826877594,
-0.0039542424492537975,
-0.05372912436723709,
0.04337071254849434,
-0.046930424869060516,
0.14199991524219513,
-0.08109129965305328,
-0.0740201324224472,
0.19311782717704773,
-0.14141198992729187,
-0.16029872000217438,
0.07878158986568451,
-0.016013095155358315,
0.010585898533463478,
0.09099086374044418,
0.1569284200668335,
0.0967722162604332,
-0.04825982451438904,
0.05445580184459686,
0.093350350856781,
-0.13786673545837402,
-0.10431396216154099,
-0.0108359195291996,
0.011295683681964874,
-0.1734706312417984,
0.03540603443980217,
0.08951817452907562,
0.07590971887111664,
-0.05806158855557442,
-0.03668428584933281,
-0.028377538546919823,
-0.0333857461810112,
0.10871516913175583,
0.033111315220594406,
0.07995153963565826,
-0.10625143349170685,
0.0049483636394143105,
0.026904085651040077,
-0.008869858458638191,
0.018392911180853844,
-0.013284390792250633,
-0.12738414108753204,
0.06627689301967621,
-0.07235392928123474,
0.011106033809483051,
-0.14846011996269226,
-0.1758899986743927,
0.031473614275455475,
0.06760512292385101,
-0.062206774950027466,
0.10246746987104416,
0.08600069582462311,
-0.01968539133667946,
-0.0042062439024448395,
-0.05479047819972038,
0.22150775790214539,
0.08420957624912262,
-0.06413225084543228,
-0.08355055749416351,
0.05786203220486641,
-0.09641031920909882,
-0.0341787114739418,
-0.09871548414230347,
0.012943299487233162,
0.11665961891412735,
0.15350371599197388,
0.06456589698791504,
0.0696137323975563,
-0.041155457496643066,
0.03924637660384178,
-0.08713148534297943,
-0.01474764198064804,
0.07326602935791016,
-0.00799576099961996,
-0.07120675593614578,
0.12947483360767365,
-0.1878480613231659,
0.4071597456932068,
0.17792394757270813,
-0.2231115847826004,
-0.015673266723752022,
-0.02353784628212452,
0.013852287083864212,
0.01458478718996048,
0.02114677056670189,
0.008185924030840397,
-0.04623984545469284,
-0.01116886269301176,
0.15580423176288605,
-0.02335749752819538,
-0.03762970119714737,
0.020537160336971283,
-0.07643087953329086,
-0.06473607569932938,
0.037336595356464386,
-0.0005800596554763615,
-0.23334620893001556,
0.19579726457595825,
0.2418530285358429,
0.03793707489967346,
0.1003696545958519,
-0.020671632140874863,
0.04212344437837601,
0.030145857483148575,
0.0025179446674883366,
0.009592757560312748,
-0.0033465144224464893,
-0.1111956313252449,
-0.03435896337032318,
0.055470626801252365,
0.0506337471306324,
0.01878334768116474,
-0.1151294857263565,
-0.028023963794112206,
0.03162233158946037,
0.04766541346907616,
0.023509476333856583,
0.06522484123706818,
0.017285961657762527,
0.08301433175802231,
-0.002817407948896289,
-0.12916651368141174,
0.12738370895385742,
-0.013581094332039356,
-0.08319280296564102,
0.1755266934633255,
-0.13657599687576294,
-0.27213791012763977,
-0.14537270367145538,
-0.20545026659965515,
-0.010569100268185139,
0.08485230058431625,
0.060311153531074524,
-0.11089320480823517,
-0.0932653546333313,
0.019974252209067345,
-0.016098445281386375,
-0.016738733276724815,
0.08151649683713913,
-0.032628268003463745,
0.07153298705816269,
-0.005337260197848082,
-0.06385580450296402,
-0.06148368492722511,
-0.008247519843280315,
-0.022365592420101166,
0.12853574752807617,
-0.09004315733909607,
0.08710922300815582,
0.15753190219402313,
-0.018444916233420372,
0.02206812985241413,
-0.037456125020980835,
0.15255513787269592,
-0.05327727645635605,
-0.006418767385184765,
0.13222435116767883,
-0.09725310653448105,
0.06480474770069122,
0.21019509434700012,
0.027446545660495758,
-0.09510031342506409,
0.04823174327611923,
-0.03686126321554184,
-0.10106606781482697,
-0.1765277236700058,
-0.10977159440517426,
-0.06943667680025101,
0.06573875993490219,
0.060382597148418427,
0.07480230182409286,
0.14258147776126862,
0.11488796770572662,
0.05105665698647499,
0.00031207146821543574,
0.007686965633183718,
0.08543910086154938,
0.16086596250534058,
0.010181078687310219,
0.15448200702667236,
-0.06320637464523315,
-0.15152716636657715,
0.0673166960477829,
0.046498384326696396,
0.1000806912779808,
0.15565045177936554,
-0.0019847217481583357,
0.006683822255581617,
0.12311019003391266,
0.18719413876533508,
0.15723173320293427,
0.03437788039445877,
-0.06865723431110382,
0.0029426799155771732,
0.009107356891036034,
-0.055845510214567184,
0.04270978644490242,
0.013546353206038475,
-0.09429333359003067,
-0.03421759977936745,
-0.10163317620754242,
0.09646040946245193,
0.08521172404289246,
0.07374359667301178,
-0.25759539008140564,
0.014967075549066067,
0.11990635842084885,
-0.01301242783665657,
-0.0725875273346901,
0.10198672860860825,
0.001322394353337586,
-0.08000880479812622,
0.08456961810588837,
-0.05472167953848839,
0.0861881822347641,
-0.02982959896326065,
0.06943631917238235,
-0.0308320801705122,
-0.08813149482011795,
0.004115123301744461,
0.07443053275346756,
-0.2364954650402069,
0.2137715071439743,
0.014133553951978683,
-0.03210868313908577,
-0.05855029076337814,
0.0017328662797808647,
0.03289517015218735,
0.22925403714179993,
0.11910178512334824,
-0.004875867627561092,
-0.2028963714838028,
-0.21543507277965546,
-0.02251306176185608,
0.015548355877399445,
0.11110672354698181,
-0.017393752932548523,
0.0066991569474339485,
-0.06346762925386429,
-0.027044590562582016,
0.014170479029417038,
-0.05648541450500488,
-0.027044810354709625,
-0.11455830186605453,
0.02404099330306053,
0.08213625848293304,
0.10321823507547379,
-0.0322553813457489,
-0.006690380163490772,
-0.13555586338043213,
0.19427260756492615,
-0.10506727546453476,
-0.02211756445467472,
-0.12999799847602844,
-0.1257033348083496,
0.037862326949834824,
-0.054595816880464554,
0.0813692957162857,
-0.06786556541919708,
0.05647601559758186,
-0.032622143626213074,
-0.17857344448566437,
0.17106378078460693,
-0.1368805319070816,
-0.033553916960954666,
-0.09683609008789062,
0.09809725731611252,
-0.07649987936019897,
-0.02115027606487274,
0.035488687455654144,
0.045699577778577805,
-0.025796666741371155,
-0.07393699884414673,
0.004503539763391018,
-0.020583797246217728,
0.021615460515022278,
0.03327745944261551,
-0.06604872643947601,
-0.09664639830589294,
0.004262495320290327,
0.024487581104040146,
0.1920362263917923,
0.2345263659954071,
-0.07804352045059204,
0.08701582998037338,
0.19322077929973602,
-0.029848413541913033,
-0.35317757725715637,
-0.04720374569296837,
-0.15904401242733002,
-0.04884794354438782,
-0.017418434843420982,
-0.05742396041750908,
0.17538832128047943,
0.03701118007302284,
-0.06446586549282074,
0.07010817527770996,
-0.07398507744073868,
-0.09564157575368881,
0.23712679743766785,
0.028098827227950096,
0.3982909023761749,
-0.12582652270793915,
-0.08233372122049332,
-0.06874469667673111,
-0.11603511869907379,
0.10376973450183868,
-0.0502709336578846,
0.029189927503466606,
0.027324622496962547,
-0.029997961595654488,
0.036156706511974335,
-0.06384582817554474,
0.11169034987688065,
-0.014665495604276657,
0.09535106271505356,
-0.12151399999856949,
-0.09317445009946823,
0.04703345149755478,
-0.05297115445137024,
0.006204760633409023,
0.031430214643478394,
0.02282123826444149,
-0.04149331524968147,
-0.03267610818147659,
-0.015677182003855705,
0.0870257243514061,
0.04499296471476555,
-0.07982941716909409,
-0.010582052171230316,
-0.016291562467813492,
-0.0051926844753324986,
-0.02479272149503231,
0.3110242486000061,
-0.007406890392303467,
0.16920243203639984,
0.14479538798332214,
0.1227525994181633,
-0.1786251813173294,
0.0742863342165947,
-0.018754228949546814,
-0.0902266651391983,
0.10643020272254944,
-0.054756272584199905,
0.11405400186777115,
0.10466306656599045,
-0.06519713997840881,
0.06835811585187912,
0.11330746114253998,
0.05208824202418327,
-0.0391228012740612,
0.16810867190361023,
-0.19966308772563934,
-0.01391957513988018,
-0.010774554684758186,
0.008383522741496563,
0.0647009015083313,
0.14355064928531647,
0.12510931491851807,
0.029575174674391747,
-0.004240388050675392,
-0.0018886930774897337,
-0.012920829467475414,
-0.03934648633003235,
0.0804961621761322,
0.05934827774763107,
0.050186578184366226,
-0.1136140450835228,
0.057620059698820114,
0.017209572717547417,
-0.14789186418056488,
-0.021266033872961998,
0.02004818618297577,
-0.16428327560424805,
-0.10287714004516602,
-0.0028313170187175274,
0.17026755213737488,
-0.09230140596628189,
-0.08462770283222198,
-0.046777766197919846,
-0.14996623992919922,
0.031738992780447006,
0.23685237765312195,
0.11397246271371841,
0.12664268910884857,
0.005291045643389225,
-0.01534995250403881,
-0.03306223452091217,
0.002788257086649537,
-0.015806084498763084,
0.08405007421970367,
-0.21374893188476562,
-0.022255321964621544,
-0.0444447286427021,
0.08087678998708725,
-0.11526697874069214,
-0.035292718559503555,
-0.1953490972518921,
0.02621348761022091,
-0.10191776603460312,
-0.04322722554206848,
-0.09645889699459076,
-0.005047681741416454,
0.033232349902391434,
-0.07502094656229019,
-0.042854707688093185,
-0.037977345287799835,
-0.09927480667829514,
0.06758632510900497,
0.027911081910133362,
0.025500277057290077,
-0.0584266297519207,
-0.05582915619015694,
0.08745945990085602,
-0.04555072262883186,
0.07872288674116135,
0.07176758348941803,
-0.07040559500455856,
0.08037012070417404,
-0.16561786830425262,
-0.08836590498685837,
0.14328888058662415,
-0.014448205940425396,
0.10458365082740784,
0.04278576746582985,
0.012972638942301273,
0.09236250817775726,
0.013338957913219929,
0.050441525876522064,
0.04269738122820854,
-0.09647955000400543,
0.03892875462770462,
-0.036136262118816376,
-0.17196717858314514,
-0.022479424253106117,
-0.0964113399386406,
0.14392651617527008,
-0.05642353743314743,
0.18677447736263275,
-0.03892840817570686,
0.04574975371360779,
-0.030985528603196144,
0.0008288167882710695,
-0.0074818432331085205,
-0.21183249354362488,
-0.057245947420597076,
-0.0327208936214447,
-0.005713791586458683,
-0.01742640510201454,
0.26248180866241455,
0.01002318225800991,
0.016205715015530586,
0.06670437753200531,
-0.0015625158557668328,
0.023625299334526062,
0.07047376036643982,
0.20619456470012665,
0.10225971788167953,
-0.0558752603828907,
-0.10657555609941483,
0.05280686169862747,
0.03873717039823532,
-0.12417341768741608,
0.04345915839076042,
0.08259660750627518,
-0.11612381041049957,
0.08895887434482574,
-0.006227423902601004,
0.014440690167248249,
-0.13037075102329254,
-0.14065462350845337,
-0.10377907752990723,
0.037035517394542694,
0.04558495059609413,
0.01992466114461422,
0.15838991105556488,
-0.017755374312400818,
0.0005363657837733626,
-0.06464708596467972,
-0.0276784785091877,
-0.2104630172252655,
-0.10185996443033218,
-0.12809935212135315,
-0.12747614085674286,
0.017324939370155334,
-0.04838615283370018,
-0.035395946353673935,
0.10606566071510315,
0.05825277045369148,
-0.00012305485142860562,
0.1508481204509735,
-0.03321610391139984,
0.02549579367041588,
0.024897277355194092,
-0.015027754008769989,
-0.01435685995966196,
0.02383086271584034,
-0.06294961273670197,
-0.14404617249965668,
-0.019336553290486336,
-0.07475697249174118,
0.00295816152356565,
-0.04009903967380524,
0.031907904893159866,
-0.10784431546926498,
-0.10122925043106079,
-0.030943909659981728,
0.06186654418706894,
-0.1128392443060875,
0.026951849460601807,
-0.002704957965761423,
0.015154106542468071,
0.05241509899497032,
0.16758757829666138,
-0.07142692804336548,
-0.06930702179670334,
-0.07240313291549683,
0.2321675419807434,
0.03879135102033615,
0.17652858793735504,
-0.030283603817224503,
-0.006728596519678831,
-0.046735428273677826,
0.2566283941268921,
0.22445161640644073,
-0.014711437746882439,
0.06091969460248947,
-0.01156927365809679,
0.014876595698297024,
0.0399865061044693,
0.13682349026203156,
0.052376314997673035,
0.19131343066692352,
-0.0437571220099926,
-0.034311480820178986,
0.00253679184243083,
-0.02890833280980587,
-0.10646890103816986,
0.062186140567064285,
0.021294180303812027,
0.002843433991074562,
-0.09607284516096115,
0.0818561539053917,
-0.10984527319669724,
0.13733768463134766,
0.058567944914102554,
-0.14357605576515198,
-0.03739192336797714,
-0.008898066356778145,
0.1631494015455246,
-0.05530363693833351,
0.06194930523633957,
-0.03368508070707321,
-0.10591402649879456,
-0.01632850430905819,
0.004050074610859156,
-0.16775666177272797,
-0.04080525040626526,
0.02856416627764702,
0.019681621342897415,
0.12473393976688385,
-0.009607510641217232,
-0.0011406916892156005,
0.09827926009893417,
0.023571781814098358,
-0.07155100256204605,
0.07564688473939896,
0.015191646292805672,
-0.08697742223739624,
-0.01613404043018818,
-0.012344286777079105,
-0.007591708097606897,
0.03513847663998604,
0.03626297786831856,
-0.22153131663799286,
0.039279431104660034,
-0.08401171118021011,
-0.07506973296403885,
-0.03340177237987518,
0.018819715827703476,
-0.027006452903151512,
0.0784551203250885,
0.02937065064907074,
0.003759315237402916,
0.007825054228305817,
-0.023880353197455406,
0.028962627053260803,
0.023775527253746986,
-0.09759825468063354,
-0.0924648642539978,
-0.12762923538684845,
-0.030340898782014847,
0.15472471714019775,
0.00037486106157302856,
-0.1261584609746933,
-0.033518072217702866,
-0.09125321358442307,
0.05063280463218689,
-0.15126559138298035,
0.06910517066717148,
0.1065027266740799,
0.054379455745220184,
-0.030121436342597008,
-0.13225363194942474,
0.02805229276418686,
0.07458813488483429,
-0.07197751104831696,
-0.12127362191677094
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | bitsoko/gumzo-tiny-00 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T16:35:54+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
59,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.049007222056388855,
0.16460949182510376,
-0.005271392408758402,
0.021910345181822777,
0.09685911983251572,
0.01403510570526123,
0.07018975168466568,
0.11002060770988464,
-0.02425350993871689,
0.11399492621421814,
0.03344893455505371,
0.09780009090900421,
0.11368958652019501,
0.1498505026102066,
-0.002398149576038122,
-0.23227156698703766,
0.04924226179718971,
-0.1249755248427391,
-0.03746527433395386,
0.1159956082701683,
0.15001481771469116,
-0.10170940309762955,
0.07611104100942612,
-0.029819702729582787,
-0.008722295984625816,
-0.032589927315711975,
-0.056551046669483185,
-0.04997202008962631,
0.051094699651002884,
0.07382578402757645,
0.06793182343244553,
0.004094683099538088,
0.09450557827949524,
-0.2669448256492615,
0.0197003111243248,
0.0730973482131958,
-0.002068581758067012,
0.07547242939472198,
0.054895199835300446,
-0.07525460422039032,
0.09282654523849487,
-0.0507965162396431,
0.1469351053237915,
0.08020289987325668,
-0.09152709692716599,
-0.19188682734966278,
-0.0887833908200264,
0.10164182633161545,
0.18469172716140747,
0.045696184039115906,
-0.022488808259367943,
0.09940612316131592,
-0.08621317893266678,
0.011039474047720432,
0.05154034495353699,
-0.06937182694673538,
-0.05223534256219864,
0.06355299055576324,
0.08018788695335388,
0.07678371667861938,
-0.12301702797412872,
-0.02094447799026966,
0.008637533523142338,
0.00831096712499857,
0.08201737701892853,
0.023290244862437248,
0.1510206013917923,
0.03883988782763481,
-0.12744688987731934,
-0.050009194761514664,
0.10665731877088547,
0.041741468012332916,
-0.04784774035215378,
-0.25138479471206665,
-0.030326439067721367,
-0.027732934802770615,
-0.029999805614352226,
-0.03873695060610771,
0.04263332113623619,
-0.0072723389603197575,
0.0826614573597908,
-0.008116158656775951,
-0.07679495960474014,
-0.03798604756593704,
0.06191713735461235,
0.060809630900621414,
0.026244111359119415,
-0.011753023602068424,
0.010934822261333466,
0.1174238994717598,
0.10631082952022552,
-0.12367359548807144,
-0.051516905426979065,
-0.06431761384010315,
-0.07867198437452316,
-0.04216236248612404,
0.03455616533756256,
0.041060756891965866,
0.049376390874385834,
0.2486443817615509,
0.017620395869016647,
0.05382118001580238,
0.03803925961256027,
0.010167144238948822,
0.06406087428331375,
0.11435336619615555,
-0.061582546681165695,
-0.09715550392866135,
-0.025186026468873024,
0.08966731280088425,
0.01176387071609497,
-0.04024789482355118,
-0.05783011019229889,
0.06293477863073349,
0.016524890437722206,
0.1202789843082428,
0.09223750233650208,
0.003793274285271764,
-0.07138240337371826,
-0.06413803994655609,
0.1937950700521469,
-0.1626761257648468,
0.04747059941291809,
0.034180231392383575,
-0.038511235266923904,
-0.0016249394975602627,
0.008853171020746231,
0.024325255304574966,
-0.021725021302700043,
0.08937039971351624,
-0.05618007108569145,
-0.041590798646211624,
-0.10983981937170029,
-0.035744234919548035,
0.03192625194787979,
0.009910091757774353,
-0.03217151761054993,
-0.031847331672906876,
-0.08444786816835403,
-0.06831640005111694,
0.09424425661563873,
-0.07356466352939606,
-0.053753651678562164,
-0.016938211396336555,
-0.07437273859977722,
0.024786023423075676,
0.01960081420838833,
0.07747352123260498,
-0.02004585787653923,
0.042900070548057556,
-0.05549933388829231,
0.06014169380068779,
0.10937028378248215,
0.033117540180683136,
-0.05445994809269905,
0.0621645413339138,
-0.2418462336063385,
0.0997670441865921,
-0.06829129904508591,
0.05325306951999664,
-0.15072302520275116,
-0.02465333603322506,
0.04913770779967308,
0.008168290369212627,
-0.010590006597340107,
0.13754788041114807,
-0.21924975514411926,
-0.027699807658791542,
0.1631394773721695,
-0.09464818984270096,
-0.07676627486944199,
0.05986984074115753,
-0.052457790821790695,
0.10692904144525528,
0.04047565534710884,
-0.026259733363986015,
0.06162377819418907,
-0.13397987186908722,
0.0005626814090646803,
-0.045883387327194214,
-0.01928110048174858,
0.15731419622898102,
0.07587230950593948,
-0.06994020938873291,
0.07348526269197464,
0.023750323802232742,
-0.023168303072452545,
-0.046913031488657,
-0.017583578824996948,
-0.1088033989071846,
0.010729904286563396,
-0.061985816806554794,
0.01937699131667614,
-0.025795195251703262,
-0.09332547336816788,
-0.028493179008364677,
-0.17521639168262482,
-0.020266273990273476,
0.08516935259103775,
-0.009352635592222214,
-0.01925206556916237,
-0.11787936836481094,
0.015734510496258736,
0.03501737862825394,
0.002549536293372512,
-0.1319509893655777,
-0.05043373629450798,
0.02751830592751503,
-0.16075198352336884,
0.033688947558403015,
-0.05403051897883415,
0.0491553395986557,
0.03133281692862511,
-0.031412381678819656,
-0.028679344803094864,
0.022094380110502243,
0.004997676704078913,
-0.014611656777560711,
-0.24550160765647888,
-0.026604164391756058,
-0.02145342156291008,
0.16796952486038208,
-0.21640902757644653,
0.0374150350689888,
0.07194960117340088,
0.15254895389080048,
0.008589224889874458,
-0.038006994873285294,
0.002335198922082782,
-0.075041763484478,
-0.03255171701312065,
-0.06050482019782066,
-0.009038056246936321,
-0.03572068363428116,
-0.05482286959886551,
0.04863523691892624,
-0.16824471950531006,
-0.029467429965734482,
0.1015508770942688,
0.06473538279533386,
-0.13604550063610077,
-0.019663551822304726,
-0.03585261106491089,
-0.042308371514081955,
-0.05517838895320892,
-0.05935737490653992,
0.10260266810655594,
0.05827045813202858,
0.04566904529929161,
-0.06485172361135483,
-0.0747392401099205,
0.0017082487465813756,
-0.019673427566885948,
-0.022536588832736015,
0.09213293343782425,
0.07581926137208939,
-0.12331884354352951,
0.09213830530643463,
0.10402927547693253,
0.08686267584562302,
0.0966128259897232,
-0.023164015263319016,
-0.08361977338790894,
-0.049845483154058456,
0.02228725142776966,
0.017598064616322517,
0.13447505235671997,
-0.007804518099874258,
0.05406574159860611,
0.04160919412970543,
-0.013909573666751385,
0.009752067737281322,
-0.09242741018533707,
0.032518286257982254,
0.03427431732416153,
-0.01857241988182068,
0.041615914553403854,
-0.039849672466516495,
0.019975949078798294,
0.09018522500991821,
0.046917494386434555,
0.04021155461668968,
0.014107138849794865,
-0.04660527780652046,
-0.11187547445297241,
0.16612006723880768,
-0.12780359387397766,
-0.23512837290763855,
-0.1463187336921692,
0.0034277087543159723,
0.03630480915307999,
-0.009390040300786495,
0.0017278295708820224,
-0.06397698074579239,
-0.11876852810382843,
-0.09194197505712509,
0.010153552517294884,
0.04896695911884308,
-0.0851091742515564,
-0.0603698305785656,
0.05686335638165474,
0.04057794436812401,
-0.14546048641204834,
0.019262617453932762,
0.04933769255876541,
-0.09224124997854233,
-0.009894786402583122,
0.08289197087287903,
0.06857553124427795,
0.18091025948524475,
0.013082148507237434,
-0.02271466888487339,
0.03428078070282936,
0.21755947172641754,
-0.13586747646331787,
0.11420658230781555,
0.1426045000553131,
-0.09194567799568176,
0.08309654146432877,
0.19839057326316833,
0.04078111797571182,
-0.10157861560583115,
0.032499175518751144,
0.018653791397809982,
-0.030491048470139503,
-0.24355553090572357,
-0.07171683013439178,
0.00034942623460665345,
-0.057900771498680115,
0.07530075311660767,
0.09018687158823013,
0.09155713021755219,
0.01583298109471798,
-0.0946493074297905,
-0.07830986380577087,
0.05305508151650429,
0.10324970632791519,
0.020061472430825233,
-0.013236436992883682,
0.09051742404699326,
-0.03375976160168648,
0.017617853358387947,
0.09066354483366013,
0.0011531224008649588,
0.17065346240997314,
0.05820678174495697,
0.18275249004364014,
0.07604338973760605,
0.07338658720254898,
0.01378361415117979,
0.01180104911327362,
0.019032908603549004,
0.02708563208580017,
-0.004741039127111435,
-0.08538748323917389,
-0.01599922962486744,
0.12008915096521378,
0.07424698024988174,
0.015674617141485214,
0.014355434104800224,
-0.04089333862066269,
0.08203015476465225,
0.17435193061828613,
-0.001506963511928916,
-0.1824604868888855,
-0.06271602213382721,
0.08220411837100983,
-0.09449198096990585,
-0.10147359222173691,
-0.02445729449391365,
0.03089604340493679,
-0.17088350653648376,
0.023070847615599632,
-0.016430631279945374,
0.11182350665330887,
-0.13931094110012054,
-0.019696295261383057,
0.0640200525522232,
0.07118809968233109,
-0.00031885437783785164,
0.05944213643670082,
-0.16128569841384888,
0.10404066741466522,
0.013166810385882854,
0.06712377816438675,
-0.09715772420167923,
0.10046469420194626,
-0.006883090827614069,
-0.013416164554655552,
0.13275203108787537,
0.008256223052740097,
-0.07161599397659302,
-0.07921489328145981,
-0.09379399567842484,
-0.009093280881643295,
0.12668752670288086,
-0.14835532009601593,
0.08585991710424423,
-0.035368360579013824,
-0.04256736859679222,
0.0022144275717437267,
-0.10755012929439545,
-0.12217973172664642,
-0.1874755620956421,
0.05520224943757057,
-0.1321607530117035,
0.039849888533353806,
-0.10649667680263519,
-0.03462952747941017,
-0.029491933062672615,
0.1882491409778595,
-0.22971367835998535,
-0.06835493445396423,
-0.15157760679721832,
-0.09785088151693344,
0.14553189277648926,
-0.04969761520624161,
0.08694402873516083,
-0.005991519894450903,
0.18016821146011353,
0.022223925217986107,
-0.021585633978247643,
0.09859558939933777,
-0.09382225573062897,
-0.1963716447353363,
-0.08180448412895203,
0.15751656889915466,
0.13459575176239014,
0.03521031513810158,
-0.0027760460507124662,
0.037876322865486145,
-0.01856307126581669,
-0.12259240448474884,
0.021658578887581825,
0.17797763645648956,
0.0652514174580574,
0.02310643345117569,
-0.026529761031270027,
-0.11104881763458252,
-0.06772379577159882,
-0.033685971051454544,
0.03064778819680214,
0.18449479341506958,
-0.0722544714808464,
0.18419069051742554,
0.143813356757164,
-0.05867353826761246,
-0.1976030021905899,
0.008879725821316242,
0.03365374729037285,
0.007196295075118542,
0.03445420414209366,
-0.20255140960216522,
0.0841677114367485,
0.00034181843511760235,
-0.05190233513712883,
0.13343381881713867,
-0.17106693983078003,
-0.15042030811309814,
0.07339101284742355,
0.03619921952486038,
-0.19460853934288025,
-0.11963265389204025,
-0.08913769572973251,
-0.05391303077340126,
-0.18051348626613617,
0.10290905088186264,
0.03496568650007248,
0.008035079576075077,
0.03376363217830658,
0.028494013473391533,
0.01669638603925705,
-0.03928735852241516,
0.1920013129711151,
-0.026591487228870392,
0.029855716973543167,
-0.08456290513277054,
-0.06990274786949158,
0.04655740037560463,
-0.05482156574726105,
0.0760476216673851,
-0.027013001963496208,
0.011612839996814728,
-0.10561433434486389,
-0.042526841163635254,
-0.029051896184682846,
0.013453613966703415,
-0.0963861495256424,
-0.08940120041370392,
-0.0490599125623703,
0.09310506284236908,
0.09519506990909576,
-0.035876575857400894,
-0.03684677556157112,
-0.07069114595651627,
0.039579302072525024,
0.18676936626434326,
0.17657315731048584,
0.04523694887757301,
-0.0789421945810318,
-0.005537794437259436,
-0.011924253776669502,
0.04352729767560959,
-0.21637341380119324,
0.06442029029130936,
0.05013522133231163,
0.017847778275609016,
0.11767403781414032,
-0.02045002020895481,
-0.1556767225265503,
-0.07006701827049255,
0.06328949332237244,
-0.06132598593831062,
-0.1951322853565216,
0.005576360039412975,
0.054395273327827454,
-0.16848263144493103,
-0.048018258064985275,
0.04364382475614548,
-0.004054433200508356,
-0.0402018167078495,
0.01867259293794632,
0.08977478742599487,
0.003425614908337593,
0.0704059898853302,
0.05869606137275696,
0.08224445581436157,
-0.10246741771697998,
0.07471306622028351,
0.08622124791145325,
-0.07954994589090347,
0.026619622483849525,
0.09149482846260071,
-0.05819176882505417,
-0.02969011478126049,
0.02704544924199581,
0.0793747529387474,
0.011502381414175034,
-0.042540501803159714,
0.011518802493810654,
-0.10228829830884933,
0.06203006953001022,
0.08760257810354233,
0.03265642002224922,
0.015443529933691025,
0.03219176456332207,
0.045628782361745834,
-0.07176384329795837,
0.1219232901930809,
0.028246978297829628,
0.015991143882274628,
-0.04067446291446686,
-0.04898078367114067,
0.024271609261631966,
-0.0303955040872097,
-0.006366716232150793,
-0.03475780412554741,
-0.0729878842830658,
-0.0171539094299078,
-0.16714228689670563,
-0.016664555296301842,
-0.04662061110138893,
0.009329318068921566,
0.03086909092962742,
-0.03788549080491066,
0.008464637212455273,
0.007407912518829107,
-0.07459274679422379,
-0.06477426737546921,
-0.022905457764863968,
0.09289900958538055,
-0.16393527388572693,
0.02335011027753353,
0.08690579235553741,
-0.12064014375209808,
0.09392421692609787,
0.01837589405477047,
-0.0037578048650175333,
0.028480252251029015,
-0.14924435317516327,
0.038928523659706116,
-0.03113253228366375,
0.014821149408817291,
0.04454975947737694,
-0.2236335128545761,
0.0009650349384173751,
-0.033828526735305786,
-0.06339430809020996,
-0.009390673600137234,
-0.036760155111551285,
-0.11370383948087692,
0.10629112273454666,
0.007970798760652542,
-0.08916810154914856,
-0.031690530478954315,
0.032128699123859406,
0.08206479996442795,
-0.0239556971937418,
0.15763959288597107,
-0.0023972811177372932,
0.0736590027809143,
-0.1675432026386261,
-0.019303109496831894,
-0.011248460970818996,
0.020926566794514656,
-0.018098697066307068,
-0.01251189224421978,
0.04078914225101471,
-0.02225574664771557,
0.18437865376472473,
-0.023570427671074867,
0.023348741233348846,
0.06592654436826706,
0.027775658294558525,
-0.025002485141158104,
0.10530006885528564,
0.05339968949556351,
0.021854043006896973,
0.02036798559129238,
0.00273964018560946,
-0.04241073876619339,
-0.023610878735780716,
-0.1998770385980606,
0.06446972489356995,
0.14037446677684784,
0.09086652100086212,
-0.017234215512871742,
0.08257289230823517,
-0.1004219725728035,
-0.11521948128938675,
0.11568495631217957,
-0.05446505919098854,
-0.004037478007376194,
-0.0672159418463707,
0.12938179075717926,
0.1446845531463623,
-0.19097456336021423,
0.06995914876461029,
-0.06848131865262985,
-0.049033988267183304,
-0.11654651165008545,
-0.1963350623846054,
-0.05714293569326401,
-0.05161691829562187,
-0.01663723587989807,
-0.046969223767519,
0.07560921460390091,
0.05719533935189247,
0.007424132898449898,
-0.0017566849710419774,
0.06332923471927643,
-0.026077456772327423,
0.00009585227962816134,
0.026813751086592674,
0.06610306352376938,
0.013093758374452591,
-0.02985633723437786,
0.017491595819592476,
-0.012147722765803337,
0.042048826813697815,
0.06357792019844055,
0.04670548066496849,
-0.030032360926270485,
0.016853880137205124,
-0.03863191977143288,
-0.10680584609508514,
0.041318636387586594,
-0.028504958376288414,
-0.08043242245912552,
0.1491626501083374,
0.02454165369272232,
0.008750278502702713,
-0.0205967016518116,
0.2416755110025406,
-0.0737907737493515,
-0.09567341208457947,
-0.1479424238204956,
0.10524045675992966,
-0.04420987144112587,
0.06244929879903793,
0.045180387794971466,
-0.10425344854593277,
0.016717668622732162,
0.12817999720573425,
0.16302813589572906,
-0.044200748205184937,
0.020526019856333733,
0.027614353224635124,
0.004152800887823105,
-0.03678637370467186,
0.0514480359852314,
0.06988705694675446,
0.1595088243484497,
-0.048713311553001404,
0.09546878933906555,
-0.0016016386216506362,
-0.09618084132671356,
-0.03802286460995674,
0.11709540337324142,
-0.018092934042215347,
0.017691975459456444,
-0.055210161954164505,
0.11857418715953827,
-0.06138255074620247,
-0.2316483110189438,
0.06108921393752098,
-0.06591550260782242,
-0.13765475153923035,
-0.02143050730228424,
0.08041442185640335,
-0.013238796964287758,
0.02708347514271736,
0.07207029312849045,
-0.07533451914787292,
0.20003929734230042,
0.037636954337358475,
-0.05420409142971039,
-0.05360380560159683,
0.08255447447299957,
-0.10376271605491638,
0.27565470337867737,
0.016520937904715538,
0.04948882386088371,
0.10317612439393997,
-0.012690499424934387,
-0.13475549221038818,
0.02108365297317505,
0.09600389003753662,
-0.0946137085556984,
0.04216265305876732,
0.19903649389743805,
0.0003853837260976434,
0.1207512691617012,
0.0790785402059555,
-0.07618726044893265,
0.049590613692998886,
-0.0941753089427948,
-0.07070460170507431,
-0.09001081436872482,
0.09455035626888275,
-0.07685617357492447,
0.14261877536773682,
0.1292559802532196,
-0.053739987313747406,
0.010677514597773552,
-0.028576120734214783,
0.04638256877660751,
0.0034859003499150276,
0.1005801111459732,
0.010024284943938255,
-0.18460705876350403,
0.02157641015946865,
0.01203901320695877,
0.1056026741862297,
-0.16518552601337433,
-0.09804878383874893,
0.042120642960071564,
0.0014211505185812712,
-0.060778699815273285,
0.12909291684627533,
0.06027422100305557,
0.04478219151496887,
-0.04292554408311844,
-0.020403601229190826,
-0.009860116057097912,
0.13677826523780823,
-0.10241927951574326,
0.0014122816501185298
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec_RTSplit0208_20
This model is a fine-tuned version of [jonatasgrosman/wav2vec2-large-xlsr-53-japanese](https://huggingface.co/jonatasgrosman/wav2vec2-large-xlsr-53-japanese) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0148
- Wer: 0.2021
- Cer: 0.1741
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
| 3.7322 | 1.0 | 120 | 3.5160 | 1.0 | 0.9489 |
| 1.4724 | 2.0 | 240 | 1.3042 | 0.9991 | 0.6951 |
| 0.8148 | 3.0 | 360 | 0.6661 | 0.8175 | 0.5193 |
| 0.527 | 4.0 | 480 | 0.4013 | 0.6019 | 0.3703 |
| 0.3803 | 5.0 | 600 | 0.2061 | 0.4203 | 0.2173 |
| 0.253 | 6.0 | 720 | 0.1038 | 0.3175 | 0.2049 |
| 0.2069 | 7.0 | 840 | 0.0560 | 0.3128 | 0.1882 |
| 0.1898 | 8.0 | 960 | 0.0380 | 0.2984 | 0.1802 |
| 0.1484 | 9.0 | 1080 | 0.0204 | 0.2139 | 0.1662 |
| 0.0758 | 10.0 | 1200 | 0.0148 | 0.2021 | 0.1741 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "jonatasgrosman/wav2vec2-large-xlsr-53-japanese", "model-index": [{"name": "wav2vec_RTSplit0208_20", "results": []}]} | automatic-speech-recognition | tndklab/wav2vec_RTSplit0208_20 | [
"transformers",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"base_model:jonatasgrosman/wav2vec2-large-xlsr-53-japanese",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:36:44+00:00 | [] | [] | TAGS
#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us
| wav2vec\_RTSplit0208\_20
========================
This model is a fine-tuned version of jonatasgrosman/wav2vec2-large-xlsr-53-japanese on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0148
* Wer: 0.2021
* Cer: 0.1741
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 6e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 4
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.14.6
* Tokenizers 0.15.0
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
80,
116,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #wav2vec2 #automatic-speech-recognition #generated_from_trainer #base_model-jonatasgrosman/wav2vec2-large-xlsr-53-japanese #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 6e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 4\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.14.6\n* Tokenizers 0.15.0"
] | [
-0.1412041336297989,
0.15137426555156708,
-0.0004950824659317732,
0.09997182339429855,
0.11816604435443878,
0.007901513949036598,
0.17537309229373932,
0.15016427636146545,
-0.04011153429746628,
0.1108313798904419,
0.11322881281375885,
0.06258971989154816,
0.05524484068155289,
0.19542798399925232,
-0.0818181112408638,
-0.22068442404270172,
0.07705570757389069,
-0.0034033337142318487,
0.010251824744045734,
0.1121915951371193,
0.07054735720157623,
-0.11850614100694656,
0.09033747017383575,
-0.007322786375880241,
-0.14293430745601654,
-0.04210640862584114,
0.016174355521798134,
-0.10974466800689697,
0.10908807069063187,
0.008706390857696533,
0.06604725122451782,
0.03539383038878441,
0.0900065004825592,
-0.1866677850484848,
0.002349936170503497,
0.016980301588773727,
0.01458936557173729,
0.07418084889650345,
0.042821869254112244,
-0.000933396047912538,
0.0035048893187195063,
-0.1137300506234169,
0.036463793367147446,
0.015336795710027218,
-0.11726072430610657,
-0.1977897733449936,
-0.07841771095991135,
0.016024377197027206,
0.098679319024086,
0.08460279554128647,
-0.020939182490110397,
0.12405982613563538,
0.0009574935538694263,
0.0794869065284729,
0.1975121796131134,
-0.3128716051578522,
-0.05420010909438133,
-0.016580859199166298,
0.03949279710650444,
0.08204091340303421,
-0.10216154903173447,
-0.018908465281128883,
0.04993114992976189,
0.021625857800245285,
0.09085845202207565,
-0.031167294830083847,
-0.03472635895013809,
-0.011195491068065166,
-0.12056951224803925,
-0.038290105760097504,
0.18850144743919373,
0.07279396057128906,
-0.06344844400882721,
-0.08026174455881119,
-0.06453180313110352,
-0.12059958279132843,
-0.054848991334438324,
-0.0070543293841183186,
0.026367703452706337,
-0.03924409672617912,
-0.09994582831859589,
-0.004911588504910469,
-0.07957334816455841,
-0.09121696650981903,
-0.017583273351192474,
0.17618964612483978,
0.01090241875499487,
0.013959892094135284,
-0.011917638592422009,
0.054277557879686356,
-0.022792238742113113,
-0.18547365069389343,
-0.022341668605804443,
0.027699081227183342,
-0.03266499191522598,
-0.013799833133816719,
-0.04344217851758003,
-0.033000387251377106,
0.04403166100382805,
0.11617182195186615,
-0.01942470110952854,
0.06578745692968369,
-0.02470904402434826,
0.0015051147202029824,
-0.08571495115756989,
0.18256285786628723,
-0.06435675173997879,
-0.0677129402756691,
0.019559592008590698,
0.12708435952663422,
0.06296100467443466,
-0.0231513399630785,
-0.09921643137931824,
-0.008808785118162632,
0.14603620767593384,
0.035404834896326065,
-0.042173828929662704,
0.05110965669155121,
-0.03869181126356125,
-0.014063776470720768,
0.058567557483911514,
-0.12158208340406418,
0.025354862213134766,
0.021827101707458496,
-0.06338157504796982,
-0.022762063890695572,
-0.012170805595815182,
0.01291244849562645,
0.01296332385390997,
0.0519440658390522,
-0.08247014880180359,
0.003685595700517297,
-0.02379593625664711,
-0.09241920709609985,
0.026958566159009933,
-0.0687103420495987,
-0.0001175150609924458,
-0.10739189386367798,
-0.1797207146883011,
-0.018069207668304443,
0.02376030571758747,
-0.04920637235045433,
-0.009401011280715466,
-0.11267255991697311,
-0.0978749543428421,
0.04726383835077286,
-0.022524166852235794,
0.03536629676818848,
-0.0796164721250534,
0.1083039864897728,
0.0794280618429184,
0.08762047439813614,
-0.04133358970284462,
0.026571081951260567,
-0.09571915119886398,
0.032445281744003296,
-0.17462952435016632,
0.07469087094068527,
-0.05457072705030441,
0.034701019525527954,
-0.12011414021253586,
-0.0671958476305008,
0.019478250294923782,
-0.022287972271442413,
0.06999751180410385,
0.1424819976091385,
-0.19131408631801605,
-0.05639734864234924,
0.19655689597129822,
-0.11995879560709,
-0.14247135818004608,
0.12762568891048431,
-0.03615688160061836,
0.03838138282299042,
0.07021597027778625,
0.22375737130641937,
0.03087463602423668,
-0.10594569891691208,
-0.040601879358291626,
-0.06359881907701492,
0.08302099257707596,
-0.036925822496414185,
0.11114954948425293,
0.0050477092154324055,
-0.0006848193006590009,
0.016178669407963753,
-0.07956434041261673,
0.03174460306763649,
-0.07065560668706894,
-0.0996895432472229,
-0.044058747589588165,
-0.10600133240222931,
0.02815171331167221,
0.016454190015792847,
0.055376626551151276,
-0.09874159842729568,
-0.07053174078464508,
0.010761037468910217,
0.10824721306562424,
-0.11714144796133041,
0.01284718792885542,
-0.10444527119398117,
0.0938587337732315,
-0.11354869604110718,
-0.02030627802014351,
-0.15474089980125427,
-0.005486346781253815,
0.05399791896343231,
0.019834648817777634,
0.014824481680989265,
-0.0752023458480835,
0.08260023593902588,
0.07641925662755966,
-0.04929106682538986,
-0.07418523728847504,
-0.004319242667406797,
0.017646290361881256,
-0.06242918223142624,
-0.17318762838840485,
-0.02871767245233059,
-0.054336290806531906,
0.16024808585643768,
-0.16467665135860443,
0.0010856473818421364,
0.009586222469806671,
0.09021752327680588,
0.04384804889559746,
-0.023363610729575157,
0.019892822951078415,
0.047984279692173004,
-0.026173073798418045,
-0.07180936634540558,
0.029512276872992516,
0.015152420848608017,
-0.1037457287311554,
0.020873865112662315,
-0.1682128757238388,
0.14848269522190094,
0.1395752727985382,
0.04242534935474396,
-0.05247782543301582,
0.020625025033950806,
-0.014322427101433277,
-0.041876472532749176,
-0.05366198346018791,
-0.01537831500172615,
0.10208909213542938,
0.008062804117798805,
0.12061847001314163,
-0.10291217267513275,
0.015420259907841682,
0.0646526888012886,
-0.027293499559164047,
-0.028270341455936432,
0.0813247412443161,
0.010808512568473816,
-0.13996154069900513,
0.13055692613124847,
0.11295907199382782,
-0.07091501355171204,
0.1261623501777649,
-0.061400000005960464,
-0.08471372723579407,
-0.049160800874233246,
0.03369266167283058,
0.03434725105762482,
0.1375427097082138,
-0.08055726438760757,
-0.02187250182032585,
0.021242210641503334,
0.022564619779586792,
-0.015945840626955032,
-0.19370488822460175,
-0.01975315250456333,
0.014006020501255989,
-0.09473686665296555,
-0.009634226560592651,
0.005931468214839697,
-0.017773766070604324,
0.09490399807691574,
-0.0005694571882486343,
-0.11420944333076477,
0.023075604811310768,
-0.015324614942073822,
-0.08736004680395126,
0.17215579748153687,
-0.0921676903963089,
-0.17300088703632355,
-0.1357395350933075,
-0.0722767785191536,
-0.056055884808301926,
0.03831305727362633,
0.05982291325926781,
-0.06652282923460007,
-0.040523629635572433,
-0.11533518880605698,
-0.04774909093976021,
0.032573483884334564,
0.0449487529695034,
0.05047421157360077,
-0.008775885216891766,
0.0660317987203598,
-0.08186756074428558,
-0.005498649552464485,
-0.01464156061410904,
-0.008005748502910137,
0.02891124226152897,
-0.000026242085368721746,
0.12651395797729492,
0.12205208092927933,
0.005935537628829479,
0.024285942316055298,
-0.03736124560236931,
0.22580985724925995,
-0.07014511525630951,
-0.018395544961094856,
0.12401127815246582,
-0.02740400657057762,
0.04592745751142502,
0.17833060026168823,
0.030612092465162277,
-0.10761554539203644,
0.0017429981380701065,
-0.04984324425458908,
-0.0159926600754261,
-0.18899597227573395,
-0.033754944801330566,
-0.04779517650604248,
0.013360806740820408,
0.10108592361211777,
0.030252007767558098,
0.014854468405246735,
0.04852636530995369,
0.021330097690224648,
0.04515378549695015,
0.004228484816849232,
0.08083099126815796,
0.09576261788606644,
0.07646510004997253,
0.10783705115318298,
-0.03230658173561096,
-0.04892025142908096,
0.032441601157188416,
0.021614739671349525,
0.20175261795520782,
0.0299892108887434,
0.193137988448143,
0.0009553404524922371,
0.15421515703201294,
0.025757189840078354,
0.08017203956842422,
0.019132642075419426,
0.009862223640084267,
-0.021082380786538124,
-0.0781867653131485,
-0.053376875817775726,
0.054635848850011826,
-0.014859071001410484,
0.060703035444021225,
-0.10492148995399475,
0.020623087882995605,
0.04996446147561073,
0.27169784903526306,
0.08805271238088608,
-0.3690784275531769,
-0.08674702048301697,
0.021065227687358856,
-0.0367366224527359,
-0.019402114674448967,
0.01681741327047348,
0.1533811390399933,
-0.0612044632434845,
0.06880920380353928,
-0.07165786623954773,
0.06362106651067734,
-0.06424785405397415,
0.019752372056245804,
0.02411559410393238,
0.04684673249721527,
0.002824755385518074,
0.0314948745071888,
-0.24140027165412903,
0.28712528944015503,
0.03586786985397339,
0.0946955755352974,
-0.05696169659495354,
-0.0038834919687360525,
0.03976857662200928,
-0.0069578527472913265,
0.11812238395214081,
-0.024606162682175636,
-0.11095675826072693,
-0.1789819896221161,
-0.13465972244739532,
0.04897531867027283,
0.10619998723268509,
-0.006434483453631401,
0.11634289473295212,
-0.013821248896420002,
-0.0443023182451725,
0.04462811350822449,
-0.024251187220215797,
-0.07929635792970657,
-0.07509298622608185,
0.009053842164576054,
0.11385448276996613,
0.04451282322406769,
-0.04915499687194824,
-0.09564180672168732,
-0.08921606838703156,
0.0902084931731224,
0.004420565441250801,
-0.006528897676616907,
-0.10501714050769806,
0.017749018967151642,
0.15001456439495087,
-0.0917128399014473,
0.052100859582424164,
0.009017262607812881,
0.11019572615623474,
0.02719823829829693,
-0.04979333281517029,
0.09063542634248734,
-0.061942748725414276,
-0.17910468578338623,
-0.05071870610117912,
0.138783261179924,
-0.007552908733487129,
0.042952146381139755,
0.021130461245775223,
0.05198457092046738,
-0.005837497301399708,
-0.0669860765337944,
0.03282371908426285,
0.02590620331466198,
0.04146227985620499,
0.019748570397496223,
-0.012347066774964333,
-0.09044408798217773,
-0.09293613582849503,
-0.02274126373231411,
0.15032701194286346,
0.29759377241134644,
-0.06566981971263885,
0.018390042707324028,
0.08737357705831528,
-0.01744837500154972,
-0.1529613435268402,
-0.0046507492661476135,
0.04321031644940376,
0.044014107435941696,
-0.005333513952791691,
-0.12244978547096252,
0.04551294073462486,
0.06094546988606453,
-0.04485302418470383,
0.07687357813119888,
-0.24799765646457672,
-0.12770286202430725,
0.0884246826171875,
0.1330942064523697,
0.1232348307967186,
-0.15262654423713684,
-0.06689915060997009,
-0.02422727830708027,
-0.1076960563659668,
0.10433056950569153,
-0.07238385826349258,
0.13326559960842133,
-0.002409221837297082,
0.06556188315153122,
0.007432775571942329,
-0.05152697116136551,
0.1498710960149765,
0.023709816858172417,
0.05323371663689613,
-0.022845882922410965,
-0.017437255010008812,
0.04780147597193718,
-0.0750357061624527,
0.06876204162836075,
-0.08563213795423508,
0.04991770163178444,
-0.06026848033070564,
-0.024500783532857895,
-0.06207650527358055,
-0.006298033520579338,
0.00395286176353693,
-0.03505624085664749,
-0.01054296363145113,
0.036420781165361404,
0.05855202302336693,
0.0031899285968393087,
0.1323307901620865,
0.011613145470619202,
0.08266790211200714,
0.14601479470729828,
0.08927149325609207,
-0.03952056169509888,
0.015870284289121628,
-0.006142308469861746,
-0.05592581257224083,
0.053551238030195236,
-0.13233886659145355,
0.04830208793282509,
0.09677808731794357,
0.01841074973344803,
0.15998423099517822,
0.04728129133582115,
-0.0495828241109848,
0.038742292672395706,
0.06929651647806168,
-0.1584930270910263,
-0.111056387424469,
0.003867720952257514,
-0.012255392968654633,
-0.11091582477092743,
0.04999949783086777,
0.13792867958545685,
-0.07000352442264557,
-0.0066223847679793835,
-0.017663979902863503,
0.021433407440781593,
-0.039054762572050095,
0.20004802942276,
0.042530614882707596,
0.05142882838845253,
-0.11011962592601776,
0.08127256482839584,
0.05678301304578781,
-0.08799626678228378,
0.048996541649103165,
0.038279805332422256,
-0.11541429907083511,
-0.023084815591573715,
-0.00001558458461659029,
0.14206252992153168,
0.00538132106885314,
-0.0756935328245163,
-0.1377873718738556,
-0.0886237695813179,
0.03491495922207832,
0.177113339304924,
0.06811617314815521,
0.036583062261343,
-0.0181216299533844,
-0.0022696226369589567,
-0.10424704849720001,
0.09346354752779007,
0.07297512143850327,
0.07519941031932831,
-0.149531751871109,
0.08094713091850281,
-0.008585318922996521,
0.027400100603699684,
-0.020543552935123444,
0.017085453495383263,
-0.10940637439489365,
0.005164341535419226,
-0.0982065349817276,
0.056565169245004654,
-0.07851643115282059,
-0.016379782930016518,
-0.0016896710731089115,
-0.08193731307983398,
-0.06113337352871895,
0.011151892133057117,
-0.08733723312616348,
-0.026363078504800797,
0.0034647583961486816,
0.04292234405875206,
-0.13643650710582733,
-0.037632495164871216,
0.022685186937451363,
-0.09767955541610718,
0.08393611013889313,
0.0864483043551445,
-0.019734136760234833,
0.04710947349667549,
-0.09520848840475082,
-0.02245686762034893,
0.08286114037036896,
0.0025889736134558916,
0.05070040002465248,
-0.14341135323047638,
-0.014357004314661026,
0.031179029494524002,
0.05052122846245766,
0.021537650376558304,
0.14844025671482086,
-0.0965467169880867,
0.00683720363304019,
-0.06784950196743011,
-0.011589626781642437,
-0.05678727105259895,
0.022133558988571167,
0.14139819145202637,
0.003112047677859664,
0.1835901290178299,
-0.09471392631530762,
0.022715279832482338,
-0.19733846187591553,
0.0012226614635437727,
-0.036960259079933167,
-0.12626180052757263,
-0.14721830189228058,
-0.02713117003440857,
0.07858846336603165,
-0.06186365336179733,
0.09447462111711502,
-0.061552248895168304,
0.06924274563789368,
0.01299452967941761,
-0.057405371218919754,
-0.0013264762237668037,
0.04017876088619232,
0.24852977693080902,
0.0576971061527729,
-0.035827282816171646,
0.07744954526424408,
0.010208655148744583,
0.0938122421503067,
0.12496055662631989,
0.12368854135274887,
0.15800440311431885,
0.03270518034696579,
0.14265680313110352,
0.08402945846319199,
-0.024530375376343727,
-0.1192057803273201,
0.05892833694815636,
-0.0690154880285263,
0.08931177854537964,
0.02568928524851799,
0.21020832657814026,
0.09831425547599792,
-0.16308072209358215,
0.004579037427902222,
-0.03712035343050957,
-0.08482547849416733,
-0.09529641270637512,
-0.060792118310928345,
-0.13080793619155884,
-0.144043430685997,
0.011156630702316761,
-0.10667292773723602,
0.034262366592884064,
0.07137389481067657,
0.01432362012565136,
0.000012607777534867637,
0.14221198856830597,
0.014257829636335373,
0.02906184270977974,
0.09490415453910828,
0.008670350536704063,
-0.04020773991942406,
0.00022622093092650175,
-0.10310152918100357,
0.024451300501823425,
0.005498264916241169,
0.056763384491205215,
-0.021193165332078934,
-0.02560819871723652,
0.06862773001194,
-0.026128707453608513,
-0.12540489435195923,
0.010948865674436092,
0.0207629706710577,
0.06055864691734314,
0.04349271580576897,
0.05639878287911415,
-0.017083214595913887,
0.02505779266357422,
0.2077161818742752,
-0.08930209279060364,
-0.07761802524328232,
-0.13242462277412415,
0.14997723698616028,
-0.014713408425450325,
-0.007720646448433399,
0.00910715851932764,
-0.10601600259542465,
0.002112677786499262,
0.1941474825143814,
0.15005312860012054,
-0.07458212226629257,
-0.0007921219221316278,
-0.027325214818120003,
-0.00626454409211874,
-0.0381433442234993,
0.065922811627388,
0.0785856544971466,
0.03481105715036392,
-0.05982352793216705,
-0.06153883412480354,
-0.057788390666246414,
-0.04130222275853157,
-0.022989094257354736,
0.04050512984395027,
-0.03199388086795807,
-0.023023828864097595,
-0.05058024078607559,
0.07841388881206512,
-0.0824628621339798,
-0.09817294776439667,
0.007028645370155573,
-0.21771271526813507,
-0.1729547381401062,
-0.0020018890500068665,
0.07483505457639694,
0.035655662417411804,
0.026528440415859222,
-0.0337730273604393,
0.02595202997326851,
0.05488588288426399,
-0.01344369351863861,
-0.057900380343198776,
-0.0579204261302948,
0.04237135499715805,
-0.0833219438791275,
0.17439395189285278,
-0.0049809738993644714,
0.06613589823246002,
0.10405856370925903,
0.08126693964004517,
-0.1076810359954834,
0.10420752316713333,
0.06104092672467232,
-0.07467813789844513,
0.055807627737522125,
0.15151943266391754,
-0.05596069246530533,
0.14149419963359833,
0.05206945538520813,
-0.1039142906665802,
0.0013270378112792969,
0.010214783251285553,
-0.02869926206767559,
-0.07482365518808365,
-0.06526384502649307,
-0.04589445888996124,
0.14720846712589264,
0.13454674184322357,
-0.06611188501119614,
0.0017617724370211363,
-0.016761446371674538,
0.055473692715168,
0.062239158898591995,
0.020190265029668808,
-0.0619213804602623,
-0.2836199402809143,
-0.017024124041199684,
0.03711088374257088,
0.02185937389731407,
-0.2431417554616928,
-0.08876391500234604,
-0.01048292312771082,
-0.04620630294084549,
-0.07439623028039932,
0.0933784618973732,
0.07889628410339355,
0.030873730778694153,
-0.05422563850879669,
-0.050494153052568436,
-0.028522413223981857,
0.17309556901454926,
-0.16356457769870758,
-0.11507130414247513
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | Prajvi/Llama2_7B_qlora_FT_bush_crisis | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:40:55+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
# Model Trained Using AutoTrain
- Problem type: Image Classification
## Validation Metricsg
loss: 1.574344515800476
f1_macro: 0.08333333333333334
f1_micro: 0.2631578947368421
f1_weighted: 0.10964912280701755
precision_macro: 0.05263157894736842
precision_micro: 0.2631578947368421
precision_weighted: 0.06925207756232686
recall_macro: 0.2
recall_micro: 0.2631578947368421
recall_weighted: 0.2631578947368421
accuracy: 0.2631578947368421
| {"tags": ["autotrain", "image-classification"], "datasets": ["autotrain-z7maa-wyroe/autotrain-data"], "widget": [{"src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg", "example_title": "Tiger"}, {"src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg", "example_title": "Teapot"}, {"src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg", "example_title": "Palace"}]} | image-classification | basavaakash002/autotrain-z7maa-wyroe | [
"transformers",
"safetensors",
"swin",
"image-classification",
"autotrain",
"dataset:autotrain-z7maa-wyroe/autotrain-data",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:45:48+00:00 | [] | [] | TAGS
#transformers #safetensors #swin #image-classification #autotrain #dataset-autotrain-z7maa-wyroe/autotrain-data #autotrain_compatible #endpoints_compatible #region-us
|
# Model Trained Using AutoTrain
- Problem type: Image Classification
## Validation Metricsg
loss: 1.574344515800476
f1_macro: 0.08333333333333334
f1_micro: 0.2631578947368421
f1_weighted: 0.10964912280701755
precision_macro: 0.05263157894736842
precision_micro: 0.2631578947368421
precision_weighted: 0.06925207756232686
recall_macro: 0.2
recall_micro: 0.2631578947368421
recall_weighted: 0.2631578947368421
accuracy: 0.2631578947368421
| [
"# Model Trained Using AutoTrain\n\n- Problem type: Image Classification",
"## Validation Metricsg\nloss: 1.574344515800476\n\nf1_macro: 0.08333333333333334\n\nf1_micro: 0.2631578947368421\n\nf1_weighted: 0.10964912280701755\n\nprecision_macro: 0.05263157894736842\n\nprecision_micro: 0.2631578947368421\n\nprecision_weighted: 0.06925207756232686\n\nrecall_macro: 0.2\n\nrecall_micro: 0.2631578947368421\n\nrecall_weighted: 0.2631578947368421\n\naccuracy: 0.2631578947368421"
] | [
"TAGS\n#transformers #safetensors #swin #image-classification #autotrain #dataset-autotrain-z7maa-wyroe/autotrain-data #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Trained Using AutoTrain\n\n- Problem type: Image Classification",
"## Validation Metricsg\nloss: 1.574344515800476\n\nf1_macro: 0.08333333333333334\n\nf1_micro: 0.2631578947368421\n\nf1_weighted: 0.10964912280701755\n\nprecision_macro: 0.05263157894736842\n\nprecision_micro: 0.2631578947368421\n\nprecision_weighted: 0.06925207756232686\n\nrecall_macro: 0.2\n\nrecall_micro: 0.2631578947368421\n\nrecall_weighted: 0.2631578947368421\n\naccuracy: 0.2631578947368421"
] | [
63,
16,
140
] | [
"passage: TAGS\n#transformers #safetensors #swin #image-classification #autotrain #dataset-autotrain-z7maa-wyroe/autotrain-data #autotrain_compatible #endpoints_compatible #region-us \n# Model Trained Using AutoTrain\n\n- Problem type: Image Classification## Validation Metricsg\nloss: 1.574344515800476\n\nf1_macro: 0.08333333333333334\n\nf1_micro: 0.2631578947368421\n\nf1_weighted: 0.10964912280701755\n\nprecision_macro: 0.05263157894736842\n\nprecision_micro: 0.2631578947368421\n\nprecision_weighted: 0.06925207756232686\n\nrecall_macro: 0.2\n\nrecall_micro: 0.2631578947368421\n\nrecall_weighted: 0.2631578947368421\n\naccuracy: 0.2631578947368421"
] | [
-0.13335736095905304,
0.1389116644859314,
-0.0007834853604435921,
0.12431363016366959,
0.12507453560829163,
0.03595146909356117,
0.10225392878055573,
0.11422539502382278,
-0.043433912098407745,
0.11709677428007126,
0.13667839765548706,
0.09872425347566605,
0.05662032216787338,
0.13071508705615997,
-0.1208576187491417,
-0.15979540348052979,
0.030714407563209534,
0.01940459944307804,
0.06387875974178314,
0.07292761653661728,
0.07111455500125885,
-0.12205232679843903,
0.1106223613023758,
-0.0194686446338892,
-0.24398767948150635,
0.014226858504116535,
0.12355797737836838,
-0.09183813631534576,
0.080501489341259,
0.09564458578824997,
0.08930782228708267,
0.016069907695055008,
0.12471862882375717,
-0.06470520794391632,
-0.023603161796927452,
0.007176673971116543,
-0.00021185306832194328,
0.05835477635264397,
0.16524222493171692,
0.007766712456941605,
-0.07376808673143387,
-0.058039166033267975,
0.052005160599946976,
0.06857742369174957,
-0.06824247539043427,
-0.11300354450941086,
-0.12188513576984406,
0.1092451736330986,
0.08226422965526581,
0.07864563167095184,
-0.0014540645061060786,
0.16347457468509674,
-0.049417976289987564,
0.08281121402978897,
0.12162337452173233,
-0.22870372235774994,
-0.06383680552244186,
0.16841301321983337,
-0.038856156170368195,
-0.07937786728143692,
-0.08341575413942337,
0.06917299330234528,
0.10878687351942062,
-0.008950998075306416,
0.024035800248384476,
-0.03284961357712746,
-0.030263738706707954,
0.007613286375999451,
-0.09169144183397293,
-0.006273187231272459,
0.16842904686927795,
0.03879460692405701,
-0.08055225759744644,
-0.10618636757135391,
-0.028271812945604324,
-0.1021607294678688,
-0.069696806371212,
-0.029707113280892372,
0.008819526992738247,
-0.05342100188136101,
-0.0705801472067833,
0.10562264919281006,
-0.04189610853791237,
-0.10127607733011246,
-0.12120461463928223,
0.10877835750579834,
-0.017011195421218872,
0.03417287394404411,
0.014650928787887096,
0.03256676718592644,
-0.03340712934732437,
-0.09015576541423798,
0.015694135800004005,
-0.0015205518575385213,
-0.09440360963344574,
-0.04953213781118393,
0.011248289607465267,
-0.04875529184937477,
0.01938013546168804,
0.13981901109218597,
0.014590887352824211,
0.06177579611539841,
-0.021249305456876755,
0.011577959172427654,
-0.06567011773586273,
0.1706569641828537,
-0.1151929721236229,
-0.09819753468036652,
0.016800906509160995,
0.020358920097351074,
0.029048176482319832,
-0.009892973117530346,
-0.08300964534282684,
-0.11200655251741409,
0.13497035205364227,
0.0264882855117321,
-0.0050240615382790565,
0.04508417472243309,
-0.06856421381235123,
-0.0294200386852026,
-0.008594931103289127,
-0.06502167135477066,
0.031792812049388885,
-0.0430319719016552,
-0.11142712831497192,
0.017644135281443596,
0.05143899470567703,
0.011879769153892994,
-0.0176661629229784,
0.13849537074565887,
-0.13230670988559723,
0.01843222975730896,
-0.05567407235503197,
-0.11642066389322281,
0.038670096546411514,
-0.0440206415951252,
-0.015184235759079456,
-0.11099246889352798,
-0.21613600850105286,
-0.04819508269429207,
-0.053666699677705765,
-0.0880521684885025,
0.004095855168998241,
-0.010900602675974369,
-0.04819903522729874,
0.08520770817995071,
0.024972151964902878,
0.09116359800100327,
-0.01025738287717104,
0.043362587690353394,
0.05098803713917732,
0.0904056578874588,
-0.03140399232506752,
0.009541825391352177,
-0.06651683151721954,
0.01837296411395073,
-0.17542299628257751,
0.018456166610121727,
-0.06466054171323776,
0.07317303866147995,
-0.08438102900981903,
-0.005960206501185894,
0.06131153926253319,
0.0026189598720520735,
0.040119320154190063,
0.17499925196170807,
-0.1418871432542801,
-0.02355097606778145,
0.08353301137685776,
-0.0987735390663147,
-0.10552598536014557,
0.1219530925154686,
0.0032660679426044226,
0.005136715713888407,
0.03788607940077782,
0.09430844336748123,
0.052371080964803696,
-0.08229506015777588,
-0.07343415915966034,
-0.05420740693807602,
-0.050781283527612686,
-0.05662589520215988,
0.07937686890363693,
-0.006642182357609272,
-0.12162666767835617,
0.021276669576764107,
0.03019184246659279,
0.021447552368044853,
-0.0841904953122139,
-0.05608448013663292,
-0.04914415255188942,
-0.06436716765165329,
0.050992101430892944,
0.02991899475455284,
0.045819271355867386,
-0.0852903500199318,
-0.054374732077121735,
0.06772071868181229,
0.09663635492324829,
0.01136092934757471,
-0.05345453694462776,
-0.1356358826160431,
0.0697636753320694,
-0.15400895476341248,
-0.031269755214452744,
-0.1519443839788437,
-0.09152746945619583,
-0.011207165196537971,
0.005235609598457813,
-0.01892179623246193,
-0.082953080534935,
0.09260158240795135,
0.046223316341638565,
-0.06576000154018402,
-0.04299696162343025,
0.08396739512681961,
0.0009139309404417872,
-0.12624888122081757,
-0.1222371906042099,
-0.012337276712059975,
0.009726868011057377,
0.19797898828983307,
-0.19233182072639465,
-0.012974729761481285,
0.053712837398052216,
0.09648352116346359,
0.051476359367370605,
-0.01734173111617565,
-0.015853235498070717,
0.03749564290046692,
-0.035268720239400864,
-0.06137396767735481,
0.048496972769498825,
-0.03240930661559105,
-0.07578951865434647,
0.03492484614253044,
-0.22964321076869965,
0.262440949678421,
0.1365712285041809,
-0.017419014126062393,
-0.0999123826622963,
-0.04768253117799759,
0.010722761042416096,
-0.06141234561800957,
-0.08153796195983887,
-0.012463708408176899,
-0.0649716928601265,
-0.02537916600704193,
0.1156022921204567,
-0.047413088381290436,
-0.030292779207229614,
0.05022900924086571,
-0.007313070353120565,
-0.07908257097005844,
0.1312451809644699,
0.008999518118798733,
-0.13422183692455292,
0.14262747764587402,
0.04559863731265068,
-0.08703611046075821,
0.11405099928379059,
0.017063317820429802,
-0.04760154336690903,
-0.06806986033916473,
-0.05184270814061165,
0.06388596445322037,
0.11754589527845383,
0.026220504194498062,
0.015735432505607605,
0.06172490492463112,
-0.0222872756421566,
-0.01996445655822754,
-0.09346898645162582,
-0.011550062336027622,
0.010494517162442207,
-0.010572277940809727,
0.04939614608883858,
0.003528750967234373,
0.04356571286916733,
0.15543706715106964,
0.007555794436484575,
-0.05564381554722786,
0.055479031056165695,
-0.0018285217229276896,
-0.08867021650075912,
0.17916761338710785,
-0.07147959619760513,
-0.17847947776317596,
-0.19557179510593414,
-0.06483256816864014,
-0.1108289286494255,
-0.0034525725059211254,
-0.012354069389402866,
-0.0834336206316948,
-0.10814011096954346,
-0.07210863381624222,
-0.019886527210474014,
0.00734645826742053,
0.00738875987008214,
0.02957179583609104,
-0.01226396206766367,
0.0722443014383316,
-0.05200519785284996,
-0.04126962274312973,
-0.005932196043431759,
-0.026022894307971,
0.09608273208141327,
0.0026881841477006674,
0.15715321898460388,
0.12498336285352707,
-0.03738241270184517,
0.03951204568147659,
-0.026081904768943787,
0.12815256416797638,
-0.026785647496581078,
0.008421017788350582,
0.19720782339572906,
0.03188452497124672,
0.01852734014391899,
0.1162194088101387,
0.014203972183167934,
-0.08607543259859085,
0.02566063590347767,
0.06558854132890701,
-0.03220371529459953,
-0.1363520324230194,
-0.12758412957191467,
0.016116591170430183,
0.059739742428064346,
0.10345818847417831,
0.02256437949836254,
0.09835100173950195,
0.12256097793579102,
0.012298434972763062,
0.09668286144733429,
-0.04513586312532425,
0.09081070125102997,
0.08970129489898682,
0.051308754831552505,
0.18703193962574005,
-0.07082065939903259,
-0.014739795587956905,
0.07899439334869385,
-0.06183239072561264,
0.08245434612035751,
0.0338808074593544,
0.05302000790834427,
-0.03171515092253685,
0.11643584817647934,
0.07975439727306366,
0.10020703822374344,
0.0683809220790863,
-0.07174939662218094,
0.03715403750538826,
-0.09468750655651093,
-0.060401562601327896,
0.035284675657749176,
-0.013843675144016743,
0.09547500312328339,
-0.13127243518829346,
0.09347659349441528,
0.04108694940805435,
0.15329043567180634,
0.054420825093984604,
-0.4413833022117615,
-0.0718417540192604,
0.03643173351883888,
0.04881865158677101,
-0.10270951688289642,
-0.02206386812031269,
0.04951032996177673,
-0.11133407056331635,
0.053175996989011765,
-0.05791350081562996,
0.09470754861831665,
-0.03447578102350235,
-0.006685248576104641,
-0.02449110709130764,
0.03918144479393959,
-0.03431959077715874,
0.027546579018235207,
-0.2012338638305664,
0.1652953177690506,
0.06868469715118408,
0.036516446620225906,
-0.08539480715990067,
0.0021113792899996042,
0.0753837451338768,
0.07305686920881271,
0.16993975639343262,
-0.012640469707548618,
-0.11378362029790878,
-0.3196440041065216,
-0.06386309862136841,
-0.04528114199638367,
0.04730058088898659,
0.006482335738837719,
0.0882972851395607,
-0.0670509785413742,
-0.029162198305130005,
0.030329754576086998,
-0.011056944727897644,
-0.12340373545885086,
-0.035188812762498856,
0.015559895895421505,
0.07815263420343399,
-0.0009110546088777483,
-0.07201138883829117,
-0.0943259596824646,
-0.08133085817098618,
0.13013066351413727,
-0.11245235055685043,
-0.02661404386162758,
-0.1907360553741455,
0.09636864811182022,
0.0857527032494545,
-0.08686043322086334,
0.05425069108605385,
-0.007687552832067013,
0.12711164355278015,
-0.004052229691296816,
-0.11347097158432007,
0.12939001619815826,
-0.09643863886594772,
-0.1153232529759407,
-0.03596153482794762,
0.031005818396806717,
-0.027365904301404953,
0.0450139120221138,
0.03532397001981735,
0.046250227838754654,
0.04274601861834526,
-0.07814384996891022,
0.06899183243513107,
0.025281986221671104,
0.15939971804618835,
0.08301931619644165,
-0.008289621211588383,
-0.13960367441177368,
-0.03290608897805214,
-0.011648924089968204,
0.09909789264202118,
0.28138071298599243,
-0.0892418697476387,
-0.05948929116129875,
0.04984069615602493,
-0.0234574843198061,
-0.2957831919193268,
0.09090621024370193,
-0.034384675323963165,
0.045927148312330246,
-0.019495997577905655,
0.010540908202528954,
0.14454634487628937,
0.188783660531044,
-0.034180425107479095,
0.03162682428956032,
-0.2380162924528122,
-0.1382056623697281,
0.16796983778476715,
0.13360927999019623,
0.12819869816303253,
-0.15046460926532745,
-0.0671684667468071,
-0.13628974556922913,
-0.15141531825065613,
0.043023888021707535,
-0.021587396040558815,
0.059669218957424164,
-0.06615184247493744,
-0.02281668782234192,
0.06050510331988335,
-0.08448860794305801,
0.13240601122379303,
-0.03067399188876152,
0.11762965470552444,
-0.0953141376376152,
-0.05421488359570503,
0.028159674257040024,
-0.07710088044404984,
0.12068916857242584,
0.031223565340042114,
0.05550181120634079,
-0.1365187168121338,
-0.018881944939494133,
-0.012464798986911774,
0.0700506791472435,
-0.046800922602415085,
-0.065359927713871,
-0.03931024298071861,
0.05975896120071411,
-0.06689782440662384,
-0.057644136250019073,
0.024408644065260887,
-0.03862040862441063,
0.06003411486744881,
0.18645191192626953,
0.09305048733949661,
-0.057838741689920425,
-0.019646020606160164,
0.01472681574523449,
-0.06775922328233719,
0.08145659416913986,
-0.038693126291036606,
0.0748472586274147,
0.12381551414728165,
0.02194143831729889,
0.13531886041164398,
0.06328987330198288,
-0.003240589052438736,
-0.02481227181851864,
0.07485663145780563,
-0.12790681421756744,
0.0010325354523956776,
-0.0409088209271431,
-0.009675813838839531,
-0.03246023505926132,
-0.02250874601304531,
0.12315483391284943,
-0.03760910406708717,
0.02413356862962246,
-0.0050312248058617115,
0.007631066720932722,
0.02547699585556984,
0.2450646311044693,
0.017457451671361923,
0.05461305007338524,
-0.09452013671398163,
0.11067105084657669,
0.05322312191128731,
-0.11262389272451401,
0.0007402468472719193,
0.04485147073864937,
-0.08266957849264145,
-0.02585407719016075,
0.040436383336782455,
0.21731747686862946,
-0.19325804710388184,
-0.05742446705698967,
-0.11535774916410446,
-0.15791763365268707,
0.04968559369444847,
0.24761119484901428,
0.09373758733272552,
0.009273241274058819,
0.04341926798224449,
-0.06815250962972641,
-0.13064567744731903,
0.08201169967651367,
0.09642405807971954,
0.06843049079179764,
-0.18150223791599274,
0.15911437571048737,
-0.05118501931428909,
-0.0026226004119962454,
-0.047064781188964844,
0.04618317633867264,
-0.17681483924388885,
-0.006093896459788084,
-0.1104121059179306,
0.06996200978755951,
-0.049681372940540314,
0.0026094731874763966,
-0.023685256019234657,
-0.004733086097985506,
-0.03672186657786369,
0.02330802008509636,
-0.036179669201374054,
-0.04588902369141579,
0.05103365331888199,
0.0459577701985836,
-0.07667393237352371,
-0.07041936367750168,
0.027386963367462158,
-0.030243946239352226,
0.049393899738788605,
0.043972015380859375,
0.08973787724971771,
-0.010874946601688862,
-0.05656520277261734,
-0.02114839293062687,
0.08357854932546616,
-0.0434424951672554,
0.06106093153357506,
-0.11480414122343063,
0.025409210473299026,
0.005878280848264694,
0.0130061199888587,
0.06074192374944687,
0.08463086187839508,
-0.11901088804006577,
-0.0010167408036068082,
-0.08590250462293625,
-0.05593834072351456,
-0.12337595224380493,
0.04737791046500206,
0.1892445981502533,
0.001220428035594523,
0.10972132533788681,
-0.1026109978556633,
-0.00036191288381814957,
-0.19642193615436554,
0.010946213267743587,
-0.04135293886065483,
-0.10320460051298141,
-0.08162891864776611,
0.02511102333664894,
0.0690954178571701,
-0.016317294910550117,
0.04184698313474655,
0.008013404905796051,
-0.006037273444235325,
0.06049860268831253,
0.09855661541223526,
-0.0024833306670188904,
0.007709651254117489,
0.19833238422870636,
0.07092055678367615,
-0.02124718949198723,
0.06425244361162186,
0.10571575164794922,
0.11733619123697281,
0.032920654863119125,
0.051913972944021225,
0.10543905198574066,
-0.08351536840200424,
0.1320190131664276,
0.05464273318648338,
-0.06313558667898178,
-0.08357607573270798,
0.09882859885692596,
-0.15121491253376007,
0.03690299764275551,
-0.05660327151417732,
0.005260472651571035,
0.19028925895690918,
-0.1080753356218338,
-0.006082620471715927,
-0.05791319161653519,
-0.05935584381222725,
-0.1377686709165573,
-0.11862426996231079,
-0.15012840926647186,
-0.040057841688394547,
0.01328703947365284,
-0.0792294591665268,
0.02372276782989502,
0.17446883022785187,
0.027679307386279106,
0.006174515001475811,
0.14622797071933746,
-0.2052328884601593,
0.016399787738919258,
0.03110211156308651,
-0.010159092955291271,
-0.06003498286008835,
-0.03592916205525398,
-0.05109844356775284,
0.04392629489302635,
0.011315582320094109,
0.02330837771296501,
-0.008809094317257404,
0.019051896408200264,
0.042615652084350586,
0.030684133991599083,
-0.10962540656328201,
-0.018168186768889427,
0.022368529811501503,
0.03715706616640091,
0.05541680008172989,
-0.003500893944874406,
0.05881669372320175,
-0.03454840928316116,
0.18791048228740692,
-0.0919618085026741,
0.0016323627205565572,
-0.08408328145742416,
0.16618989408016205,
0.004735098220407963,
0.08707711845636368,
-0.00005582227822742425,
-0.06559290736913681,
0.08801434189081192,
0.15241411328315735,
0.09896717220544815,
-0.039978910237550735,
-0.04154748469591141,
-0.015479637309908867,
-0.017820069566369057,
-0.06984778493642807,
0.09050577878952026,
0.020594662055373192,
0.04146885871887207,
-0.04469791054725647,
0.06313355267047882,
-0.02162550389766693,
-0.06214427947998047,
-0.01737087592482567,
0.05869792029261589,
5.102317004457291e-7,
0.04165583476424217,
-0.04695412889122963,
0.06157394126057625,
0.017519207671284676,
0.08876834064722061,
0.15301425755023956,
-0.11498237401247025,
-0.14541855454444885,
0.03323272243142128,
0.014358775690197945,
-0.041275959461927414,
0.07124044746160507,
-0.06408686935901642,
0.020786000415682793,
-0.0606737956404686,
-0.014748531393706799,
-0.15750567615032196,
-0.12054766714572906,
-0.02341954968869686,
0.13193616271018982,
0.2881123125553131,
0.0048997970297932625,
0.054274048656225204,
0.13362367451190948,
-0.026779336854815483,
-0.18606126308441162,
0.08931062370538712,
-0.029790308326482773,
-0.12396212667226791,
0.10207215696573257,
0.15368841588497162,
-0.026028456166386604,
0.19476710259914398,
-0.0004426687955856323,
-0.20213864743709564,
0.009588764980435371,
-0.019227098673582077,
-0.02178170345723629,
-0.05643894150853157,
0.032717470079660416,
-0.06212611868977547,
0.15439382195472717,
0.1279255896806717,
-0.023147011175751686,
-0.006907738745212555,
-0.05154537409543991,
0.0474972166121006,
0.021846896037459373,
0.04163600131869316,
0.01335784699767828,
-0.11839797347784042,
0.07579461485147476,
-0.11296864598989487,
-0.01972503587603569,
-0.29663482308387756,
-0.07041583210229874,
-0.05787850171327591,
-0.07347404211759567,
-0.07782746106386185,
0.0949511006474495,
0.13375407457351685,
0.043187450617551804,
-0.03907344490289688,
-0.2311243712902069,
0.046032752841711044,
0.1344984918832779,
-0.06980665773153305,
-0.11919830739498138
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | token-classification | sahillihas/merged-model-distil-bert-finetuned-ner | [
"transformers",
"safetensors",
"distilbert",
"token-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:46:30+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
49,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07577246427536011,
0.160971537232399,
-0.0037754587829113007,
0.025772565975785255,
0.11726456880569458,
0.008670859038829803,
0.07505112886428833,
0.10672148317098618,
-0.015180999413132668,
0.12574146687984467,
0.037852395325899124,
0.1016918197274208,
0.10858695209026337,
0.19223041832447052,
-0.0040494236163794994,
-0.20760315656661987,
0.0604725256562233,
-0.11741560697555542,
0.010131745599210262,
0.12369410693645477,
0.13811981678009033,
-0.11246351152658463,
0.06881340593099594,
-0.04209144040942192,
-0.01980483904480934,
-0.035129256546497345,
-0.06221386045217514,
-0.052587710320949554,
0.06488312780857086,
0.05438476800918579,
0.06559453904628754,
0.021042533218860626,
0.0840018093585968,
-0.28357821702957153,
0.019257022067904472,
0.0766754001379013,
0.002646347275003791,
0.06359819322824478,
0.07652021199464798,
-0.07693555951118469,
0.09431595355272293,
-0.059984829276800156,
0.15352454781532288,
0.07422540336847305,
-0.09718401730060577,
-0.18205241858959198,
-0.08698616921901703,
0.10743562132120132,
0.1786193698644638,
0.05337109789252281,
-0.03495080769062042,
0.14324238896369934,
-0.07000283896923065,
0.017874419689178467,
0.06666957587003708,
-0.0738137811422348,
-0.05611572042107582,
0.06171523407101631,
0.07419267296791077,
0.09766413271427155,
-0.13268914818763733,
-0.009328355081379414,
0.039757777005434036,
0.01763407699763775,
0.11028990149497986,
0.017873499542474747,
0.12974926829338074,
0.03081592544913292,
-0.1462060660123825,
-0.059808772057294846,
0.10660259425640106,
0.03442228585481644,
-0.05801061540842056,
-0.24857386946678162,
-0.00732324318960309,
-0.03398886322975159,
-0.03147710859775543,
-0.045992717146873474,
0.04087558388710022,
-0.027321036905050278,
0.08885861188173294,
0.0030806211289018393,
-0.06717013567686081,
-0.05033329501748085,
0.09465020149946213,
0.06679657101631165,
0.026782603934407234,
-0.028168529272079468,
0.01867647096514702,
0.11993368715047836,
0.10629972815513611,
-0.11282733827829361,
-0.06030019000172615,
-0.06428343802690506,
-0.08693956583738327,
-0.04836411029100418,
0.03905215859413147,
0.0691521167755127,
0.05319429561495781,
0.20513659715652466,
-0.005697339307516813,
0.04798254370689392,
0.03319632261991501,
0.012943265959620476,
0.06999380886554718,
0.07388506829738617,
-0.057592663913965225,
-0.13561639189720154,
-0.032188303768634796,
0.1180354431271553,
0.0034096194431185722,
-0.029926057904958725,
-0.03362307325005531,
0.061736419796943665,
0.04901990294456482,
0.12567868828773499,
0.06872507929801941,
0.016199197620153427,
-0.07887519896030426,
-0.05266522243618965,
0.18039943277835846,
-0.15842315554618835,
0.026145054027438164,
0.01778857596218586,
-0.04808545112609863,
-0.03221158683300018,
0.01776694506406784,
0.010511110536754131,
-0.027442574501037598,
0.09513456374406815,
-0.064456507563591,
-0.0437372550368309,
-0.10800766944885254,
-0.05169657990336418,
0.03205830603837967,
-0.02027302049100399,
-0.02823185920715332,
-0.039158232510089874,
-0.12495775520801544,
-0.0780569463968277,
0.07112475484609604,
-0.06458895653486252,
-0.06230264902114868,
-0.035103172063827515,
-0.06434695422649384,
0.01328826043754816,
-0.0016070485580712557,
0.11751557141542435,
-0.02924034371972084,
0.05098489671945572,
-0.058480363339185715,
0.06735116988420486,
0.13876435160636902,
0.030480407178401947,
-0.06721892952919006,
0.06733744591474533,
-0.21017982065677643,
0.10650269687175751,
-0.0903177410364151,
0.030309993773698807,
-0.16417838633060455,
-0.020649932324886322,
0.03423565998673439,
0.03357144445180893,
-0.011650003492832184,
0.14477767050266266,
-0.17939403653144836,
-0.035702068358659744,
0.18933741748332977,
-0.12515437602996826,
-0.092041015625,
0.05768411606550217,
-0.06046192720532417,
0.13482485711574554,
0.05646788328886032,
-0.021687285974621773,
0.05586513131856918,
-0.13862864673137665,
-0.02437570132315159,
-0.06216869503259659,
-0.014887474477291107,
0.1528262197971344,
0.059848666191101074,
-0.046563770622015,
0.02752561867237091,
0.017916126176714897,
-0.026528963819146156,
-0.04967709258198738,
-0.03546646237373352,
-0.09774158149957657,
0.00822865217924118,
-0.07969193160533905,
0.02241206355392933,
-0.022354070097208023,
-0.09176263213157654,
-0.03786236792802811,
-0.15299317240715027,
0.01120391022413969,
0.09777288138866425,
-0.0018325879937037826,
-0.030514128506183624,
-0.09985145181417465,
-0.0009975290158763528,
0.014818643219769001,
-0.006299570668488741,
-0.15255902707576752,
-0.05487200245261192,
0.02453494630753994,
-0.16442841291427612,
0.027572009712457657,
-0.04693130776286125,
0.03860992193222046,
0.04300114884972572,
-0.04640088230371475,
-0.03466949611902237,
0.020573051646351814,
0.021380405873060226,
-0.02527492307126522,
-0.25819748640060425,
-0.015081126242876053,
-0.050718042999506,
0.1735834777355194,
-0.24959886074066162,
0.04765407368540764,
0.0635317862033844,
0.12164347618818283,
0.008648045361042023,
-0.043875813484191895,
0.03864365443587303,
-0.0540064312517643,
-0.03614342212677002,
-0.06847652792930603,
-0.006085617933422327,
-0.03506658226251602,
-0.04601104184985161,
0.036002401262521744,
-0.17769227921962738,
-0.026747822761535645,
0.11292517185211182,
0.07372257113456726,
-0.16646650433540344,
-0.07437358796596527,
-0.03605179488658905,
-0.05963003262877464,
-0.07967717945575714,
-0.05352560430765152,
0.08457031846046448,
0.047888897359371185,
0.050600916147232056,
-0.06766360253095627,
-0.05917840451002121,
0.012786190956830978,
-0.014223156496882439,
-0.030130434781312943,
0.08920878916978836,
0.06863047182559967,
-0.13044358789920807,
0.10779033601284027,
0.07565328478813171,
0.07318258285522461,
0.10535803437232971,
0.004109293222427368,
-0.09151213616132736,
-0.019049784168601036,
0.030527541413903236,
0.01623775251209736,
0.14909477531909943,
-0.058906275779008865,
0.037276990711688995,
0.042408693581819534,
-0.025859607383608818,
0.007144377566874027,
-0.09663669019937515,
0.020623620599508286,
0.027944007888436317,
-0.010041053406894207,
0.02522680163383484,
-0.05456255003809929,
0.016582027077674866,
0.10647457838058472,
0.0315246656537056,
0.029977189376950264,
0.01468842476606369,
-0.04244951531291008,
-0.1256953328847885,
0.17814458906650543,
-0.09753371775150299,
-0.24721357226371765,
-0.12567918002605438,
-0.009092925116419792,
0.03918062895536423,
-0.012046195566654205,
0.01987382024526596,
-0.05762042477726936,
-0.10939344018697739,
-0.1005193367600441,
0.03179878741502762,
0.06615057587623596,
-0.08552238345146179,
-0.0674377977848053,
0.05327526107430458,
0.04219662770628929,
-0.1277419775724411,
0.020916761830449104,
0.041558679193258286,
-0.07218944281339645,
0.008345745503902435,
0.056684307754039764,
0.07901664078235626,
0.1805250197649002,
0.011102218180894852,
-0.018453268334269524,
0.012500734999775887,
0.21901123225688934,
-0.1455126702785492,
0.09525687992572784,
0.13985903561115265,
-0.06373906135559082,
0.08310941606760025,
0.20266622304916382,
0.03211383894085884,
-0.10105792433023453,
0.03751940280199051,
0.03433764725923538,
-0.03450850397348404,
-0.24166366457939148,
-0.07809934765100479,
0.0032382020726799965,
-0.06378763914108276,
0.10437320917844772,
0.08564607799053192,
0.10547725111246109,
0.04683990776538849,
-0.11285781115293503,
-0.06474319845438004,
0.05201273784041405,
0.11905202269554138,
-0.028264760971069336,
-0.0015284299151971936,
0.09661765396595001,
-0.023179147392511368,
0.024211812764406204,
0.091147281229496,
0.024589717388153076,
0.18455083668231964,
0.04622681811451912,
0.13505716621875763,
0.0910676121711731,
0.06324100494384766,
0.014886823482811451,
0.021261626854538918,
0.01765025220811367,
0.02958594262599945,
-0.018347864970564842,
-0.08319743722677231,
-0.009293846786022186,
0.13596801459789276,
0.024112429469823837,
0.04030708968639374,
0.006161263212561607,
-0.048370931297540665,
0.07493776082992554,
0.17692242562770844,
0.014188401401042938,
-0.22553029656410217,
-0.06700198352336884,
0.07436924427747726,
-0.07410255074501038,
-0.11821255087852478,
-0.01875862292945385,
0.02883634716272354,
-0.18362410366535187,
0.03783649206161499,
-0.027923477813601494,
0.10164659470319748,
-0.11413569748401642,
-0.021443994715809822,
0.03836014121770859,
0.05620839074254036,
-0.032449860125780106,
0.07187992334365845,
-0.18838849663734436,
0.13979743421077728,
0.009049986489117146,
0.06722430139780045,
-0.1007412001490593,
0.08134089410305023,
0.017924414947628975,
0.005374805070459843,
0.16363629698753357,
-0.0033102943561971188,
-0.061859820038080215,
-0.10252897441387177,
-0.08680899441242218,
-0.015643585473299026,
0.09927050769329071,
-0.12645161151885986,
0.09391359239816666,
-0.00760313356295228,
-0.03509579226374626,
-0.0025038165040314198,
-0.13591554760932922,
-0.13381120562553406,
-0.17948365211486816,
0.04688849300146103,
-0.12670443952083588,
0.04449006915092468,
-0.1076912134885788,
-0.056078001856803894,
-0.041555143892765045,
0.19199959933757782,
-0.21791574358940125,
-0.08265700191259384,
-0.15180857479572296,
-0.06470925360918045,
0.11797353625297546,
-0.04333622008562088,
0.08231177926063538,
0.009503866545855999,
0.19813112914562225,
-0.0016607206780463457,
-0.004083174746483564,
0.0959770604968071,
-0.09965015947818756,
-0.2107795774936676,
-0.09851331263780594,
0.13872341811656952,
0.13654115796089172,
0.04202498495578766,
0.0035957242362201214,
0.026746563613414764,
-0.003834358649328351,
-0.1127103865146637,
0.03198874741792679,
0.15448687970638275,
0.11144908517599106,
0.038192201405763626,
-0.028366046026349068,
-0.13728831708431244,
-0.10052017867565155,
-0.04788326099514961,
0.009005418047308922,
0.19273622334003448,
-0.07073298841714859,
0.1647789627313614,
0.1582319587469101,
-0.06154803931713104,
-0.2109813690185547,
0.03008965775370598,
0.03392798826098442,
-0.003125392831861973,
0.04911292716860771,
-0.2020125389099121,
0.08147795498371124,
0.014267290011048317,
-0.05861039087176323,
0.12597185373306274,
-0.17893800139427185,
-0.1463114470243454,
0.0894082561135292,
0.07762204855680466,
-0.19969920814037323,
-0.1310804784297943,
-0.09587812423706055,
-0.04523595795035362,
-0.09975142776966095,
0.09141786396503448,
-0.006663811393082142,
0.00584616232663393,
0.03217587620019913,
0.01619444414973259,
0.016584929078817368,
-0.04955518618226051,
0.19393765926361084,
-0.0010097876656800508,
0.049837395548820496,
-0.07284197211265564,
-0.07579720765352249,
0.03499779477715492,
-0.07016132026910782,
0.08607465773820877,
-0.017582224681973457,
0.006899731699377298,
-0.11493121087551117,
-0.06381088495254517,
-0.04385467618703842,
0.0321214497089386,
-0.08662021160125732,
-0.09713311493396759,
-0.04749152809381485,
0.105962373316288,
0.0914415493607521,
-0.03657814860343933,
-0.06360591948032379,
-0.0900130569934845,
0.052277419716119766,
0.22165511548519135,
0.18128447234630585,
0.0717686340212822,
-0.07471393048763275,
-0.008312922902405262,
-0.021178819239139557,
0.05892517790198326,
-0.20861060917377472,
0.049327246844768524,
0.03792690858244896,
0.03388948366045952,
0.12989014387130737,
-0.025939077138900757,
-0.16148367524147034,
-0.052209001034498215,
0.05540921539068222,
-0.07436220347881317,
-0.16007542610168457,
0.006253175903111696,
0.08246254920959473,
-0.1577574759721756,
-0.040928490459918976,
0.0384182333946228,
-0.029118487611413002,
-0.031075680628418922,
0.0023355213925242424,
0.08367416262626648,
0.021480288356542587,
0.10684667527675629,
0.06465408951044083,
0.10775714367628098,
-0.10543544590473175,
0.06850111484527588,
0.08449427038431168,
-0.10566192865371704,
0.03645234927535057,
0.059928081929683685,
-0.06519834697246552,
-0.03475077450275421,
0.03517557680606842,
0.08707848936319351,
0.02939283289015293,
-0.06929907202720642,
0.006287590134888887,
-0.10996777564287186,
0.06547177582979202,
0.13484768569469452,
0.04042261838912964,
0.011329763568937778,
0.040748391300439835,
0.030845293775200844,
-0.09934883564710617,
0.12121784687042236,
0.05189082771539688,
0.03800792247056961,
-0.0530884675681591,
-0.017324361950159073,
0.03913010656833649,
-0.019996602088212967,
-0.017436852678656578,
-0.03976694867014885,
-0.07124674320220947,
-0.011837370693683624,
-0.16637970507144928,
0.020963486284017563,
-0.06557925045490265,
0.011502030305564404,
0.015106898732483387,
-0.03105640783905983,
0.00648743100464344,
0.012552679516375065,
-0.07388464361429214,
-0.045671455562114716,
-0.00442871730774641,
0.10947413742542267,
-0.16812850534915924,
0.008818828500807285,
0.08374110609292984,
-0.12818290293216705,
0.08439380675554276,
0.0020208086352795362,
-0.0032075955532491207,
0.02065693773329258,
-0.1317322701215744,
0.05960160121321678,
-0.006254117004573345,
0.006569796707481146,
0.03298337012529373,
-0.21374841034412384,
0.0022494937293231487,
-0.05006609857082367,
-0.06266371160745621,
-0.0007477401522919536,
-0.03451014682650566,
-0.11394216865301132,
0.10471237450838089,
0.014338361099362373,
-0.07764562219381332,
-0.021158380433917046,
0.048064179718494415,
0.10854631662368774,
-0.04941081255674362,
0.1473926305770874,
-0.01730068400502205,
0.059951625764369965,
-0.1836392730474472,
-0.02037329412996769,
-0.018616020679473877,
0.016425784677267075,
-0.037237949669361115,
-0.006127772852778435,
0.051674678921699524,
-0.01703222282230854,
0.2173253446817398,
-0.022195780649781227,
0.026851661503314972,
0.06280741095542908,
-0.002322237938642502,
-0.013443386182188988,
0.09835761785507202,
0.046587999910116196,
0.013757294043898582,
0.022333940491080284,
0.010798705741763115,
-0.04187694564461708,
-0.00982733629643917,
-0.13689175248146057,
0.07585512101650238,
0.16693128645420074,
0.08301392197608948,
-0.00876491330564022,
0.04851881042122841,
-0.10903714597225189,
-0.10301291197538376,
0.09611163288354874,
-0.03860481455922127,
-0.017003152519464493,
-0.050830189138650894,
0.13978728652000427,
0.15471777319908142,
-0.19259576499462128,
0.06503742933273315,
-0.0670342966914177,
-0.05641358345746994,
-0.10491608083248138,
-0.17902258038520813,
-0.06081055849790573,
-0.037293195724487305,
-0.01373780332505703,
-0.06118905171751976,
0.06020566076040268,
0.10252559930086136,
0.014830772764980793,
0.008825747296214104,
0.082810178399086,
-0.027097169309854507,
0.007841658778488636,
0.04027826339006424,
0.06163087487220764,
0.015406538732349873,
-0.06096850335597992,
0.008726970292627811,
0.0014340170891955495,
0.03696046769618988,
0.053261034190654755,
0.03251050412654877,
-0.012918577529489994,
0.006573197897523642,
-0.023761101067066193,
-0.10426638275384903,
0.04047070071101189,
-0.02524152584373951,
-0.052286889404058456,
0.15217694640159607,
0.022632325068116188,
-0.005827770568430424,
-0.022215187549591064,
0.2338925302028656,
-0.06722013652324677,
-0.0785137191414833,
-0.14054499566555023,
0.14897066354751587,
-0.03910091519355774,
0.05552518740296364,
0.046295974403619766,
-0.105179063975811,
0.03945622220635414,
0.13950063288211823,
0.14286132156848907,
-0.04025404527783394,
0.012145495042204857,
0.009853771887719631,
0.0038128900341689587,
-0.028584260493516922,
0.052456315606832504,
0.05106692388653755,
0.12483679503202438,
-0.06253138184547424,
0.09782347828149796,
-0.006969998124986887,
-0.09410524368286133,
-0.025307200849056244,
0.13257914781570435,
0.002008436480537057,
0.023570919409394264,
-0.0815717875957489,
0.12813320755958557,
-0.05384290590882301,
-0.2535746693611145,
0.07024679332971573,
-0.06422019749879837,
-0.15186867117881775,
-0.019949471578001976,
0.020098339766263962,
-0.004076872952282429,
0.023379439488053322,
0.06391935795545578,
-0.0649435967206955,
0.15813666582107544,
0.03603892773389816,
-0.06333170086145401,
-0.07421127706766129,
0.07901927083730698,
-0.07848652452230453,
0.3060082495212555,
0.007198201958090067,
0.05301123857498169,
0.09581229835748672,
-0.03854295611381531,
-0.13927392661571503,
0.028419770300388336,
0.09097424894571304,
-0.05137800797820091,
0.061749134212732315,
0.2061481475830078,
-0.010625248774886131,
0.11590193957090378,
0.07372608780860901,
-0.08649783581495285,
0.04505147039890289,
-0.09321032464504242,
-0.09480153769254684,
-0.08939364552497864,
0.09279339760541916,
-0.055773381143808365,
0.1508018523454666,
0.1253325492143631,
-0.048127513378858566,
0.008756249211728573,
-0.019953599199652672,
0.058552540838718414,
0.0011799887288361788,
0.11399417370557785,
0.028718046844005585,
-0.19564524292945862,
0.02838771604001522,
-0.0018802561098709702,
0.1010337695479393,
-0.23878240585327148,
-0.0895518958568573,
0.04254518821835518,
-0.0008608286152593791,
-0.060450565069913864,
0.12522582709789276,
0.05044018104672432,
0.043433722108602524,
-0.0508544035255909,
-0.04669932648539543,
-0.008167625404894352,
0.1615993231534958,
-0.10094329714775085,
-0.0017090580658987164
] |
null | null | null |

MHENN cat (generated by pixlr)
the primary base of all models is mistral-instruct-v0.1
this is a continuous fine-tune of mistral and is MHENN4 fine-tuned further. all models are continuous finetunes and started at the original MHENN version
| {"language": ["en"], "license": "cc", "tags": ["biology", "chemistry", "code", "text-generation-inference"], "datasets": ["netcat420/MHENN5"], "pipeline_tag": "text-generation"} | text-generation | netcat420/MHENN5-GGUF | [
"gguf",
"biology",
"chemistry",
"code",
"text-generation-inference",
"text-generation",
"en",
"dataset:netcat420/MHENN5",
"license:cc",
"region:us"
] | 2024-02-08T16:49:15+00:00 | [] | [
"en"
] | TAGS
#gguf #biology #chemistry #code #text-generation-inference #text-generation #en #dataset-netcat420/MHENN5 #license-cc #region-us
|
!image/png
MHENN cat (generated by pixlr)
the primary base of all models is mistral-instruct-v0.1
this is a continuous fine-tune of mistral and is MHENN4 fine-tuned further. all models are continuous finetunes and started at the original MHENN version
| [] | [
"TAGS\n#gguf #biology #chemistry #code #text-generation-inference #text-generation #en #dataset-netcat420/MHENN5 #license-cc #region-us \n"
] | [
51
] | [
"passage: TAGS\n#gguf #biology #chemistry #code #text-generation-inference #text-generation #en #dataset-netcat420/MHENN5 #license-cc #region-us \n"
] | [
-0.016221115365624428,
0.1238027885556221,
-0.004739625379443169,
-0.044104333966970444,
-0.018346818163990974,
0.030840421095490456,
0.13281035423278809,
0.11633479595184326,
0.13845060765743256,
-0.0003858512209262699,
0.16400760412216187,
0.020464975386857986,
-0.006401718594133854,
0.0920034646987915,
0.0023803601507097483,
-0.09592459350824356,
0.011132396757602692,
-0.010952835902571678,
-0.09813324362039566,
0.04900188371539116,
0.06520470231771469,
-0.030652707442641258,
0.11246567964553833,
-0.06861812621355057,
-0.06558401882648468,
-0.007103933487087488,
0.03436315059661865,
-0.04998191446065903,
0.12163586914539337,
0.031017299741506577,
0.024680281057953835,
0.08952276408672333,
-0.06758309155702591,
-0.23567040264606476,
0.044440075755119324,
-0.07522372901439667,
-0.11760903149843216,
0.11734045296907425,
0.03906969726085663,
-0.0026718757580965757,
0.19929930567741394,
0.02327505685389042,
-0.09341663122177124,
0.11337790638208389,
-0.13757705688476562,
0.03914762660861015,
-0.035480987280607224,
-0.051294952630996704,
-0.10184437036514282,
0.032832685858011246,
0.02598545327782631,
0.022008316591382027,
-0.07946090400218964,
0.029273396357893944,
0.15773817896842957,
-0.2707078456878662,
0.06406226754188538,
0.28627336025238037,
0.1087658628821373,
0.10343225300312042,
-0.07797892391681671,
0.12280594557523727,
0.025645151734352112,
-0.04835188761353493,
-0.02105650119483471,
-0.09337383508682251,
-0.016219226643443108,
0.0738174244761467,
-0.05011913180351257,
0.0026848111301660538,
0.28270140290260315,
-0.016271214932203293,
0.021543346345424652,
0.040542151778936386,
-0.08263133466243744,
-0.014357651583850384,
-0.005261463578790426,
0.021606219932436943,
0.0028351356741040945,
0.08436392992734909,
0.10322818160057068,
-0.05924849584698677,
-0.08885549753904343,
-0.021193111315369606,
-0.12628498673439026,
0.06651677936315536,
-0.07835836708545685,
0.028591608628630638,
-0.12321890145540237,
-0.008530806750059128,
-0.02965325303375721,
-0.06394637376070023,
-0.026318244636058807,
-0.05873574689030647,
0.08378542959690094,
-0.0025667084846645594,
0.0034285897854715586,
-0.0021661650389432907,
0.23733235895633698,
0.023197265341877937,
-0.13525091111660004,
0.04075044393539429,
-0.034059859812259674,
0.10795879364013672,
0.0524970144033432,
-0.12919607758522034,
-0.02587972953915596,
0.02880020998418331,
0.06609899550676346,
-0.10584130883216858,
-0.05252813920378685,
-0.050883788615465164,
-0.10301447659730911,
-0.015143177472054958,
0.019530804827809334,
0.07210540771484375,
0.0259800236672163,
-0.027658207342028618,
-0.006743696518242359,
-0.0117034325376153,
0.06271951645612717,
-0.00728458259254694,
-0.01610819809138775,
0.13663487136363983,
0.024552861228585243,
-0.00481015769764781,
-0.06791198998689651,
0.02333308383822441,
0.039165593683719635,
-0.07912008464336395,
-0.17755931615829468,
-0.0751681849360466,
-0.006993853487074375,
-0.019501078873872757,
0.08550111949443817,
0.09062555432319641,
0.03456203639507294,
-0.08539406955242157,
-0.07941177487373352,
0.01495546754449606,
0.04231831803917885,
-0.03659367561340332,
0.02928725630044937,
0.03385315090417862,
-0.03676292300224304,
0.044977352023124695,
-0.017976412549614906,
0.04005223512649536,
-0.09093999862670898,
0.07971753925085068,
-0.07390619814395905,
0.05974924936890602,
-0.16550688445568085,
-0.006721277721226215,
-0.08289837092161179,
0.05775761604309082,
0.001658217515796423,
-0.022167421877384186,
-0.12407092750072479,
0.052703484892845154,
-0.03881790488958359,
-0.03002600185573101,
-0.05905066058039665,
-0.05387145280838013,
-0.003966398071497679,
0.18959513306617737,
-0.08302263170480728,
-0.090504951775074,
0.11521093547344208,
-0.01633444055914879,
-0.12051025032997131,
0.025644946843385696,
0.023673715069890022,
0.012172484770417213,
0.03441055119037628,
0.32033658027648926,
0.0509444922208786,
-0.06258250027894974,
-0.00833246298134327,
0.15133044123649597,
0.023193195462226868,
-0.11507949978113174,
0.10297660529613495,
0.001297518378123641,
-0.13332577049732208,
0.03044590912759304,
0.014141879044473171,
0.07178879529237747,
-0.059765223413705826,
-0.09893783181905746,
0.00006691217276966199,
0.008642952889204025,
0.009272697381675243,
-0.04012777656316757,
0.0841357633471489,
-0.04729427024722099,
0.06502688676118851,
-0.06586921960115433,
0.01172174233943224,
-0.0055734519846737385,
0.010220830328762531,
0.006012409459799528,
0.1540450006723404,
0.08368253707885742,
0.025642501190304756,
-0.05344076082110405,
-0.09820256382226944,
0.037594523280858994,
-0.04217902198433876,
0.04201631620526314,
-0.004780206363648176,
0.006388242822140455,
0.0373503677546978,
-0.009122681804001331,
0.07896435260772705,
-0.014130004681646824,
0.021695492789149284,
-0.04208473861217499,
-0.18594910204410553,
0.09970095008611679,
-0.011836672201752663,
0.07474903762340546,
-0.14554515480995178,
-0.0034799000713974237,
0.21944738924503326,
0.004027981776744127,
-0.018113965168595314,
-0.03700917586684227,
0.05948156490921974,
-0.048070430755615234,
-0.03636956214904785,
0.028164898976683617,
0.08834368735551834,
-0.03969418630003929,
-0.1803702414035797,
0.07110278308391571,
-0.019435590133070946,
0.15013206005096436,
0.11821234226226807,
-0.1470348984003067,
0.08381456881761551,
-0.16645365953445435,
-0.056636132299900055,
0.04089213162660599,
0.03583185747265816,
0.0004973368486389518,
-0.01583036221563816,
-0.059895601123571396,
0.03636340796947479,
-0.061696652323007584,
0.009374726563692093,
0.018412763252854347,
-0.017391959205269814,
-0.04141020029783249,
0.11850176751613617,
0.12356588244438171,
-0.11604075133800507,
0.14489243924617767,
0.17266657948493958,
0.14458444714546204,
0.22483669221401215,
-0.018995506688952446,
-0.010419189929962158,
-0.023222722113132477,
0.004597872961312532,
-0.035527583211660385,
0.1337641030550003,
-0.118270143866539,
0.04278586059808731,
0.06711100041866302,
0.03350680321455002,
0.1064692959189415,
-0.10644888132810593,
-0.0883464589715004,
-0.061046794056892395,
-0.05332091823220253,
-0.07810142636299133,
-0.011860722675919533,
-0.07434389740228653,
0.14672526717185974,
0.010601299814879894,
-0.056109145283699036,
0.035612139850854874,
-0.002097743097692728,
-0.13421018421649933,
0.2804914712905884,
-0.12892276048660278,
-0.19788485765457153,
-0.099059097468853,
0.0006926135392859578,
-0.05953652039170265,
0.02356567606329918,
0.01373133435845375,
-0.09919602423906326,
-0.022917134687304497,
-0.024273794144392014,
-0.04290785640478134,
-0.08926507085561752,
-0.006298506632447243,
-0.013562068343162537,
0.07064664363861084,
-0.057049255818128586,
-0.11136846244335175,
-0.0517142154276371,
-0.021889299154281616,
0.02787886932492256,
0.14483025670051575,
-0.10444004088640213,
0.09960386157035828,
0.17474566400051117,
0.0056801107712090015,
-0.017657699063420296,
-0.05213159695267677,
0.282275915145874,
-0.04505401477217674,
-0.018621686846017838,
0.062027789652347565,
0.053325992077589035,
0.04648401588201523,
0.0718025490641594,
0.09583976119756699,
-0.12262662500143051,
0.015099829062819481,
-0.05128327012062073,
-0.11586473137140274,
-0.3042396903038025,
-0.11541830748319626,
-0.08415443450212479,
0.13556872308254242,
-0.10579896718263626,
0.05976872146129608,
0.2107986956834793,
0.00455491105094552,
0.023704396560788155,
-0.012660733424127102,
0.007237887475639582,
0.013199927285313606,
0.1345032900571823,
-0.022294213995337486,
0.009529072791337967,
-0.012591059319674969,
0.029301924630999565,
0.11803290992975235,
0.15398059785366058,
0.1501573771238327,
0.20606599748134613,
0.25042420625686646,
0.07748863101005554,
0.022612314671278,
0.11228056252002716,
0.030373377725481987,
-0.028415706008672714,
-0.0027752115856856108,
-0.02904517576098442,
-0.09264213591814041,
0.05287958309054375,
0.04291348159313202,
-0.11308416724205017,
-0.17520800232887268,
0.014525090344250202,
-0.05121215432882309,
0.08937516808509827,
-0.08962414413690567,
0.04239796847105026,
-0.011775363236665726,
0.04023202881217003,
0.09137767553329468,
0.08419973403215408,
-0.02491173893213272,
0.10210184752941132,
0.03993679955601692,
-0.047712743282318115,
-0.001328612444922328,
-0.02774903178215027,
0.07644391804933548,
0.040108609944581985,
0.0836169421672821,
-0.07937736809253693,
-0.13334093987941742,
-0.026491811498999596,
0.1298777163028717,
-0.3014437258243561,
0.24689419567584991,
0.037269674241542816,
-0.08748982846736908,
-0.05507713183760643,
-0.0625634491443634,
-0.031339772045612335,
0.12104426324367523,
0.13536566495895386,
0.09075427800416946,
-0.13942116498947144,
-0.03129452094435692,
-0.11746026575565338,
0.030194411054253578,
0.10389906167984009,
0.007805141620337963,
-0.016401631757616997,
0.038868289440870285,
0.031648240983486176,
-0.007841578684747219,
0.053448207676410675,
-0.17179450392723083,
-0.06057571619749069,
0.13586068153381348,
-0.028246792033314705,
0.04415978118777275,
-0.029216615483164787,
-0.01100489404052496,
0.011784864589571953,
0.20330512523651123,
-0.3171374201774597,
-0.06133411079645157,
-0.12602481245994568,
-0.04483227804303169,
0.05413725599646568,
-0.050411518663167953,
-0.016117190942168236,
-0.03006265126168728,
-0.03434659540653229,
-0.12310726195573807,
-0.23617671430110931,
0.19122456014156342,
-0.044249143451452255,
-0.08593367040157318,
-0.04834254831075668,
0.14333388209342957,
0.0005731327692046762,
0.12757866084575653,
0.019860759377479553,
0.03826369345188141,
-0.09837817400693893,
-0.15292982757091522,
0.11417632550001144,
-0.11090333014726639,
-0.04501255974173546,
-0.023507971316576004,
-0.11334193497896194,
-0.003718484891578555,
0.036924172192811966,
-0.07727467268705368,
0.23925769329071045,
0.29084521532058716,
-0.005096937529742718,
0.1823001205921173,
0.27507519721984863,
-0.14538782835006714,
-0.24580204486846924,
-0.018343161791563034,
-0.1651555448770523,
-0.031861498951911926,
-0.03918563947081566,
-0.33861470222473145,
0.09712637215852737,
0.13781921565532684,
-0.03694858402013779,
0.2748470902442932,
-0.26132893562316895,
-0.10054344683885574,
0.01222703605890274,
0.007237367331981659,
0.3419612646102905,
-0.21144451200962067,
-0.11603666096925735,
-0.06513828039169312,
-0.1898423284292221,
0.24129553139209747,
-0.1143738403916359,
0.10064928233623505,
-0.03237977996468544,
0.02396467886865139,
-0.030596164986491203,
-0.07128838449716568,
0.1747552454471588,
0.07378000020980835,
0.026966480538249016,
0.011905455961823463,
-0.1449420303106308,
0.1821395456790924,
0.11082306504249573,
-0.021957335993647575,
-0.07014746963977814,
0.012333258986473083,
-0.17133478820323944,
0.019463062286376953,
-0.08664590865373611,
0.15460224449634552,
-0.04377417638897896,
-0.15566283464431763,
-0.14115183055400848,
0.0260586179792881,
-0.031534481793642044,
-0.004453498404473066,
0.24111324548721313,
-0.04026948660612106,
0.1319984793663025,
0.19115324318408966,
0.11394626647233963,
-0.1523655652999878,
0.10124056041240692,
0.04281030595302582,
-0.03584253415465355,
0.0486263781785965,
-0.19168581068515778,
-0.03351970762014389,
0.1499890387058258,
0.04281656816601753,
0.04459967091679573,
0.08367227762937546,
-0.06707806140184402,
0.01321006752550602,
0.11068189144134521,
-0.26207754015922546,
-0.06449555605649948,
-0.08966286480426788,
-0.008918320760130882,
0.06624829024076462,
0.00623059319332242,
0.11002833396196365,
0.005764349363744259,
-0.01635845936834812,
-0.006785365752875805,
0.05499672144651413,
-0.0779227539896965,
0.11004547029733658,
0.14116305112838745,
0.03138231858611107,
-0.1001719981431961,
0.018754133954644203,
0.06333979964256287,
0.029789062216877937,
-0.012105606496334076,
0.13937069475650787,
-0.11666101217269897,
-0.07984291762113571,
-0.09160231798887253,
0.08449683338403702,
-0.19597913324832916,
0.0178676787763834,
-0.07167703658342361,
-0.1206507459282875,
0.029267340898513794,
0.18059268593788147,
0.03315553069114685,
0.11826761066913605,
0.005426662042737007,
-0.03049197979271412,
0.09653694927692413,
0.026339882984757423,
-0.23906295001506805,
0.039154909551143646,
0.01783420331776142,
-0.010957246646285057,
-0.06923726946115494,
0.08859849721193314,
-0.08790260553359985,
-0.031020892783999443,
-0.2131602019071579,
-0.027687089517712593,
-0.03979065269231796,
-0.04011128470301628,
-0.08875452727079391,
-0.08480491489171982,
-0.07245317101478577,
0.012498890981078148,
-0.04137305170297623,
0.013751511462032795,
-0.06749176234006882,
0.0018609694670885801,
-0.06912892311811447,
0.09470707923173904,
-0.10985460877418518,
-0.019862886518239975,
0.060827210545539856,
0.010137319564819336,
0.10233072936534882,
0.060558103024959564,
0.05719166249036789,
0.08523079007863998,
-0.25380662083625793,
0.008540756069123745,
0.06301940232515335,
0.0542362779378891,
-0.029538361355662346,
-0.08134771883487701,
-0.03739827871322632,
0.008819217793643475,
0.016331633552908897,
0.05882049724459648,
-0.0687306597828865,
-0.09324274212121964,
-0.09125349670648575,
-0.14798757433891296,
-0.10211686044931412,
-0.004092260729521513,
-0.042749956250190735,
0.12159188091754913,
-0.06006694957613945,
0.01421920582652092,
0.016728827729821205,
-0.002955499105155468,
-0.10241873562335968,
0.03679869696497917,
0.025947030633687973,
-0.15565943717956543,
-0.18377813696861267,
-0.04382495582103729,
0.012717794626951218,
-0.027153542265295982,
0.3071739971637726,
0.018800072371959686,
-0.13464955985546112,
0.057895660400390625,
0.07717162370681763,
0.07916785031557083,
0.036810290068387985,
0.2873179316520691,
0.12214276939630508,
-0.036883946508169174,
-0.004350453149527311,
0.0698120966553688,
-0.016407454386353493,
0.05422041937708855,
0.1988236904144287,
0.06398841738700867,
0.02760128118097782,
-0.009531398303806782,
0.036621857434511185,
-0.048483069986104965,
0.007475048303604126,
0.03890637308359146,
0.04330817237496376,
0.02807278372347355,
-0.03380899131298065,
0.04817608743906021,
0.18889471888542175,
-0.15195055305957794,
0.08048873394727707,
-0.03781668841838837,
-0.10068557411432266,
-0.13294202089309692,
-0.07990942895412445,
-0.04863140732049942,
-0.06676718592643738,
-0.016988888382911682,
-0.17432823777198792,
0.0009456602274440229,
0.044601477682590485,
0.056219447404146194,
-0.031218646094202995,
0.05765550956130028,
0.055218979716300964,
-0.04923626035451889,
0.05840783566236496,
-0.018311668187379837,
-0.0032786759547889233,
-0.137933611869812,
-0.04903699830174446,
0.009374438785016537,
-0.05905408784747124,
-0.0299224890768528,
0.04121382161974907,
0.012905307114124298,
-0.015148365870118141,
-0.09607250988483429,
-0.06722058355808258,
-0.0746963620185852,
-0.01584736257791519,
-0.03319759666919708,
0.11259081214666367,
-0.02126235142350197,
-0.04908528923988342,
0.08165793120861053,
0.12244649231433868,
0.024937864392995834,
0.02757955528795719,
0.07492898404598236,
0.10143864154815674,
0.0009980006143450737,
0.1336364597082138,
-0.024333221837878227,
-0.005602519027888775,
-0.014818217605352402,
0.24591867625713348,
0.26666349172592163,
-0.15127348899841309,
-0.019750166684389114,
0.023462433367967606,
0.039819251745939255,
0.14734472334384918,
0.13997140526771545,
0.0671372339129448,
0.24966847896575928,
-0.08341238647699356,
-0.035497721284627914,
-0.10038480907678604,
-0.03954329341650009,
-0.17806509137153625,
0.06975771486759186,
0.13910290598869324,
-0.07589858025312424,
-0.07818760722875595,
0.08485772460699081,
-0.1614205241203308,
0.0540921650826931,
-0.05768756940960884,
-0.15166844427585602,
-0.0055720508098602295,
-0.052986230701208115,
-0.038485366851091385,
-0.006921469233930111,
0.017178285866975784,
-0.07359275966882706,
-0.026673704385757446,
-0.08866230398416519,
0.07808958739042282,
-0.26522618532180786,
-0.00118987204041332,
0.07230005413293839,
-0.07096701860427856,
0.09572350978851318,
-0.02603987790644169,
0.10190670192241669,
0.06946782022714615,
-0.048522498458623886,
-0.07270181179046631,
0.053220394998788834,
0.07320166379213333,
-0.018514709547162056,
-0.045802026987075806,
-0.016885723918676376,
0.04403424635529518,
-0.007698014844208956,
0.07934968173503876,
0.06891944259405136,
0.005223965272307396,
0.18429023027420044,
-0.04170680418610573,
0.002970488043501973,
0.0029950872994959354,
-0.09741651266813278,
0.05168628692626953,
0.03225894272327423,
-0.0432075634598732,
-0.021879717707633972,
-0.09287658333778381,
-0.008304721675813198,
-0.0008950300980359316,
-0.07422282546758652,
-0.07279957830905914,
-0.0928192213177681,
-0.057742632925510406,
0.1007789596915245,
0.026183869689702988,
-0.2165503203868866,
0.017828363925218582,
-0.1282229721546173,
0.15750962495803833,
-0.1391647756099701,
0.08234654366970062,
0.12041380256414413,
-0.06266535073518753,
-0.010221803560853004,
-0.09730897098779678,
0.028730064630508423,
0.005675000604242086,
-0.06328702718019485,
-0.07849529385566711
] |
null | null | null |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
```
***
Vanilla Quantization by [nold](https://huggingface.co/nold), Model by [mayflowergmbh](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo). Created using [llm-quantizer](https://github.com/Nold360/llm-quantizer) Pipeline - 4bc844478df79ecfd72815473b30ae09499e179d
| {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | null | nold/Wiedervereinigung-7b-dpo-GGUF | [
"gguf",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"region:us"
] | 2024-02-08T16:49:32+00:00 | [] | [
"de",
"en"
] | TAGS
#gguf #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
*
Vanilla Quantization by nold, Model by mayflowergmbh. Created using llm-quantizer Pipeline - 4bc844478df79ecfd72815473b30ae09499e179d
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage\n\n\n\n*\n\nVanilla Quantization by nold, Model by mayflowergmbh. Created using llm-quantizer Pipeline - 4bc844478df79ecfd72815473b30ae09499e179d"
] | [
"TAGS\n#gguf #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage\n\n\n\n*\n\nVanilla Quantization by nold, Model by mayflowergmbh. Created using llm-quantizer Pipeline - 4bc844478df79ecfd72815473b30ae09499e179d"
] | [
156,
148,
120,
4,
55
] | [
"passage: TAGS\n#gguf #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage\n\n\n\n*\n\nVanilla Quantization by nold, Model by mayflowergmbh. Created using llm-quantizer Pipeline - 4bc844478df79ecfd72815473b30ae09499e179d"
] | [
-0.08230505138635635,
0.0661981999874115,
-0.0038462476804852486,
0.044075097888708115,
0.027671251446008682,
0.019233494997024536,
0.0901309996843338,
0.09297683089971542,
0.09559749811887741,
0.08149673789739609,
-0.022987650707364082,
-0.007784143555909395,
0.07206140458583832,
0.08223453164100647,
0.03906045854091644,
-0.24433718621730804,
0.08084246516227722,
-0.04772130399942398,
-0.018650436773896217,
0.046994518488645554,
0.12734049558639526,
-0.059273045510053635,
0.07175431400537491,
-0.00345247657969594,
-0.03562639281153679,
0.04293343052268028,
-0.055182091891765594,
-0.00227110437117517,
0.0754384845495224,
0.07600485533475876,
0.04771967604756355,
0.007747499272227287,
0.007033248897641897,
-0.15658541023731232,
0.00989589560776949,
0.007433215156197548,
-0.02572060562670231,
0.04316142946481705,
0.11290599405765533,
-0.012749074026942253,
0.11631334573030472,
-0.16536030173301697,
-0.007127850782126188,
0.05245106667280197,
-0.10842445492744446,
-0.09994804859161377,
-0.18514880537986755,
0.17265214025974274,
0.0981261283159256,
0.04487089812755585,
-0.04870328679680824,
0.06450343132019043,
-0.01283809170126915,
0.05492271110415459,
0.18404263257980347,
-0.21515217423439026,
-0.04921948164701462,
0.18397754430770874,
-0.0038878503255546093,
-0.015339571051299572,
-0.06574897468090057,
0.03682037070393562,
0.028906892985105515,
0.0032167304307222366,
-0.038268063217401505,
-0.05373748391866684,
0.20979620516300201,
-0.029954878613352776,
-0.14070604741573334,
-0.006064483895897865,
0.04817050322890282,
0.08150315284729004,
-0.06914368271827698,
-0.16824974119663239,
-0.00005838110882905312,
-0.04398896172642708,
-0.04972187057137489,
-0.11884430795907974,
0.013232116587460041,
-0.026747172698378563,
0.08289545774459839,
-0.06749159097671509,
-0.046450573951005936,
0.04219890013337135,
-0.023493673652410507,
0.15915334224700928,
0.0029454936739057302,
-0.03262558579444885,
0.02697853185236454,
0.062136437743902206,
-0.18201909959316254,
-0.09960221499204636,
-0.06041514500975609,
-0.07027272135019302,
-0.10832048952579498,
-0.05139278247952461,
-0.02672179415822029,
-0.03583158552646637,
-0.0211323369294405,
0.18403759598731995,
-0.043019384145736694,
0.0642460435628891,
0.08900227397680283,
-0.01807299256324768,
0.024518145248293877,
0.13590234518051147,
-0.08527596294879913,
-0.14306552708148956,
-0.017393512651324272,
0.002732150722295046,
0.044319264590740204,
0.027122238650918007,
-0.034277308732271194,
-0.04043794050812721,
0.009743897244334221,
0.035852424800395966,
0.01361438725143671,
0.0339425653219223,
-0.08414427191019058,
-0.050243329256772995,
0.19649779796600342,
-0.11861972510814667,
0.03527740016579628,
-0.03103097714483738,
-0.06672099977731705,
0.09496037662029266,
0.02241726778447628,
0.0070290896110236645,
-0.044208597391843796,
0.15446577966213226,
-0.04188095033168793,
-0.025985335931181908,
-0.030661525204777718,
-0.058357883244752884,
0.04920751601457596,
-0.020152516663074493,
-0.041258495301008224,
-0.09154333174228668,
-0.07879520952701569,
-0.10404882580041885,
0.030865773558616638,
-0.02774801105260849,
-0.0007447574753314257,
-0.00316027388907969,
-0.04273518919944763,
-0.021410521119832993,
0.03639916330575943,
-0.03777042403817177,
-0.00020835849863942713,
-0.03636665269732475,
-0.07606226205825806,
0.030000602826476097,
-0.08547979593276978,
0.02412455901503563,
-0.08106761425733566,
0.04567677527666092,
-0.3074672818183899,
0.09401128441095352,
-0.11355097591876984,
0.013725575059652328,
-0.16024768352508545,
-0.07014013826847076,
-0.01271369494497776,
0.0025991916190832853,
0.04985981434583664,
0.13437117636203766,
-0.13732387125492096,
-0.0726780891418457,
0.08848182111978531,
-0.10447333753108978,
-0.04127005115151405,
0.10818520188331604,
-0.007680641487240791,
0.03554141893982887,
0.11520795524120331,
0.11204150319099426,
0.20881855487823486,
-0.09057637304067612,
-0.11141064763069153,
-0.05233079940080643,
0.041477836668491364,
0.1595696210861206,
0.028518907725811005,
-0.04861973226070404,
0.04391065984964371,
0.042720843106508255,
-0.08311664313077927,
-0.016809487715363503,
0.01386397983878851,
-0.055837541818618774,
-0.01569998636841774,
-0.007148903328925371,
0.0204305537045002,
-0.030993321910500526,
-0.00225193053483963,
-0.02021850273013115,
-0.09267231076955795,
0.0456828847527504,
0.11474914848804474,
0.023063674569129944,
0.0001255346869584173,
-0.046098705381155014,
0.08392710983753204,
0.0015918610151857138,
0.016555538401007652,
-0.1217111200094223,
-0.1458091139793396,
0.014140255749225616,
-0.24240289628505707,
0.04220712557435036,
0.03349653631448746,
0.07715558260679245,
0.01535447221249342,
-0.07224669307470322,
-0.029856910929083824,
-0.04340348020195961,
-0.004001551773399115,
0.0017551230266690254,
-0.10229711979627609,
-0.08104001730680466,
-0.03457856550812721,
0.17948177456855774,
-0.06369201093912125,
-0.007102349307388067,
0.047408849000930786,
0.20604208111763,
0.040207087993621826,
-0.05164994299411774,
-0.029365215450525284,
0.03753908723592758,
0.043732818216085434,
-0.025395411998033524,
-0.0031073903664946556,
-0.03846167027950287,
-0.12718485295772552,
0.04523405432701111,
-0.06220902502536774,
-0.03350803628563881,
0.06548537313938141,
0.08463367074728012,
-0.08492764830589294,
-0.03851380571722984,
-0.027344442903995514,
-0.033266741782426834,
-0.004008131567388773,
-0.08755967766046524,
0.07543648034334183,
0.057320620864629745,
0.011089518666267395,
-0.04235684871673584,
-0.056102629750967026,
-0.0051950267516076565,
0.020523931831121445,
-0.09562923014163971,
0.11213168501853943,
-0.06982012093067169,
-0.17992036044597626,
0.0750274807214737,
0.14622940123081207,
0.05127828195691109,
0.12566837668418884,
0.00019447767408564687,
-0.039485909044742584,
-0.09521537274122238,
0.0225612111389637,
-0.012642906978726387,
0.06621808558702469,
0.020286377519369125,
0.043340735137462616,
0.03163086622953415,
0.026791604235768318,
0.029871925711631775,
-0.030136801302433014,
0.04453228414058685,
0.011976024135947227,
-0.02005777508020401,
0.035142648965120316,
0.0812213197350502,
-0.0012772250920534134,
0.07597579061985016,
0.04804181680083275,
-0.010314546525478363,
-0.031078271567821503,
-0.03763475641608238,
-0.08685600012540817,
0.1123267337679863,
-0.09211300313472748,
-0.13176727294921875,
-0.1705230176448822,
-0.006923781242221594,
-0.14633376896381378,
-0.03978287801146507,
0.03159940242767334,
-0.053407926112413406,
-0.08642128109931946,
-0.0661877691745758,
0.13693548738956451,
0.02182941697537899,
-0.05767035111784935,
-0.044414814561605453,
-0.040699418634176254,
0.03080487996339798,
-0.12512977421283722,
-0.03970514237880707,
-0.009146020747721195,
-0.0035184817388653755,
-0.01139584556221962,
0.007087996229529381,
0.060705021023750305,
0.034120459109544754,
0.029435189440846443,
-0.03156502544879913,
-0.030992481857538223,
0.24797046184539795,
-0.11406886577606201,
0.10998792946338654,
0.06347774714231491,
-0.01716259866952896,
0.06743977963924408,
0.1410948783159256,
0.05881202965974808,
-0.06495487689971924,
-0.006013025529682636,
0.07035719603300095,
-0.012408793903887272,
-0.18271200358867645,
-0.14439718425273895,
-0.0580071397125721,
-0.02109546959400177,
0.02665717527270317,
0.048812445253133774,
0.008127860724925995,
0.0392625629901886,
-0.08915930241346359,
-0.02624484896659851,
0.036247286945581436,
0.0619446337223053,
0.12411288172006607,
0.007285361178219318,
0.053174879401922226,
-0.05866597592830658,
-0.04587456211447716,
0.09478633850812912,
-0.01544978003948927,
0.1386660486459732,
0.026049742475152016,
0.11254632472991943,
0.03500010818243027,
0.06768845021724701,
0.004285241011530161,
0.07753511518239975,
0.03677171468734741,
-0.03022422082722187,
-0.006035996600985527,
-0.10592222958803177,
0.03801392391324043,
0.0769306942820549,
0.06101841852068901,
0.0017711305990815163,
-0.030051935464143753,
-0.025677738711237907,
0.08146803081035614,
0.23431698977947235,
0.04646538570523262,
-0.15238183736801147,
-0.08763974159955978,
0.03187350928783417,
-0.002383083337917924,
0.005634739063680172,
-0.047379955649375916,
-0.02374374493956566,
-0.11674916744232178,
0.14116880297660828,
-0.03333625569939613,
0.07306768745183945,
-0.0371527336537838,
-0.0015028773341327906,
0.01692022569477558,
0.05309753492474556,
0.0036606118083000183,
0.04737497866153717,
-0.15315552055835724,
0.10832154750823975,
0.014111307449638844,
0.058370549231767654,
-0.038380809128284454,
0.049318913370370865,
0.03806482255458832,
-0.0007536906632594764,
0.13669106364250183,
0.05197558179497719,
-0.07648709416389465,
-0.08561763167381287,
-0.06704910844564438,
-0.027836095541715622,
0.11715442687273026,
-0.09256036579608917,
0.08506038784980774,
0.0036716589238494635,
-0.02942567504942417,
-0.06747187674045563,
0.10171542316675186,
-0.1633705347776413,
-0.12303369492292404,
0.1063452810049057,
-0.03571411222219467,
0.041766200214624405,
-0.05822112411260605,
-0.0493696965277195,
-0.20067070424556732,
0.18999718129634857,
-0.023365305736660957,
-0.038143716752529144,
-0.10296402126550674,
0.006261076312512159,
0.18726396560668945,
-0.11052516847848892,
0.05004175752401352,
0.007769464049488306,
0.08411538600921631,
-0.05550133064389229,
-0.09494791179895401,
0.021478133276104927,
-0.08810429275035858,
-0.11672063916921616,
0.03365663066506386,
0.15160398185253143,
0.07067090272903442,
0.015656057745218277,
0.016349781304597855,
0.03916226699948311,
0.06222109869122505,
-0.13673287630081177,
0.02076168917119503,
0.14372707903385162,
-0.022138681262731552,
0.1205955371260643,
-0.06421913951635361,
-0.09319619089365005,
-0.08127584308385849,
-0.015001548454165459,
0.04199814796447754,
0.3002086281776428,
-0.054939620196819305,
0.07344271242618561,
0.08453861624002457,
-0.046599600464105606,
-0.1629296988248825,
0.015746736899018288,
0.07290489226579666,
0.034735437482595444,
0.025306330993771553,
-0.09915652871131897,
0.11867987364530563,
0.1346604973077774,
-0.012154631316661835,
0.013423374854028225,
-0.2466493844985962,
-0.11850413680076599,
0.05730994790792465,
-0.022003989666700363,
-0.026539944112300873,
-0.07403889298439026,
-0.08433690667152405,
-0.047723568975925446,
-0.1792062222957611,
0.05220179632306099,
0.002756307367235422,
0.016645759344100952,
0.016257714480161667,
0.04295114055275917,
0.0576053224503994,
-0.01312936469912529,
0.20681563019752502,
0.08539973944425583,
0.04670539125800133,
-0.07105137407779694,
-0.04496637359261513,
0.02286422625184059,
-0.05896851792931557,
0.12912632524967194,
0.06302998960018158,
0.015198764391243458,
-0.10232464224100113,
0.0057438332587480545,
-0.07203610241413116,
0.0829903706908226,
-0.07862526923418045,
-0.03706818446516991,
-0.09043271094560623,
0.13369856774806976,
0.03587161377072334,
-0.008966083638370037,
0.08774729073047638,
-0.02928287163376808,
0.03739361837506294,
0.09688448905944824,
0.05554451420903206,
0.020445605739951134,
-0.08281299471855164,
0.010944717563688755,
-0.034187398850917816,
0.05716309696435928,
0.03335196524858475,
0.04453592002391815,
0.13774459064006805,
0.0013269874034449458,
0.13428224623203278,
-0.022699527442455292,
-0.138581782579422,
-0.0605802945792675,
0.09805168956518173,
-0.12383836507797241,
-0.21572594344615936,
-0.0319746769964695,
-0.017082007601857185,
-0.02003122866153717,
0.013660648837685585,
0.20097610354423523,
-0.03886917233467102,
-0.035089269280433655,
0.027892665937542915,
0.05224016308784485,
-0.07428330183029175,
0.13619329035282135,
-0.01524782832711935,
0.023625144734978676,
-0.07779724895954132,
0.09738370776176453,
0.06047085300087929,
-0.20842888951301575,
-0.024418694898486137,
0.04669163003563881,
-0.0679430142045021,
-0.062027499079704285,
-0.06045025587081909,
0.11207004636526108,
-0.11848466843366623,
-0.024319002404808998,
-0.007846927270293236,
-0.10652878135442734,
0.026692479848861694,
0.04169171676039696,
0.055091194808483124,
0.04138587787747383,
0.034280575811862946,
0.009867762215435505,
0.004170876927673817,
0.044884875416755676,
0.12715519964694977,
0.08733366429805756,
-0.08777783811092377,
0.024416495114564896,
-0.0773172676563263,
0.011060984805226326,
-0.008184853009879589,
-0.0015866814646869898,
-0.12119533121585846,
-0.06794126331806183,
-0.1191670149564743,
-0.016317754983901978,
-0.097144715487957,
0.0026869424618780613,
-0.016660068184137344,
0.0004976900527253747,
0.0036793153267353773,
-0.006920567713677883,
-0.08456059545278549,
-0.0658738762140274,
-0.03766410052776337,
0.07975701987743378,
-0.11251162737607956,
0.011831764131784439,
0.043340153992176056,
-0.05189117044210434,
0.08126039803028107,
0.03332556411623955,
0.03721082955598831,
0.025303522124886513,
-0.00234447349794209,
0.00422879122197628,
0.018192177638411522,
-0.0025633471086621284,
0.03280448541045189,
-0.1742238849401474,
-0.01526977214962244,
-0.009109999053180218,
-0.0469994992017746,
0.01522783376276493,
-0.07149869948625565,
-0.11104925721883774,
-0.018441013991832733,
0.007968107238411903,
-0.03961640223860741,
-0.0858447477221489,
0.029963504523038864,
0.08368641138076782,
0.0471210852265358,
0.12942053377628326,
-0.03902847319841385,
0.03888131305575371,
-0.12235388904809952,
-0.027165260165929794,
-0.00337430601939559,
-0.042391929775476456,
0.04514678567647934,
-0.03551601245999336,
0.02847931534051895,
-0.014436093159019947,
0.08633583039045334,
-0.03725148364901543,
-0.02945937030017376,
0.04355907067656517,
-0.05002434179186821,
-0.07657687366008759,
0.02259151265025139,
0.07588682323694229,
0.020371584221720695,
-0.014154759235680103,
-0.08315335959196091,
-0.0012210736749693751,
-0.060568153858184814,
0.02031254582107067,
0.08618708699941635,
0.15812815725803375,
0.08620477467775345,
0.06752781569957733,
0.04839190095663071,
-0.0530608668923378,
-0.0011592513183131814,
0.022681955248117447,
-0.007669140584766865,
0.0736071988940239,
-0.036159370094537735,
-0.0041753267869353294,
0.05953095108270645,
-0.22939994931221008,
0.1496705859899521,
-0.016473866999149323,
-0.009511861950159073,
-0.049655791372060776,
-0.13322819769382477,
-0.057011928409338,
-0.0487997941672802,
-0.0052420878782868385,
-0.09401299804449081,
0.020522335544228554,
0.0031311241909861565,
0.05318816006183624,
-0.02112257294356823,
0.15097597241401672,
-0.242802232503891,
-0.08804459869861603,
0.12710928916931152,
0.01460300199687481,
-0.022015158087015152,
0.04311415180563927,
-0.0012394189834594727,
-0.027656005695462227,
0.053266242146492004,
0.026778602972626686,
0.037709444761276245,
0.03738611191511154,
-0.02257530391216278,
-0.036679819226264954,
-0.061500586569309235,
-0.00489058019593358,
0.004398271441459656,
0.0050781178288161755,
0.11427018791437149,
0.02589121088385582,
-0.003278349991887808,
-0.04211294278502464,
0.148456409573555,
-0.05162584036588669,
-0.05079231411218643,
-0.11940518766641617,
0.16680201888084412,
0.029940368607640266,
0.058537986129522324,
-0.0033869766630232334,
-0.11427579820156097,
0.02843492478132248,
0.034175388514995575,
0.18782368302345276,
0.03925425931811333,
0.016284165903925896,
0.013694128952920437,
0.010151738300919533,
0.0018349234014749527,
0.03224439173936844,
0.006665799301117659,
0.2235625982284546,
-0.03392384946346283,
0.17816300690174103,
-0.0432233102619648,
-0.047020964324474335,
-0.01830262877047062,
0.14195387065410614,
-0.007906012237071991,
-0.008355372585356236,
-0.05397706851363182,
0.12922322750091553,
-0.04579142481088638,
-0.2158907800912857,
0.026554925367236137,
-0.11047399789094925,
-0.15026980638504028,
-0.026532191783189774,
-0.021137407049536705,
0.09276730567216873,
0.12569285929203033,
-0.00009102699550567195,
-0.04202064126729965,
0.1919136792421341,
0.0034573243465274572,
-0.03659699112176895,
-0.1383049488067627,
0.07840006798505783,
0.039149146527051926,
0.11161525547504425,
0.03036588244140148,
0.06211865693330765,
0.10635489970445633,
-0.014155654236674309,
-0.12665459513664246,
-0.002839411608874798,
0.029203858226537704,
-0.11794775724411011,
-0.0018156560836359859,
0.16500066220760345,
-0.017972107976675034,
0.08171810954809189,
0.045998506247997284,
-0.07310955971479416,
0.035278771072626114,
0.09888055175542831,
-0.0022786587942391634,
-0.09263771027326584,
0.15167036652565002,
-0.09429812431335449,
0.13950563967227936,
0.23271402716636658,
-0.024279877543449402,
0.031877633184194565,
-0.039828162640333176,
0.05592124164104462,
0.059132952243089676,
0.08693300932645798,
-0.001420993241481483,
-0.14195342361927032,
0.057221490889787674,
-0.03684935346245766,
0.040995318442583084,
-0.11562268435955048,
-0.09231892973184586,
0.036276813596487045,
0.008960967883467674,
0.011430397629737854,
0.12309540808200836,
0.04831520467996597,
-0.0005774216842837632,
-0.0036279400810599327,
-0.13374987244606018,
-0.01466161385178566,
0.07312078773975372,
-0.05114613100886345,
-0.039348967373371124
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"language": ["my"], "library_name": "transformers", "tags": ["text-generation-inference"], "base_model": "mistralai/Mistral-7B-v0.1", "pipeline_tag": "text-generation"} | text-generation | Pplus/mistral-health-faq | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"my",
"arxiv:1910.09700",
"base_model:mistralai/Mistral-7B-v0.1",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:53:12+00:00 | [
"1910.09700"
] | [
"my"
] | TAGS
#transformers #safetensors #mistral #text-generation #text-generation-inference #my #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #text-generation-inference #my #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
74,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #text-generation-inference #my #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.07007022202014923,
0.17216657102108002,
-0.0044931978918612,
0.009951205924153328,
0.1131531223654747,
0.01096577662974596,
0.08301664888858795,
0.11091405898332596,
-0.017744656652212143,
0.13544400036334991,
0.027204709127545357,
0.0907534658908844,
0.11184169352054596,
0.18117913603782654,
-0.004374158103018999,
-0.22582994401454926,
0.06418736279010773,
-0.11617618054151535,
0.015118241310119629,
0.12464745342731476,
0.14238113164901733,
-0.09939833730459213,
0.06758810579776764,
-0.036237556487321854,
-0.00847383588552475,
-0.0301818810403347,
-0.0681241825222969,
-0.05691312998533249,
0.06426545232534409,
0.06862884759902954,
0.06025746464729309,
0.027254221960902214,
0.07589249312877655,
-0.30748265981674194,
0.020479226484894753,
0.0755985751748085,
0.004736555274575949,
0.0660804808139801,
0.09297988563776016,
-0.06783685088157654,
0.12824349105358124,
-0.05548833683133125,
0.14444255828857422,
0.07492650300264359,
-0.09439358860254288,
-0.18775145709514618,
-0.07577677071094513,
0.08664055168628693,
0.16314123570919037,
0.06213073432445526,
-0.03767126798629761,
0.1368471086025238,
-0.07709400355815887,
0.01745004765689373,
0.0951375812292099,
-0.08371777087450027,
-0.06298305839300156,
0.03925740346312523,
0.07888459414243698,
0.09285017102956772,
-0.1269974261522293,
-0.006057520397007465,
0.03999495878815651,
0.01758549176156521,
0.09068720787763596,
0.020391998812556267,
0.12334398925304413,
0.02139975130558014,
-0.1372612565755844,
-0.0593627467751503,
0.13676847517490387,
0.033188458532094955,
-0.04385112226009369,
-0.23196908831596375,
-0.008587258867919445,
-0.03676225244998932,
-0.03305363282561302,
-0.04521414265036583,
0.032500091940164566,
-0.02931971289217472,
0.09178682416677475,
0.010864448733627796,
-0.06584383547306061,
-0.04802355170249939,
0.08901771157979965,
0.06800875067710876,
0.028233278542757034,
-0.030495978891849518,
0.01608407497406006,
0.11591225117444992,
0.08912530541419983,
-0.12737871706485748,
-0.06457502394914627,
-0.0609903559088707,
-0.0921572893857956,
-0.04878385365009308,
0.03485938161611557,
0.08922164142131805,
0.06468601524829865,
0.18931348621845245,
-0.027914652600884438,
0.057908795773983,
0.03639543429017067,
0.011834617704153061,
0.05957827717065811,
0.05115082859992981,
-0.06416898965835571,
-0.1386060267686844,
-0.05250956490635872,
0.1188293993473053,
0.003456867765635252,
-0.021914470940828323,
-0.0237666554749012,
0.050104979425668716,
0.059296365827322006,
0.12223809957504272,
0.07050306349992752,
0.015567705035209656,
-0.0729752779006958,
-0.03959690406918526,
0.18003545701503754,
-0.15482020378112793,
0.012381725013256073,
0.023959007114171982,
-0.05113372206687927,
-0.02997114509344101,
0.003178349696099758,
0.011684882454574108,
-0.029027173295617104,
0.10634123533964157,
-0.06836062669754028,
-0.033616676926612854,
-0.10634177178144455,
-0.05229957774281502,
0.032634224742650986,
0.0023613926023244858,
-0.03421846032142639,
-0.03262892737984657,
-0.10544931888580322,
-0.07711925357580185,
0.06270017474889755,
-0.07204209268093109,
-0.07996293902397156,
-0.03229460120201111,
-0.06876879930496216,
0.01486404798924923,
0.0002548435877542943,
0.12428608536720276,
-0.03278392180800438,
0.036232464015483856,
-0.04389956220984459,
0.06568844616413116,
0.1334300935268402,
0.02972482517361641,
-0.06120942905545235,
0.06939606368541718,
-0.2016572505235672,
0.10966432839632034,
-0.10155929625034332,
0.028649024665355682,
-0.17428885400295258,
-0.03757236897945404,
0.015466705895960331,
0.024316225200891495,
-0.01214563474059105,
0.14113181829452515,
-0.18038971722126007,
-0.04048161208629608,
0.17529131472110748,
-0.1301843374967575,
-0.09400205314159393,
0.05691378563642502,
-0.043602678924798965,
0.11423614621162415,
0.04334885627031326,
-0.028929773718118668,
0.08229176700115204,
-0.13379287719726562,
-0.015990247949957848,
-0.05165699124336243,
0.019206436350941658,
0.1611340343952179,
0.07194012403488159,
-0.060496870428323746,
0.029713748022913933,
0.019016290083527565,
-0.04383252188563347,
-0.05135297402739525,
-0.043560903519392014,
-0.0978030413389206,
0.010351533070206642,
-0.06678906083106995,
0.013440856710076332,
-0.016217121854424477,
-0.0798160508275032,
-0.03391555696725845,
-0.15091665089130402,
0.005175219848752022,
0.10159014910459518,
0.009244544431567192,
-0.030563080683350563,
-0.08848219364881516,
0.004710964392870665,
0.026217307895421982,
-0.014112339355051517,
-0.1643935739994049,
-0.0704331025481224,
0.030449124053120613,
-0.17085792124271393,
0.01253602560609579,
-0.04533807188272476,
0.036921288818120956,
0.03759137541055679,
-0.05006613954901695,
-0.0169720146805048,
0.00892031379044056,
0.013210953213274479,
-0.017225008457899094,
-0.2490008920431137,
-0.017066052183508873,
-0.05404505878686905,
0.17701205611228943,
-0.23433353006839752,
0.04612627997994423,
0.07548178732395172,
0.1281396448612213,
0.0012473975075408816,
-0.04735951125621796,
0.03987478092312813,
-0.04553639143705368,
-0.033707525581121445,
-0.06379081308841705,
0.00481377961114049,
-0.0359448678791523,
-0.041427213698625565,
0.030441666021943092,
-0.18181093037128448,
-0.015902213752269745,
0.09740739315748215,
0.07919257134199142,
-0.1655759960412979,
-0.08645899593830109,
-0.03326718136668205,
-0.06301309913396835,
-0.09190620481967926,
-0.054805826395750046,
0.11487458646297455,
0.048859089612960815,
0.04560605809092522,
-0.07613297551870346,
-0.05512377992272377,
0.006562414113432169,
-0.0023015830665826797,
-0.04069887101650238,
0.10103511810302734,
0.07136524468660355,
-0.10866732895374298,
0.10353238135576248,
0.05708415433764458,
0.06237545236945152,
0.1002880185842514,
-0.00007064964302117005,
-0.09633834660053253,
-0.013897167518734932,
0.04325821250677109,
0.0059199645183980465,
0.122816301882267,
-0.08842988312244415,
0.02685580402612686,
0.044696956872940063,
-0.032585062086582184,
0.01932637020945549,
-0.10239683091640472,
0.02132706344127655,
0.023719890043139458,
-0.017367253080010414,
0.02572731114923954,
-0.04240568354725838,
0.019186943769454956,
0.10660936683416367,
0.02979191765189171,
0.02250889129936695,
0.004275399725884199,
-0.044923100620508194,
-0.1269157975912094,
0.1781591922044754,
-0.09617447108030319,
-0.25258758664131165,
-0.12530945241451263,
0.014719489961862564,
0.037887345999479294,
-0.026203157380223274,
0.018268613144755363,
-0.05262831225991249,
-0.09911014884710312,
-0.0993904322385788,
0.022795064374804497,
0.05274422839283943,
-0.09232530742883682,
-0.037059418857097626,
0.05619334802031517,
0.03720129653811455,
-0.12143722921609879,
0.031004147604107857,
0.05316212773323059,
-0.06760138273239136,
0.002292343880981207,
0.054185185581445694,
0.077271968126297,
0.17408515512943268,
0.026298170909285545,
-0.012244116514921188,
0.011016701348125935,
0.2362947314977646,
-0.14931544661521912,
0.09147528558969498,
0.14036309719085693,
-0.04276866465806961,
0.08585868775844574,
0.22114619612693787,
0.03406374529004097,
-0.09148888289928436,
0.042979560792446136,
0.03869602456688881,
-0.027051430195569992,
-0.23789075016975403,
-0.07563423365354538,
-0.0062853265553712845,
-0.08982802927494049,
0.1010618656873703,
0.09721110016107559,
0.11412711441516876,
0.05037218704819679,
-0.11309226602315903,
-0.0799541175365448,
0.06477019190788269,
0.11566304415464401,
-0.013344850391149521,
0.008980720303952694,
0.09896431863307953,
-0.0280318446457386,
0.018983203917741776,
0.07995910942554474,
0.015489152632653713,
0.19934868812561035,
0.04322971776127815,
0.1356404721736908,
0.08792906254529953,
0.06491667032241821,
0.019378261640667915,
0.013966104947030544,
0.023625172674655914,
0.03074001520872116,
-0.016531452536582947,
-0.09206534177064896,
-0.0019642736297100782,
0.13143789768218994,
0.024012312293052673,
0.036462876945734024,
0.01839156076312065,
-0.04132327064871788,
0.06566638499498367,
0.19778576493263245,
0.014198499731719494,
-0.2241695523262024,
-0.05855998024344444,
0.0779438316822052,
-0.07737105339765549,
-0.11225864291191101,
-0.006669743452221155,
0.03595970198512077,
-0.1826065182685852,
0.055689968168735504,
-0.02543032355606556,
0.1040540263056755,
-0.098399817943573,
-0.02506032958626747,
0.05872899293899536,
0.06888821721076965,
-0.03339259326457977,
0.07716517895460129,
-0.2004852592945099,
0.15839076042175293,
0.004238042514771223,
0.061132971197366714,
-0.09730493277311325,
0.08817940205335617,
0.02049981616437435,
0.01998664252460003,
0.16277162730693817,
-0.006233964115381241,
-0.07969716936349869,
-0.06710035353899002,
-0.07333749532699585,
-0.00851830467581749,
0.10374880582094193,
-0.119022898375988,
0.08728116750717163,
-0.009845673106610775,
-0.033534735441207886,
-0.008647039532661438,
-0.11049225181341171,
-0.14500078558921814,
-0.17863692343235016,
0.06486384570598602,
-0.10681203752756119,
0.04576342925429344,
-0.11316319555044174,
-0.06685791909694672,
-0.04438050091266632,
0.177337646484375,
-0.20108501613140106,
-0.09153132885694504,
-0.14666935801506042,
-0.07529181241989136,
0.1350918859243393,
-0.04595456272363663,
0.06873366981744766,
0.0030155333224684,
0.2071489542722702,
0.00417616730555892,
0.000595996854826808,
0.07922636717557907,
-0.09489859640598297,
-0.21072575449943542,
-0.08102815598249435,
0.14811863005161285,
0.12189377099275589,
0.04148714616894722,
-0.008190339431166649,
0.024378053843975067,
-0.0031244028359651566,
-0.11677618324756622,
0.012768102809786797,
0.14990170300006866,
0.0895787701010704,
0.03665802255272865,
-0.036793217062950134,
-0.15181754529476166,
-0.11253168433904648,
-0.045346830040216446,
0.015278095379471779,
0.18255352973937988,
-0.070163294672966,
0.1547125279903412,
0.1534806191921234,
-0.060473520308732986,
-0.19479450583457947,
0.03185559809207916,
0.05695788934826851,
-0.016955113038420677,
0.028245683759450912,
-0.21447153389453888,
0.07511495798826218,
0.03288934752345085,
-0.055963773280382156,
0.13512617349624634,
-0.18937912583351135,
-0.1436237096786499,
0.08063424378633499,
0.06978213787078857,
-0.2315143197774887,
-0.14839693903923035,
-0.10886243730783463,
-0.051572367548942566,
-0.12794552743434906,
0.07710566371679306,
0.008850512094795704,
0.0008830989827401936,
0.03528085723519325,
0.0250087957829237,
0.01630784571170807,
-0.046906646341085434,
0.20955048501491547,
0.0020430502481758595,
0.038992878049612045,
-0.0715077742934227,
-0.07674292474985123,
0.028848381713032722,
-0.05751539394259453,
0.0663011446595192,
-0.021192258223891258,
0.0073353080078959465,
-0.10886449366807938,
-0.059305425733327866,
-0.06568103283643723,
0.035433199256658554,
-0.08981285989284515,
-0.08954839408397675,
-0.05353783816099167,
0.10278448462486267,
0.08859456330537796,
-0.03576088324189186,
-0.0412970595061779,
-0.10614104568958282,
0.09167167544364929,
0.2134261280298233,
0.1891546994447708,
0.07615596055984497,
-0.080021433532238,
0.0011183436727151275,
-0.01876096986234188,
0.05365345999598503,
-0.1975526213645935,
0.04296958073973656,
0.050431568175554276,
0.035165105015039444,
0.13235437870025635,
-0.024213729426264763,
-0.1503533273935318,
-0.04391707479953766,
0.058047402650117874,
-0.059608906507492065,
-0.1805514395236969,
-0.0059903888031840324,
0.10045134276151657,
-0.1622408777475357,
-0.05036443844437599,
0.03251100704073906,
-0.02649686299264431,
-0.030590277165174484,
0.002945733955129981,
0.08646061271429062,
0.02590290457010269,
0.12058506906032562,
0.05032283812761307,
0.11979082226753235,
-0.10713419318199158,
0.0742834284901619,
0.08200490474700928,
-0.09405406564474106,
0.020458171144127846,
0.08504309505224228,
-0.06727524101734161,
-0.036187492311000824,
0.01820012368261814,
0.06389964371919632,
0.042645927518606186,
-0.07021811604499817,
0.0028932266868650913,
-0.11859320849180222,
0.0681084468960762,
0.14435605704784393,
0.03161701187491417,
-0.0038415840826928616,
0.05028916150331497,
0.015496023930609226,
-0.0850900262594223,
0.10899036377668381,
0.04021456465125084,
0.03565603122115135,
-0.0535336397588253,
-0.0015395901864394546,
0.03303303197026253,
-0.0180622860789299,
-0.0170016810297966,
-0.03820356726646423,
-0.06679307669401169,
-0.008154413662850857,
-0.15656228363513947,
0.032507557421922684,
-0.09030274301767349,
0.007973136380314827,
0.016889072954654694,
-0.040304601192474365,
-0.009464473463594913,
0.009600612334907055,
-0.08098160475492477,
-0.041689153760671616,
-0.008696116507053375,
0.1168205663561821,
-0.1535230427980423,
0.018481019884347916,
0.09162583202123642,
-0.12737098336219788,
0.08006944507360458,
-0.0014931897167116404,
-0.007888589054346085,
0.021528344601392746,
-0.14086826145648956,
0.05843579024076462,
-0.004117744974792004,
0.007637839764356613,
0.022557614371180534,
-0.21338015794754028,
0.005586002487689257,
-0.0496617928147316,
-0.049058202654123306,
-0.002663566265255213,
-0.04568304866552353,
-0.11893048882484436,
0.08990754932165146,
0.003580483142286539,
-0.08129750937223434,
-0.018480706959962845,
0.04276391491293907,
0.12282580137252808,
-0.052702080458402634,
0.13983815908432007,
-0.021727459505200386,
0.06252221763134003,
-0.180588036775589,
-0.01657816581428051,
-0.016596540808677673,
0.019065063446760178,
-0.023052049800753593,
-0.0033404789865016937,
0.05848023295402527,
-0.016081420704722404,
0.21039538085460663,
-0.04198361188173294,
0.03565425053238869,
0.06855200231075287,
-0.004131033550947905,
-0.0164280254393816,
0.0888247862458229,
0.05538523569703102,
0.0351928286254406,
0.016887199133634567,
0.02425265684723854,
-0.05003520846366882,
-0.02976570837199688,
-0.13282382488250732,
0.08092635869979858,
0.17945943772792816,
0.07149085402488708,
-0.006156144198030233,
0.05559084191918373,
-0.1264938861131668,
-0.06027240678668022,
0.11694290488958359,
-0.030709294602274895,
-0.008311858400702477,
-0.05723434314131737,
0.1348825991153717,
0.17173373699188232,
-0.18521584570407867,
0.0682738795876503,
-0.06612344831228256,
-0.05502048879861832,
-0.11512868851423264,
-0.15780095756053925,
-0.06656219810247421,
-0.03713955730199814,
-0.006657142657786608,
-0.07599110156297684,
0.058146361261606216,
0.10814768075942993,
0.01649489998817444,
0.012411192990839481,
0.08115821331739426,
-0.023476915434002876,
-0.017222139984369278,
0.056713175028562546,
0.05987583100795746,
0.03374504670500755,
-0.07288883626461029,
0.004022928886115551,
0.01519998162984848,
0.03514529764652252,
0.05022862181067467,
0.027062498033046722,
-0.027844702824950218,
0.00266394205391407,
-0.020648643374443054,
-0.10066942125558853,
0.04150643199682236,
-0.025371480733156204,
-0.059392500668764114,
0.1425405889749527,
0.021591631695628166,
-0.001896145404316485,
-0.02174076810479164,
0.2362985461950302,
-0.07640518248081207,
-0.08484780043363571,
-0.14615602791309357,
0.14522431790828705,
-0.048668526113033295,
0.0541016161441803,
0.053670741617679596,
-0.09649816155433655,
0.023890214040875435,
0.12918691337108612,
0.14267030358314514,
-0.024175671860575676,
0.012749777175486088,
0.01700887829065323,
0.0074583557434380054,
-0.020442897453904152,
0.03548114374279976,
0.0545295812189579,
0.11312546581029892,
-0.06511163711547852,
0.09795662760734558,
-0.0026150604244321585,
-0.07668136805295944,
-0.02761838585138321,
0.11868252605199814,
0.001529296743683517,
0.018759965896606445,
-0.0797269269824028,
0.13996590673923492,
-0.06425666064023972,
-0.2453904002904892,
0.06402044743299484,
-0.06147700920701027,
-0.16456370055675507,
-0.021682903170585632,
0.015881149098277092,
0.0013109506107866764,
0.024318749085068703,
0.06509236246347427,
-0.06498810648918152,
0.16337989270687103,
0.042915429919958115,
-0.0702371820807457,
-0.08443083614110947,
0.0819341316819191,
-0.08810044825077057,
0.2804708778858185,
0.0032089892774820328,
0.046412281692028046,
0.09941036999225616,
-0.04573097452521324,
-0.13503609597682953,
0.046525660902261734,
0.10299289971590042,
-0.06600441038608551,
0.06489510089159012,
0.19861240684986115,
-0.007146832533180714,
0.11715326458215714,
0.08334899693727493,
-0.08566517382860184,
0.0474541075527668,
-0.06844794005155563,
-0.078901506960392,
-0.09705349802970886,
0.09238405525684357,
-0.05988804250955582,
0.14773398637771606,
0.1369680017232895,
-0.047565218061208725,
-0.004905369598418474,
-0.02382342331111431,
0.05518682301044464,
0.0009391483617946506,
0.12553124129772186,
0.01690833643078804,
-0.1990947425365448,
0.028391404077410698,
-0.020028050988912582,
0.103692427277565,
-0.23527520895004272,
-0.07436610013246536,
0.046738263219594955,
-0.01355393510311842,
-0.055957842618227005,
0.11939758062362671,
0.049826230853796005,
0.04208337143063545,
-0.048016853630542755,
-0.08991681039333344,
0.0031305148731917143,
0.17133277654647827,
-0.1144007220864296,
0.0017134913941845298
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# deepseek-7B-instr-1.5-26k
This model is a fine-tuned version of [deepseek-ai/deepseek-coder-7b-instruct-v1.5](https://huggingface.co/deepseek-ai/deepseek-coder-7b-instruct-v1.5) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- PEFT 0.4.0
- PEFT 0.4.0
- Transformers 4.30.0
- Pytorch 2.0.0.post200
- Datasets 2.10.1
- Tokenizers 0.13.3
| {"license": "other", "library_name": "peft", "tags": ["generated_from_trainer"], "model-index": [{"name": "deepseek-7B-instr-1.5-26k", "results": []}]} | null | zzz99/deepseek-7B-instr-1.5-26k | [
"peft",
"generated_from_trainer",
"license:other",
"region:us"
] | 2024-02-08T16:56:01+00:00 | [] | [] | TAGS
#peft #generated_from_trainer #license-other #region-us
|
# deepseek-7B-instr-1.5-26k
This model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
### Framework versions
- PEFT 0.4.0
- PEFT 0.4.0
- Transformers 4.30.0
- Pytorch 2.0.0.post200
- Datasets 2.10.1
- Tokenizers 0.13.3
| [
"# deepseek-7B-instr-1.5-26k\n\nThis model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 8\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- PEFT 0.4.0\n- PEFT 0.4.0\n- Transformers 4.30.0\n- Pytorch 2.0.0.post200\n- Datasets 2.10.1\n- Tokenizers 0.13.3"
] | [
"TAGS\n#peft #generated_from_trainer #license-other #region-us \n",
"# deepseek-7B-instr-1.5-26k\n\nThis model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 8\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- PEFT 0.4.0\n- PEFT 0.4.0\n- Transformers 4.30.0\n- Pytorch 2.0.0.post200\n- Datasets 2.10.1\n- Tokenizers 0.13.3"
] | [
21,
50,
6,
12,
8,
3,
112,
4,
45
] | [
"passage: TAGS\n#peft #generated_from_trainer #license-other #region-us \n# deepseek-7B-instr-1.5-26k\n\nThis model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 8\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1### Training results### Framework versions\n\n- PEFT 0.4.0\n- PEFT 0.4.0\n- Transformers 4.30.0\n- Pytorch 2.0.0.post200\n- Datasets 2.10.1\n- Tokenizers 0.13.3"
] | [
-0.12516607344150543,
0.11699464172124863,
-0.0013523565139621496,
0.08319231867790222,
0.12099869549274445,
0.013572370633482933,
0.07988762110471725,
0.14275899529457092,
-0.08982892334461212,
0.11610516160726547,
0.09452543407678604,
0.03378542140126228,
0.057419974356889725,
0.18345585465431213,
0.00019948487170040607,
-0.2997017800807953,
0.006033783312886953,
-0.018719607964158058,
-0.07248083502054214,
0.10273519158363342,
0.1185317412018776,
-0.08863529562950134,
0.07619612663984299,
0.005798323079943657,
-0.13916508853435516,
0.02322978526353836,
-0.06101967394351959,
-0.0404403991997242,
0.06279759109020233,
0.0135916443541646,
0.07198463380336761,
-0.032662875950336456,
0.10298746824264526,
-0.20964699983596802,
0.004583204630762339,
0.07710078358650208,
0.05358850955963135,
0.10135839879512787,
0.0669344887137413,
0.004581238143146038,
0.0878845825791359,
-0.1198405921459198,
0.07553964108228683,
0.03437655046582222,
-0.09789162874221802,
-0.1960291862487793,
-0.08583511412143707,
0.10010721534490585,
0.11875686794519424,
0.08734540641307831,
-0.005238850135356188,
0.11900533735752106,
-0.06746440380811691,
0.039630450308322906,
0.20991450548171997,
-0.307170569896698,
-0.09156642109155655,
0.06445503234863281,
-0.00872751884162426,
0.09399957954883575,
-0.12050409615039825,
-0.022708574309945107,
0.06961194425821304,
0.03901952505111694,
0.07046125829219818,
0.014651581645011902,
0.03479551523923874,
-0.02504768595099449,
-0.1230211928486824,
-0.03637249767780304,
0.21181604266166687,
0.09239604324102402,
-0.05723957717418671,
-0.1253625899553299,
-0.032167959958314896,
-0.20038115978240967,
-0.027238842099905014,
-0.026912080124020576,
0.03915204852819443,
-0.025947727262973785,
-0.03302107751369476,
-0.04876494035124779,
-0.07710946351289749,
-0.06155303865671158,
0.013038097880780697,
0.11542387306690216,
0.05621371045708656,
0.014304104261100292,
0.0018985334318131208,
0.11337284743785858,
-0.028734605759382248,
-0.11339739710092545,
-0.03519729897379875,
-0.027353549376130104,
-0.04261330887675285,
-0.051887162029743195,
-0.06585769355297089,
0.027494149282574654,
-0.016076361760497093,
0.15573421120643616,
-0.07174995541572571,
0.0730968788266182,
0.04989733546972275,
0.02414710447192192,
-0.02804434671998024,
0.12403274327516556,
-0.047763969749212265,
-0.025411173701286316,
0.0051944442093372345,
0.09235367923974991,
0.0235188789665699,
0.004716871306300163,
-0.09527641534805298,
-0.016908323392271996,
0.05991854891180992,
0.03425249084830284,
-0.06979810446500778,
-0.015053756535053253,
-0.04861665889620781,
-0.07546573132276535,
0.03398888185620308,
-0.10551660507917404,
0.031587738543748856,
-0.008367802947759628,
-0.09626685827970505,
-0.009752478450536728,
-0.001391229103319347,
0.015316098928451538,
-0.03454401716589928,
0.0642441138625145,
-0.07952804118394852,
0.006055269856005907,
-0.08293061703443527,
-0.03402845934033394,
0.006577314343303442,
-0.058251067996025085,
-0.005911587737500668,
-0.07667455822229385,
-0.1731981337070465,
-0.04903257638216019,
0.04372648894786835,
-0.07912391424179077,
-0.1017230749130249,
-0.04544420912861824,
-0.047002874314785004,
-0.005612330045551062,
-0.005695079918950796,
0.09549602121114731,
-0.03952664136886597,
0.07193327695131302,
-0.006583521142601967,
0.022187771275639534,
0.013135231100022793,
0.04151390492916107,
-0.07181273400783539,
0.032861944288015366,
-0.12710100412368774,
0.07988054305315018,
-0.07879875600337982,
0.022961726412177086,
-0.1410549134016037,
-0.11966085433959961,
-0.09421353787183762,
-0.0023039064835757017,
0.06634058803319931,
0.1246546283364296,
-0.1340954303741455,
-0.038717977702617645,
0.13746577501296997,
-0.05324963107705116,
-0.11070626229047775,
0.09970636665821075,
-0.026497701182961464,
0.04902453348040581,
0.03893650323152542,
0.18988549709320068,
0.12534771859645844,
-0.09360382705926895,
0.0030407756567001343,
0.007448106072843075,
0.11543311923742294,
-0.0226940605789423,
0.0978817269206047,
-0.0651789978146553,
0.03453554958105087,
0.02000829391181469,
-0.028341371566057205,
-0.009781915694475174,
-0.08256632834672928,
-0.0811067596077919,
-0.043809060007333755,
-0.09730823338031769,
0.02482057921588421,
0.007952690124511719,
0.09148833155632019,
-0.0757615715265274,
-0.09765291959047318,
0.06948000937700272,
0.1419915407896042,
-0.03724091127514839,
0.013285433873534203,
-0.058052465319633484,
0.07303079962730408,
-0.07311039417982101,
-0.04172695800662041,
-0.16946499049663544,
-0.05837273597717285,
0.05606597289443016,
-0.05153419449925423,
0.040207501500844955,
0.0209362730383873,
0.0712745264172554,
0.060719992965459824,
-0.05489218980073929,
-0.038025446236133575,
-0.10870090872049332,
-0.04122213274240494,
-0.09126079082489014,
-0.1429375857114792,
-0.07229990512132645,
-0.020199112594127655,
0.15700994431972504,
-0.2198643684387207,
0.010600727051496506,
-0.05189764127135277,
0.15429404377937317,
0.03659876063466072,
-0.08990506082773209,
0.01468991581350565,
0.05441562831401825,
0.004197623580694199,
-0.0933711901307106,
0.043689366430044174,
0.0002380295773036778,
-0.09332987666130066,
-0.09976164996623993,
-0.11150380969047546,
0.12414487451314926,
0.06096787005662918,
0.03626081347465515,
-0.07200253754854202,
-0.06753283739089966,
-0.07896627485752106,
-0.027364114299416542,
-0.05681030824780464,
0.028482595458626747,
0.17301097512245178,
0.009015307761728764,
0.12989157438278198,
-0.09787659347057343,
-0.05819297209382057,
0.017529936507344246,
-0.011312581598758698,
-0.006335787009447813,
0.08435018360614777,
0.06226669251918793,
-0.057759445160627365,
0.07842166721820831,
0.10229300707578659,
-0.0388554148375988,
0.1380177140235901,
-0.06985705345869064,
-0.12284614145755768,
-0.028404435142874718,
0.05105656012892723,
-0.009524405933916569,
0.11889355629682541,
-0.06659074872732162,
0.012662936933338642,
0.03523028641939163,
0.05666247010231018,
0.05379620939493179,
-0.15919220447540283,
-0.021551193669438362,
0.02939784899353981,
-0.03515017032623291,
0.006712837610393763,
0.008932160213589668,
0.010697806254029274,
0.07668963819742203,
0.03791133314371109,
0.015724405646324158,
0.005460922606289387,
-0.005879736039787531,
-0.07391509413719177,
0.171491339802742,
-0.11653531342744827,
-0.1347491294145584,
-0.1654636263847351,
0.06587105989456177,
-0.09468265622854233,
-0.0162555743008852,
0.020618660375475883,
-0.079002246260643,
-0.022968759760260582,
-0.09768671542406082,
0.01662088744342327,
-0.08022204786539078,
0.002696617739275098,
0.06552322953939438,
0.007936405949294567,
0.1183205246925354,
-0.11623679101467133,
0.02103310078382492,
-0.017925944179296494,
-0.08240774273872375,
-0.02967902459204197,
0.024194223806262016,
0.07312270998954773,
0.10589363425970078,
0.02865508943796158,
0.056902531534433365,
-0.04230232536792755,
0.25637394189834595,
-0.08942895382642746,
-0.020930463448166847,
0.142739400267601,
0.045705076307058334,
0.04143515229225159,
0.08693791925907135,
0.049581319093704224,
-0.11966770887374878,
0.025961723178625107,
0.07912083715200424,
-0.013341008685529232,
-0.19946299493312836,
-0.04225056990981102,
-0.03510644659399986,
-0.0641399621963501,
0.12880806624889374,
0.06039361283183098,
0.015563955530524254,
0.053146544843912125,
-0.02506263554096222,
0.04879840463399887,
-0.04097817838191986,
0.0956323966383934,
0.07347274571657181,
0.050309620797634125,
0.08704962581396103,
-0.037953346967697144,
-0.00711479876190424,
0.08021499216556549,
0.009665746241807938,
0.23773445188999176,
-0.0029334286227822304,
0.09606260806322098,
0.03603134676814079,
0.1830814927816391,
-0.020370375365018845,
0.03723280131816864,
0.015521863475441933,
-0.025713052600622177,
-0.004472886677831411,
-0.05980784073472023,
-0.0549619197845459,
0.043081849813461304,
0.01677231676876545,
0.08933128416538239,
-0.10639100521802902,
0.01813145913183689,
0.00994008220732212,
0.28378620743751526,
0.06503459066152573,
-0.2787324786186218,
-0.10021916031837463,
0.011950315907597542,
-0.025341792032122612,
-0.08141099661588669,
0.05704478546977043,
0.0954824686050415,
-0.11566615104675293,
0.04067061096429825,
-0.06898435205221176,
0.09149936586618423,
-0.03409574553370476,
0.005184775684028864,
0.06095678359270096,
0.13057585060596466,
0.0013789826771244407,
0.0833495482802391,
-0.18788999319076538,
0.196742981672287,
0.009256367571651936,
0.10846353322267532,
-0.03717612475156784,
0.028282983228564262,
0.006433072034269571,
0.061052531003952026,
0.10157594829797745,
0.008525011129677296,
-0.001667667180299759,
-0.15602654218673706,
-0.1015176773071289,
0.023838398978114128,
0.12547436356544495,
-0.11710191518068314,
0.07473831623792648,
-0.03181777149438858,
0.028029482811689377,
0.011129731312394142,
-0.013720597140491009,
-0.12991631031036377,
-0.11474084109067917,
0.009115979075431824,
-0.008080692030489445,
-0.010431909933686256,
-0.09311901032924652,
-0.08913803845643997,
-0.040472403168678284,
0.12704887986183167,
-0.023437393829226494,
-0.029340187087655067,
-0.13331875205039978,
0.09252263605594635,
0.12859341502189636,
-0.05982222035527229,
0.022562634199857712,
0.008942539803683758,
0.12630802392959595,
0.05160552263259888,
-0.058241937309503555,
0.0586838461458683,
-0.05450543016195297,
-0.16415457427501678,
-0.028207438066601753,
0.15951110422611237,
0.01170831173658371,
0.03967779129743576,
-0.027077404782176018,
0.011347545310854912,
0.005281689576804638,
-0.10365466773509979,
0.03309725597500801,
0.10117781162261963,
0.03790945187211037,
0.05679956451058388,
-0.08000616729259491,
0.14424024522304535,
-0.033123988658189774,
-0.009819968603551388,
0.10947131365537643,
0.269849956035614,
-0.08646753430366516,
0.048502031713724136,
0.0023921625688672066,
-0.05271763727068901,
-0.12784823775291443,
0.038358718156814575,
0.147935152053833,
0.042637210339307785,
0.047207511961460114,
-0.21880924701690674,
0.10697508603334427,
0.13876697421073914,
-0.02582935430109501,
0.0756845772266388,
-0.33887672424316406,
-0.12100081890821457,
0.07059015333652496,
0.0919303372502327,
0.044267427176237106,
-0.13426867127418518,
-0.05846184492111206,
-0.00560618145391345,
-0.06426306813955307,
0.06385292112827301,
-0.05453558638691902,
0.11528030782938004,
-0.02454410493373871,
0.08097343146800995,
0.027364198118448257,
-0.0020022087264806032,
0.1613992154598236,
0.00034194387262687087,
0.08746729046106339,
-0.030879870057106018,
0.07752940058708191,
0.01423604041337967,
-0.07976994663476944,
0.05402127280831337,
-0.014903377741575241,
0.058345142751932144,
-0.21315371990203857,
-0.01630953513085842,
-0.06601209938526154,
0.04925653710961342,
-0.05807015299797058,
-0.049675289541482925,
-0.0539715550839901,
0.05271154269576073,
0.0446876622736454,
-0.0056878491304814816,
0.13777004182338715,
0.01735069788992405,
0.12757593393325806,
0.0741695836186409,
0.05422702059149742,
0.007803990505635738,
-0.1404692381620407,
-0.025099890306591988,
-0.01780662313103676,
0.05986950173974037,
-0.16663722693920135,
-0.007229727692902088,
0.10848436504602432,
0.027109812945127487,
0.1212250143289566,
0.025746963918209076,
-0.0630268082022667,
-0.004129263572394848,
0.051996421068906784,
-0.07740606367588043,
-0.16425076127052307,
-0.04256431385874748,
0.03282317891716957,
-0.15113286674022675,
0.022720640525221825,
0.10510396957397461,
-0.06030641868710518,
-0.0232755895704031,
-0.004496915731579065,
0.001549547421745956,
-0.04087812826037407,
0.13907060027122498,
0.0521073192358017,
0.06333605200052261,
-0.07685327529907227,
0.14329349994659424,
0.04545480012893677,
-0.04268673434853554,
0.05068742483854294,
0.06685077399015427,
-0.06563059240579605,
-0.010960487648844719,
0.05664701759815216,
0.12495743483304977,
0.0062880730256438255,
-0.026469649747014046,
-0.07865297794342041,
-0.07772713154554367,
0.06941677629947662,
0.01278393529355526,
0.06217416748404503,
-0.024766717106103897,
-0.04911782592535019,
0.029384233057498932,
-0.12206050008535385,
0.07371938228607178,
0.042706236243247986,
0.06804373860359192,
-0.18589286506175995,
0.0842166319489479,
-0.016534069553017616,
0.007116052322089672,
-0.004739133641123772,
-0.007262894883751869,
-0.07150151580572128,
-0.027059346437454224,
-0.09772371500730515,
-0.039612650871276855,
-0.02992813102900982,
0.013406707905232906,
-0.00403441209346056,
-0.03669734299182892,
-0.0647064670920372,
0.03299132362008095,
-0.09190638363361359,
-0.06765855103731155,
0.0007756357081234455,
0.07383562624454498,
-0.09805642813444138,
0.02447526529431343,
0.0017740470357239246,
-0.09349694103002548,
0.08146729320287704,
0.05566084012389183,
0.040430791676044464,
0.03282642364501953,
-0.02179214544594288,
0.039974145591259,
0.025018932297825813,
0.00216946704313159,
0.048325374722480774,
-0.0635363906621933,
-0.009735562838613987,
-0.038387734442949295,
0.03022690862417221,
0.011243767105042934,
0.028621835634112358,
-0.13993915915489197,
-0.07762403786182404,
-0.04089908301830292,
-0.02454747073352337,
-0.05799194052815437,
0.03927535563707352,
0.07025960087776184,
0.05502120032906532,
0.14108480513095856,
-0.06842862069606781,
0.02953573688864708,
-0.21718937158584595,
-0.03201969712972641,
-0.004855106119066477,
0.012757088989019394,
-0.07816106826066971,
-0.048245299607515335,
0.07284041494131088,
-0.04173947498202324,
0.09063302725553513,
-0.032013263553380966,
0.06238287687301636,
0.0325767807662487,
-0.06660634279251099,
0.01736709289252758,
0.002864856505766511,
0.2151271253824234,
0.069370798766613,
-0.005763240624219179,
0.09745053946971893,
-0.009857650846242905,
0.06964104622602463,
0.09731132537126541,
0.14178653061389923,
0.17214521765708923,
-0.041452132165431976,
0.08099161088466644,
0.044873856008052826,
-0.10318611562252045,
-0.11812839657068253,
0.12889857590198517,
0.030902571976184845,
0.0929686427116394,
-0.03568175435066223,
0.09483759105205536,
0.10642926394939423,
-0.1854497492313385,
0.030610963702201843,
-0.030041266232728958,
-0.09116590768098831,
-0.12080439180135727,
-0.09236668795347214,
-0.08147643506526947,
-0.1703941822052002,
0.0322062149643898,
-0.13973261415958405,
0.017973890528082848,
0.07655072957277298,
0.01966041885316372,
-0.003699688473716378,
0.17520442605018616,
-0.0640786662697792,
-0.00847852323204279,
0.08839216083288193,
0.001467366935685277,
0.010797901079058647,
-0.06915517896413803,
-0.09137535095214844,
0.03489866480231285,
-0.04886912927031517,
0.08799587190151215,
-0.06224292516708374,
-0.023907184600830078,
0.02105504274368286,
-0.006920868065208197,
-0.07952049374580383,
0.03715640306472778,
0.0060727414675056934,
0.017980210483074188,
0.08051647245883942,
0.03835034742951393,
-0.032339490950107574,
-0.06330712884664536,
0.3027225434780121,
-0.08559391647577286,
-0.012346452102065086,
-0.15049540996551514,
0.18433506786823273,
0.0535014308989048,
-0.0009774763602763414,
0.03873644769191742,
-0.10843447595834732,
-0.03448968008160591,
0.13995885848999023,
0.13376891613006592,
-0.062306977808475494,
-0.029563231393694878,
0.03420707955956459,
-0.017804866656661034,
-0.07675386220216751,
0.11845536530017853,
0.10226316004991531,
-0.008350122720003128,
-0.049218375235795975,
0.008368856273591518,
-0.005464243236929178,
-0.019247988238930702,
-0.09245823323726654,
0.08875387907028198,
0.027131594717502594,
-0.007769857998937368,
-0.04716897010803223,
0.0954156294465065,
-0.03582560271024704,
-0.09592349082231522,
0.06217767298221588,
-0.17222067713737488,
-0.2018786072731018,
-0.03400587663054466,
0.011319128796458244,
0.006062902510166168,
0.07161881029605865,
-0.01906955987215042,
-0.008197755552828312,
0.1412009745836258,
-0.02210281416773796,
-0.04191746562719345,
-0.14323072135448456,
0.07918611168861389,
-0.0793868899345398,
0.21346282958984375,
0.004872489254921675,
0.012933208607137203,
0.09561746567487717,
0.008828926831483841,
-0.117454893887043,
0.019777532666921616,
0.05683048814535141,
-0.014324221760034561,
-0.005799987353384495,
0.1662532091140747,
-0.033248551189899445,
0.11165951192378998,
0.048569437116384506,
-0.14629383385181427,
-0.04824242368340492,
-0.05733959749341011,
-0.019657008349895477,
-0.0863097682595253,
0.013875024393200874,
-0.07732638716697693,
0.14770197868347168,
0.20260675251483917,
-0.05382109433412552,
-0.029839908704161644,
-0.04649529606103897,
0.07333169877529144,
0.06512913107872009,
0.08499287068843842,
-0.011261305771768093,
-0.1980987787246704,
-0.006731657311320305,
0.025856781750917435,
-0.0027307122945785522,
-0.2376796156167984,
-0.086336150765419,
0.047704871743917465,
-0.053181201219558716,
-0.04654703661799431,
0.12245933711528778,
0.03244679793715477,
0.024514401331543922,
-0.04398943856358528,
-0.10405305027961731,
-0.05054599046707153,
0.1413375586271286,
-0.1438271552324295,
-0.06095093861222267
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SciBERT_AsymmetricLoss_25K_bs64_P1_N1
This model is a fine-tuned version of [allenai/scibert_scivocab_uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 67.0896
- Accuracy: 0.9945
- Precision: 0.7586
- Recall: 0.6438
- F1: 0.6965
- Hamming: 0.0055
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 25000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Hamming |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 83.6475 | 0.16 | 5000 | 79.3653 | 0.9938 | 0.7361 | 0.5667 | 0.6404 | 0.0062 |
| 75.8712 | 0.32 | 10000 | 72.7250 | 0.9942 | 0.7513 | 0.6068 | 0.6714 | 0.0058 |
| 72.4202 | 0.47 | 15000 | 69.4174 | 0.9944 | 0.7568 | 0.6237 | 0.6838 | 0.0056 |
| 70.0693 | 0.63 | 20000 | 67.8098 | 0.9945 | 0.7561 | 0.6385 | 0.6923 | 0.0055 |
| 68.9765 | 0.79 | 25000 | 67.0896 | 0.9945 | 0.7586 | 0.6438 | 0.6965 | 0.0055 |
### Framework versions
- Transformers 4.35.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.7.1
- Tokenizers 0.14.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy", "precision", "recall", "f1"], "base_model": "allenai/scibert_scivocab_uncased", "model-index": [{"name": "SciBERT_AsymmetricLoss_25K_bs64_P1_N1", "results": []}]} | text-classification | bdpc/SciBERT_AsymmetricLoss_25K_bs64_P1_N1 | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:allenai/scibert_scivocab_uncased",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:58:16+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us
| SciBERT\_AsymmetricLoss\_25K\_bs64\_P1\_N1
==========================================
This model is a fine-tuned version of allenai/scibert\_scivocab\_uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 67.0896
* Accuracy: 0.9945
* Precision: 0.7586
* Recall: 0.6438
* F1: 0.6965
* Hamming: 0.0055
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 25000
### Training results
### Framework versions
* Transformers 4.35.0.dev0
* Pytorch 2.0.1+cu118
* Datasets 2.7.1
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
61,
116,
4,
36
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000### Training results### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
-0.12239006906747818,
0.07038096338510513,
-0.002270608674734831,
0.13089551031589508,
0.17182838916778564,
0.03552595153450966,
0.11841200292110443,
0.13127560913562775,
-0.06969314068555832,
0.038499392569065094,
0.12407459318637848,
0.1446039080619812,
0.022002389654517174,
0.13452577590942383,
-0.05503028258681297,
-0.2885366678237915,
-0.0033677059691399336,
0.01731622777879238,
-0.06730091571807861,
0.13224023580551147,
0.0828372985124588,
-0.1337146908044815,
0.09499790519475937,
-0.020193813368678093,
-0.16191041469573975,
-0.0009068193612620234,
0.01633271761238575,
-0.044335950165987015,
0.14709243178367615,
0.01964900828897953,
0.12271825969219208,
0.022447623312473297,
0.11982916295528412,
-0.17441865801811218,
0.008429979905486107,
0.04743145778775215,
0.02328234352171421,
0.09247520565986633,
0.05412723496556282,
0.00020402234804350883,
0.11227516084909439,
-0.09098335355520248,
0.058341994881629944,
0.020313246175646782,
-0.13214179873466492,
-0.1700744330883026,
-0.07841253280639648,
0.02907271310687065,
0.07433215528726578,
0.08933878690004349,
-0.0061609200201928616,
0.10206861793994904,
-0.08811420947313309,
0.10985638201236725,
0.22064051032066345,
-0.2594301402568817,
-0.0766574814915657,
0.010001081973314285,
0.004469216335564852,
0.09542322158813477,
-0.10710377246141434,
-0.01143158320337534,
0.034490350633859634,
0.03448912128806114,
0.12150989472866058,
-0.02728152461349964,
-0.12987226247787476,
0.022906633093953133,
-0.13800175487995148,
-0.02804391272366047,
0.09836933761835098,
0.023764215409755707,
-0.043362807482481,
-0.027673223987221718,
-0.0765603631734848,
-0.17501822113990784,
-0.04179663583636284,
-0.012468980625271797,
0.03498247265815735,
-0.04469481483101845,
-0.07365084439516068,
-0.026723813265562057,
-0.08419449627399445,
-0.0973527580499649,
-0.04072478413581848,
0.17177926003932953,
0.03978623077273369,
0.012730268761515617,
-0.0018546836217865348,
0.12453312426805496,
0.01509181223809719,
-0.15217764675617218,
0.00950139295309782,
0.013917750678956509,
-0.049028120934963226,
-0.04589609429240227,
-0.057228997349739075,
-0.016281330958008766,
-0.01917402632534504,
0.12961876392364502,
-0.029023857787251472,
0.05072631686925888,
0.03436003252863884,
0.016388587653636932,
-0.10456088930368423,
0.19624429941177368,
-0.05854874104261398,
-0.022884616628289223,
-0.007876993156969547,
0.0894712433218956,
0.007368528749793768,
-0.033018648624420166,
-0.10883118957281113,
-0.0166451595723629,
0.10293465107679367,
0.023269997909665108,
-0.08344634622335434,
0.06950521469116211,
-0.04303796589374542,
-0.033930059522390366,
0.019838472828269005,
-0.09911400079727173,
0.0360114648938179,
0.01767577975988388,
-0.10162538290023804,
-0.018144438043236732,
0.024868860840797424,
0.004757971037179232,
-0.015042777173221111,
0.1282130479812622,
-0.09785988926887512,
0.03428512066602707,
-0.09385675191879272,
-0.12153982371091843,
-0.0053965444676578045,
-0.09725828468799591,
0.0011557049583643675,
-0.0817437693476677,
-0.19228290021419525,
-0.028452470898628235,
0.03871012106537819,
-0.052222736179828644,
-0.0348193496465683,
-0.06804528087377548,
-0.06677358597517014,
0.02783213183283806,
-0.007976556196808815,
0.12797744572162628,
-0.05714099481701851,
0.11634469777345657,
0.03988274186849594,
0.07020349055528641,
-0.027476532384753227,
0.05806878209114075,
-0.10281573981046677,
0.010806224308907986,
-0.1829170435667038,
0.06699594855308533,
-0.05433490499854088,
0.04267386719584465,
-0.08098269253969193,
-0.10431589186191559,
0.024051155894994736,
0.0008766418322920799,
0.09290977567434311,
0.1409061849117279,
-0.1830790638923645,
-0.07253091037273407,
0.1420736014842987,
-0.06886310875415802,
-0.090731680393219,
0.11803504824638367,
-0.06576776504516602,
0.03177472949028015,
0.06343913078308105,
0.15235337615013123,
0.08405910432338715,
-0.0665288120508194,
0.016191352158784866,
-0.009675825014710426,
0.08389129489660263,
-0.011932277120649815,
0.07144893705844879,
0.020486149936914444,
-0.000908686954062432,
0.025170424953103065,
-0.0617491789162159,
0.04150257632136345,
-0.10790619999170303,
-0.09412276744842529,
-0.035276785492897034,
-0.10416639596223831,
0.08382663130760193,
0.07279685884714127,
0.07795432955026627,
-0.08594062924385071,
-0.08733929693698883,
0.057237446308135986,
0.09898535907268524,
-0.06270033121109009,
0.01929069682955742,
-0.06239078938961029,
0.054203860461711884,
-0.022769588977098465,
-0.025389688089489937,
-0.18733346462249756,
-0.045162271708250046,
0.020382581278681755,
0.04846292361617088,
0.02281942404806614,
0.0016226792940869927,
0.08511372655630112,
0.08827467262744904,
-0.06484182924032211,
-0.031042365357279778,
-0.040745727717876434,
-0.00862665381282568,
-0.14019912481307983,
-0.20115338265895844,
-0.06440245360136032,
-0.01868603751063347,
0.13572026789188385,
-0.21429120004177094,
0.03430602699518204,
-0.013535602018237114,
0.08657269924879074,
0.020001845434308052,
-0.024251582100987434,
-0.032936129719018936,
0.07709255069494247,
-0.03667707368731499,
-0.057209283113479614,
0.06744059920310974,
-0.013141147792339325,
-0.08329112827777863,
-0.050432052463293076,
-0.11685102432966232,
0.13448618352413177,
0.10413527488708496,
-0.08861696720123291,
-0.09657712280750275,
0.004747803322970867,
-0.05052494630217552,
-0.03206455707550049,
-0.05012806877493858,
0.02545151673257351,
0.17179572582244873,
0.005685955286026001,
0.15303145349025726,
-0.056193869560956955,
-0.03603309392929077,
0.020509283989667892,
-0.007702842820435762,
0.042179323732852936,
0.13229165971279144,
0.10557302832603455,
-0.08586869388818741,
0.12952619791030884,
0.10975980758666992,
-0.08752638846635818,
0.14262709021568298,
-0.024030623957514763,
-0.08278664201498032,
-0.01359491515904665,
-0.026771798729896545,
0.0010569716105237603,
0.10394185036420822,
-0.11336454749107361,
-0.017819305881857872,
0.011675954796373844,
0.014026987366378307,
-0.003758769016712904,
-0.20600396394729614,
-0.0349532887339592,
0.041023433208465576,
-0.05383365973830223,
-0.03309403732419014,
-0.013819299638271332,
0.00809128675609827,
0.12128349393606186,
0.011991294100880623,
-0.09089910238981247,
0.01355469785630703,
-0.00276798103004694,
-0.06932799518108368,
0.20992186665534973,
-0.07254642993211746,
-0.12312544137239456,
-0.12119962275028229,
-0.05350416153669357,
-0.05827227607369423,
0.02897077426314354,
0.03125174343585968,
-0.09345889091491699,
-0.009478596970438957,
-0.06180248409509659,
0.024914832785725594,
0.005248423665761948,
0.04262742027640343,
-0.024328241124749184,
-0.01973292976617813,
0.05876430496573448,
-0.08981389552354813,
-0.0028476088773459196,
-0.05926596373319626,
-0.08142127841711044,
0.03406904265284538,
0.03101099655032158,
0.11546559631824493,
0.1509244740009308,
-0.023641230538487434,
0.012300890870392323,
-0.02422642521560192,
0.22516395151615143,
-0.07099036127328873,
-0.0068967039696872234,
0.138851135969162,
-0.026212040334939957,
0.04785362631082535,
0.13172578811645508,
0.0631052702665329,
-0.08356857299804688,
0.010682963766157627,
0.04271089658141136,
-0.0341249518096447,
-0.21104124188423157,
-0.037000760436058044,
-0.05317976698279381,
0.005707957316190004,
0.10244151949882507,
0.013822940178215504,
0.02043784223496914,
0.07110318541526794,
0.016179269179701805,
0.048033326864242554,
-0.03800271078944206,
0.05945992097258568,
0.09588494896888733,
0.044698722660541534,
0.13814692199230194,
-0.02504982426762581,
-0.06687279045581818,
0.042679667472839355,
-0.023041626438498497,
0.1997382640838623,
-0.026216614991426468,
0.12034111469984055,
0.014289258979260921,
0.15840868651866913,
-0.009299556724727154,
0.08283321559429169,
0.011236663907766342,
-0.03521028533577919,
-0.01213148608803749,
-0.04027840495109558,
-0.06489378213882446,
0.02492157556116581,
-0.0615946464240551,
0.06076880544424057,
-0.15200024843215942,
0.01523659285157919,
0.04764033481478691,
0.2705806791782379,
0.06676151603460312,
-0.32958221435546875,
-0.10971777141094208,
0.007234081160277128,
-0.036199044436216354,
-0.03388427942991257,
0.008762715384364128,
0.12918280065059662,
-0.10107716917991638,
0.026911530643701553,
-0.06965263187885284,
0.0912308543920517,
-0.06432180106639862,
0.04941486194729805,
0.07924848794937134,
0.09748173505067825,
-0.0077408477663993835,
0.07018333673477173,
-0.27493512630462646,
0.2770404517650604,
0.013825390487909317,
0.0627499520778656,
-0.0700584277510643,
-0.00786398071795702,
0.05056403577327728,
0.08633213490247726,
0.060020167380571365,
-0.012451590038836002,
-0.04552178084850311,
-0.2327864170074463,
-0.07576194405555725,
0.02321651019155979,
0.11330311000347137,
-0.0644078478217125,
0.11162886023521423,
-0.04210285469889641,
-0.0032250939402729273,
0.05830661579966545,
-0.04423123970627785,
-0.05232122540473938,
-0.08034133911132812,
0.0038433547597378492,
0.008150003850460052,
-0.007344100158661604,
-0.0489417240023613,
-0.12082503736019135,
-0.07118731737136841,
0.15340112149715424,
-0.03301874175667763,
-0.03455265611410141,
-0.13572381436824799,
0.09249578416347504,
0.09975622594356537,
-0.0914166271686554,
0.027698101475834846,
0.014259293675422668,
0.07630632817745209,
0.03966827318072319,
-0.0584312304854393,
0.11153434216976166,
-0.06539176404476166,
-0.2048051506280899,
-0.0641174241900444,
0.10474395751953125,
0.045213837176561356,
0.07100849598646164,
-0.021890288218855858,
0.025506693869829178,
-0.019763877615332603,
-0.0837186798453331,
0.026365172117948532,
-0.010994844138622284,
0.05372616648674011,
0.04464171826839447,
-0.06631428748369217,
0.045672476291656494,
-0.06296247988939285,
-0.008350170217454433,
0.15975536406040192,
0.24500009417533875,
-0.10207007080316544,
0.02417619898915291,
0.030164232477545738,
-0.06615900993347168,
-0.2066960632801056,
0.04072249308228493,
0.07441367954015732,
0.024937978014349937,
0.04037139564752579,
-0.20382609963417053,
0.08771143108606339,
0.09525907039642334,
-0.011051145382225513,
0.08907002955675125,
-0.30238470435142517,
-0.1271258443593979,
0.10659535974264145,
0.1258370280265808,
0.12308255583047867,
-0.14216050505638123,
-0.02015303634107113,
-0.011090027168393135,
-0.08133064955472946,
0.09111294150352478,
-0.06461554765701294,
0.1279584765434265,
-0.029369210824370384,
0.08897161483764648,
0.0302097350358963,
-0.04464256018400192,
0.11791153997182846,
0.016575243324041367,
0.09271527081727982,
-0.048984382301568985,
-0.04559503123164177,
0.022436564788222313,
-0.04936091974377632,
0.007371949963271618,
-0.07676497101783752,
0.046593062579631805,
-0.11720437556505203,
-0.017861295491456985,
-0.08944916725158691,
0.021612970158457756,
-0.03606710955500603,
-0.06448612362146378,
-0.016325149685144424,
0.0374222993850708,
0.0463714674115181,
-0.006653534714132547,
0.11898743361234665,
-0.014422859996557236,
0.1480257511138916,
0.10997170209884644,
0.0865267738699913,
-0.018907129764556885,
-0.020895663648843765,
0.0013944103848189116,
-0.005069685634225607,
0.05545942485332489,
-0.12528328597545624,
0.03395261988043785,
0.14881139993667603,
0.03429792821407318,
0.1277073621749878,
0.08298145234584808,
-0.0016531600849702954,
0.008662156760692596,
0.052623432129621506,
-0.1730736345052719,
-0.058394577354192734,
-0.01513443049043417,
-0.06459499895572662,
-0.1320958137512207,
0.04858201742172241,
0.11238663643598557,
-0.06819935142993927,
-0.01308648195117712,
-0.010966382920742035,
0.011705826036632061,
-0.03743020072579384,
0.21776831150054932,
0.05902910605072975,
0.059291381388902664,
-0.10401014983654022,
0.06399628520011902,
0.04900115728378296,
-0.05636313557624817,
0.007586529944092035,
0.09954381734132767,
-0.09021449834108353,
-0.03634560480713844,
0.06135982647538185,
0.15154771506786346,
-0.07580685615539551,
-0.018328489735722542,
-0.15647806227207184,
-0.11130106449127197,
0.07111833989620209,
0.17828921973705292,
0.09962186217308044,
0.020853176712989807,
-0.06655143946409225,
0.022309882566332817,
-0.12215378135442734,
0.09659730643033981,
0.05762745067477226,
0.07719870656728745,
-0.14473845064640045,
0.18751679360866547,
-0.02199823595583439,
0.04957422986626625,
-0.02999516949057579,
0.02244112640619278,
-0.10935354232788086,
0.010224604047834873,
-0.13072331249713898,
-0.03427128493785858,
-0.0065178032964468,
0.0005157325067557395,
-0.011834277771413326,
-0.0751551166176796,
-0.04827370122075081,
0.0018287680577486753,
-0.11437570303678513,
-0.02571658417582512,
0.01943645440042019,
0.03479974716901779,
-0.12090522795915604,
-0.04035745933651924,
0.03265730291604996,
-0.08077806979417801,
0.0791250541806221,
0.059578001499176025,
0.020609606057405472,
0.05064117908477783,
-0.12445028126239777,
0.012717261910438538,
0.03291705623269081,
0.009428560733795166,
0.06306450068950653,
-0.0664442628622055,
0.0022305322345346212,
-0.03386590629816055,
0.062340330332517624,
0.021519776433706284,
0.11060437560081482,
-0.11858026683330536,
0.029840100556612015,
-0.011216999031603336,
-0.06925348192453384,
-0.06505866348743439,
0.05867297947406769,
0.07389768958091736,
0.03140941262245178,
0.16329561173915863,
-0.08925597369670868,
0.051601339131593704,
-0.2181098908185959,
0.00008990098285721615,
-0.008830444887280464,
-0.120360367000103,
-0.13204564154148102,
-0.07951638847589493,
0.08901864290237427,
-0.04739021137356758,
0.09151442348957062,
0.018703293055295944,
0.09766723960638046,
0.02487986907362938,
-0.033752284944057465,
0.033930208534002304,
0.03136355057358742,
0.18021716177463531,
0.049268703907728195,
-0.04822346568107605,
0.07613780349493027,
0.05260326713323593,
0.09048043936491013,
0.14555074274539948,
0.22101272642612457,
0.12525056302547455,
-0.01419612392783165,
0.0909113958477974,
0.04209940880537033,
-0.07237182557582855,
-0.16593821346759796,
0.004391070455312729,
-0.051212284713983536,
0.0855998545885086,
-0.045730944722890854,
0.214377298951149,
0.047232963144779205,
-0.16798417270183563,
0.047059111297130585,
-0.04942953959107399,
-0.10331861674785614,
-0.12448010593652725,
-0.013456725515425205,
-0.08117780089378357,
-0.1410934180021286,
-0.0017057594377547503,
-0.10980669409036636,
0.03821501508355141,
0.10636288672685623,
0.00930856354534626,
-0.009974460117518902,
0.16052278876304626,
0.021347688511013985,
0.037782177329063416,
0.07069765776395798,
0.012359456159174442,
-0.0037504786159843206,
-0.07913175225257874,
-0.09391181915998459,
-0.0043456340208649635,
-0.013917669653892517,
0.03180159255862236,
-0.07397375255823135,
-0.05759376287460327,
0.031400784850120544,
-0.006047637667506933,
-0.11018330603837967,
0.014917339198291302,
0.013548837043344975,
0.07078913599252701,
0.06232479214668274,
0.009646967984735966,
0.01151877362281084,
-0.021598177030682564,
0.2456178516149521,
-0.0954652726650238,
-0.04430177062749863,
-0.10939180850982666,
0.29019132256507874,
0.03431282192468643,
0.0016141736414283514,
0.026468364521861076,
-0.07295914739370346,
-0.010271442122757435,
0.2286510020494461,
0.17699094116687775,
-0.1165318489074707,
-0.008708895184099674,
0.00492211664095521,
-0.007045229431241751,
-0.004108157008886337,
0.10807367414236069,
0.09717002511024475,
0.037984974682331085,
-0.09937871992588043,
-0.04219667613506317,
-0.025488393381237984,
-0.031447846442461014,
-0.030896805226802826,
0.07915139943361282,
0.026698444038629532,
0.018573476001620293,
-0.05945734679698944,
0.043708670884370804,
-0.07175356894731522,
-0.10994996875524521,
0.06808049231767654,
-0.21573501825332642,
-0.17456208169460297,
-0.027590222656726837,
0.06229446455836296,
0.02155875228345394,
0.07249417155981064,
-0.024299105629324913,
-0.006704878993332386,
0.07940018177032471,
-0.016732007265090942,
-0.06199353188276291,
-0.1070980429649353,
0.10544995218515396,
-0.09495073556900024,
0.19682423770427704,
-0.047200337052345276,
0.048412907868623734,
0.11637755483388901,
0.04999631643295288,
-0.07480812817811966,
0.043046023696660995,
0.05958407372236252,
-0.09588882327079773,
0.02586553804576397,
0.13978761434555054,
-0.048779454082250595,
0.08047907054424286,
0.03550424054265022,
-0.1517646163702011,
0.01805322989821434,
-0.07355251908302307,
-0.07428668439388275,
-0.030563239008188248,
-0.04522649943828583,
-0.03618873283267021,
0.14076092839241028,
0.24955657124519348,
-0.029349206015467644,
0.013129165396094322,
-0.07485652714967728,
0.003950111567974091,
0.054614175111055374,
0.054860472679138184,
-0.06460306793451309,
-0.23843276500701904,
0.002485887613147497,
0.0697617158293724,
-0.010654392652213573,
-0.23470166325569153,
-0.08844705671072006,
0.004091604147106409,
-0.05367844179272652,
-0.08337385207414627,
0.09824232757091522,
0.055510640144348145,
0.04235042631626129,
-0.04358693212270737,
-0.06732012331485748,
-0.06624514609575272,
0.16375784575939178,
-0.15720592439174652,
-0.08453965187072754
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SciBERT_AsymmetricLoss_25K_bs64_P4_N1
This model is a fine-tuned version of [allenai/scibert_scivocab_uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 30.2502
- Accuracy: 0.9871
- Precision: 0.4247
- Recall: 0.8998
- F1: 0.5770
- Hamming: 0.0129
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 25000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Hamming |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 36.6287 | 0.16 | 5000 | 34.9978 | 0.9852 | 0.3863 | 0.8728 | 0.5355 | 0.0148 |
| 33.8929 | 0.32 | 10000 | 32.4942 | 0.9857 | 0.3958 | 0.8901 | 0.5480 | 0.0143 |
| 32.5419 | 0.47 | 15000 | 31.3170 | 0.9867 | 0.4162 | 0.8941 | 0.5680 | 0.0133 |
| 31.565 | 0.63 | 20000 | 30.6092 | 0.9869 | 0.4201 | 0.8975 | 0.5723 | 0.0131 |
| 31.105 | 0.79 | 25000 | 30.2502 | 0.9871 | 0.4247 | 0.8998 | 0.5770 | 0.0129 |
### Framework versions
- Transformers 4.35.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.7.1
- Tokenizers 0.14.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy", "precision", "recall", "f1"], "base_model": "allenai/scibert_scivocab_uncased", "model-index": [{"name": "SciBERT_AsymmetricLoss_25K_bs64_P4_N1", "results": []}]} | text-classification | bdpc/SciBERT_AsymmetricLoss_25K_bs64_P4_N1 | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:allenai/scibert_scivocab_uncased",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T16:58:18+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us
| SciBERT\_AsymmetricLoss\_25K\_bs64\_P4\_N1
==========================================
This model is a fine-tuned version of allenai/scibert\_scivocab\_uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 30.2502
* Accuracy: 0.9871
* Precision: 0.4247
* Recall: 0.8998
* F1: 0.5770
* Hamming: 0.0129
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 25000
### Training results
### Framework versions
* Transformers 4.35.0.dev0
* Pytorch 2.0.1+cu118
* Datasets 2.7.1
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
61,
116,
4,
36
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 25000### Training results### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
-0.12239006906747818,
0.07038096338510513,
-0.002270608674734831,
0.13089551031589508,
0.17182838916778564,
0.03552595153450966,
0.11841200292110443,
0.13127560913562775,
-0.06969314068555832,
0.038499392569065094,
0.12407459318637848,
0.1446039080619812,
0.022002389654517174,
0.13452577590942383,
-0.05503028258681297,
-0.2885366678237915,
-0.0033677059691399336,
0.01731622777879238,
-0.06730091571807861,
0.13224023580551147,
0.0828372985124588,
-0.1337146908044815,
0.09499790519475937,
-0.020193813368678093,
-0.16191041469573975,
-0.0009068193612620234,
0.01633271761238575,
-0.044335950165987015,
0.14709243178367615,
0.01964900828897953,
0.12271825969219208,
0.022447623312473297,
0.11982916295528412,
-0.17441865801811218,
0.008429979905486107,
0.04743145778775215,
0.02328234352171421,
0.09247520565986633,
0.05412723496556282,
0.00020402234804350883,
0.11227516084909439,
-0.09098335355520248,
0.058341994881629944,
0.020313246175646782,
-0.13214179873466492,
-0.1700744330883026,
-0.07841253280639648,
0.02907271310687065,
0.07433215528726578,
0.08933878690004349,
-0.0061609200201928616,
0.10206861793994904,
-0.08811420947313309,
0.10985638201236725,
0.22064051032066345,
-0.2594301402568817,
-0.0766574814915657,
0.010001081973314285,
0.004469216335564852,
0.09542322158813477,
-0.10710377246141434,
-0.01143158320337534,
0.034490350633859634,
0.03448912128806114,
0.12150989472866058,
-0.02728152461349964,
-0.12987226247787476,
0.022906633093953133,
-0.13800175487995148,
-0.02804391272366047,
0.09836933761835098,
0.023764215409755707,
-0.043362807482481,
-0.027673223987221718,
-0.0765603631734848,
-0.17501822113990784,
-0.04179663583636284,
-0.012468980625271797,
0.03498247265815735,
-0.04469481483101845,
-0.07365084439516068,
-0.026723813265562057,
-0.08419449627399445,
-0.0973527580499649,
-0.04072478413581848,
0.17177926003932953,
0.03978623077273369,
0.012730268761515617,
-0.0018546836217865348,
0.12453312426805496,
0.01509181223809719,
-0.15217764675617218,
0.00950139295309782,
0.013917750678956509,
-0.049028120934963226,
-0.04589609429240227,
-0.057228997349739075,
-0.016281330958008766,
-0.01917402632534504,
0.12961876392364502,
-0.029023857787251472,
0.05072631686925888,
0.03436003252863884,
0.016388587653636932,
-0.10456088930368423,
0.19624429941177368,
-0.05854874104261398,
-0.022884616628289223,
-0.007876993156969547,
0.0894712433218956,
0.007368528749793768,
-0.033018648624420166,
-0.10883118957281113,
-0.0166451595723629,
0.10293465107679367,
0.023269997909665108,
-0.08344634622335434,
0.06950521469116211,
-0.04303796589374542,
-0.033930059522390366,
0.019838472828269005,
-0.09911400079727173,
0.0360114648938179,
0.01767577975988388,
-0.10162538290023804,
-0.018144438043236732,
0.024868860840797424,
0.004757971037179232,
-0.015042777173221111,
0.1282130479812622,
-0.09785988926887512,
0.03428512066602707,
-0.09385675191879272,
-0.12153982371091843,
-0.0053965444676578045,
-0.09725828468799591,
0.0011557049583643675,
-0.0817437693476677,
-0.19228290021419525,
-0.028452470898628235,
0.03871012106537819,
-0.052222736179828644,
-0.0348193496465683,
-0.06804528087377548,
-0.06677358597517014,
0.02783213183283806,
-0.007976556196808815,
0.12797744572162628,
-0.05714099481701851,
0.11634469777345657,
0.03988274186849594,
0.07020349055528641,
-0.027476532384753227,
0.05806878209114075,
-0.10281573981046677,
0.010806224308907986,
-0.1829170435667038,
0.06699594855308533,
-0.05433490499854088,
0.04267386719584465,
-0.08098269253969193,
-0.10431589186191559,
0.024051155894994736,
0.0008766418322920799,
0.09290977567434311,
0.1409061849117279,
-0.1830790638923645,
-0.07253091037273407,
0.1420736014842987,
-0.06886310875415802,
-0.090731680393219,
0.11803504824638367,
-0.06576776504516602,
0.03177472949028015,
0.06343913078308105,
0.15235337615013123,
0.08405910432338715,
-0.0665288120508194,
0.016191352158784866,
-0.009675825014710426,
0.08389129489660263,
-0.011932277120649815,
0.07144893705844879,
0.020486149936914444,
-0.000908686954062432,
0.025170424953103065,
-0.0617491789162159,
0.04150257632136345,
-0.10790619999170303,
-0.09412276744842529,
-0.035276785492897034,
-0.10416639596223831,
0.08382663130760193,
0.07279685884714127,
0.07795432955026627,
-0.08594062924385071,
-0.08733929693698883,
0.057237446308135986,
0.09898535907268524,
-0.06270033121109009,
0.01929069682955742,
-0.06239078938961029,
0.054203860461711884,
-0.022769588977098465,
-0.025389688089489937,
-0.18733346462249756,
-0.045162271708250046,
0.020382581278681755,
0.04846292361617088,
0.02281942404806614,
0.0016226792940869927,
0.08511372655630112,
0.08827467262744904,
-0.06484182924032211,
-0.031042365357279778,
-0.040745727717876434,
-0.00862665381282568,
-0.14019912481307983,
-0.20115338265895844,
-0.06440245360136032,
-0.01868603751063347,
0.13572026789188385,
-0.21429120004177094,
0.03430602699518204,
-0.013535602018237114,
0.08657269924879074,
0.020001845434308052,
-0.024251582100987434,
-0.032936129719018936,
0.07709255069494247,
-0.03667707368731499,
-0.057209283113479614,
0.06744059920310974,
-0.013141147792339325,
-0.08329112827777863,
-0.050432052463293076,
-0.11685102432966232,
0.13448618352413177,
0.10413527488708496,
-0.08861696720123291,
-0.09657712280750275,
0.004747803322970867,
-0.05052494630217552,
-0.03206455707550049,
-0.05012806877493858,
0.02545151673257351,
0.17179572582244873,
0.005685955286026001,
0.15303145349025726,
-0.056193869560956955,
-0.03603309392929077,
0.020509283989667892,
-0.007702842820435762,
0.042179323732852936,
0.13229165971279144,
0.10557302832603455,
-0.08586869388818741,
0.12952619791030884,
0.10975980758666992,
-0.08752638846635818,
0.14262709021568298,
-0.024030623957514763,
-0.08278664201498032,
-0.01359491515904665,
-0.026771798729896545,
0.0010569716105237603,
0.10394185036420822,
-0.11336454749107361,
-0.017819305881857872,
0.011675954796373844,
0.014026987366378307,
-0.003758769016712904,
-0.20600396394729614,
-0.0349532887339592,
0.041023433208465576,
-0.05383365973830223,
-0.03309403732419014,
-0.013819299638271332,
0.00809128675609827,
0.12128349393606186,
0.011991294100880623,
-0.09089910238981247,
0.01355469785630703,
-0.00276798103004694,
-0.06932799518108368,
0.20992186665534973,
-0.07254642993211746,
-0.12312544137239456,
-0.12119962275028229,
-0.05350416153669357,
-0.05827227607369423,
0.02897077426314354,
0.03125174343585968,
-0.09345889091491699,
-0.009478596970438957,
-0.06180248409509659,
0.024914832785725594,
0.005248423665761948,
0.04262742027640343,
-0.024328241124749184,
-0.01973292976617813,
0.05876430496573448,
-0.08981389552354813,
-0.0028476088773459196,
-0.05926596373319626,
-0.08142127841711044,
0.03406904265284538,
0.03101099655032158,
0.11546559631824493,
0.1509244740009308,
-0.023641230538487434,
0.012300890870392323,
-0.02422642521560192,
0.22516395151615143,
-0.07099036127328873,
-0.0068967039696872234,
0.138851135969162,
-0.026212040334939957,
0.04785362631082535,
0.13172578811645508,
0.0631052702665329,
-0.08356857299804688,
0.010682963766157627,
0.04271089658141136,
-0.0341249518096447,
-0.21104124188423157,
-0.037000760436058044,
-0.05317976698279381,
0.005707957316190004,
0.10244151949882507,
0.013822940178215504,
0.02043784223496914,
0.07110318541526794,
0.016179269179701805,
0.048033326864242554,
-0.03800271078944206,
0.05945992097258568,
0.09588494896888733,
0.044698722660541534,
0.13814692199230194,
-0.02504982426762581,
-0.06687279045581818,
0.042679667472839355,
-0.023041626438498497,
0.1997382640838623,
-0.026216614991426468,
0.12034111469984055,
0.014289258979260921,
0.15840868651866913,
-0.009299556724727154,
0.08283321559429169,
0.011236663907766342,
-0.03521028533577919,
-0.01213148608803749,
-0.04027840495109558,
-0.06489378213882446,
0.02492157556116581,
-0.0615946464240551,
0.06076880544424057,
-0.15200024843215942,
0.01523659285157919,
0.04764033481478691,
0.2705806791782379,
0.06676151603460312,
-0.32958221435546875,
-0.10971777141094208,
0.007234081160277128,
-0.036199044436216354,
-0.03388427942991257,
0.008762715384364128,
0.12918280065059662,
-0.10107716917991638,
0.026911530643701553,
-0.06965263187885284,
0.0912308543920517,
-0.06432180106639862,
0.04941486194729805,
0.07924848794937134,
0.09748173505067825,
-0.0077408477663993835,
0.07018333673477173,
-0.27493512630462646,
0.2770404517650604,
0.013825390487909317,
0.0627499520778656,
-0.0700584277510643,
-0.00786398071795702,
0.05056403577327728,
0.08633213490247726,
0.060020167380571365,
-0.012451590038836002,
-0.04552178084850311,
-0.2327864170074463,
-0.07576194405555725,
0.02321651019155979,
0.11330311000347137,
-0.0644078478217125,
0.11162886023521423,
-0.04210285469889641,
-0.0032250939402729273,
0.05830661579966545,
-0.04423123970627785,
-0.05232122540473938,
-0.08034133911132812,
0.0038433547597378492,
0.008150003850460052,
-0.007344100158661604,
-0.0489417240023613,
-0.12082503736019135,
-0.07118731737136841,
0.15340112149715424,
-0.03301874175667763,
-0.03455265611410141,
-0.13572381436824799,
0.09249578416347504,
0.09975622594356537,
-0.0914166271686554,
0.027698101475834846,
0.014259293675422668,
0.07630632817745209,
0.03966827318072319,
-0.0584312304854393,
0.11153434216976166,
-0.06539176404476166,
-0.2048051506280899,
-0.0641174241900444,
0.10474395751953125,
0.045213837176561356,
0.07100849598646164,
-0.021890288218855858,
0.025506693869829178,
-0.019763877615332603,
-0.0837186798453331,
0.026365172117948532,
-0.010994844138622284,
0.05372616648674011,
0.04464171826839447,
-0.06631428748369217,
0.045672476291656494,
-0.06296247988939285,
-0.008350170217454433,
0.15975536406040192,
0.24500009417533875,
-0.10207007080316544,
0.02417619898915291,
0.030164232477545738,
-0.06615900993347168,
-0.2066960632801056,
0.04072249308228493,
0.07441367954015732,
0.024937978014349937,
0.04037139564752579,
-0.20382609963417053,
0.08771143108606339,
0.09525907039642334,
-0.011051145382225513,
0.08907002955675125,
-0.30238470435142517,
-0.1271258443593979,
0.10659535974264145,
0.1258370280265808,
0.12308255583047867,
-0.14216050505638123,
-0.02015303634107113,
-0.011090027168393135,
-0.08133064955472946,
0.09111294150352478,
-0.06461554765701294,
0.1279584765434265,
-0.029369210824370384,
0.08897161483764648,
0.0302097350358963,
-0.04464256018400192,
0.11791153997182846,
0.016575243324041367,
0.09271527081727982,
-0.048984382301568985,
-0.04559503123164177,
0.022436564788222313,
-0.04936091974377632,
0.007371949963271618,
-0.07676497101783752,
0.046593062579631805,
-0.11720437556505203,
-0.017861295491456985,
-0.08944916725158691,
0.021612970158457756,
-0.03606710955500603,
-0.06448612362146378,
-0.016325149685144424,
0.0374222993850708,
0.0463714674115181,
-0.006653534714132547,
0.11898743361234665,
-0.014422859996557236,
0.1480257511138916,
0.10997170209884644,
0.0865267738699913,
-0.018907129764556885,
-0.020895663648843765,
0.0013944103848189116,
-0.005069685634225607,
0.05545942485332489,
-0.12528328597545624,
0.03395261988043785,
0.14881139993667603,
0.03429792821407318,
0.1277073621749878,
0.08298145234584808,
-0.0016531600849702954,
0.008662156760692596,
0.052623432129621506,
-0.1730736345052719,
-0.058394577354192734,
-0.01513443049043417,
-0.06459499895572662,
-0.1320958137512207,
0.04858201742172241,
0.11238663643598557,
-0.06819935142993927,
-0.01308648195117712,
-0.010966382920742035,
0.011705826036632061,
-0.03743020072579384,
0.21776831150054932,
0.05902910605072975,
0.059291381388902664,
-0.10401014983654022,
0.06399628520011902,
0.04900115728378296,
-0.05636313557624817,
0.007586529944092035,
0.09954381734132767,
-0.09021449834108353,
-0.03634560480713844,
0.06135982647538185,
0.15154771506786346,
-0.07580685615539551,
-0.018328489735722542,
-0.15647806227207184,
-0.11130106449127197,
0.07111833989620209,
0.17828921973705292,
0.09962186217308044,
0.020853176712989807,
-0.06655143946409225,
0.022309882566332817,
-0.12215378135442734,
0.09659730643033981,
0.05762745067477226,
0.07719870656728745,
-0.14473845064640045,
0.18751679360866547,
-0.02199823595583439,
0.04957422986626625,
-0.02999516949057579,
0.02244112640619278,
-0.10935354232788086,
0.010224604047834873,
-0.13072331249713898,
-0.03427128493785858,
-0.0065178032964468,
0.0005157325067557395,
-0.011834277771413326,
-0.0751551166176796,
-0.04827370122075081,
0.0018287680577486753,
-0.11437570303678513,
-0.02571658417582512,
0.01943645440042019,
0.03479974716901779,
-0.12090522795915604,
-0.04035745933651924,
0.03265730291604996,
-0.08077806979417801,
0.0791250541806221,
0.059578001499176025,
0.020609606057405472,
0.05064117908477783,
-0.12445028126239777,
0.012717261910438538,
0.03291705623269081,
0.009428560733795166,
0.06306450068950653,
-0.0664442628622055,
0.0022305322345346212,
-0.03386590629816055,
0.062340330332517624,
0.021519776433706284,
0.11060437560081482,
-0.11858026683330536,
0.029840100556612015,
-0.011216999031603336,
-0.06925348192453384,
-0.06505866348743439,
0.05867297947406769,
0.07389768958091736,
0.03140941262245178,
0.16329561173915863,
-0.08925597369670868,
0.051601339131593704,
-0.2181098908185959,
0.00008990098285721615,
-0.008830444887280464,
-0.120360367000103,
-0.13204564154148102,
-0.07951638847589493,
0.08901864290237427,
-0.04739021137356758,
0.09151442348957062,
0.018703293055295944,
0.09766723960638046,
0.02487986907362938,
-0.033752284944057465,
0.033930208534002304,
0.03136355057358742,
0.18021716177463531,
0.049268703907728195,
-0.04822346568107605,
0.07613780349493027,
0.05260326713323593,
0.09048043936491013,
0.14555074274539948,
0.22101272642612457,
0.12525056302547455,
-0.01419612392783165,
0.0909113958477974,
0.04209940880537033,
-0.07237182557582855,
-0.16593821346759796,
0.004391070455312729,
-0.051212284713983536,
0.0855998545885086,
-0.045730944722890854,
0.214377298951149,
0.047232963144779205,
-0.16798417270183563,
0.047059111297130585,
-0.04942953959107399,
-0.10331861674785614,
-0.12448010593652725,
-0.013456725515425205,
-0.08117780089378357,
-0.1410934180021286,
-0.0017057594377547503,
-0.10980669409036636,
0.03821501508355141,
0.10636288672685623,
0.00930856354534626,
-0.009974460117518902,
0.16052278876304626,
0.021347688511013985,
0.037782177329063416,
0.07069765776395798,
0.012359456159174442,
-0.0037504786159843206,
-0.07913175225257874,
-0.09391181915998459,
-0.0043456340208649635,
-0.013917669653892517,
0.03180159255862236,
-0.07397375255823135,
-0.05759376287460327,
0.031400784850120544,
-0.006047637667506933,
-0.11018330603837967,
0.014917339198291302,
0.013548837043344975,
0.07078913599252701,
0.06232479214668274,
0.009646967984735966,
0.01151877362281084,
-0.021598177030682564,
0.2456178516149521,
-0.0954652726650238,
-0.04430177062749863,
-0.10939180850982666,
0.29019132256507874,
0.03431282192468643,
0.0016141736414283514,
0.026468364521861076,
-0.07295914739370346,
-0.010271442122757435,
0.2286510020494461,
0.17699094116687775,
-0.1165318489074707,
-0.008708895184099674,
0.00492211664095521,
-0.007045229431241751,
-0.004108157008886337,
0.10807367414236069,
0.09717002511024475,
0.037984974682331085,
-0.09937871992588043,
-0.04219667613506317,
-0.025488393381237984,
-0.031447846442461014,
-0.030896805226802826,
0.07915139943361282,
0.026698444038629532,
0.018573476001620293,
-0.05945734679698944,
0.043708670884370804,
-0.07175356894731522,
-0.10994996875524521,
0.06808049231767654,
-0.21573501825332642,
-0.17456208169460297,
-0.027590222656726837,
0.06229446455836296,
0.02155875228345394,
0.07249417155981064,
-0.024299105629324913,
-0.006704878993332386,
0.07940018177032471,
-0.016732007265090942,
-0.06199353188276291,
-0.1070980429649353,
0.10544995218515396,
-0.09495073556900024,
0.19682423770427704,
-0.047200337052345276,
0.048412907868623734,
0.11637755483388901,
0.04999631643295288,
-0.07480812817811966,
0.043046023696660995,
0.05958407372236252,
-0.09588882327079773,
0.02586553804576397,
0.13978761434555054,
-0.048779454082250595,
0.08047907054424286,
0.03550424054265022,
-0.1517646163702011,
0.01805322989821434,
-0.07355251908302307,
-0.07428668439388275,
-0.030563239008188248,
-0.04522649943828583,
-0.03618873283267021,
0.14076092839241028,
0.24955657124519348,
-0.029349206015467644,
0.013129165396094322,
-0.07485652714967728,
0.003950111567974091,
0.054614175111055374,
0.054860472679138184,
-0.06460306793451309,
-0.23843276500701904,
0.002485887613147497,
0.0697617158293724,
-0.010654392652213573,
-0.23470166325569153,
-0.08844705671072006,
0.004091604147106409,
-0.05367844179272652,
-0.08337385207414627,
0.09824232757091522,
0.055510640144348145,
0.04235042631626129,
-0.04358693212270737,
-0.06732012331485748,
-0.06624514609575272,
0.16375784575939178,
-0.15720592439174652,
-0.08453965187072754
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# tinystarcoder-rlhf-model
This model is a fine-tuned version of [bigcode/tiny_starcoder_py](https://huggingface.co/bigcode/tiny_starcoder_py) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 2
- eval_batch_size: 1
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| {"license": "bigcode-openrail-m", "tags": ["generated_from_trainer"], "base_model": "bigcode/tiny_starcoder_py", "model-index": [{"name": "tinystarcoder-rlhf-model", "results": []}]} | text-generation | vedantpalit/tinystarcoder-rlhf-model | [
"transformers",
"safetensors",
"gpt_bigcode",
"text-generation",
"generated_from_trainer",
"base_model:bigcode/tiny_starcoder_py",
"license:bigcode-openrail-m",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T16:59:00+00:00 | [] | [] | TAGS
#transformers #safetensors #gpt_bigcode #text-generation #generated_from_trainer #base_model-bigcode/tiny_starcoder_py #license-bigcode-openrail-m #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# tinystarcoder-rlhf-model
This model is a fine-tuned version of bigcode/tiny_starcoder_py on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 2
- eval_batch_size: 1
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.17.0
- Tokenizers 0.15.1
| [
"# tinystarcoder-rlhf-model\n\nThis model is a fine-tuned version of bigcode/tiny_starcoder_py on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 2\n- eval_batch_size: 1\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #safetensors #gpt_bigcode #text-generation #generated_from_trainer #base_model-bigcode/tiny_starcoder_py #license-bigcode-openrail-m #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# tinystarcoder-rlhf-model\n\nThis model is a fine-tuned version of bigcode/tiny_starcoder_py on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 2\n- eval_batch_size: 1\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
83,
39,
6,
12,
8,
3,
105,
4,
33
] | [
"passage: TAGS\n#transformers #safetensors #gpt_bigcode #text-generation #generated_from_trainer #base_model-bigcode/tiny_starcoder_py #license-bigcode-openrail-m #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# tinystarcoder-rlhf-model\n\nThis model is a fine-tuned version of bigcode/tiny_starcoder_py on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 2\n- eval_batch_size: 1\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 100\n- num_epochs: 1### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1"
] | [
-0.08394484221935272,
0.15147535502910614,
-0.004756434820592403,
0.09890836477279663,
0.10344575345516205,
0.015683069825172424,
0.09299421310424805,
0.1685793399810791,
-0.1124889999628067,
0.09003719687461853,
0.06714519113302231,
0.006853069178760052,
0.08182850480079651,
0.14517252147197723,
-0.008520033210515976,
-0.2457135021686554,
-0.018146926537156105,
-0.011275538243353367,
-0.05308143422007561,
0.09252729266881943,
0.1386830359697342,
-0.07887037098407745,
0.06606096774339676,
0.007157470565289259,
-0.10444852709770203,
0.01158046256750822,
-0.05541741102933884,
-0.07543579488992691,
0.06815985590219498,
-0.038266897201538086,
0.053072623908519745,
0.008839412592351437,
0.10397235304117203,
-0.17714326083660126,
-0.008614310063421726,
0.07558406889438629,
0.04115469753742218,
0.09017868340015411,
0.07301055639982224,
0.003933479078114033,
0.0644744411110878,
-0.1794646829366684,
0.09756266325712204,
0.03397148475050926,
-0.06391476839780807,
-0.19573108851909637,
-0.09370198100805283,
0.11857035011053085,
0.074152372777462,
0.0755961537361145,
0.026699399575591087,
0.10892711579799652,
-0.08114376664161682,
0.062262821942567825,
0.23692189157009125,
-0.2161748707294464,
-0.03333713859319687,
0.040182191878557205,
0.062126267701387405,
0.06285948306322098,
-0.10045298933982849,
-0.0431860089302063,
0.02868630923330784,
0.017483297735452652,
0.09632978588342667,
0.016317635774612427,
-0.037344954907894135,
-0.013869548216462135,
-0.10766448080539703,
-0.037720225751399994,
0.1451922208070755,
0.010768112726509571,
-0.043732400983572006,
-0.1613970547914505,
-0.06257867813110352,
-0.1825442761182785,
-0.015110494568943977,
-0.04590534418821335,
0.047591060400009155,
-0.040484216064214706,
-0.015068181790411472,
-0.060069549828767776,
-0.07447069883346558,
-0.054471202194690704,
0.012698207050561905,
0.11566021293401718,
0.04355612397193909,
0.012988190166652203,
-0.050689756870269775,
0.10495049506425858,
0.0026692859828472137,
-0.16447316110134125,
-0.06014242768287659,
-0.0013379796873778105,
-0.032243888825178146,
-0.04691155254840851,
-0.021012190729379654,
0.03948727622628212,
0.0005769905401393771,
0.1919005662202835,
-0.06155506148934364,
0.0417567640542984,
0.014567520469427109,
-0.0006361011764965951,
-0.009068922139704227,
0.13060647249221802,
-0.04645051062107086,
-0.04541802033782005,
0.05165380239486694,
0.09662626683712006,
0.06372292339801788,
-0.01190990675240755,
-0.08468655496835709,
-0.060319866985082626,
0.0871523916721344,
0.09401096403598785,
-0.00977327674627304,
-0.015347104519605637,
-0.04492852836847305,
-0.028803182765841484,
0.09847959876060486,
-0.14278344810009003,
0.057870052754879,
0.009908368811011314,
-0.09758932888507843,
-0.0015459361020475626,
-0.009895825758576393,
-0.001169942319393158,
-0.03426014631986618,
0.07298900932073593,
-0.05440390855073929,
-0.017292410135269165,
-0.053843237459659576,
-0.017177965492010117,
0.04443107917904854,
-0.10841000825166702,
0.004573357291519642,
-0.05060072988271713,
-0.20075586438179016,
-0.023190079256892204,
0.007988660596311092,
-0.09828711301088333,
-0.05682462081313133,
-0.022746199741959572,
-0.12233315408229828,
0.027718432247638702,
-0.009250392206013203,
0.08241419494152069,
-0.03336453437805176,
0.06764362752437592,
0.01130069512873888,
0.047275032848119736,
0.05827809497714043,
0.03891393914818764,
-0.11343264579772949,
0.02248399518430233,
-0.12009941786527634,
0.07773338258266449,
-0.0571504645049572,
-0.01665034517645836,
-0.10704132914543152,
-0.11633837223052979,
-0.00298565486446023,
-0.04218944534659386,
0.049388010054826736,
0.13660314679145813,
-0.14174729585647583,
-0.02325228601694107,
0.17793595790863037,
-0.08675937354564667,
-0.08306676894426346,
0.10067855566740036,
-0.02504703402519226,
0.005402753129601479,
0.04544472694396973,
0.11448244750499725,
0.11802888661623001,
-0.11908837407827377,
-0.049974843859672546,
0.03846096992492676,
0.04844826087355614,
0.008976904675364494,
0.12392764538526535,
0.010456264019012451,
0.1061505675315857,
0.005332575645297766,
-0.07356832921504974,
-0.03338075801730156,
-0.04607734456658363,
-0.07993502169847488,
-0.06447526812553406,
-0.06428296864032745,
0.04342576488852501,
0.027005240321159363,
0.037602368742227554,
-0.09858434647321701,
-0.14404647052288055,
0.05191684886813164,
0.1532398760318756,
-0.027543656527996063,
0.01867334358394146,
-0.08418961614370346,
0.04227710887789726,
-0.0495419017970562,
-0.0336514376103878,
-0.1492992341518402,
-0.07612375915050507,
0.059621453285217285,
-0.07667067646980286,
0.008031398057937622,
-0.017469462007284164,
0.07377634197473526,
0.06717128306627274,
-0.0419732928276062,
-0.04286561533808708,
-0.09687528759241104,
-0.019583286717534065,
-0.07556235790252686,
-0.15259143710136414,
-0.08550558984279633,
-0.03604497015476227,
0.16481703519821167,
-0.15829119086265564,
0.028370236977934837,
0.07600303739309311,
0.1420050412416458,
0.051281411200761795,
-0.04960852488875389,
0.011155495420098305,
0.012165739201009274,
-0.017085546627640724,
-0.11070108413696289,
0.016004452481865883,
-0.0017751685809344053,
-0.06495169550180435,
-0.02328791469335556,
-0.1324257254600525,
0.06931528449058533,
0.07472791522741318,
0.111770398914814,
-0.0736437439918518,
0.016785239800810814,
-0.0656878799200058,
-0.02677524834871292,
-0.056572191417217255,
-0.03796682506799698,
0.17316405475139618,
0.023358982056379318,
0.11806876212358475,
-0.08923900872468948,
-0.11764098703861237,
-0.002481641247868538,
0.01576765440404415,
-0.027619503438472748,
0.07676240056753159,
0.0012154302094131708,
-0.07671574503183365,
0.12685607373714447,
0.09182500094175339,
-0.01888337917625904,
0.12189843505620956,
-0.054686449468135834,
-0.08186034113168716,
-0.03288408741354942,
0.019679296761751175,
0.00426073744893074,
0.07185382395982742,
-0.05160736292600632,
0.016111917793750763,
0.04476488009095192,
0.016701869666576385,
0.03342260420322418,
-0.13209237158298492,
-0.0031027013901621103,
0.058228615671396255,
-0.04542889446020126,
0.007233893498778343,
-0.025222886353731155,
0.0359012745320797,
0.08528997749090195,
0.045743267983198166,
0.0002861715911421925,
0.05265166610479355,
-0.011894584633409977,
-0.06184254214167595,
0.14565928280353546,
-0.1016223356127739,
-0.1442236602306366,
-0.15157577395439148,
0.08045762777328491,
-0.03936136141419411,
-0.017861727625131607,
0.014610879123210907,
-0.08037664741277695,
-0.033659204840660095,
-0.10874065011739731,
-0.0343652106821537,
-0.04504433646798134,
-0.0270908884704113,
0.07113005220890045,
0.039796728640794754,
0.050926901400089264,
-0.14594650268554688,
0.0033677704632282257,
0.022707607597112656,
-0.1113927960395813,
-0.02151506207883358,
0.0739649161696434,
0.049710944294929504,
0.1057107001543045,
-0.02912200428545475,
0.008753761649131775,
-0.03442792221903801,
0.16410551965236664,
-0.08033396303653717,
0.030502626672387123,
0.19239184260368347,
0.018855378031730652,
0.06751085817813873,
0.09761802852153778,
0.027435289695858955,
-0.05092655494809151,
-0.004536820575594902,
0.02292734570801258,
-0.03224669396877289,
-0.2310270071029663,
-0.04323940724134445,
-0.024130139499902725,
-0.023766862228512764,
0.08978529274463654,
0.07096388190984726,
-0.006718350574374199,
0.055355481803417206,
-0.04723820835351944,
0.02407335676252842,
-0.005417304113507271,
0.09583848714828491,
0.04013912379741669,
0.0555756650865078,
0.06967587769031525,
-0.03435666486620903,
-0.03609497845172882,
0.0819368064403534,
0.03569629788398743,
0.13770753145217896,
-0.07046355307102203,
0.18147875368595123,
-0.016015715897083282,
0.11858726292848587,
-0.02452143095433712,
0.06101470813155174,
-0.022475330159068108,
0.01975703053176403,
0.008743589743971825,
-0.08430840075016022,
-0.03342839702963829,
0.041118718683719635,
-0.04381928965449333,
0.07127714157104492,
-0.08396513015031815,
0.05031420662999153,
0.014701136387884617,
0.24456091225147247,
0.04946830868721008,
-0.29772305488586426,
-0.06900821626186371,
0.04159638658165932,
-0.013701907359063625,
-0.050166502594947815,
0.018364129588007927,
0.13518814742565155,
-0.13132242858409882,
0.07225845754146576,
-0.03201037272810936,
0.07061007618904114,
-0.06047101691365242,
-0.020834075286984444,
0.03859108313918114,
0.10830829292535782,
0.024713043123483658,
0.09947972744703293,
-0.1779138445854187,
0.10568619519472122,
0.012356951832771301,
0.11798463016748428,
-0.1128959134221077,
0.06650035083293915,
0.008402775041759014,
0.014118810184299946,
0.1294160634279251,
-0.013591097667813301,
-0.0601363405585289,
-0.17122364044189453,
-0.13469646871089935,
0.010428803972899914,
0.13157770037651062,
-0.0853809267282486,
0.08406711369752884,
-0.04870127886533737,
-0.016318798065185547,
0.022393163293600082,
-0.09452337771654129,
-0.089886873960495,
-0.12656940519809723,
0.03465860337018967,
0.0218243096023798,
-0.0316302664577961,
-0.09933177381753922,
-0.09360779821872711,
-0.024213051423430443,
0.16602610051631927,
-0.07679657638072968,
-0.09498590975999832,
-0.1405206322669983,
0.017865421250462532,
0.1593356877565384,
-0.07436514645814896,
0.03624076023697853,
0.02265157550573349,
0.17307107150554657,
0.031271643936634064,
-0.08600810915231705,
0.04283636063337326,
-0.08094005286693573,
-0.22578822076320648,
-0.028558578342199326,
0.18306733667850494,
0.013442830182611942,
0.055612962692976,
0.006432106718420982,
0.045076947659254074,
-0.0035113536287099123,
-0.08627456426620483,
0.013947899453341961,
0.0710134506225586,
0.06894384324550629,
0.06387761980295181,
-0.04933994263410568,
0.025745118036866188,
0.014317088760435581,
0.01629549451172352,
0.16579145193099976,
0.22508247196674347,
-0.09144040197134018,
0.13587632775306702,
0.061394210904836655,
-0.05819522961974144,
-0.1897144466638565,
0.026637813076376915,
0.08971098810434341,
0.010473000817000866,
0.03729548677802086,
-0.1859888881444931,
0.12030358612537384,
0.12486188858747482,
-0.040119390934705734,
0.08096406608819962,
-0.3284792900085449,
-0.1110936626791954,
0.017289385199546814,
0.03889606148004532,
-0.022058723494410515,
-0.1370343416929245,
-0.05756385996937752,
-0.032524921000003815,
-0.15414465963840485,
0.12268048524856567,
-0.07785823196172714,
0.09751047939062119,
-0.03249194473028183,
0.05184556543827057,
0.03496868163347244,
-0.056082144379615784,
0.12330829352140427,
0.011410937644541264,
0.050271112471818924,
-0.06991633772850037,
0.005427563097327948,
0.14370547235012054,
-0.07816080003976822,
0.09885240346193314,
-0.06341676414012909,
0.06513447314500809,
-0.1741187870502472,
-0.011235843412578106,
-0.058188337832689285,
0.06221992149949074,
-0.053867921233177185,
-0.05816227197647095,
-0.06621336191892624,
0.04895821586251259,
0.06527548283338547,
-0.03695027902722359,
0.0930933877825737,
0.06563447415828705,
0.09195343405008316,
0.12066623568534851,
0.1137348935008049,
-0.008441684767603874,
-0.09125377982854843,
-0.005413351114839315,
-0.013031055219471455,
0.04687946289777756,
-0.0980270579457283,
0.048278406262397766,
0.11332843452692032,
0.062435463070869446,
0.07994219660758972,
0.0417603962123394,
-0.05615850165486336,
0.013518943451344967,
0.03352905064821243,
-0.08187501132488251,
-0.12302949279546738,
-0.027115896344184875,
-0.00010410531103843823,
-0.1447584182024002,
0.03824503347277641,
0.10061270743608475,
-0.044263944029808044,
-0.03617604821920395,
-0.018515489995479584,
0.057947587221860886,
-0.005332488566637039,
0.16163098812103271,
0.06127887964248657,
0.07330459356307983,
-0.08760534226894379,
0.1377808302640915,
0.08168421685695648,
-0.062166035175323486,
0.07837507873773575,
0.10741470754146576,
-0.06102409586310387,
-0.012593820691108704,
0.09023275971412659,
0.07260917127132416,
0.01572570577263832,
-0.03448035567998886,
-0.06839100271463394,
-0.08826923370361328,
0.04974411427974701,
0.03762580826878548,
0.043672382831573486,
0.005510127637535334,
-0.022657934576272964,
0.04355524852871895,
-0.16608324646949768,
0.12053480744361877,
0.017920363694429398,
0.08033760637044907,
-0.14420679211616516,
0.09936481714248657,
-0.014671621844172478,
0.045918311923742294,
-0.01837795041501522,
0.03603023663163185,
-0.0616946704685688,
-0.02130625769495964,
-0.058548640459775925,
0.013944356702268124,
-0.03299042955040932,
0.006217719987034798,
-0.03582986444234848,
-0.024943025782704353,
-0.03011714294552803,
0.06388850510120392,
-0.05775227025151253,
-0.0816456750035286,
-0.03687434643507004,
0.04410677030682564,
-0.15365707874298096,
0.0007333611720241606,
0.026668768376111984,
-0.11492826789617538,
0.11834399402141571,
0.037492770701646805,
0.019131643697619438,
-0.031808026134967804,
-0.08304214477539062,
-0.02289559505879879,
0.009157116524875164,
0.017305435612797737,
0.04422009363770485,
-0.1125340536236763,
0.028372803702950478,
-0.020921951159834862,
-0.01556643657386303,
0.013456979766488075,
0.07231058180332184,
-0.13235695660114288,
0.005197592079639435,
-0.05882444977760315,
-0.031130829825997353,
-0.06508861482143402,
0.043982427567243576,
0.11719769984483719,
-0.022901752963662148,
0.14959529042243958,
-0.043774303048849106,
0.0682380199432373,
-0.19346477091312408,
-0.03204946964979172,
-0.008337944746017456,
-0.04055317863821983,
-0.08481472730636597,
0.022072240710258484,
0.07423101365566254,
-0.03840840235352516,
0.12336964160203934,
-0.012657723389565945,
0.04356101527810097,
0.042322952300310135,
-0.01813465729355812,
0.040585052222013474,
0.02274729683995247,
0.18257597088813782,
0.07099947333335876,
-0.05131705477833748,
0.07796406745910645,
-0.046981364488601685,
0.07145297527313232,
-0.03994390740990639,
0.15196460485458374,
0.15900668501853943,
-0.031869497150182724,
0.06981301307678223,
0.0777265727519989,
-0.11030550301074982,
-0.1865164041519165,
0.03702300786972046,
-0.003963193856179714,
0.0683845579624176,
-0.050211772322654724,
0.16087864339351654,
0.1082458347082138,
-0.15877121686935425,
0.04162154719233513,
-0.03555358946323395,
-0.06466174870729446,
-0.12122877687215805,
-0.06737632304430008,
-0.08484161645174026,
-0.07481300085783005,
0.009337246417999268,
-0.11567498743534088,
0.02534599043428898,
0.10208790749311447,
0.010986393317580223,
-0.003817459102720022,
0.15910795331001282,
-0.04751815274357796,
0.014775623567402363,
0.06428439170122147,
0.02760043926537037,
0.008179032243788242,
-0.01658249832689762,
-0.06417075544595718,
0.04699978232383728,
0.03680845722556114,
0.11077156662940979,
-0.012787887826561928,
0.015018614009022713,
-0.0008295344305224717,
-0.027659693732857704,
-0.0831754058599472,
0.02309097722172737,
0.0091854901984334,
0.03645041212439537,
0.022326748818159103,
0.023240892216563225,
0.008240167051553726,
-0.04253740608692169,
0.24426564574241638,
-0.06514456123113632,
-0.06602231413125992,
-0.08568411320447922,
0.14181523025035858,
0.04211290180683136,
-0.01126346830278635,
0.07121045887470245,
-0.12926025688648224,
-0.037092916667461395,
0.10515400022268295,
0.15790890157222748,
-0.019660599529743195,
-0.01019237469881773,
0.010322865098714828,
-0.015573523938655853,
-0.03873766213655472,
0.10454771667718887,
0.08815320581197739,
0.08474013954401016,
-0.058431778103113174,
0.04774891212582588,
0.02741023525595665,
-0.03846423700451851,
-0.10576566308736801,
0.10559991747140884,
-0.03357768803834915,
0.01234275009483099,
-0.011733445338904858,
0.039737213402986526,
0.0015878919512033463,
-0.14861471951007843,
0.05897849053144455,
-0.18458107113838196,
-0.18813776969909668,
0.008702831342816353,
0.09590877592563629,
-0.019291970878839493,
0.05054694041609764,
0.04576687514781952,
-0.02140423282980919,
0.08662143349647522,
-0.0055348994210362434,
-0.07409334927797318,
-0.04858573526144028,
0.06005789712071419,
-0.09133744239807129,
0.2710112929344177,
-0.0002786808181554079,
0.06772257387638092,
0.11339615285396576,
-0.003541517537087202,
-0.2044481486082077,
-0.004234706982970238,
0.07456027716398239,
-0.046059828251600266,
0.03619524464011192,
0.17616575956344604,
-0.04171779006719589,
0.052895065397024155,
0.06366066634654999,
-0.11540348827838898,
-0.05410347878932953,
0.021543225273489952,
0.006811733357608318,
-0.11130239814519882,
-0.009293349459767342,
-0.04961462318897247,
0.14925794303417206,
0.15768782794475555,
-0.06257748603820801,
-0.019639937207102776,
-0.07614559680223465,
0.04182226210832596,
0.056826066225767136,
0.08914963901042938,
0.014896423555910587,
-0.20106366276741028,
0.023236265406012535,
0.07131977379322052,
0.0635722205042839,
-0.2180442065000534,
-0.07666079699993134,
0.04745936393737793,
-0.02996215596795082,
-0.09114408493041992,
0.0579543262720108,
0.05771220847964287,
0.03275777027010918,
-0.018725724890828133,
-0.08429817855358124,
-0.06227031722664833,
0.11995308101177216,
-0.19492100179195404,
-0.017671946436166763
] |
null | null | transformers |
# Nordavind-7B
A model trained on top of <https://huggingface.co/norallm/normistral-7b-warm> on various Norwegian instruct data with a context length of 4096. The name - *nordavind* - a cold wind from the north, is derived from Mistral. See https://en.wikipedia.org/wiki/Mistral_(wind)!
Runtime: ~20 hours on a V100 32GB. Final loss: `1.301`.
Prompt format:
```python
<s>{system_prompt} [INST] {inst} [/INST] \\n {res} </s>
```
Inference prompt:
```python
<s>{system_prompt} [INST] {inst} [/INST] \\n
```
## Examples:
*generate("Lag en skikkelig norsk setning", tokens=50, first_sent=True)*
> `Jeg liker å gå på tur i skogen.`
*generate("Lag en liste med tre forskjellige bruksområder (inkludert eksempler) for kunstig intelligens.")*
>1. Automatisering av oppgaver: AI kan brukes til å automatisere oppgaver som å svare på kundehenvendelser, behandle medisinske bilder og diagnostisere sykdommer, og til og med kjøre biler.
>2. Forbedre kundeservice: AI kan brukes for å forbedre kundeservicen ved å gi personlige anbefalinger, gi raskere svar og gi mer nøyaktige svar.
>3. Forbedre sikkerhet: AI kan brukes i sikkerhetssystemer for å oppdage og forhindre cyberangrep, oppdage svindel og oppdage mistenkelige aktiviteter.
*generate("Hvem tror du vinner valget i Bergen?", tokens=60)*
>Det er vanskelig å forutsi hvem som vinner valget i Bergen, men det er sannsynlig at enten Arbeiderpartiet eller Høyre vil vinne. Arbeiderpartiet har hatt makten i Bergen i mange år, men Høyre har hatt en sterk tilstedeværelse i byen de siste årene.
*generate("Forklar dyp læring til en femteklassing")*
>Dyp læring er en type maskinlæring som bruker nevrale nettverk for å lære komplekse mønstre i data. Det innebærer å trene en datamaskin til å gjenkjenne mønstre i data og ta beslutninger basert på disse mønstrene.
## Model loading
```python
from transformers import pipeline
pipe = pipeline("text-generation", model="tollefj/nordavind-7b-instruct-warm")
```
## Inference
```python
system_prompt = 'Du er "Nordavind", en hjelpsom assistent.'
def make_prompt(inst, res=None):
if not res:
return f"""<s>{system_prompt} [INST] {inst} [/INST] \\n"""
return f"""<s>{system_prompt} [INST] {inst} [/INST] \\n {res} </s>"""
def postprocess(output, first_sent=False):
output = output.split("\\n")[-1].strip()
# ignore hashtags as we often see #no_output
output = output.split("#")[0].strip()
# ignore incomplete sentences
if not output.endswith("."):
output = output.rsplit(".", 1)[0] + "."
if first_sent:
return output.split(".")[0] + "."
return output
def generate(prompt, tokens=100, first_sent=False, sample=False, temperature=1.0):
prompt = make_prompt(prompt)
output = pipe(
prompt,
max_length=tokens,
do_sample=sample,
temperature=temperature,
)
output = output[0]["generated_text"]
output = postprocess(output, first_sent=first_sent)
print(output)
```
# Training details
The model was fine-tuned in an 4bit BitsAndBytes config.
```python
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_compute_dtype=getattr(torch, "float16"),
bnb_4bit_use_double_quant=False,
)
```
with the following LoRa-configuration:
```python
config = LoraConfig(
target_modules=[
"q_proj",
"k_proj",
"v_proj",
"o_proj",
"gate_proj",
"up_proj",
"down_proj",
"lm_head",
],
bias="none",
lora_dropout=0.05,
task_type="CAUSAL_LM",
)
``` | {"language": ["no"], "datasets": ["tollefj/nor-instruct-combined"]} | text-generation | tollefj/nordavind-7b-instruct-warm | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"no",
"dataset:tollefj/nor-instruct-combined",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:03:44+00:00 | [] | [
"no"
] | TAGS
#transformers #safetensors #mistral #text-generation #no #dataset-tollefj/nor-instruct-combined #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Nordavind-7B
A model trained on top of <URL on various Norwegian instruct data with a context length of 4096. The name - *nordavind* - a cold wind from the north, is derived from Mistral. See URL
Runtime: ~20 hours on a V100 32GB. Final loss: '1.301'.
Prompt format:
Inference prompt:
## Examples:
*generate("Lag en skikkelig norsk setning", tokens=50, first_sent=True)*
> 'Jeg liker å gå på tur i skogen.'
*generate("Lag en liste med tre forskjellige bruksområder (inkludert eksempler) for kunstig intelligens.")*
>1. Automatisering av oppgaver: AI kan brukes til å automatisere oppgaver som å svare på kundehenvendelser, behandle medisinske bilder og diagnostisere sykdommer, og til og med kjøre biler.
>2. Forbedre kundeservice: AI kan brukes for å forbedre kundeservicen ved å gi personlige anbefalinger, gi raskere svar og gi mer nøyaktige svar.
>3. Forbedre sikkerhet: AI kan brukes i sikkerhetssystemer for å oppdage og forhindre cyberangrep, oppdage svindel og oppdage mistenkelige aktiviteter.
*generate("Hvem tror du vinner valget i Bergen?", tokens=60)*
>Det er vanskelig å forutsi hvem som vinner valget i Bergen, men det er sannsynlig at enten Arbeiderpartiet eller Høyre vil vinne. Arbeiderpartiet har hatt makten i Bergen i mange år, men Høyre har hatt en sterk tilstedeværelse i byen de siste årene.
*generate("Forklar dyp læring til en femteklassing")*
>Dyp læring er en type maskinlæring som bruker nevrale nettverk for å lære komplekse mønstre i data. Det innebærer å trene en datamaskin til å gjenkjenne mønstre i data og ta beslutninger basert på disse mønstrene.
## Model loading
## Inference
# Training details
The model was fine-tuned in an 4bit BitsAndBytes config.
with the following LoRa-configuration:
| [
"# Nordavind-7B\n\nA model trained on top of <URL on various Norwegian instruct data with a context length of 4096. The name - *nordavind* - a cold wind from the north, is derived from Mistral. See URL\nRuntime: ~20 hours on a V100 32GB. Final loss: '1.301'.\n\nPrompt format:\n\nInference prompt:",
"## Examples:\n\n*generate(\"Lag en skikkelig norsk setning\", tokens=50, first_sent=True)*\n\n> 'Jeg liker å gå på tur i skogen.'\n\n*generate(\"Lag en liste med tre forskjellige bruksområder (inkludert eksempler) for kunstig intelligens.\")*\n\n>1. Automatisering av oppgaver: AI kan brukes til å automatisere oppgaver som å svare på kundehenvendelser, behandle medisinske bilder og diagnostisere sykdommer, og til og med kjøre biler.\n>2. Forbedre kundeservice: AI kan brukes for å forbedre kundeservicen ved å gi personlige anbefalinger, gi raskere svar og gi mer nøyaktige svar.\n>3. Forbedre sikkerhet: AI kan brukes i sikkerhetssystemer for å oppdage og forhindre cyberangrep, oppdage svindel og oppdage mistenkelige aktiviteter.\n\n\n*generate(\"Hvem tror du vinner valget i Bergen?\", tokens=60)*\n>Det er vanskelig å forutsi hvem som vinner valget i Bergen, men det er sannsynlig at enten Arbeiderpartiet eller Høyre vil vinne. Arbeiderpartiet har hatt makten i Bergen i mange år, men Høyre har hatt en sterk tilstedeværelse i byen de siste årene.\n\n*generate(\"Forklar dyp læring til en femteklassing\")*\n>Dyp læring er en type maskinlæring som bruker nevrale nettverk for å lære komplekse mønstre i data. Det innebærer å trene en datamaskin til å gjenkjenne mønstre i data og ta beslutninger basert på disse mønstrene.",
"## Model loading",
"## Inference",
"# Training details\n\nThe model was fine-tuned in an 4bit BitsAndBytes config.\n\nwith the following LoRa-configuration:"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #no #dataset-tollefj/nor-instruct-combined #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Nordavind-7B\n\nA model trained on top of <URL on various Norwegian instruct data with a context length of 4096. The name - *nordavind* - a cold wind from the north, is derived from Mistral. See URL\nRuntime: ~20 hours on a V100 32GB. Final loss: '1.301'.\n\nPrompt format:\n\nInference prompt:",
"## Examples:\n\n*generate(\"Lag en skikkelig norsk setning\", tokens=50, first_sent=True)*\n\n> 'Jeg liker å gå på tur i skogen.'\n\n*generate(\"Lag en liste med tre forskjellige bruksområder (inkludert eksempler) for kunstig intelligens.\")*\n\n>1. Automatisering av oppgaver: AI kan brukes til å automatisere oppgaver som å svare på kundehenvendelser, behandle medisinske bilder og diagnostisere sykdommer, og til og med kjøre biler.\n>2. Forbedre kundeservice: AI kan brukes for å forbedre kundeservicen ved å gi personlige anbefalinger, gi raskere svar og gi mer nøyaktige svar.\n>3. Forbedre sikkerhet: AI kan brukes i sikkerhetssystemer for å oppdage og forhindre cyberangrep, oppdage svindel og oppdage mistenkelige aktiviteter.\n\n\n*generate(\"Hvem tror du vinner valget i Bergen?\", tokens=60)*\n>Det er vanskelig å forutsi hvem som vinner valget i Bergen, men det er sannsynlig at enten Arbeiderpartiet eller Høyre vil vinne. Arbeiderpartiet har hatt makten i Bergen i mange år, men Høyre har hatt en sterk tilstedeværelse i byen de siste årene.\n\n*generate(\"Forklar dyp læring til en femteklassing\")*\n>Dyp læring er en type maskinlæring som bruker nevrale nettverk for å lære komplekse mønstre i data. Det innebærer å trene en datamaskin til å gjenkjenne mønstre i data og ta beslutninger basert på disse mønstrene.",
"## Model loading",
"## Inference",
"# Training details\n\nThe model was fine-tuned in an 4bit BitsAndBytes config.\n\nwith the following LoRa-configuration:"
] | [
66,
85,
326,
4,
4,
32
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #no #dataset-tollefj/nor-instruct-combined #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Nordavind-7B\n\nA model trained on top of <URL on various Norwegian instruct data with a context length of 4096. The name - *nordavind* - a cold wind from the north, is derived from Mistral. See URL\nRuntime: ~20 hours on a V100 32GB. Final loss: '1.301'.\n\nPrompt format:\n\nInference prompt:## Examples:\n\n*generate(\"Lag en skikkelig norsk setning\", tokens=50, first_sent=True)*\n\n> 'Jeg liker å gå på tur i skogen.'\n\n*generate(\"Lag en liste med tre forskjellige bruksområder (inkludert eksempler) for kunstig intelligens.\")*\n\n>1. Automatisering av oppgaver: AI kan brukes til å automatisere oppgaver som å svare på kundehenvendelser, behandle medisinske bilder og diagnostisere sykdommer, og til og med kjøre biler.\n>2. Forbedre kundeservice: AI kan brukes for å forbedre kundeservicen ved å gi personlige anbefalinger, gi raskere svar og gi mer nøyaktige svar.\n>3. Forbedre sikkerhet: AI kan brukes i sikkerhetssystemer for å oppdage og forhindre cyberangrep, oppdage svindel og oppdage mistenkelige aktiviteter.\n\n\n*generate(\"Hvem tror du vinner valget i Bergen?\", tokens=60)*\n>Det er vanskelig å forutsi hvem som vinner valget i Bergen, men det er sannsynlig at enten Arbeiderpartiet eller Høyre vil vinne. Arbeiderpartiet har hatt makten i Bergen i mange år, men Høyre har hatt en sterk tilstedeværelse i byen de siste årene.\n\n*generate(\"Forklar dyp læring til en femteklassing\")*\n>Dyp læring er en type maskinlæring som bruker nevrale nettverk for å lære komplekse mønstre i data. Det innebærer å trene en datamaskin til å gjenkjenne mønstre i data og ta beslutninger basert på disse mønstrene.## Model loading## Inference"
] | [
-0.02992171235382557,
0.07532121241092682,
-0.012389962561428547,
0.045963436365127563,
0.04285994544625282,
-0.0004136826319154352,
0.0870424136519432,
0.07531016319990158,
0.05972912907600403,
0.10475395619869232,
0.0906565934419632,
0.07493271678686142,
0.05696523189544678,
0.09319797903299332,
0.035255882889032364,
-0.1489371359348297,
0.09802654385566711,
-0.08951196074485779,
0.044517967849969864,
0.05111860856413841,
0.0713048055768013,
-0.05437278747558594,
0.05488138273358345,
-0.02882847748696804,
0.0005497324164025486,
-0.006992138456553221,
-0.006715202704071999,
-0.005745678208768368,
0.09892215579748154,
0.05247662961483002,
0.028248758986592293,
-0.0053673842921853065,
0.04421716555953026,
-0.15492628514766693,
0.015773523598909378,
0.028896206989884377,
-0.012544549070298672,
0.054902613162994385,
0.004481928888708353,
-0.00055683747632429,
0.11248163133859634,
-0.10442901402711868,
0.037675485014915466,
0.008607660420238972,
-0.05610445886850357,
-0.02681935392320156,
-0.039432983845472336,
0.007018972188234329,
0.08692070096731186,
0.04734798148274422,
-0.04571525380015373,
0.1336071491241455,
-0.11388662457466125,
0.04709704965353012,
0.005319759249687195,
-0.19929054379463196,
-0.04660125449299812,
0.02095375582575798,
-0.01747678406536579,
0.021100614219903946,
-0.04813794419169426,
0.024034593254327774,
0.03039410524070263,
0.0077962446957826614,
0.013318827375769615,
0.03519171103835106,
0.03378334268927574,
0.012768731452524662,
-0.07156006246805191,
-0.032586470246315,
0.13523836433887482,
0.045191604644060135,
-0.030268194153904915,
-0.14587640762329102,
-0.03653308004140854,
0.004686467815190554,
0.00857231579720974,
-0.09299295395612717,
-0.013806420378386974,
-0.018920352682471275,
0.06996036320924759,
0.0027549562510102987,
-0.05780413746833801,
-0.025776229798793793,
-0.041219353675842285,
0.07049567997455597,
0.0193612203001976,
-0.006806473713368177,
0.0877169668674469,
0.04160493612289429,
0.045645784586668015,
-0.046857088804244995,
-0.024218516424298286,
-0.023470276966691017,
-0.18366505205631256,
-0.03673919662833214,
-0.013278063386678696,
-0.044804688543081284,
0.08337625861167908,
0.1737775355577469,
0.15110357105731964,
0.017467642202973366,
-0.019318100064992905,
0.014692277647554874,
0.08256698399782181,
0.00828153919428587,
-0.09607682377099991,
-0.05458805337548256,
-0.01479417271912098,
0.01918056607246399,
0.06336141377687454,
-0.06395705044269562,
-0.0013630568282678723,
-0.01376791950315237,
0.08390284329652786,
0.049855295568704605,
0.021774234250187874,
0.0738401785492897,
-0.06234848126769066,
-0.015541142784059048,
0.10104868561029434,
-0.10324084013700485,
-0.0006452980451285839,
0.03580664470791817,
-0.050659965723752975,
0.00155476457439363,
-0.014624400064349174,
0.05269567295908928,
-0.07434984296560287,
0.039978936314582825,
-0.005248486530035734,
-0.035605113953351974,
-0.03013245202600956,
-0.05176057294011116,
0.05764135345816612,
0.028995782136917114,
-0.04023229330778122,
-0.11908967792987823,
-0.10894507169723511,
-0.04539267346262932,
0.06663589179515839,
-0.07035921514034271,
-0.02246404066681862,
-0.08864854276180267,
-0.03315301612019539,
0.00616854103282094,
-0.023465193808078766,
0.044174157083034515,
-0.03317354619503021,
0.02143498696386814,
-0.12308453768491745,
0.05716822296380997,
0.019719185307621956,
-0.012063559144735336,
-0.07693424820899963,
-0.014784581959247589,
-0.03943219408392906,
0.10740067809820175,
-0.08481797575950623,
-0.010130267590284348,
-0.09428171813488007,
-0.00021515270054806024,
-0.005676781293004751,
0.014166540466248989,
0.053665608167648315,
0.11107800155878067,
-0.17373374104499817,
0.020921412855386734,
0.0032637459225952625,
-0.09577351063489914,
-0.08096381276845932,
0.16948501765727997,
-0.016228336840867996,
0.05024353414773941,
0.10491681098937988,
0.12725290656089783,
0.05291888490319252,
-0.05471370741724968,
-0.060356784611940384,
-0.06232978776097298,
-0.12387561053037643,
0.059995271265506744,
0.042495038360357285,
0.06645068526268005,
-0.08849813044071198,
0.03921154513955116,
0.002091478556394577,
-0.003774363547563553,
0.0023282216861844063,
-0.05741870030760765,
0.02230514958500862,
-0.05328225716948509,
0.03555692359805107,
0.0431717187166214,
-0.008124813437461853,
-0.010418740101158619,
-0.06366406381130219,
-0.12551750242710114,
0.04230343550443649,
-0.04418472945690155,
-0.0012679146602749825,
-0.12718041241168976,
0.023894041776657104,
-0.002409653039649129,
0.04287285357713699,
-0.08412859588861465,
-0.14074163138866425,
0.0447886623442173,
-0.08687420934438705,
0.016301747411489487,
0.10014750808477402,
0.045802418142557144,
0.050124235451221466,
0.012575287371873856,
0.0009244916727766395,
0.08452009409666061,
-0.013487699441611767,
-0.041787512600421906,
-0.13234470784664154,
0.011981191113591194,
-0.040295056998729706,
0.07970687747001648,
-0.09367657452821732,
-0.00565062090754509,
0.006538360379636288,
0.11135794222354889,
-0.017497435212135315,
-0.014159797690808773,
0.007605694700032473,
-0.04151134938001633,
-0.011273703537881374,
-0.005205878522247076,
0.01842319406569004,
-0.02959607169032097,
-0.0958920270204544,
0.07977158576250076,
-0.06995830684900284,
-0.04079620912671089,
0.004624009598046541,
0.06761550903320312,
-0.042104147374629974,
-0.012891488149762154,
-0.04935760051012039,
-0.03681551292538643,
-0.05485788732767105,
-0.0794888287782669,
0.13814593851566315,
0.0771603137254715,
0.015696074813604355,
-0.016680661588907242,
-0.029921358451247215,
-0.011593637987971306,
-0.008815736509859562,
-0.01663752645254135,
0.06878501176834106,
-0.014407120645046234,
-0.11808829754590988,
0.0011785832466557622,
-0.03508855402469635,
-0.008586813695728779,
0.11487407237291336,
0.012706867419183254,
-0.04092029854655266,
-0.06307701766490936,
0.06592744588851929,
0.03335875645279884,
0.043068964034318924,
-0.05093660578131676,
0.02214687317609787,
0.04115573316812515,
-0.030840791761875153,
0.00042420547106303275,
-0.00258717592805624,
0.020937373861670494,
-0.054012835025787354,
-0.02245868183672428,
0.06899496912956238,
0.03653382137417793,
-0.015243892557919025,
0.07874348014593124,
0.004476678092032671,
-0.050177037715911865,
0.03471262380480766,
-0.04731794819235802,
-0.08391439914703369,
0.10097050666809082,
-0.04165402799844742,
-0.14517512917518616,
-0.09588265419006348,
0.02900787442922592,
-0.052127592265605927,
0.00881150085479021,
0.009678194299340248,
-0.07525523006916046,
-0.10631148517131805,
-0.09693217277526855,
0.010647481307387352,
0.1446288377046585,
-0.05707480385899544,
-0.040394432842731476,
0.016962016001343727,
0.04818326234817505,
-0.07226282358169556,
-0.002615760313346982,
-0.041732363402843475,
-0.048711713403463364,
0.033708471804857254,
0.022346684709191322,
0.06164506450295448,
0.059839650988578796,
0.04027717933058739,
-0.020570529624819756,
0.03181133419275284,
0.1261330246925354,
-0.03713624179363251,
0.08514554798603058,
0.026279577985405922,
-0.021932609379291534,
0.039463236927986145,
0.17727577686309814,
0.049457043409347534,
-0.07835284620523453,
0.0552729070186615,
0.06115955114364624,
-0.010231834836304188,
-0.17733199894428253,
-0.10225392132997513,
-0.01755947805941105,
0.023343287408351898,
0.10337822884321213,
0.05384195223450661,
-0.010087087750434875,
0.0597711019217968,
-0.026065006852149963,
0.0035597539972513914,
0.06828876584768295,
0.04061878100037575,
0.04939268156886101,
0.0034063002094626427,
0.05515044555068016,
-0.02386460267007351,
-0.06138120964169502,
0.05881570652127266,
-0.027944963425397873,
0.08375688642263412,
-0.01512877643108368,
0.10004851222038269,
0.05782315880060196,
0.09608543664216995,
-0.02937537431716919,
0.08341288566589355,
0.06244523823261261,
0.010426354594528675,
-0.007889954373240471,
-0.05250715836882591,
-0.135830819606781,
0.07744232565164566,
0.04854774475097656,
-0.0428556427359581,
0.04319239407777786,
0.083411805331707,
0.07709489017724991,
0.14191024005413055,
0.030457288026809692,
-0.16557759046554565,
-0.04234513267874718,
0.055343568325042725,
-0.06709671020507812,
-0.057755835354328156,
0.01756773516535759,
0.09240022301673889,
-0.0866096019744873,
-0.0017743982607498765,
-0.05566766858100891,
0.031996361911296844,
-0.09391214698553085,
0.0318768210709095,
-0.05377061665058136,
0.04129381105303764,
0.02125578373670578,
0.11055601388216019,
-0.15558075904846191,
0.07744811475276947,
-0.01010037399828434,
0.08683441579341888,
-0.10435537248849869,
0.028577134013175964,
0.010226885788142681,
-0.013882442377507687,
0.15231378376483917,
0.00931544415652752,
-0.08792993426322937,
-0.0748823806643486,
-0.12697890400886536,
0.04135541990399361,
0.01600475050508976,
-0.0835283175110817,
0.04054420813918114,
-0.00323438155464828,
0.007802966516464949,
-0.06148770451545715,
-0.0071467068046331406,
-0.11253093928098679,
-0.1293981671333313,
0.05608720704913139,
0.01635260321199894,
0.1013699322938919,
-0.01947754994034767,
-0.025681478902697563,
-0.13708631694316864,
0.18722853064537048,
-0.09100330621004105,
-0.033342231065034866,
-0.09204864501953125,
0.07095398008823395,
0.0716012716293335,
-0.03748096525669098,
-0.05309971421957016,
0.011598332785069942,
0.12027116864919662,
-0.03326400741934776,
-0.0010348166106268764,
0.023950738832354546,
-0.07702523469924927,
-0.11352933198213577,
-0.046329647302627563,
0.06941987574100494,
0.10293963551521301,
0.029813343659043312,
0.04519181698560715,
0.012765682302415371,
-0.005772222299128771,
-0.09189247339963913,
0.020903103053569794,
0.08888481557369232,
-0.05873941257596016,
0.044257696717977524,
-0.05228510871529579,
-0.058785077184438705,
-0.07406150549650192,
-0.004226867575198412,
0.08789236098527908,
0.11594720929861069,
-0.033713240176439285,
0.09458080679178238,
0.18943628668785095,
-0.014296043664216995,
-0.21924883127212524,
-0.11714691668748856,
0.07899008691310883,
0.025063635781407356,
-0.00808207131922245,
-0.09098692983388901,
0.15232324600219727,
0.06202588602900505,
-0.014352560043334961,
-0.113409623503685,
-0.07971467077732086,
-0.09282644093036652,
0.0412258505821228,
0.036819398403167725,
0.04537566751241684,
0.011407481506466866,
-0.020929131656885147,
-0.03842545300722122,
0.027880584821105003,
0.1116442009806633,
-0.015480770729482174,
0.03015981800854206,
0.02653704211115837,
0.040643010288476944,
0.06867220997810364,
-0.047911178320646286,
0.09547653794288635,
0.04997266083955765,
0.05055619776248932,
-0.10766807943582535,
0.011133226566016674,
0.05799255147576332,
-0.05873624235391617,
0.13745345175266266,
-0.0322599820792675,
-0.0059548597782850266,
-0.042632460594177246,
-0.04068249464035034,
-0.0399228073656559,
0.06482134014368057,
-0.017873283475637436,
-0.04221424460411072,
-0.03702927380800247,
0.10235141217708588,
0.11382520198822021,
0.03770475462079048,
0.07381107658147812,
-0.006805785465985537,
-0.0016377550782635808,
0.07287368178367615,
0.15803667902946472,
-0.001193327596411109,
0.044314734637737274,
0.009408493526279926,
-0.017993668094277382,
0.05750831589102745,
-0.04589781537652016,
0.07350169122219086,
0.0988365039229393,
-0.04511671140789986,
0.08204097300767899,
-0.028158819302916527,
-0.08763987571001053,
-0.032663047313690186,
0.062014710158109665,
-0.11024084687232971,
-0.16206693649291992,
0.03212263435125351,
0.046787187457084656,
-0.029493281617760658,
-0.03979412838816643,
0.10111239552497864,
-0.05196761339902878,
0.029797717928886414,
0.04656580835580826,
0.05935470759868622,
0.015696613118052483,
0.08875231444835663,
-0.030921336263418198,
0.05946778878569603,
-0.07458408176898956,
0.06890729814767838,
0.08688972890377045,
-0.09046753495931625,
0.09386580437421799,
-0.007545516826212406,
-0.09643270820379257,
-0.0833311527967453,
-0.05664590746164322,
0.13098324835300446,
0.09066101908683777,
-0.06604907661676407,
0.02247498370707035,
-0.04428013041615486,
0.015637826174497604,
-0.05214357003569603,
0.0028017049189656973,
0.04057801887392998,
-0.029065312817692757,
-0.014160249382257462,
-0.05310266464948654,
0.10673785209655762,
0.04575001448392868,
-0.009895055554807186,
0.00852592196315527,
0.009913607500493526,
-0.017620055004954338,
-0.006288567557930946,
-0.033583834767341614,
-0.007075883448123932,
-0.01143251545727253,
-0.020067207515239716,
-0.023743171244859695,
0.011571616865694523,
-0.05656757205724716,
0.01546116080135107,
0.04439246281981468,
-0.011553307995200157,
-0.014393582940101624,
0.0010831363033503294,
-0.029012545943260193,
-0.007684133946895599,
0.019212093204259872,
0.06669417768716812,
-0.09896234422922134,
-0.044812094420194626,
0.04210824891924858,
-0.0654878243803978,
0.09667382389307022,
0.006170318461954594,
-0.04130103066563606,
0.036126479506492615,
-0.12961523234844208,
0.04293559491634369,
-0.011591155081987381,
0.07398322969675064,
0.014080413617193699,
-0.1270880550146103,
-0.03852934017777443,
0.029231734573841095,
-0.045268721878528595,
0.028578178957104683,
0.042647019028663635,
-0.05178777873516083,
0.09252389520406723,
0.04531119018793106,
-0.06731165200471878,
-0.08311871439218521,
0.022022798657417297,
-0.07377858459949493,
0.006502967327833176,
0.17284855246543884,
-0.046897731721401215,
0.005291426088660955,
-0.058284394443035126,
0.0028809031937271357,
0.046194400638341904,
-0.016602352261543274,
-0.0615449920296669,
-0.043136116117239,
0.03940000757575035,
-0.06606221944093704,
0.0875881165266037,
-0.037189193069934845,
-0.04337218403816223,
0.0259420033544302,
-0.006738260388374329,
-0.09237346053123474,
0.021484196186065674,
-0.009993067942559719,
0.026832591742277145,
0.012722962535917759,
-0.03252529352903366,
-0.09543269127607346,
-0.03142309933900833,
-0.039696235209703445,
0.0451810248196125,
0.04718765616416931,
0.06525536626577377,
0.033754948526620865,
0.032497797161340714,
-0.0239693783223629,
-0.003625148208811879,
0.14887475967407227,
-0.1318327933549881,
0.051056694239377975,
-0.019145112484693527,
0.05477005988359451,
0.06010095402598381,
-0.050482992082834244,
0.01614338532090187,
-0.010331838391721249,
-0.05803069472312927,
-0.05346718803048134,
-0.2463373988866806,
-0.05997229740023613,
0.007256549317389727,
0.04038799926638603,
-0.055783264338970184,
0.02432173863053322,
0.027174029499292374,
0.0532723031938076,
-0.011636477895081043,
0.10981108248233795,
-0.05092217028141022,
0.03382834792137146,
0.03734121471643448,
-0.019816648215055466,
-0.021734753623604774,
0.05451642721891403,
-0.022984327748417854,
-0.03482161834836006,
0.08189237862825394,
0.03666992112994194,
0.06735314428806305,
0.11128395050764084,
-0.02128453366458416,
-0.03457336872816086,
-0.07839812338352203,
0.006000207271426916,
-0.04724646359682083,
-0.06048116460442543,
0.007308062631636858,
0.0995718315243721,
-0.010206645354628563,
-0.010170884430408478,
0.11920441687107086,
0.000616796372924,
-0.06225423142313957,
-0.1787489801645279,
0.08074238151311874,
-0.015903430059552193,
0.05039341002702713,
-0.02077993005514145,
-0.04039192199707031,
0.010238328017294407,
0.10326960682868958,
0.06195814907550812,
-0.016393518075346947,
-0.02578476071357727,
-0.008739655837416649,
-0.009152168408036232,
-0.004890160169452429,
0.0783749371767044,
-0.018551528453826904,
0.06499797105789185,
-0.02451005019247532,
0.07542647421360016,
-0.04563574120402336,
-0.044666726142168045,
-0.034767840057611465,
0.11429260671138763,
-0.025819886475801468,
0.01049365010112524,
-0.08229777961969376,
0.05732249096035957,
-0.07318755984306335,
-0.14762072265148163,
-0.03862224146723747,
0.0037641366943717003,
-0.08886279910802841,
-0.00405406067147851,
0.04024698585271835,
0.007476701866835356,
0.0792769342660904,
0.02010595053434372,
-0.007928709499537945,
0.07672158628702164,
-0.0308461245149374,
-0.06148694083094597,
0.06560496985912323,
-0.0020273588597774506,
0.011283733882009983,
0.20355060696601868,
0.041939206421375275,
0.09568920731544495,
0.09246329963207245,
-0.05975901708006859,
-0.1012122631072998,
0.07167940586805344,
0.04686431586742401,
-0.05017752945423126,
0.0627966970205307,
0.12241940200328827,
0.045414820313453674,
0.07850082963705063,
0.09995830804109573,
-0.045543428510427475,
0.004203717224299908,
0.020646344870328903,
0.00856807641685009,
-0.08176456391811371,
0.12164030969142914,
-0.0069979229010641575,
0.11078941822052002,
0.10913322865962982,
0.009008493274450302,
0.007124488707631826,
-0.07364370673894882,
0.036928799003362656,
-0.06552077829837799,
0.0851142555475235,
-0.006617127452045679,
-0.17990145087242126,
0.029635347425937653,
-0.01656869426369667,
0.06850206851959229,
-0.09218178689479828,
-0.06248903647065163,
-0.026037108153104782,
-0.0029985324945300817,
0.022141549736261368,
0.10034403204917908,
0.046330373734235764,
0.000657947501167655,
-0.02132398821413517,
-0.07619437575340271,
0.031313251703977585,
0.07160325348377228,
-0.0260330680757761,
-0.041391972452402115
] |
null | null | transformers | # MiquMaid-v2-70B 3bpw
## Description
Exllama quant of [NeverSleep/MiquMaid-v2-70B](https://huggingface.co/NeverSleep/MiquMaid-v2-70B)
## Other quants:
EXL2: [4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-4bpw-exl2), [3.5bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-3.5bpw-exl2), [3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-3bpw-exl2), [2.4bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-2.4bpw-exl2), [2.3bpw](https://huggingface.co/Kooten/MiquMaid-v2-70B-2.3bpw-exl2)
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
[2bit Imatrix GGUF](https://huggingface.co/Kooten/MiquMaid-v2-70B-Imatrix-GGUF)
## Prompt format: Alpaca
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Contact
Kooten on discord
[ko-fi.com/kooten](https://ko-fi.com/kooten) | {"license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences", "nsfw"]} | text-generation | Kooten/MiquMaid-v2-70B-3.5bpw-exl2 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"not-for-all-audiences",
"nsfw",
"conversational",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:06:44+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| # MiquMaid-v2-70B 3bpw
## Description
Exllama quant of NeverSleep/MiquMaid-v2-70B
## Other quants:
EXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw
2.4bpw is probably the most you can fit in a 24gb card
GGUF:
2bit Imatrix GGUF
## Prompt format: Alpaca
## Contact
Kooten on discord
URL | [
"# MiquMaid-v2-70B 3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# MiquMaid-v2-70B 3bpw",
"## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B",
"## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF",
"## Prompt format: Alpaca",
"## Contact\nKooten on discord\n\nURL"
] | [
75,
14,
21,
60,
8,
7
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #not-for-all-audiences #nsfw #conversational #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MiquMaid-v2-70B 3bpw## Description\nExllama quant of NeverSleep/MiquMaid-v2-70B## Other quants:\nEXL2: 4bpw, 3.5bpw, 3bpw, 2.4bpw, 2.3bpw\n\n2.4bpw is probably the most you can fit in a 24gb card\n\nGGUF:\n2bit Imatrix GGUF## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL"
] | [
-0.055794719606637955,
0.037058692425489426,
-0.00551636703312397,
0.08587703108787537,
0.02601953037083149,
0.02203146554529667,
0.1505025327205658,
0.12238593399524689,
-0.0006199769559316337,
0.009135269559919834,
0.04752225801348686,
0.11139299720525742,
0.04638947546482086,
0.16380546987056732,
-0.054107118397951126,
-0.09724549949169159,
0.029823653399944305,
0.03208458423614502,
0.08016858249902725,
0.11589667201042175,
0.06529995054006577,
-0.0320008210837841,
0.05554889887571335,
-0.04611354321241379,
-0.10763328522443771,
-0.009882854297757149,
-0.02762743830680847,
-0.06662817299365997,
0.04721532762050629,
0.08100058138370514,
0.04063035547733307,
0.10513613373041153,
-0.03328100964426994,
-0.18256843090057373,
0.028376445174217224,
0.037196286022663116,
-0.06107915937900543,
0.02666592039167881,
0.029348738491535187,
-0.08317127823829651,
0.006035763770341873,
0.00454889889806509,
-0.06754196435213089,
0.07220575958490372,
-0.09121168404817581,
-0.07362551242113113,
-0.049151573330163956,
0.11858851462602615,
0.0733235627412796,
0.028863783925771713,
-0.0032604560256004333,
0.09735222905874252,
-0.0121019147336483,
0.08273068070411682,
0.23265093564987183,
-0.25577181577682495,
-0.02006392553448677,
0.08974011987447739,
0.026635553687810898,
0.09407830983400345,
-0.07273215055465698,
0.032397378236055374,
0.024459265172481537,
-0.028171688318252563,
0.07989612966775894,
-0.0209616981446743,
0.061923135071992874,
-0.03253960609436035,
-0.09032288938760757,
0.008913268335163593,
0.14756210148334503,
0.038891758769750595,
-0.062055930495262146,
-0.05218642204999924,
-0.08001535385847092,
-0.10742942243814468,
-0.05002063140273094,
0.043527182191610336,
0.009708556346595287,
-0.012216011993587017,
-0.04670985788106918,
-0.0010901977075263858,
-0.1079021617770195,
-0.046482257544994354,
-0.12932981550693512,
0.17389343678951263,
0.0073859067633748055,
0.028146548196673393,
0.009776992723345757,
0.034906063228845596,
-0.0836302638053894,
-0.08541147410869598,
-0.08843536674976349,
-0.06146755442023277,
0.06935381889343262,
-0.0003020302392542362,
0.00649604294449091,
0.08042702078819275,
0.08897026628255844,
0.1442209780216217,
-0.0703231543302536,
0.08767212927341461,
0.008488049730658531,
0.06491420418024063,
-0.029906831681728363,
-0.11966456472873688,
-0.018528785556554794,
-0.053485702723264694,
0.053682953119277954,
0.030843643471598625,
0.0846916139125824,
-0.022111795842647552,
-0.10224611312150955,
0.00514326011762023,
-0.0006818854017183185,
0.04166772961616516,
0.01862618513405323,
0.05822315067052841,
-0.04990827664732933,
-0.02471364289522171,
0.19090896844863892,
-0.06272641569375992,
0.017801465466618538,
0.057164598256349564,
0.02836984023451805,
0.049390897154808044,
0.06680581718683243,
-0.010729063302278519,
-0.022755343466997147,
0.052349817007780075,
-0.030281927436590195,
0.016005510464310646,
-0.017433829605579376,
-0.057888586074113846,
0.0835120677947998,
-0.03825972229242325,
-0.02134828083217144,
-0.18694104254245758,
-0.15212693810462952,
0.08920330554246902,
-0.00722495885565877,
-0.043543919920921326,
0.017435133457183838,
0.006063202396035194,
-0.03183622658252716,
-0.0060494206845760345,
-0.052008915692567825,
0.016196902841329575,
-0.07665175944566727,
0.08720386028289795,
0.08712700009346008,
0.1318335235118866,
-0.11594341695308685,
-0.017479216679930687,
-0.04657600447535515,
0.04717026278376579,
-0.0853913351893425,
-0.0027906561736017466,
-0.12123213708400726,
-0.02799489162862301,
-0.06409980356693268,
-0.030117226764559746,
-0.05123929679393768,
0.045704763382673264,
-0.009797452948987484,
0.09483790397644043,
-0.06832786649465561,
-0.04342836141586304,
0.26220571994781494,
-0.14577728509902954,
-0.10856323689222336,
0.11465399712324142,
0.05336647853255272,
-0.06419917941093445,
0.048586271703243256,
0.08411403745412827,
0.1323460340499878,
-0.11166200041770935,
-0.10522672533988953,
0.07038141787052155,
-0.00970124825835228,
-0.09239289909601212,
0.0889875516295433,
0.08049222081899643,
0.00850773137062788,
0.029073791578412056,
0.13412660360336304,
0.026922175660729408,
0.008904177695512772,
-0.057666171342134476,
-0.06398695707321167,
-0.08606843650341034,
0.0820099264383316,
-0.016900915652513504,
-0.03963853791356087,
-0.12502865493297577,
-0.09472314268350601,
0.013765457086265087,
0.07063763588666916,
-0.018079254776239395,
-0.01699911057949066,
-0.13973155617713928,
0.11186779290437698,
-0.11084051430225372,
0.028221851214766502,
-0.04685581102967262,
-0.030165914446115494,
-0.0359673872590065,
-0.02127806842327118,
0.04883222281932831,
0.06878245621919632,
0.09319710731506348,
0.07141590863466263,
-0.0995154082775116,
0.03242332115769386,
0.12366412580013275,
0.025114310905337334,
-0.003982612397521734,
-0.08367832005023956,
0.05582946911454201,
-0.024932915344834328,
0.14123304188251495,
-0.09158356487751007,
0.046571023762226105,
0.15510426461696625,
0.08549287170171738,
0.004489276558160782,
-0.042968813329935074,
0.06918752193450928,
0.013916977681219578,
-0.04594044014811516,
0.0033482613507658243,
0.07445769011974335,
-0.008756941184401512,
-0.009768081828951836,
0.12005212903022766,
-0.2587744891643524,
0.18446321785449982,
0.182372584939003,
-0.02516406588256359,
-0.031198015436530113,
-0.056735798716545105,
0.014379963278770447,
-0.018485093489289284,
-0.018439672887325287,
-0.06349343061447144,
0.02431151270866394,
-0.022643670439720154,
0.13255839049816132,
-0.10029217600822449,
-0.0495423823595047,
0.056594423949718475,
-0.0548754446208477,
-0.0096862418577075,
0.043943483382463455,
0.07697691768407822,
-0.12199513614177704,
0.149742990732193,
0.18101711571216583,
-0.07339601218700409,
0.12538166344165802,
0.026565277948975563,
-0.009576302953064442,
-0.033542126417160034,
0.045995939522981644,
0.036264099180698395,
0.07603953778743744,
-0.07118269056081772,
0.05635423585772514,
0.05627565085887909,
-0.013182253576815128,
0.03219657763838768,
-0.13800227642059326,
-0.03504409268498421,
0.039374031126499176,
-0.09086860716342926,
-0.06738387793302536,
0.004174193367362022,
-0.012668007053434849,
0.09573137760162354,
-0.05158466845750809,
0.015149026177823544,
0.026696188375353813,
0.04596295952796936,
-0.0811554417014122,
0.14553742110729218,
-0.09576793760061264,
-0.3124021887779236,
-0.08458353579044342,
-0.1290571540594101,
-0.05821377784013748,
0.022490575909614563,
0.09255395084619522,
-0.05344387888908386,
-0.0439242422580719,
-0.1047719269990921,
-0.06361774355173111,
-0.0056802695617079735,
-0.003833507653325796,
0.012078561820089817,
-0.010372775606811047,
0.0728408545255661,
-0.08969549834728241,
-0.008826829493045807,
0.08141878992319107,
-0.048503290861845016,
0.16609102487564087,
-0.011386152356863022,
0.09205219149589539,
0.11651858687400818,
-0.004199695773422718,
-0.025573141872882843,
-0.034996118396520615,
0.12297538667917252,
-0.05350892245769501,
0.06652691215276718,
0.12710869312286377,
-0.01150542963296175,
0.06713458895683289,
0.08232244849205017,
0.0320449024438858,
-0.06462686508893967,
-0.04956880956888199,
0.017348336055874825,
-0.12025006115436554,
-0.13407625257968903,
-0.04092860594391823,
-0.04234806448221207,
0.15821059048175812,
0.04265715181827545,
0.06196301057934761,
0.04643281176686287,
0.1278086006641388,
-0.06708541512489319,
-0.01597411558032036,
0.06713785231113434,
0.08346005529165268,
0.12067695707082748,
0.023249128833413124,
0.15782779455184937,
-0.08508652448654175,
-0.016618158668279648,
0.11880792677402496,
0.11169899255037308,
0.025900309905409813,
-0.007269761059433222,
0.10141203552484512,
0.07180818915367126,
0.021298279985785484,
0.11315414309501648,
0.0743105411529541,
-0.0022617466747760773,
-0.047692105174064636,
-0.034170001745224,
-0.06661652028560638,
0.0006767046288587153,
0.024223079904913902,
-0.0022236269433051348,
-0.03974836692214012,
0.028123879805207253,
-0.03491645306348801,
0.08211523294448853,
0.025116048753261566,
0.08061923086643219,
-0.21902379393577576,
-0.03757256641983986,
0.07676295191049576,
-0.007673356682062149,
-0.03788398951292038,
0.028545696288347244,
0.08326412737369537,
0.03109794668853283,
0.1663588434457779,
-0.017530007287859917,
0.06978920847177505,
-0.07932668179273605,
0.022961843758821487,
-0.037056177854537964,
0.06361158192157745,
-0.01097548846155405,
0.10104384273290634,
-0.27172383666038513,
0.0542665459215641,
0.039233505725860596,
0.01408074889332056,
-0.050473641604185104,
-0.026981519535183907,
0.012601666152477264,
-0.0028275002259761095,
0.08812874555587769,
0.011452347040176392,
0.07567279040813446,
-0.14082768559455872,
-0.08665996789932251,
0.01299256831407547,
0.07514341920614243,
0.052487023174762726,
0.049140699207782745,
0.035737160593271255,
-0.05252374708652496,
-0.0025517549365758896,
-0.03884303942322731,
-0.04861748218536377,
-0.11059124767780304,
0.09687034040689468,
0.15538153052330017,
-0.031068667769432068,
-0.023341260850429535,
-0.025269130244851112,
-0.21113848686218262,
0.2168864905834198,
-0.12145423889160156,
-0.08864093571901321,
-0.04917607456445694,
-0.08461973816156387,
-0.002955604810267687,
-0.021017704159021378,
0.03674638271331787,
-0.039554499089717865,
0.11417778581380844,
-0.03824997320771217,
-0.0682920515537262,
0.06353332847356796,
-0.12234485894441605,
-0.12536613643169403,
-0.03249586373567581,
0.14601539075374603,
-0.09700530022382736,
0.02606894262135029,
0.03096768446266651,
-0.010003983974456787,
-0.01267858874052763,
-0.1314631849527359,
0.09149650484323502,
0.022751331329345703,
-0.020579833537340164,
0.019007747992873192,
0.008471752516925335,
-0.0456121563911438,
-0.00426630349829793,
-0.04845684766769409,
0.11827517300844193,
0.3377668261528015,
-0.03185639530420303,
0.10258221626281738,
0.1201733872294426,
-0.037739720195531845,
-0.2336094081401825,
-0.1510925143957138,
-0.09103754162788391,
-0.08046480268239975,
-0.02699376828968525,
-0.07716339826583862,
0.053471941500902176,
0.12744560837745667,
-0.04027595743536949,
0.19772842526435852,
-0.13224950432777405,
-0.1106499508023262,
0.019527878612279892,
0.11084029078483582,
0.2560846507549286,
-0.17666950821876526,
-0.06042444333434105,
-0.07543893158435822,
-0.17581211030483246,
0.18876908719539642,
-0.14222246408462524,
0.10704601556062698,
-0.02479296736419201,
0.05173082649707794,
-0.016261322423815727,
-0.08113130927085876,
0.11432833224534988,
-0.06530508399009705,
0.0021991038229316473,
-0.12762698531150818,
0.0699155330657959,
0.04063810035586357,
-0.018215203657746315,
0.02120107039809227,
-0.23195922374725342,
-0.010340007953345776,
-0.0658649280667305,
-0.0557309128344059,
-0.03929542750120163,
0.0006204773089848459,
-0.04375394433736801,
-0.039568446576595306,
-0.0477343387901783,
-0.009384268894791603,
0.012407281436026096,
-0.034045446664094925,
-0.028769811615347862,
-0.07181419432163239,
0.09065404534339905,
0.30380991101264954,
0.10013653337955475,
-0.07689516991376877,
-0.06243501231074333,
-0.06469909101724625,
-0.09729354083538055,
0.007882685400545597,
-0.1467600017786026,
0.05872216448187828,
0.03727081045508385,
0.0043907444924116135,
0.06567860394716263,
0.041395239531993866,
-0.03551694005727768,
0.02648034319281578,
0.151759535074234,
-0.14206430315971375,
-0.02392647974193096,
0.0050039938651025295,
0.10753510892391205,
0.007772146258503199,
0.019195962697267532,
0.13839368522167206,
-0.01822209544479847,
0.01694544404745102,
0.0467451848089695,
0.023811403661966324,
-0.0012931037927046418,
0.11705449968576431,
0.03843652084469795,
0.07678496837615967,
-0.11330214887857437,
0.06489766389131546,
-0.03713605925440788,
-0.041807979345321655,
-0.004402770195156336,
0.067186638712883,
-0.09016749262809753,
-0.1291780173778534,
0.005980282090604305,
-0.04567454755306244,
-0.05536511167883873,
-0.06192367896437645,
-0.03732465207576752,
-0.13649630546569824,
0.028472784906625748,
0.16401052474975586,
0.013465787284076214,
-0.023570172488689423,
0.08403871208429337,
0.03423906862735748,
-0.07721805572509766,
0.08450780808925629,
-0.02024892531335354,
0.0932949110865593,
-0.12562046945095062,
0.04778832197189331,
-0.0018510225927457213,
-0.008328593336045742,
-0.046998344361782074,
0.03575984761118889,
-0.10576926171779633,
-0.03422940522432327,
-0.1484963744878769,
0.1001758798956871,
-0.10631293803453445,
0.002615515375509858,
-0.016775276511907578,
0.013426712714135647,
-0.020057665184140205,
-0.0424758680164814,
-0.0681799128651619,
0.011686165817081928,
0.021916724741458893,
-0.01995811238884926,
-0.10387775301933289,
-0.027028711512684822,
0.024782640859484673,
-0.029213130474090576,
0.05646292120218277,
0.025037720799446106,
-0.05364381521940231,
-0.007000060752034187,
-0.19439636170864105,
-0.01861538179218769,
0.09080087393522263,
0.05435383692383766,
-0.038621991872787476,
0.010392874479293823,
0.04416031390428543,
0.05270203948020935,
0.026781700551509857,
0.03861422836780548,
0.09255000948905945,
-0.06873907148838043,
0.028321364894509315,
-0.07983972877264023,
-0.06263894587755203,
-0.03307188302278519,
-0.012866174802184105,
0.13620442152023315,
-0.06790950894355774,
0.11497485637664795,
-0.07176326960325241,
-0.011156494729220867,
-0.13461895287036896,
0.024997614324092865,
0.033791542053222656,
-0.17117220163345337,
-0.045310184359550476,
-0.03299915790557861,
0.02220051735639572,
-0.03386075422167778,
0.1808503121137619,
-0.08092465996742249,
-0.13628876209259033,
0.030344394966959953,
-0.08750427514314651,
-0.027497343719005585,
-0.009999354369938374,
0.2225428819656372,
0.06151311844587326,
-0.006169435102492571,
-0.11728879064321518,
0.018290292471647263,
0.07512237876653671,
-0.126974955201149,
0.08097714185714722,
0.10971301794052124,
-0.0835871547460556,
0.05248398333787918,
0.05294673889875412,
-0.06126713752746582,
-0.05817796662449837,
0.006229124031960964,
-0.15757393836975098,
0.07732334733009338,
-0.0007101144292391837,
0.07597989588975906,
0.11896641552448273,
-0.03569675609469414,
0.0009565629879944026,
-0.05815242603421211,
-0.022728459909558296,
-0.1186361163854599,
-0.058300744742155075,
-0.09636622667312622,
-0.09013380855321884,
0.017210861667990685,
-0.0691295862197876,
0.008830285631120205,
0.055902812629938126,
0.059296850115060806,
0.02003251016139984,
0.14423906803131104,
-0.019910454750061035,
0.00389664014801383,
-0.010406107641756535,
-0.0254911407828331,
-0.031260401010513306,
0.061606697738170624,
-0.05516728758811951,
0.06500161439180374,
-0.02569907158613205,
0.03691266104578972,
0.031030362471938133,
0.037743307650089264,
0.01503006648272276,
-0.06809692829847336,
-0.09648462384939194,
-0.04216599836945534,
0.03633902966976166,
0.06706123054027557,
0.12487156689167023,
-0.026939352974295616,
0.00036624312633648515,
0.019807880744338036,
0.08095617592334747,
-0.018272878602147102,
-0.10736820101737976,
-0.014991619624197483,
0.15311604738235474,
-0.08616484701633453,
0.06251084804534912,
-0.02078184112906456,
-0.02555062063038349,
0.02438984252512455,
0.2443608045578003,
0.1663966029882431,
-0.03641454502940178,
0.05749938637018204,
0.014484323561191559,
0.019892482087016106,
0.06607931107282639,
0.0886022225022316,
0.0625302642583847,
0.17859956622123718,
-0.026972521096467972,
-0.002393130911514163,
0.041147224605083466,
-0.003998247906565666,
-0.07661376893520355,
0.05952990800142288,
-0.021633319556713104,
0.013268732465803623,
-0.06949619203805923,
0.001243487000465393,
-0.07429052144289017,
0.040628451853990555,
0.0009124776115640998,
-0.10635393857955933,
-0.054657433182001114,
0.0020435776095837355,
0.08159732073545456,
0.035445019602775574,
-0.017489628866314888,
-0.020547043532133102,
-0.01478720922023058,
-0.018833398818969727,
-0.014896908774971962,
-0.16964401304721832,
0.06449147313833237,
-0.034376613795757294,
0.003608871717005968,
0.05955912172794342,
-0.035402242094278336,
0.13209189474582672,
0.10875898599624634,
-0.027017896994948387,
-0.058813080191612244,
0.17661026120185852,
0.005975014064460993,
-0.17364946007728577,
0.040736936032772064,
0.12640008330345154,
-0.04211036488413811,
0.02658778429031372,
0.09020702540874481,
-0.056052617728710175,
-0.04376968741416931,
0.0949697345495224,
-0.06953398138284683,
-0.0125097231939435,
0.04082071781158447,
-0.034664105623960495,
0.07812070101499557,
0.05610956996679306,
-0.039043277502059937,
-0.0448882095515728,
-0.015740925446152687,
0.07260306924581528,
-0.03749755024909973,
-0.05722919851541519,
-0.01678759604692459,
-0.18302753567695618,
-0.042082346975803375,
0.07044994086027145,
0.1295129507780075,
-0.31317228078842163,
-0.011929132975637913,
-0.08842283487319946,
-0.00029699987499043345,
-0.1469092071056366,
0.05319574102759361,
0.1847960501909256,
0.03221810609102249,
-0.06581797450780869,
-0.10988670587539673,
-0.01626383140683174,
0.09527785331010818,
-0.0937410444021225,
-0.08179105073213577
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2.dev0 | {"library_name": "peft", "base_model": "openai/whisper-medium"} | null | GuideU/whisper-medium-peft | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:openai/whisper-medium",
"region:us"
] | 2024-02-08T17:09:10+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-medium #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2.dev0 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2.dev0"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-medium #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2.dev0"
] | [
37,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
14
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-openai/whisper-medium #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2.dev0"
] | [
-0.11009866744279861,
0.2024371325969696,
-0.003206209046766162,
0.03271740302443504,
0.09118996560573578,
0.01663760282099247,
0.05659909173846245,
0.12370318919420242,
-0.02453320473432541,
0.111717090010643,
0.06940724700689316,
0.09495196491479874,
0.10706774890422821,
0.21882233023643494,
0.005430983379483223,
-0.20161078870296478,
0.03062301315367222,
-0.09585263580083847,
-0.005208972375839949,
0.12259761244058609,
0.1433263123035431,
-0.09692972898483276,
0.0772864893078804,
-0.013327867724001408,
-0.0038842286448925734,
-0.03462498262524605,
-0.07317504286766052,
-0.03270827233791351,
0.04217591509222984,
0.046634119004011154,
0.054846905171871185,
-0.001632028492167592,
0.08431585133075714,
-0.26679325103759766,
0.019222689792513847,
0.04782154783606529,
-0.004945284686982632,
0.08679599314928055,
0.09928152710199356,
-0.044094935059547424,
0.12621305882930756,
-0.029993176460266113,
0.14152145385742188,
0.08018402755260468,
-0.09098219126462936,
-0.2321607768535614,
-0.06780530512332916,
0.08639897406101227,
0.17797859013080597,
0.07539986073970795,
-0.04217426851391792,
0.13048596680164337,
-0.08637699484825134,
0.018287068232893944,
0.042356036603450775,
-0.09433645755052567,
-0.0689825788140297,
0.050973303616046906,
0.10072700679302216,
0.0537245087325573,
-0.13192376494407654,
-0.03038507141172886,
0.025542687624692917,
0.033788491040468216,
0.07712707668542862,
0.014484375715255737,
0.15027053654193878,
0.022871432825922966,
-0.14623862504959106,
-0.04286900907754898,
0.13670222461223602,
0.026032548397779465,
-0.03852787986397743,
-0.22724835574626923,
0.001529703033156693,
-0.08042699098587036,
-0.026709336787462234,
-0.045652322471141815,
0.03831268101930618,
0.003471496282145381,
0.10472959280014038,
-0.0252132136374712,
-0.08977772295475006,
-0.013851462863385677,
0.09049975872039795,
0.057591576129198074,
0.02474396303296089,
-0.01849929615855217,
0.007976575754582882,
0.1233440712094307,
0.0538969412446022,
-0.13283517956733704,
-0.06260012090206146,
-0.07263471186161041,
-0.047148197889328,
-0.040849532932043076,
0.040800560265779495,
0.033856235444545746,
0.058973655104637146,
0.2534539997577667,
-0.04208846390247345,
0.06013455241918564,
0.0630035251379013,
0.020091090351343155,
0.04737419635057449,
0.09410442411899567,
-0.058251358568668365,
-0.15362310409545898,
-0.013485423289239407,
0.09679124504327774,
-0.0005560808931477368,
-0.024919619783759117,
-0.04632657393813133,
0.044966019690036774,
0.033624596893787384,
0.10795962810516357,
0.09797848016023636,
-0.007575538009405136,
-0.0761645957827568,
-0.052682019770145416,
0.20625486969947815,
-0.15268760919570923,
0.04134632647037506,
0.02161458320915699,
-0.013273223303258419,
-0.047189388424158096,
0.014851560816168785,
0.016398582607507706,
-0.02696417085826397,
0.1000782772898674,
-0.0663708820939064,
-0.040644917637109756,
-0.11481954902410507,
-0.025142092257738113,
0.032559316605329514,
0.008839539252221584,
-0.030496904626488686,
-0.03246923163533211,
-0.0655660405755043,
-0.09341175854206085,
0.10268495976924896,
-0.061980314552783966,
-0.058284200727939606,
-0.027954116463661194,
-0.08972311019897461,
0.02129172906279564,
0.027600616216659546,
0.0886649563908577,
-0.026499373838305473,
0.04017249122262001,
-0.013393272645771503,
0.06304483860731125,
0.08294795453548431,
0.03322071582078934,
-0.0728275254368782,
0.06519324332475662,
-0.19936253130435944,
0.08671455830335617,
-0.08336646854877472,
0.03106561489403248,
-0.1587802767753601,
-0.01773025095462799,
0.005489597097039223,
0.021313171833753586,
0.033772874623537064,
0.16349025070667267,
-0.20131449401378632,
-0.033698536455631256,
0.1625654250383377,
-0.10868725180625916,
-0.11860693991184235,
0.04061909019947052,
-0.04561535269021988,
0.14895962178707123,
0.022006284445524216,
-0.005507924128323793,
0.09887097030878067,
-0.14825734496116638,
-0.02532867155969143,
-0.019422758370637894,
-0.0026927099097520113,
0.10152779519557953,
0.08852628618478775,
-0.0857844427227974,
0.027540525421500206,
0.014840609394013882,
-0.04615878686308861,
-0.02697797119617462,
-0.0472409687936306,
-0.10624409466981888,
0.0064543248154222965,
-0.08032941818237305,
0.022391676902770996,
-0.007382849231362343,
-0.08235899358987808,
-0.012921100482344627,
-0.1626535952091217,
-0.03835228458046913,
0.08471313863992691,
0.009188815951347351,
-0.01974017173051834,
-0.10182208567857742,
0.04455062747001648,
-0.03024916723370552,
-0.019499123096466064,
-0.14880192279815674,
-0.022324327379465103,
0.017342153936624527,
-0.13800902664661407,
0.004288440104573965,
-0.11964302510023117,
0.0679396241903305,
0.015246276743710041,
-0.06130421906709671,
-0.03892730548977852,
-0.006771598942577839,
0.0065277256071567535,
-0.04932733625173569,
-0.24407243728637695,
-0.0229876097291708,
-0.05410248041152954,
0.1561938226222992,
-0.22360195219516754,
0.03812039643526077,
0.051377709954977036,
0.13001714646816254,
0.0051629492081701756,
-0.0623646005988121,
0.03172328695654869,
-0.06711845844984055,
-0.025466570630669594,
-0.07407180964946747,
-0.007078366354107857,
-0.005271645728498697,
-0.04614658281207085,
0.01688758097589016,
-0.1241086944937706,
-0.03365376591682434,
0.09813281893730164,
0.07030590623617172,
-0.15916158258914948,
-0.01893945038318634,
-0.04436181113123894,
-0.06113912910223007,
-0.08331375569105148,
-0.0571521557867527,
0.11029870063066483,
0.05243945121765137,
0.039340369403362274,
-0.07477232068777084,
-0.06951816380023956,
0.008418791927397251,
-0.023719090968370438,
-0.021986473351716995,
0.11305508762598038,
0.07015880197286606,
-0.11025155335664749,
0.09662238508462906,
0.07120772451162338,
0.033990997821092606,
0.08122898638248444,
-0.027229713276028633,
-0.10351946204900742,
-0.027974385768175125,
0.04967174306511879,
0.016339648514986038,
0.15686221420764923,
-0.06688867509365082,
0.053951483219861984,
0.04580030217766762,
-0.03763902187347412,
0.045170098543167114,
-0.0951809510588646,
0.010320063680410385,
0.010588224977254868,
-0.012268835678696632,
0.017750857397913933,
-0.019856834784150124,
0.009806704707443714,
0.08485019952058792,
0.050270188599824905,
0.04045475646853447,
0.024510113522410393,
-0.03227609768509865,
-0.13088376820087433,
0.18756933510303497,
-0.09848044812679291,
-0.2368582785129547,
-0.15648193657398224,
0.060589104890823364,
0.056129127740859985,
-0.01686626859009266,
0.02268805541098118,
-0.05732046440243721,
-0.1066744476556778,
-0.08282996714115143,
-0.0015978855080902576,
0.03106439858675003,
-0.055582206696271896,
-0.063987135887146,
0.04574095085263252,
0.04801351949572563,
-0.11690964549779892,
0.03542613610625267,
0.0572783499956131,
-0.018838563933968544,
0.0030222926288843155,
0.058811869472265244,
0.08392565697431564,
0.18140165507793427,
-0.0049103000201284885,
-0.0030762588139623404,
0.050625890493392944,
0.276137113571167,
-0.15898942947387695,
0.11078847944736481,
0.12469034641981125,
-0.06991945952177048,
0.07997562736272812,
0.19113734364509583,
0.034052927047014236,
-0.10188113898038864,
0.02831374853849411,
0.02665877155959606,
-0.030039019882678986,
-0.2654711902141571,
-0.0518592894077301,
-0.01572928950190544,
-0.08510659635066986,
0.07807961106300354,
0.09032213687896729,
0.07695518434047699,
0.03864472359418869,
-0.07157575339078903,
-0.08782824873924255,
0.035125214606523514,
0.09993012994527817,
-0.02044498547911644,
0.007299587596207857,
0.08347291499376297,
-0.03636764734983444,
0.013268891721963882,
0.0979384034872055,
-0.013029022142291069,
0.16253091394901276,
0.04864763841032982,
0.10921364277601242,
0.0823088064789772,
0.086046002805233,
-0.0016927534015849233,
0.027841240167617798,
0.012308957055211067,
0.024365678429603577,
0.01545211486518383,
-0.08616873621940613,
0.024395445361733437,
0.11133791506290436,
0.034814607352018356,
0.03356733173131943,
0.01319130603224039,
-0.03411559388041496,
0.0514051578938961,
0.17941512167453766,
0.010384206660091877,
-0.20285914838314056,
-0.08107996731996536,
0.061339326202869415,
-0.07809831202030182,
-0.13518983125686646,
-0.015136094763875008,
0.03763682767748833,
-0.16720987856388092,
0.029630286619067192,
-0.03975967317819595,
0.09987382590770721,
-0.08712952584028244,
-0.03981887176632881,
0.09728891402482986,
0.06899993866682053,
-0.025770334526896477,
0.05472307652235031,
-0.18765847384929657,
0.12896108627319336,
0.026221822947263718,
0.07138883322477341,
-0.08470454812049866,
0.10136586427688599,
0.0038224072195589542,
0.001941873924806714,
0.16831141710281372,
0.0027702832594513893,
-0.061441369354724884,
-0.06337907910346985,
-0.097203329205513,
-0.01389650721102953,
0.09655178338289261,
-0.13133913278579712,
0.0676209107041359,
-0.018025092780590057,
-0.029476499184966087,
-0.004371267277747393,
-0.0809553936123848,
-0.12178321927785873,
-0.17113831639289856,
0.05660977587103844,
-0.09705130010843277,
0.025761017575860023,
-0.09568924456834793,
-0.061186064034700394,
0.0035622408613562584,
0.17403623461723328,
-0.2123050093650818,
-0.10166607052087784,
-0.15061740577220917,
-0.08943579345941544,
0.15979807078838348,
-0.04532547667622566,
0.08330436050891876,
-0.0009239431237801909,
0.16457463800907135,
0.012002703733742237,
-0.009404581971466541,
0.10059209913015366,
-0.0922110453248024,
-0.19625182449817657,
-0.05747678130865097,
0.16754567623138428,
0.13366706669330597,
0.03595029190182686,
-0.01497281901538372,
0.028062202036380768,
-0.04846571758389473,
-0.12053200602531433,
0.025715794414281845,
0.14651928842067719,
0.06557150930166245,
-0.013246169313788414,
-0.024673573672771454,
-0.10213647037744522,
-0.058535050600767136,
-0.04496318846940994,
-0.000271237367996946,
0.18773308396339417,
-0.07589530944824219,
0.16338887810707092,
0.11650621891021729,
-0.05624350905418396,
-0.20781683921813965,
0.04608511924743652,
0.0547085702419281,
0.01428757980465889,
0.03855567425489426,
-0.1947295367717743,
0.08816302567720413,
-0.002484177937731147,
-0.07337980717420578,
0.16430360078811646,
-0.17164799571037292,
-0.14066950976848602,
0.0986516997218132,
0.03467428684234619,
-0.221357062458992,
-0.1406555473804474,
-0.10340195149183273,
-0.0204685740172863,
-0.11865565180778503,
0.049556195735931396,
-0.002514638938009739,
0.010927721858024597,
0.02587680146098137,
0.010316438972949982,
0.026906289160251617,
-0.04808894172310829,
0.20290213823318481,
-0.031034843996167183,
0.010442646220326424,
-0.05301577225327492,
-0.07923763245344162,
0.02449740096926689,
-0.04850103706121445,
0.10227593034505844,
-0.007302489131689072,
0.027132811024785042,
-0.15227559208869934,
-0.04190750792622566,
-0.05175783857703209,
0.031091609969735146,
-0.09283854067325592,
-0.08684138208627701,
-0.04307285696268082,
0.09177820384502411,
0.09692409634590149,
-0.026455501094460487,
0.002735197776928544,
-0.08842956274747849,
0.07237176597118378,
0.20328128337860107,
0.19540368020534515,
0.0673445612192154,
-0.0665844976902008,
0.021039098501205444,
-0.029899196699261665,
0.047898609191179276,
-0.23221103847026825,
0.04022745415568352,
0.056692276149988174,
0.021701494231820107,
0.08425480127334595,
-0.01113090105354786,
-0.1566295027732849,
-0.06986991316080093,
0.0843244343996048,
-0.05097663030028343,
-0.16972899436950684,
-0.024742592126131058,
0.04111665487289429,
-0.20407931506633759,
-0.03856145963072777,
0.02273906208574772,
-0.022370249032974243,
-0.03584175184369087,
0.023235144093632698,
0.08023678511381149,
-0.017230235040187836,
0.10885428637266159,
0.08365701138973236,
0.09376563131809235,
-0.10289199650287628,
0.0781165137887001,
0.07410477101802826,
-0.0444173701107502,
0.029042528942227364,
0.11548705399036407,
-0.04903494939208031,
-0.033712029457092285,
0.0776403620839119,
0.087586909532547,
0.03461514413356781,
-0.054930076003074646,
0.01341523788869381,
-0.05983353406190872,
0.061362724751234055,
0.11390124261379242,
0.02798430807888508,
-0.003471667878329754,
0.05852166563272476,
0.03264214098453522,
-0.09175065904855728,
0.11115634441375732,
0.056876327842473984,
0.01687372848391533,
-0.046828240156173706,
-0.034685246646404266,
-0.00582516985014081,
-0.016895564272999763,
-0.0193549282848835,
-0.0073980591259896755,
-0.08912167698144913,
-0.006853221915662289,
-0.09585588425397873,
0.026455292478203773,
-0.07429184019565582,
0.009478707797825336,
0.027017662301659584,
-0.050912436097860336,
0.0019426238723099232,
0.008463377133011818,
-0.07261147350072861,
-0.04933904483914375,
-0.014948287047445774,
0.08334089815616608,
-0.1307406723499298,
0.03569316864013672,
0.07361753284931183,
-0.10428391396999359,
0.07386817783117294,
-0.006275589112192392,
0.005691129714250565,
0.006024680100381374,
-0.1596011519432068,
0.05492672696709633,
-0.020406130701303482,
-0.012377656996250153,
0.018198953941464424,
-0.20413579046726227,
-0.00724787637591362,
-0.0465419702231884,
-0.06142492592334747,
0.00788462907075882,
-0.021011238917708397,
-0.12247146666049957,
0.0959572121500969,
0.0008864665287546813,
-0.063669353723526,
-0.019644584506750107,
0.037210576236248016,
0.10085419565439224,
-0.025979185476899147,
0.1338614672422409,
-0.026602977886795998,
0.07514169812202454,
-0.17479518055915833,
-0.008498976938426495,
-0.01264677569270134,
0.03891533613204956,
-0.027397572994232178,
-0.019134197384119034,
0.05989483743906021,
-0.02328432723879814,
0.17925292253494263,
-0.020975660532712936,
0.06667487323284149,
0.057393498718738556,
0.013725213706493378,
0.01170370727777481,
0.08392732590436935,
0.05704644322395325,
0.00034836301347240806,
-0.00427555525675416,
0.03230053558945656,
-0.00959516130387783,
-0.03922975808382034,
-0.16322222352027893,
0.06583553552627563,
0.14941278100013733,
0.05264207348227501,
0.02224845439195633,
0.02692834660410881,
-0.1112549751996994,
-0.08361829817295074,
0.11817898601293564,
-0.01807986944913864,
-0.03348846733570099,
-0.06777885556221008,
0.1682516634464264,
0.14215874671936035,
-0.19584153592586517,
0.07468510419130325,
-0.04927406087517738,
-0.048824477940797806,
-0.13674022257328033,
-0.17000776529312134,
-0.061463937163352966,
-0.047424692660570145,
-0.020518729463219643,
-0.06607957184314728,
0.04851391911506653,
0.05097969248890877,
0.0042974818497896194,
-0.013050520792603493,
0.1118624210357666,
0.005412214435636997,
-0.022934196516871452,
0.05071772634983063,
0.0626746267080307,
0.033446792513132095,
-0.08643098920583725,
0.008642468601465225,
-0.0008420732920058072,
0.01561922300606966,
0.06577242910861969,
0.019390005618333817,
-0.05587013065814972,
0.019618025049567223,
-0.017686422914266586,
-0.11614995449781418,
0.04118560254573822,
-0.015379027463495731,
-0.040191102772951126,
0.14555253088474274,
0.032228391617536545,
0.00936981663107872,
-0.02181495912373066,
0.2284102588891983,
-0.07834140211343765,
-0.07076483219861984,
-0.1495603322982788,
0.07301045954227448,
-0.06328897178173065,
0.03271631896495819,
0.0314510203897953,
-0.11771969497203827,
0.01453650277107954,
0.15985040366649628,
0.13032564520835876,
-0.008840114809572697,
0.013647024519741535,
0.04246244579553604,
0.005071018356829882,
-0.03291143476963043,
0.020294995978474617,
0.04968753084540367,
0.15129049122333527,
-0.06876696646213531,
0.06457379460334778,
-0.010652754455804825,
-0.08086130768060684,
-0.017649903893470764,
0.10228952765464783,
-0.002693018177524209,
0.0011343740625306964,
-0.06971137225627899,
0.14117014408111572,
-0.08513681590557098,
-0.22404254972934723,
0.06276452541351318,
-0.07106048613786697,
-0.1470157504081726,
-0.046128761023283005,
0.022739365696907043,
-0.011136319488286972,
0.011263154447078705,
0.0774403065443039,
-0.051264770328998566,
0.17222900688648224,
0.043993398547172546,
-0.05115639790892601,
-0.08587399125099182,
0.05609995499253273,
-0.141911119222641,
0.2806292176246643,
0.0200900100171566,
0.04187998175621033,
0.10264251381158829,
-0.01733429729938507,
-0.14160606265068054,
0.012304436415433884,
0.10760311037302017,
-0.07158973813056946,
0.05635349452495575,
0.1711389124393463,
0.0010433071292936802,
0.1288049966096878,
0.058055851608514786,
-0.055035579949617386,
0.03553353250026703,
-0.09254972636699677,
-0.04957038536667824,
-0.10847853124141693,
0.08378443121910095,
-0.08495595306158066,
0.16126281023025513,
0.12579724192619324,
-0.06842607259750366,
-0.007818365469574928,
-0.01991927996277809,
0.08144511282444,
0.00816399697214365,
0.10904333740472794,
0.013884330168366432,
-0.18571335077285767,
0.035715438425540924,
0.0071507403627038,
0.10372865945100784,
-0.19752784073352814,
-0.05860366299748421,
0.042490433901548386,
-0.018033182248473167,
-0.07894746214151382,
0.11886858940124512,
0.045162100344896317,
0.0341024249792099,
-0.042450662702322006,
-0.05220695585012436,
0.009793782606720924,
0.14423930644989014,
-0.11108020693063736,
-0.006787123158574104
] |
null | null | transformers |
# AIFT-polyglot-ko-1.3b-ao-instruct-v0.91
베이스 모델 : EleutherAI/polyglot-ko-1.3b
학습 데이터 : 자체 제작한 Open Orca 스타일 데이터셋 약 48,000건 (중복 제거 및 데이터 분포 조정)
학습 방법 : Full finetuning
epoch : 3
## ko-lm-evaluation-harness(5-shot)
|kobest_boolq|kobest_copa|kobest_hellaswag|pawsx_ko|
|--|--|--|--|
|0.5398860398860399|0.71|0.436|0.476|
## Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2+cu121
- Datasets 2.0.0
- Tokenizers 0.15.0 | {"license": "cc-by-nc-4.0"} | text-generation | mu0gum/AIFT-polyglot-ko-1.3b-ao-instruct-v0.91 | [
"transformers",
"safetensors",
"gpt_neox",
"text-generation",
"license:cc-by-nc-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:09:40+00:00 | [] | [] | TAGS
#transformers #safetensors #gpt_neox #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| AIFT-polyglot-ko-1.3b-ao-instruct-v0.91
=======================================
베이스 모델 : EleutherAI/polyglot-ko-1.3b
학습 데이터 : 자체 제작한 Open Orca 스타일 데이터셋 약 48,000건 (중복 제거 및 데이터 분포 조정)
학습 방법 : Full finetuning
epoch : 3
ko-lm-evaluation-harness(5-shot)
--------------------------------
Framework versions
------------------
* Transformers 4.36.2
* Pytorch 2.1.2+cu121
* Datasets 2.0.0
* Tokenizers 0.15.0
| [] | [
"TAGS\n#transformers #safetensors #gpt_neox #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
61
] | [
"passage: TAGS\n#transformers #safetensors #gpt_neox #text-generation #license-cc-by-nc-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.015875736251473427,
0.06792441755533218,
-0.004260574001818895,
-0.003573999274522066,
0.09302686154842377,
0.007972216233611107,
0.15617983043193817,
0.1009492352604866,
-0.0495191290974617,
-0.008622605353593826,
0.1769961565732956,
0.18198569118976593,
-0.015316423960030079,
0.11442672461271286,
-0.07727409154176712,
-0.1629425287246704,
0.07098323106765747,
0.030820608139038086,
0.029123421758413315,
0.09611809253692627,
0.09364017099142075,
-0.03651152923703194,
0.07950315624475479,
-0.0336429737508297,
-0.15559576451778412,
-0.0034996995236724615,
0.0763092190027237,
-0.14335785806179047,
0.09810535609722137,
0.06109551712870598,
0.09796357154846191,
0.09954842925071716,
-0.0354967825114727,
-0.16072992980480194,
0.01546885073184967,
0.02651660144329071,
-0.10524118691682816,
0.04567767679691315,
0.07098478823900223,
-0.025863422080874443,
0.10609933733940125,
0.017986752092838287,
-0.03669922798871994,
0.07661309093236923,
-0.13340726494789124,
-0.09139983355998993,
-0.07848293334245682,
0.022423643618822098,
0.0781165063381195,
0.07692290097475052,
-0.0024318373762071133,
0.13906948268413544,
-0.0560675784945488,
0.07386516034603119,
0.07116516679525375,
-0.3775385618209839,
0.018751472234725952,
0.13614347577095032,
0.06811447441577911,
0.05308680236339569,
-0.026811860501766205,
0.10581197589635849,
0.0764341726899147,
-0.01766723208129406,
0.05564224347472191,
-0.062236037105321884,
-0.056995294988155365,
0.03960440680384636,
-0.08396530896425247,
-0.03591637313365936,
0.22946777939796448,
-0.030930640175938606,
0.012745416723191738,
-0.0630120113492012,
-0.05514655262231827,
-0.0033475137315690517,
-0.029083989560604095,
0.011939236894249916,
-0.02319134585559368,
0.09509684890508652,
0.00767021719366312,
-0.03771902993321419,
-0.14328469336032867,
-0.03587660193443298,
-0.1630532294511795,
0.10533306747674942,
-0.005020496901124716,
0.0367196761071682,
-0.09610164165496826,
0.06484699249267578,
-0.04330690950155258,
-0.07405602186918259,
-0.026138463988900185,
-0.08700596541166306,
0.08894500136375427,
-0.034255627542734146,
-0.0345153771340847,
0.0034097086172550917,
0.12670321762561798,
0.15751653909683228,
-0.022536851465702057,
0.0028703780844807625,
-0.10557136684656143,
0.08989471942186356,
-0.06196356564760208,
-0.0008039474487304688,
0.022676173597574234,
-0.0011653725523501635,
0.11892679333686829,
-0.11042793840169907,
0.07916740328073502,
-0.043822888284921646,
-0.14785729348659515,
-0.018349651247262955,
0.0055917371064424515,
0.13466738164424896,
0.013689950108528137,
0.09850327670574188,
-0.03124234452843666,
0.059009671211242676,
0.21154886484146118,
-0.043589264154434204,
0.00361501588486135,
-0.014686137437820435,
0.08221596479415894,
-0.00703246146440506,
0.02354220300912857,
0.028675979003310204,
-0.05785486474633217,
0.05475493520498276,
-0.07524936646223068,
-0.051845427602529526,
-0.021648358553647995,
-0.06022633984684944,
0.07347924262285233,
-0.03156989440321922,
0.023720398545265198,
-0.19839827716350555,
-0.1889348328113556,
0.04069524630904198,
-0.01990375854074955,
-0.011854826472699642,
-0.037967219948768616,
-0.012644313275814056,
-0.04974636062979698,
0.03688441962003708,
-0.08012954890727997,
-0.06451497226953506,
-0.08505716919898987,
0.08339550346136093,
-0.015187600627541542,
0.04268661141395569,
-0.16610850393772125,
0.023927202448248863,
-0.11425803601741791,
-0.007118675392121077,
-0.06816098093986511,
0.027088617905974388,
-0.06277585029602051,
0.13572534918785095,
-0.04562985897064209,
-0.02065426856279373,
-0.058291930705308914,
0.04683077335357666,
-0.04373355954885483,
0.1677791327238083,
-0.07659433037042618,
-0.06721650809049606,
0.2661617398262024,
-0.13946214318275452,
-0.20669354498386383,
0.10493091493844986,
0.024658722802996635,
0.005621774587780237,
0.11478817462921143,
0.14788773655891418,
0.03692333772778511,
-0.05322267487645149,
0.013618793338537216,
0.10389456152915955,
-0.1023234948515892,
-0.1356758177280426,
0.006095150485634804,
-0.037248652428388596,
-0.1714916080236435,
0.028513513505458832,
0.005575274582952261,
0.07011395692825317,
-0.015216976404190063,
-0.04686193913221359,
-0.06336840242147446,
-0.022492948919534683,
0.03135162219405174,
-0.03875114023685455,
0.05064034089446068,
-0.11623278260231018,
-0.044608358293771744,
-0.07374400645494461,
-0.015223292633891106,
-0.018475282937288284,
0.03159451484680176,
-0.0837363600730896,
0.07665333896875381,
-0.04468084126710892,
0.06223265454173088,
-0.07009455561637878,
-0.12474779039621353,
0.015008268877863884,
0.08021419495344162,
0.027173342183232307,
0.014335712417960167,
0.03182602301239967,
0.015934092923998833,
-0.03549407795071602,
0.005273176822811365,
0.18119539320468903,
0.019904833287000656,
-0.021456729620695114,
-0.08794846385717392,
0.11055481433868408,
-0.04000182822346687,
0.038301557302474976,
-0.1474364846944809,
0.028935015201568604,
0.1278601884841919,
0.05931323394179344,
-0.012121004983782768,
0.05237482115626335,
-0.02971651591360569,
0.018399981781840324,
-0.07349790632724762,
-0.008849740028381348,
0.08719821274280548,
0.04395877197384834,
-0.12102939933538437,
0.22971059381961823,
-0.20978282392024994,
0.22945232689380646,
0.21562804281711578,
-0.19029328227043152,
-0.002291277749463916,
-0.06055017188191414,
0.01703605428338051,
0.008282307535409927,
0.03807663545012474,
-0.03330976888537407,
0.009400800801813602,
-0.038331497460603714,
0.15861396491527557,
-0.07770883291959763,
-0.026943163946270943,
0.02287987619638443,
-0.03903284668922424,
-0.02887120097875595,
0.03525786101818085,
0.11718220263719559,
-0.18113969266414642,
0.18644040822982788,
0.2566486895084381,
0.017288167029619217,
0.11590734869241714,
-0.020337563008069992,
0.010525204241275787,
0.040689606219530106,
0.04066481813788414,
0.044807299971580505,
-0.03911454603075981,
-0.06699611991643906,
0.007127000018954277,
0.06696028262376785,
0.024676529690623283,
0.060781288892030716,
-0.15408416092395782,
-0.06674940884113312,
-0.017471564933657646,
-0.0587126798927784,
-0.017107442021369934,
0.0756961926817894,
-0.020684514194726944,
0.10922964662313461,
-0.042616505175828934,
-0.037186428904533386,
0.13813842833042145,
0.012875895947217941,
-0.12527363002300262,
0.180256649851799,
-0.15214373171329498,
-0.2760857939720154,
-0.17953771352767944,
-0.13331466913223267,
-0.0367591418325901,
0.07381676137447357,
0.12386907637119293,
-0.06514368206262589,
-0.07127518951892853,
-0.07354099303483963,
-0.03411320969462395,
-0.02288845181465149,
0.023240964859724045,
-0.04035535827279091,
0.04613136127591133,
-0.037709061056375504,
-0.11137937009334564,
-0.04135577008128166,
0.03403418883681297,
-0.10537485778331757,
0.13554608821868896,
-0.07957988232374191,
0.10750098526477814,
0.16068819165229797,
0.013331354595720768,
-0.007426455616950989,
-0.08063692599534988,
0.12824392318725586,
-0.07760006934404373,
-0.012611713260412216,
0.18509256839752197,
-0.019661307334899902,
0.05548827722668648,
0.11280113458633423,
0.020338812842965126,
-0.10334822535514832,
0.045209795236587524,
-0.07268919050693512,
-0.10354217141866684,
-0.21830298006534576,
-0.11934605985879898,
-0.09334360808134079,
0.08755974471569061,
0.04708457738161087,
0.07760902494192123,
0.18560563027858734,
0.08894463628530502,
-0.05691071227192879,
0.035354554653167725,
0.07944868505001068,
0.1109585389494896,
0.2522401809692383,
-0.00162787816952914,
0.13730557262897491,
-0.11244110018014908,
-0.07570648193359375,
0.10258850455284119,
0.061577554792165756,
0.11229012161493301,
0.11134403944015503,
0.1224270835518837,
0.04962128773331642,
0.09408516436815262,
0.16790685057640076,
0.1133037731051445,
0.04855963960289955,
-0.03517252951860428,
-0.012196213006973267,
-0.05271274596452713,
0.0023207291960716248,
0.04888153448700905,
-0.09712223708629608,
-0.1678878515958786,
-0.007723464630544186,
-0.10383259505033493,
0.09082846343517303,
0.058297257870435715,
0.04500590264797211,
-0.2776898145675659,
0.024884790182113647,
0.08635592460632324,
0.03891945257782936,
-0.08423604816198349,
0.08984600752592087,
-0.0011324180522933602,
-0.04208039864897728,
0.14409169554710388,
-0.04618876427412033,
0.08190100640058517,
-0.010948301292955875,
0.05327041074633598,
-0.030651668086647987,
-0.04780147969722748,
-0.002397296018898487,
0.11093255877494812,
-0.2867632210254669,
0.1749102622270584,
0.028691578656435013,
0.018376220017671585,
-0.0730835571885109,
0.0021109117660671473,
0.010383482091128826,
0.20389121770858765,
0.15272358059883118,
-0.022006558254361153,
-0.16683736443519592,
-0.05092480778694153,
-0.0420130230486393,
0.04102875292301178,
0.1169663518667221,
0.0008615820552222431,
-0.00983872078359127,
-0.033705003559589386,
-0.007869700901210308,
0.03189156576991081,
-0.06276766955852509,
-0.07454617321491241,
-0.17968083918094635,
0.05845177173614502,
0.13556665182113647,
0.10288438946008682,
-0.04758039861917496,
0.012691372074186802,
-0.16545429825782776,
0.21273791790008545,
-0.19886305928230286,
-0.06102028861641884,
-0.09782598912715912,
-0.10430463403463364,
0.009820027276873589,
-0.025657162070274353,
0.07969769835472107,
-0.05379030853509903,
0.04614865779876709,
-0.07133958488702774,
-0.18360966444015503,
0.12143389135599136,
-0.1170133650302887,
-0.07040712982416153,
-0.033304374665021896,
0.1318398416042328,
-0.09586362540721893,
-0.007925568148493767,
0.05101296678185463,
0.040650319308042526,
-0.047776173800230026,
-0.13350218534469604,
0.019554244354367256,
-0.013521900400519371,
0.057986754924058914,
-0.010411947965621948,
-0.09526223689317703,
-0.07598970830440521,
0.008489053696393967,
-0.06727813184261322,
0.23210135102272034,
0.2820625603199005,
-0.063027024269104,
0.1374441236257553,
0.1711452305316925,
-0.10035507380962372,
-0.3379229009151459,
-0.08063624054193497,
-0.18898345530033112,
-0.07436835765838623,
-0.03134903311729431,
-0.12211208045482635,
0.08012701570987701,
0.07752654701471329,
-0.05822116136550903,
0.16946236789226532,
-0.1426265686750412,
-0.10255585610866547,
0.11955316364765167,
0.0035280627198517323,
0.2790551781654358,
-0.16766896843910217,
-0.10526716709136963,
-0.10453645884990692,
-0.08739271759986877,
0.17812007665634155,
-0.08564294129610062,
0.0741075947880745,
0.01186076644808054,
-0.006044826470315456,
-0.005852845031768084,
-0.055415112525224686,
0.09902096539735794,
-0.057509779930114746,
0.05684974789619446,
-0.12410257756710052,
0.05157855525612831,
0.12606194615364075,
0.005077750887721777,
0.03226790949702263,
-0.151239812374115,
0.01515447348356247,
0.015820277854800224,
-0.04550417512655258,
-0.03205495327711105,
0.08902449160814285,
0.008286191150546074,
-0.05643158406019211,
-0.04899320751428604,
-0.05606469511985779,
0.0012182962382212281,
-0.03332222253084183,
0.20786674320697784,
-0.046633873134851456,
0.16305558383464813,
0.11626877635717392,
0.1362091302871704,
-0.1383957713842392,
0.11236870288848877,
-0.020965153351426125,
-0.10542184114456177,
0.0699729174375534,
-0.12520024180412292,
0.06001553684473038,
0.07651414722204208,
-0.05399452522397041,
0.06592380255460739,
0.10149753093719482,
0.010504057630896568,
-0.0058835288509726524,
0.17048533260822296,
-0.21712957322597504,
-0.027841897681355476,
-0.04468731954693794,
0.04313308373093605,
0.07866048067808151,
0.09473986178636551,
0.1620844304561615,
-0.015544242225587368,
0.001481342245824635,
0.012155916541814804,
0.027010971680283546,
-0.04953674599528313,
0.03895360231399536,
0.018709653988480568,
0.0049772681668400764,
-0.11753838509321213,
0.07633733004331589,
0.021474463865160942,
-0.12199857085943222,
0.0024703454691916704,
0.07897571474313736,
-0.15915900468826294,
-0.12469859421253204,
-0.02271478809416294,
0.07181016355752945,
-0.156537264585495,
-0.0926901176571846,
-0.02251989021897316,
-0.14750556647777557,
0.048373084515333176,
0.17396965622901917,
0.03671041131019592,
0.08714442700147629,
0.042702414095401764,
-0.022168554365634918,
-0.06192021816968918,
0.013455421663820744,
-0.1173214390873909,
0.06740880012512207,
-0.09965146332979202,
0.03295274078845978,
-0.02560277283191681,
0.031547147780656815,
-0.08395134657621384,
0.021892143413424492,
-0.12608109414577484,
0.0003414917446207255,
-0.06741814315319061,
0.008619471453130245,
-0.11397039145231247,
-0.02096758969128132,
0.0016162191750481725,
0.012214076705276966,
-0.037397272884845734,
-0.026739615947008133,
-0.0779520571231842,
0.012347370386123657,
-0.030438603833317757,
0.05337870493531227,
-0.09572375565767288,
-0.01978202350437641,
0.03950776904821396,
-0.022960565984249115,
0.12157205492258072,
0.020053010433912277,
-0.06695373356342316,
0.10669680684804916,
-0.2281760722398758,
-0.02739614248275757,
0.12611421942710876,
0.018189627677202225,
-0.018842989578843117,
0.056869082152843475,
0.019167013466358185,
0.12612859904766083,
-0.03298935294151306,
0.06164347007870674,
-0.027469109743833542,
-0.10134735703468323,
0.005009813699871302,
-0.05225798115134239,
-0.08473224192857742,
-0.027974102646112442,
-0.053285978734493256,
0.0951649621129036,
-0.0283129271119833,
0.15928736329078674,
-0.08658003807067871,
0.0015982245095074177,
-0.04233749955892563,
0.014432087540626526,
-0.01200929470360279,
-0.1842130720615387,
-0.11348370462656021,
-0.03951694816350937,
0.014225421473383904,
0.01748904213309288,
0.3155052661895752,
-0.008207025937736034,
-0.07072759419679642,
0.06939960271120071,
0.029779626056551933,
0.03808148577809334,
0.01966366358101368,
0.32170113921165466,
0.10220780223608017,
-0.03518247604370117,
-0.12916293740272522,
0.0385720357298851,
0.031898610293865204,
-0.11481058597564697,
0.0768415629863739,
0.05001695081591606,
-0.057655759155750275,
0.08270741254091263,
0.02893637679517269,
-0.03978033736348152,
-0.04630865156650543,
-0.052226077765226364,
-0.049269381910562515,
0.07481169700622559,
0.0021648406982421875,
0.03532717749476433,
0.19295921921730042,
-0.02999711036682129,
-0.022041475400328636,
-0.02618403360247612,
-0.041587669402360916,
-0.17897571623325348,
-0.184970885515213,
-0.10462644696235657,
-0.1543276011943817,
0.03121074102818966,
-0.08962878584861755,
0.031447675079107285,
0.04943243786692619,
0.05078345537185669,
-0.054631344974040985,
0.08281046152114868,
-0.014329242520034313,
-0.052746713161468506,
0.04657139256596565,
-0.03852191939949989,
0.026036912575364113,
-0.0017314543947577477,
-0.07270800322294235,
-0.04574187844991684,
-0.04130619764328003,
-0.00933757796883583,
0.05340241640806198,
0.020133333280682564,
0.0860057845711708,
-0.13258397579193115,
-0.0653938576579094,
-0.038260407745838165,
0.08719656616449356,
-0.017378726974129677,
0.11841363459825516,
0.0025310772471129894,
-0.04905865341424942,
0.09556856751441956,
0.14174483716487885,
-0.03354538977146149,
-0.11209102720022202,
-0.02920149825513363,
0.17688313126564026,
0.016777511686086655,
0.11187730729579926,
-0.026336584240198135,
-0.005438379943370819,
0.017388928681612015,
0.3035591244697571,
0.2709004282951355,
-0.03343981131911278,
0.03128137066960335,
-0.04556027054786682,
0.0233965702354908,
0.07979070395231247,
0.14330004155635834,
0.0675305724143982,
0.21356138586997986,
-0.04014052823185921,
-0.06675122678279877,
-0.01717163622379303,
0.0428086556494236,
-0.11868636310100555,
0.05642444267868996,
-0.017197763547301292,
-0.06292377412319183,
-0.016860581934452057,
0.10756975412368774,
-0.124013751745224,
0.12735669314861298,
-0.028046347200870514,
-0.07458663731813431,
0.005404497496783733,
0.02975502796471119,
0.09861481934785843,
-0.04216795414686203,
0.02073851227760315,
-0.03311353921890259,
-0.07175879925489426,
-0.00475742993876338,
-0.007904637604951859,
-0.18394966423511505,
0.020349040627479553,
-0.009962019510567188,
0.017269672825932503,
0.11532521992921829,
0.004933062009513378,
0.07041039317846298,
0.09284470975399017,
0.02974783256649971,
-0.08884172141551971,
0.14371007680892944,
-0.0039965263567864895,
-0.07052398473024368,
0.06930533796548843,
-0.047673389315605164,
-0.02082948014140129,
-0.0037012435495853424,
0.06228780746459961,
-0.09473217278718948,
0.07300903648138046,
0.008882936090230942,
-0.09975136071443558,
-0.03959985077381134,
-0.00924383383244276,
-0.08098772913217545,
0.0959787592291832,
0.007070253603160381,
-0.019126050174236298,
-0.0026237028650939465,
-0.035258252173662186,
0.02009262703359127,
0.005121667869389057,
-0.06699046492576599,
-0.006899187341332436,
-0.12245488166809082,
-0.051456600427627563,
0.16723278164863586,
0.03918656334280968,
-0.21800144016742706,
0.012990894727408886,
-0.1102394238114357,
0.052274033427238464,
-0.1980532705783844,
0.06837881356477737,
0.1652859002351761,
-0.009850753471255302,
-0.03664639964699745,
-0.13490453362464905,
0.033964768052101135,
0.05329315736889839,
-0.0688016340136528,
-0.1060384213924408
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# result
This model is a fine-tuned version of [microsoft/MiniLM-L12-H384-uncased](https://huggingface.co/microsoft/MiniLM-L12-H384-uncased) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 12
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2.0
### Training results
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu118
- Datasets 2.16.1
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "base_model": "microsoft/MiniLM-L12-H384-uncased", "model-index": [{"name": "result", "results": []}]} | question-answering | glzhangzhi/nlp_hw_6 | [
"transformers",
"safetensors",
"bert",
"question-answering",
"generated_from_trainer",
"base_model:microsoft/MiniLM-L12-H384-uncased",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:13:41+00:00 | [] | [] | TAGS
#transformers #safetensors #bert #question-answering #generated_from_trainer #base_model-microsoft/MiniLM-L12-H384-uncased #license-mit #endpoints_compatible #region-us
|
# result
This model is a fine-tuned version of microsoft/MiniLM-L12-H384-uncased on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 12
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2.0
### Training results
### Framework versions
- Transformers 4.38.0.dev0
- Pytorch 2.2.0+cu118
- Datasets 2.16.1
- Tokenizers 0.15.0
| [
"# result\n\nThis model is a fine-tuned version of microsoft/MiniLM-L12-H384-uncased on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 3e-05\n- train_batch_size: 12\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 2.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #safetensors #bert #question-answering #generated_from_trainer #base_model-microsoft/MiniLM-L12-H384-uncased #license-mit #endpoints_compatible #region-us \n",
"# result\n\nThis model is a fine-tuned version of microsoft/MiniLM-L12-H384-uncased on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 3e-05\n- train_batch_size: 12\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 2.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.0"
] | [
61,
35,
6,
12,
8,
3,
90,
4,
38
] | [
"passage: TAGS\n#transformers #safetensors #bert #question-answering #generated_from_trainer #base_model-microsoft/MiniLM-L12-H384-uncased #license-mit #endpoints_compatible #region-us \n# result\n\nThis model is a fine-tuned version of microsoft/MiniLM-L12-H384-uncased on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 3e-05\n- train_batch_size: 12\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 2.0### Training results### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.2.0+cu118\n- Datasets 2.16.1\n- Tokenizers 0.15.0"
] | [
-0.10214880108833313,
0.038422491401433945,
-0.0006832705112174153,
0.07103057205677032,
0.1610812395811081,
0.022495539858937263,
0.1223580464720726,
0.10244910418987274,
-0.10454361140727997,
0.04806065559387207,
0.06888917833566666,
0.0418655164539814,
0.008730003610253334,
0.0906883254647255,
-0.04178168252110481,
-0.24036821722984314,
0.019059140235185623,
-0.0036809120792895555,
-0.13626116514205933,
0.07245688885450363,
0.10383597016334534,
-0.13201084733009338,
0.0716099888086319,
0.024811802431941032,
-0.17881378531455994,
0.03752271831035614,
-0.008299706503748894,
-0.029708880931138992,
0.10360810160636902,
0.03957541286945343,
0.10192509740591049,
0.004319877829402685,
0.1388636976480484,
-0.18423420190811157,
0.0011875334894284606,
0.07802800089120865,
0.03720594570040703,
0.05658683553338051,
0.02692246064543724,
0.004839879460632801,
0.10465031117200851,
-0.11632019281387329,
0.07522691041231155,
0.05442969501018524,
-0.07060839235782623,
-0.23057162761688232,
-0.10001618415117264,
0.03246162459254265,
0.07515023648738861,
0.10505327582359314,
0.00022042531054466963,
0.17824268341064453,
-0.12528830766677856,
0.05928489938378334,
0.2387647181749344,
-0.25276297330856323,
-0.06450394541025162,
0.0878344476222992,
0.04151374101638794,
0.0443127304315567,
-0.0891718938946724,
-0.007728128228336573,
0.0879971981048584,
0.033633675426244736,
0.0670761987566948,
-0.01588681899011135,
-0.09187664091587067,
0.017258260399103165,
-0.12332729995250702,
0.004042926710098982,
0.190829336643219,
0.05325450748205185,
-0.042437437921762466,
-0.05146053433418274,
-0.04991326853632927,
-0.060979731380939484,
-0.011178561486303806,
-0.033950187265872955,
0.06740991771221161,
-0.06989073753356934,
-0.0865161195397377,
-0.047263193875551224,
-0.08369532972574234,
-0.09174416959285736,
-0.0008305201190523803,
0.16150404512882233,
0.03875277191400528,
0.026666415855288506,
-0.0671214908361435,
0.10872568935155869,
-0.011125467717647552,
-0.11425681412220001,
-0.04582978039979935,
-0.0006302384426817298,
-0.029156360775232315,
-0.044250912964344025,
-0.08642439544200897,
0.01947627030313015,
0.023395227268338203,
0.14368940889835358,
-0.10738735646009445,
0.06676941365003586,
0.045521337538957596,
0.010502365417778492,
-0.057564668357372284,
0.14047861099243164,
-0.05420905351638794,
-0.005461333319544792,
0.007382431998848915,
0.0797632560133934,
0.027923783287405968,
-0.00025876439758576453,
-0.09443984925746918,
0.008801563642919064,
0.067241370677948,
0.046510495245456696,
-0.0896945372223854,
0.03905551880598068,
-0.03537694364786148,
-0.018484197556972504,
-0.041409652680158615,
-0.11623415350914001,
0.02851884439587593,
0.009439583867788315,
-0.08699244260787964,
-0.010490169748663902,
0.046375371515750885,
0.020788803696632385,
-0.0007008735556155443,
0.09742896258831024,
-0.10301322489976883,
0.010954032652080059,
-0.11466877907514572,
-0.07109640538692474,
0.0014784384984523058,
-0.02719767391681671,
0.013694324530661106,
-0.08413492888212204,
-0.19301845133304596,
-0.02859393134713173,
0.04122644290328026,
-0.02935119904577732,
0.0009448144119232893,
-0.03458463400602341,
-0.08135440945625305,
-0.004469151142984629,
-0.009399347007274628,
0.10465843975543976,
-0.04878551885485649,
0.0606670156121254,
0.06387732923030853,
0.04349224641919136,
-0.04732764512300491,
0.02450139820575714,
-0.08634042739868164,
0.02961854450404644,
-0.10217160731554031,
0.03766296058893204,
-0.09383503347635269,
0.0742998719215393,
-0.059985995292663574,
-0.11610202491283417,
0.004337682388722897,
0.010698660276830196,
0.0656406581401825,
0.10535605251789093,
-0.16358719766139984,
-0.04442281648516655,
0.14983385801315308,
-0.09396520256996155,
-0.12796391546726227,
0.0602373406291008,
-0.040137577801942825,
0.07459233701229095,
0.06723199039697647,
0.12352278083562851,
0.056249313056468964,
-0.13611000776290894,
-0.0011061514960601926,
0.008634136989712715,
0.08310756832361221,
0.014282956719398499,
0.04547214135527611,
-0.031159527599811554,
0.006839044392108917,
0.03209824860095978,
-0.07393486797809601,
-0.042343076318502426,
-0.10657711327075958,
-0.08348580449819565,
-0.07336563616991043,
-0.10614318400621414,
0.016638679429888725,
0.04414007067680359,
0.07236693054437637,
-0.09703487902879715,
-0.09511911869049072,
0.2041196972131729,
0.13226626813411713,
-0.06128927692770958,
0.01629459112882614,
-0.0870768204331398,
0.01768735982477665,
0.006150342058390379,
-0.03807626664638519,
-0.18931704759597778,
-0.11179124563932419,
0.003989957273006439,
-0.07381200045347214,
0.02329348586499691,
0.04963250458240509,
0.06036645546555519,
0.07424338907003403,
-0.06528036296367645,
0.001068975543603301,
-0.10673975944519043,
0.012482468970119953,
-0.08389434963464737,
-0.18423502147197723,
-0.06286052614450455,
-0.006382693536579609,
0.1585920751094818,
-0.23141439259052277,
0.02961554192006588,
0.019376195967197418,
0.14635267853736877,
0.03786120563745499,
-0.028690697625279427,
-0.04496893286705017,
0.06033540517091751,
-0.005868155974894762,
-0.08670860528945923,
0.05798641964793205,
0.005104715935885906,
-0.09092660248279572,
-0.1148117333650589,
-0.13737145066261292,
0.06882695853710175,
0.10985159128904343,
-0.011732704937458038,
-0.051557399332523346,
0.0032072472386062145,
-0.0637250766158104,
-0.030995171517133713,
-0.07070329785346985,
-0.02412986196577549,
0.12784580886363983,
-0.014929844066500664,
0.11604791134595871,
-0.08423350751399994,
-0.06905780732631683,
-0.011501673609018326,
-0.0345185212790966,
0.014588369987905025,
0.053538594394922256,
0.05632138252258301,
-0.09813104569911957,
0.09289920330047607,
0.10195086896419525,
-0.0926830843091011,
0.1675693541765213,
-0.06271053105592728,
-0.06438159197568893,
-0.020005609840154648,
-0.014416038990020752,
-0.0033217703457921743,
0.14453667402267456,
-0.1016850620508194,
0.019224882125854492,
0.010200497694313526,
0.03630368039011955,
0.05270387604832649,
-0.20936870574951172,
-0.005905556958168745,
0.008645953610539436,
-0.030692726373672485,
-0.03705540671944618,
-0.012781567871570587,
0.00944675225764513,
0.0688643604516983,
0.017698004841804504,
0.039018940180540085,
0.027457429096102715,
-0.023512346670031548,
-0.0892520546913147,
0.21158467233181,
-0.11854993551969528,
-0.09813470393419266,
-0.09792592376470566,
0.06418347358703613,
-0.06743919849395752,
-0.020133059471845627,
0.03851955756545067,
-0.10590626299381256,
-0.05602722987532616,
-0.07691032439470291,
0.025280334055423737,
0.005675371736288071,
-0.028832823038101196,
0.04771803691983223,
0.03151979297399521,
0.11652285605669022,
-0.1606932133436203,
0.003154870355501771,
-0.03437739610671997,
-0.1240125298500061,
-0.0014469448942691088,
0.03923337161540985,
0.09339454770088196,
0.11450957506895065,
-0.007566667627543211,
0.029240574687719345,
-0.01992928236722946,
0.2481480985879898,
-0.06406798213720322,
-0.060432907193899155,
0.1442948579788208,
0.02826729230582714,
0.03081500716507435,
0.09574409574270248,
0.055585235357284546,
-0.1358165591955185,
0.04830106720328331,
0.09087132662534714,
-0.014139268547296524,
-0.21684125065803528,
-0.04800941050052643,
-0.028402673080563545,
-0.07093991339206696,
0.04690800607204437,
0.030847152695059776,
-0.009520129300653934,
0.03777344897389412,
-0.013987399637699127,
0.04746566340327263,
-0.03692372888326645,
0.0701870247721672,
0.11494417488574982,
0.04196624085307121,
0.11032753437757492,
-0.04010232537984848,
-0.014253976754844189,
0.053775884211063385,
-0.035489484667778015,
0.29239732027053833,
-0.008413058705627918,
0.06794369220733643,
0.08347869664430618,
0.1310511827468872,
-0.026189427822828293,
0.04523588344454765,
0.020475022494792938,
-0.04314497113227844,
0.01986878551542759,
-0.07384487986564636,
-0.01512174867093563,
0.037309180945158005,
-0.07815510779619217,
0.10009293258190155,
-0.12487389892339706,
0.047456443309783936,
0.05677187070250511,
0.2612195611000061,
0.04003234580159187,
-0.2830120623111725,
-0.10697314143180847,
0.026356417685747147,
-0.014545240439474583,
-0.039712484925985336,
0.028662487864494324,
0.12703350186347961,
-0.1293497085571289,
0.06868695467710495,
-0.04825795441865921,
0.08723873645067215,
0.025330159813165665,
0.03190944343805313,
0.06240351125597954,
0.12511928379535675,
0.0035651212092489004,
0.08139390498399734,
-0.2396443784236908,
0.22754956781864166,
0.016494469717144966,
0.13422791659832,
-0.037375688552856445,
0.004452277906239033,
0.04469151794910431,
0.11863266676664352,
0.0572412870824337,
-0.0013656256487593055,
-0.048093825578689575,
-0.1547793596982956,
0.011601577512919903,
0.06731986254453659,
0.13318593800067902,
-0.01973378285765648,
0.10328990966081619,
-0.05888216570019722,
0.0202812347561121,
0.0696888267993927,
-0.03792921453714371,
-0.19565391540527344,
-0.09315094351768494,
0.003875878406688571,
-0.007424152921885252,
-0.00011804953101091087,
-0.12418307363986969,
-0.1033705398440361,
-0.02762106992304325,
0.12245707958936691,
0.006388081703335047,
-0.047256194055080414,
-0.12480371445417404,
0.09660779684782028,
0.08435748517513275,
-0.046121541410684586,
0.029732830822467804,
0.029090411961078644,
0.11603604257106781,
0.059533920139074326,
-0.0738874077796936,
0.07233765721321106,
-0.07041454315185547,
-0.19893483817577362,
-0.030835090205073357,
0.10401535034179688,
0.05088351294398308,
0.046937860548496246,
-0.01916981115937233,
0.009211286902427673,
0.013020549900829792,
-0.10536099225282669,
-0.022922566160559654,
0.08774611353874207,
0.07506265491247177,
0.06976808607578278,
-0.07679405808448792,
0.057275477796792984,
0.001922758063301444,
-0.023670807480812073,
0.10530870407819748,
0.2331165373325348,
-0.07960878312587738,
0.021278418600559235,
0.04283362254500389,
-0.06532935053110123,
-0.2134310007095337,
0.11629029363393784,
0.09315105527639389,
0.012208239175379276,
0.058652859181165695,
-0.14719611406326294,
0.15134310722351074,
0.12398602068424225,
-0.02693081460893154,
0.08860896527767181,
-0.2947426736354828,
-0.1455158293247223,
0.06863383948802948,
0.13883492350578308,
0.03990592435002327,
-0.1372455507516861,
-0.03455016389489174,
-0.07893376052379608,
-0.18694829940795898,
0.11589314043521881,
-0.1896868646144867,
0.10186078399419785,
-0.0051863668486475945,
0.07110118865966797,
0.012853848747909069,
-0.035007692873477936,
0.16189387440681458,
-0.0009413247462362051,
0.13002093136310577,
-0.04647762328386307,
0.0070865098387002945,
0.10082369297742844,
-0.057765357196331024,
0.017897579818964005,
-0.04197122901678085,
0.07509622722864151,
-0.06020943447947502,
-0.024169355630874634,
-0.08432064205408096,
0.07166735827922821,
-0.0769958570599556,
-0.06294915825128555,
-0.0647415816783905,
0.05890950933098793,
0.030011886730790138,
-0.027378443628549576,
0.09068423509597778,
-0.013624308630824089,
0.13794231414794922,
0.09414967894554138,
0.12476159632205963,
-0.09260019659996033,
-0.04231204837560654,
0.01593708246946335,
-0.0069047315046191216,
0.07846075296401978,
-0.15368986129760742,
0.04620605334639549,
0.12182224541902542,
0.0636211708188057,
0.1303240954875946,
0.0640406608581543,
-0.0549180693924427,
0.01414407417178154,
0.05559385567903519,
-0.08764007687568665,
-0.1579446941614151,
-0.02351156249642372,
-0.026004305109381676,
-0.11394856870174408,
0.0791887640953064,
0.10698030889034271,
-0.0636972039937973,
0.008642604574561119,
-0.02328290045261383,
-0.013312747702002525,
-0.05394488573074341,
0.18055766820907593,
0.06297222524881363,
0.0727628692984581,
-0.09561605751514435,
0.14063504338264465,
0.021002739667892456,
-0.04681648313999176,
0.01479082927107811,
0.04833195358514786,
-0.09415635466575623,
-0.02306775376200676,
0.054821744561195374,
0.16636870801448822,
-0.05879760906100273,
-0.0765114575624466,
-0.13472869992256165,
-0.12552805244922638,
0.014696762897074223,
0.1372544914484024,
0.08633221685886383,
-0.02262985147535801,
-0.010592338629066944,
0.06041600555181503,
-0.12352509796619415,
0.0905432254076004,
0.0067392378114163876,
0.06084965169429779,
-0.16695035994052887,
0.14196406304836273,
0.027996590360999107,
0.028706632554531097,
-0.015141511335968971,
0.02163749188184738,
-0.1130470335483551,
-0.010069848038256168,
-0.21958273649215698,
-0.029998041689395905,
-0.011579000391066074,
0.0014160530408844352,
0.010440998710691929,
-0.044970620423555374,
-0.07604894787073135,
0.07047256082296371,
-0.08912953734397888,
-0.050569087266922,
0.02771027572453022,
0.05108199641108513,
-0.10542500019073486,
0.008599142543971539,
0.033920757472515106,
-0.08174082636833191,
0.058203842490911484,
0.052574217319488525,
0.021375609561800957,
0.07509591430425644,
-0.10383488982915878,
-0.02230888046324253,
0.030385281890630722,
0.026618143543601036,
0.07858540862798691,
-0.08048967272043228,
-0.03769182786345482,
-0.014995203353464603,
0.09879207611083984,
0.012070965021848679,
0.04855629801750183,
-0.11284012347459793,
-0.03268251568078995,
-0.05080920457839966,
-0.03376750648021698,
-0.05508153885602951,
0.0026769412215799093,
0.06902297586202621,
0.03212518244981766,
0.16868796944618225,
-0.08419795334339142,
0.004305054433643818,
-0.19849471747875214,
-0.03897259756922722,
0.001572975073941052,
-0.02823009341955185,
-0.12043587118387222,
-0.016908841207623482,
0.0711449533700943,
-0.06230387091636658,
0.1372201144695282,
-0.04691586270928383,
0.11393195390701294,
0.04860926792025566,
-0.0587882474064827,
-0.009708311408758163,
0.012669166550040245,
0.2325936108827591,
0.044424135237932205,
0.021192926913499832,
0.08189241588115692,
0.014822975732386112,
0.059278592467308044,
0.017156444489955902,
0.26656922698020935,
0.1488897055387497,
-0.08561196178197861,
0.09011638909578323,
0.06858715415000916,
-0.09461060911417007,
-0.136133074760437,
0.07527840882539749,
0.008374429307878017,
0.09163209050893784,
-0.065658800303936,
0.1020503044128418,
0.13822847604751587,
-0.17603528499603271,
0.03265255689620972,
-0.10307025909423828,
-0.09678604453802109,
-0.13245083391666412,
0.014879093505442142,
-0.08026392757892609,
-0.19858752191066742,
0.023947108536958694,
-0.13848929107189178,
0.022499805316329002,
0.13160288333892822,
0.0019936058670282364,
0.0036749807186424732,
0.21274122595787048,
-0.08331172913312912,
0.014847522601485252,
0.020016251131892204,
-0.007978975772857666,
-0.016718029975891113,
-0.07286106050014496,
-0.061899345368146896,
0.039521224796772,
-0.02510322444140911,
0.037571020424366,
-0.04456143081188202,
-0.06070277839899063,
0.011681163683533669,
0.0029943615663796663,
-0.054822687059640884,
0.028871918097138405,
0.046818867325782776,
0.036165837198495865,
0.01565561257302761,
0.03395644947886467,
-0.0001659768749959767,
-0.03345523402094841,
0.27693992853164673,
-0.11327768862247467,
-0.12802578508853912,
-0.12884585559368134,
0.24828189611434937,
0.04347173497080803,
0.012621107511222363,
0.057801254093647,
-0.1297813355922699,
-0.017538901418447495,
0.19141842424869537,
0.13213536143302917,
-0.08253628760576248,
-0.00826248712837696,
-0.017843492329120636,
-0.024366052821278572,
-0.1065385639667511,
0.11576984822750092,
0.10763959586620331,
0.04460736736655235,
-0.038590822368860245,
-0.04743880406022072,
0.011644752696156502,
-0.03272220864892006,
-0.07801144570112228,
0.06916404515504837,
0.02245890349149704,
-0.0003790800110436976,
-0.021772537380456924,
0.0450969859957695,
-0.0025578804779797792,
-0.1584234982728958,
0.09620247781276703,
-0.14715592563152313,
-0.15231826901435852,
-0.026824316009879112,
0.07487809658050537,
-0.028258686885237694,
0.0744873508810997,
-0.041179485619068146,
-0.02890796586871147,
0.13648205995559692,
-0.014680073596537113,
-0.052340324968099594,
-0.1351405680179596,
0.09796242415904999,
-0.08463121950626373,
0.2278449982404709,
-0.017432793974876404,
0.07299866527318954,
0.11293245851993561,
0.018033957108855247,
-0.10357566922903061,
0.07042152434587479,
0.0698477029800415,
-0.12889550626277924,
0.000747131125535816,
0.11145851016044617,
-0.04871426522731781,
0.10464200377464294,
0.0365363210439682,
-0.19305098056793213,
0.0033370417077094316,
0.06679482758045197,
-0.046331942081451416,
-0.08010809123516083,
-0.025719331577420235,
-0.08605212718248367,
0.1293260008096695,
0.2032650262117386,
-0.049889225512742996,
0.012774749659001827,
-0.05449087172746658,
0.08406045287847519,
0.05331132933497429,
0.08015140146017075,
-0.021828800439834595,
-0.2377006560564041,
0.061527419835329056,
0.045743074268102646,
-0.034162506461143494,
-0.23069749772548676,
-0.07719779014587402,
0.0672038272023201,
-0.05492941662669182,
-0.02384638600051403,
0.08861972391605377,
0.09773287922143936,
0.04076176509261131,
-0.046091560274362564,
-0.16633130609989166,
-0.0761164054274559,
0.17163290083408356,
-0.1445247381925583,
-0.0858091190457344
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-cola
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8508
- Matthews Correlation: 0.5222
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Matthews Correlation |
|:-------------:|:-----:|:----:|:---------------:|:--------------------:|
| 0.5175 | 1.0 | 535 | 0.4641 | 0.4737 |
| 0.3441 | 2.0 | 1070 | 0.5276 | 0.4519 |
| 0.2263 | 3.0 | 1605 | 0.6166 | 0.4820 |
| 0.1825 | 4.0 | 2140 | 0.7947 | 0.5189 |
| 0.1226 | 5.0 | 2675 | 0.8508 | 0.5222 |
### Framework versions
- Transformers 4.37.2
- Pytorch 2.1.0.post100
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["matthews_correlation"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": []}]} | text-classification | sebastiansson/distilbert-base-uncased-finetuned-cola | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:14:25+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
| distilbert-base-uncased-finetuned-cola
======================================
This model is a fine-tuned version of distilbert-base-uncased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8508
* Matthews Correlation: 0.5222
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.37.2
* Pytorch 2.1.0.post100
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0.post100\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0.post100\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
72,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0.post100\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.09918540716171265,
0.10952845215797424,
-0.0027998662553727627,
0.11115968227386475,
0.13909825682640076,
0.011455338448286057,
0.1575334221124649,
0.11805421859025955,
-0.07272949814796448,
0.037622738629579544,
0.1271652728319168,
0.13514123857021332,
0.013705095276236534,
0.124543696641922,
-0.08126663416624069,
-0.2173682600259781,
0.006084616295993328,
0.02843414433300495,
-0.06125222519040108,
0.11731503903865814,
0.09899840503931046,
-0.1212853267788887,
0.08469711244106293,
-0.012436360120773315,
-0.16581344604492188,
0.013953817076981068,
0.017963789403438568,
-0.05330635979771614,
0.12172436714172363,
0.02645292691886425,
0.12399032711982727,
0.03406988084316254,
0.08207383006811142,
-0.18286378681659698,
0.009388027712702751,
0.06540092080831528,
-0.00575599167495966,
0.08265705406665802,
0.04475090652704239,
-0.010617990978062153,
0.07749440521001816,
-0.09260192513465881,
0.061280205845832825,
0.018355585634708405,
-0.1253901571035385,
-0.2355220913887024,
-0.0846601203083992,
0.03453722968697548,
0.09807407855987549,
0.08083675056695938,
-0.011348223313689232,
0.12143838405609131,
-0.05108213052153587,
0.09517359733581543,
0.20144297182559967,
-0.3085339665412903,
-0.0570409819483757,
0.04580366238951683,
0.017588207498192787,
0.09171521663665771,
-0.09742079675197601,
-0.020200280472636223,
0.0584707036614418,
0.02619449980556965,
0.13123184442520142,
-0.02700990065932274,
-0.05215401574969292,
-0.0017104905564337969,
-0.14391030371189117,
-0.02140060067176819,
0.17440462112426758,
0.05330505222082138,
-0.0475928969681263,
-0.04972748085856438,
-0.07222861796617508,
-0.10667619109153748,
-0.03893764689564705,
-0.010204669088125229,
0.049656931310892105,
-0.018241245299577713,
-0.0576329380273819,
-0.021319322288036346,
-0.09590031206607819,
-0.061858851462602615,
-0.05085647106170654,
0.1398085653781891,
0.03847574442625046,
0.01021587010473013,
-0.012409732677042484,
0.10017907619476318,
-0.022071432322263718,
-0.14696824550628662,
0.023150475695729256,
0.016935821622610092,
0.014422710984945297,
-0.04454508796334267,
-0.05497660115361214,
-0.09024739265441895,
0.015588556416332722,
0.1618066430091858,
-0.04698273912072182,
0.054370008409023285,
0.004391283728182316,
0.045632582157850266,
-0.09627922624349594,
0.1588800549507141,
-0.03533352538943291,
-0.03311815857887268,
0.0285312682390213,
0.08498013019561768,
0.055885199457407,
-0.013608837500214577,
-0.12684418261051178,
0.04166419059038162,
0.10638432204723358,
0.01748322695493698,
-0.0480012521147728,
0.068929523229599,
-0.058113448321819305,
-0.02000315859913826,
0.040773943066596985,
-0.09896226227283478,
0.025595372542738914,
0.0003400439163669944,
-0.061831194907426834,
-0.03870568424463272,
0.03422771021723747,
0.018077731132507324,
-0.0037857890129089355,
0.10712700337171555,
-0.07603821903467178,
0.012083617970347404,
-0.08124494552612305,
-0.12681415677070618,
0.01296938769519329,
-0.08717121928930283,
0.021338235586881638,
-0.10927107185125351,
-0.19151651859283447,
-0.016707461327314377,
0.06784743070602417,
-0.026968063786625862,
-0.03733532130718231,
-0.057145483791828156,
-0.07186254113912582,
0.01560496911406517,
-0.015743296593427658,
0.06994020193815231,
-0.06494370847940445,
0.09569893032312393,
0.03456460312008858,
0.06501862406730652,
-0.05727163702249527,
0.04419717192649841,
-0.1078069657087326,
0.03440326079726219,
-0.1759592592716217,
0.03435272350907326,
-0.07142141461372375,
0.07165180891752243,
-0.08480407297611237,
-0.07308702915906906,
0.003060003276914358,
-0.003666185075417161,
0.06729752570390701,
0.09536062926054001,
-0.16849209368228912,
-0.0621507503092289,
0.1533018946647644,
-0.08815863728523254,
-0.1433558464050293,
0.13205084204673767,
-0.058060385286808014,
0.04890841618180275,
0.06818928569555283,
0.19010397791862488,
0.07364527136087418,
-0.08395706117153168,
0.0118927713483572,
0.003826362080872059,
0.0641862004995346,
-0.03754108026623726,
0.07101979851722717,
-0.003312636399641633,
-0.008031697943806648,
0.017139311879873276,
-0.04887978360056877,
0.044280461966991425,
-0.07503286749124527,
-0.08799202740192413,
-0.045090630650520325,
-0.10309763252735138,
0.056196048855781555,
0.0546409972012043,
0.0670943632721901,
-0.11266086995601654,
-0.08981997519731522,
0.06926826387643814,
0.07351595908403397,
-0.07154957950115204,
0.02785249426960945,
-0.06294240802526474,
0.08611220866441727,
-0.060000527650117874,
-0.012665865011513233,
-0.16311505436897278,
-0.03162907436490059,
0.020147565752267838,
0.001003387849777937,
0.019483475014567375,
0.0004081057559233159,
0.07101048529148102,
0.07534157484769821,
-0.07065591216087341,
-0.033910248428583145,
-0.013974918983876705,
0.015209258534014225,
-0.12256777286529541,
-0.19632624089717865,
-0.012132162228226662,
-0.034605253487825394,
0.14026345312595367,
-0.2210167944431305,
0.05323110520839691,
-0.007290673442184925,
0.09310513734817505,
0.03717559948563576,
-0.011758444830775261,
-0.036543890833854675,
0.062157075852155685,
-0.047274988144636154,
-0.06910045444965363,
0.06397224217653275,
0.011218882165849209,
-0.0943567082285881,
-0.04624650999903679,
-0.14442019164562225,
0.18113601207733154,
0.13368168473243713,
-0.08707200735807419,
-0.06834060698747635,
0.008390668779611588,
-0.03753599524497986,
-0.025667166337370872,
-0.04235432296991348,
0.004037561360746622,
0.12172063440084457,
-0.011115927249193192,
0.15713268518447876,
-0.08990471810102463,
-0.038037966936826706,
0.01749979704618454,
-0.04859010875225067,
0.011719519272446632,
0.11101803928613663,
0.08353177458047867,
-0.10557500272989273,
0.1464804708957672,
0.20560124516487122,
-0.09581980854272842,
0.12530189752578735,
-0.049720872193574905,
-0.053551964461803436,
-0.02300531230866909,
0.003443800611421466,
0.012261972762644291,
0.10344847291707993,
-0.10853395611047745,
0.005330473184585571,
0.016126105561852455,
0.017864083871245384,
0.012637985870242119,
-0.21089178323745728,
-0.021102793514728546,
0.0404253825545311,
-0.048292096704244614,
0.0021909892093390226,
-0.025135794654488564,
-0.007134533952921629,
0.09846746921539307,
-0.009996266104280949,
-0.08519577234983444,
0.04910226911306381,
-0.003810721915215254,
-0.07659651339054108,
0.20359647274017334,
-0.09514002501964569,
-0.15176674723625183,
-0.12737657129764557,
-0.06145340949296951,
-0.06014768406748772,
0.030733399093151093,
0.0689171627163887,
-0.06591558456420898,
-0.04690474271774292,
-0.11263532191514969,
-0.008560564368963242,
0.022724425420165062,
0.021844172850251198,
0.018514463678002357,
-0.0022553782910108566,
0.09111488610506058,
-0.10449259728193283,
-0.006617920473217964,
-0.03655526041984558,
-0.048297781497240067,
0.03610340133309364,
0.02830566093325615,
0.10387900471687317,
0.1434391438961029,
-0.02461906149983406,
-0.0014482594560831785,
-0.024072963744401932,
0.22838033735752106,
-0.05563629791140556,
-0.004471868276596069,
0.1346987634897232,
-0.021905450150370598,
0.058032602071762085,
0.13342882692813873,
0.06175020709633827,
-0.09909850358963013,
0.019758054986596107,
0.037890367209911346,
-0.033501267433166504,
-0.21930700540542603,
-0.03624303638935089,
-0.03721863031387329,
0.002109420020133257,
0.10216720402240753,
0.03319579362869263,
0.028553983196616173,
0.06795760244131088,
0.01474855374544859,
0.07577842473983765,
-0.007449786644428968,
0.0742892324924469,
0.11239592730998993,
0.040772728621959686,
0.13443857431411743,
-0.05129411071538925,
-0.050176411867141724,
0.045494195073843,
-0.010403979569673538,
0.19735902547836304,
0.020672250539064407,
0.1271076202392578,
0.05270140990614891,
0.15882396697998047,
-0.006613150238990784,
0.061684682965278625,
-0.012890028767287731,
-0.03297553211450577,
-0.018569309264421463,
-0.04905761778354645,
-0.025752048939466476,
0.0374843068420887,
-0.08359900116920471,
0.053947485983371735,
-0.10284333676099777,
-0.0015621122438460588,
0.06211623549461365,
0.22456888854503632,
0.05989588424563408,
-0.3149779438972473,
-0.08690163493156433,
0.03068653680384159,
-0.022421235218644142,
-0.021597113460302353,
0.031252436339855194,
0.10811489820480347,
-0.05026739835739136,
0.04710332676768303,
-0.07224394381046295,
0.09194858372211456,
-0.034556251019239426,
0.042978014796972275,
0.052048832178115845,
0.08577791601419449,
-0.009673984721302986,
0.07306557893753052,
-0.2902931272983551,
0.25650736689567566,
0.019269423559308052,
0.0677768737077713,
-0.05101475864648819,
0.0036771241575479507,
0.035993997007608414,
0.09048552066087723,
0.06837793439626694,
-0.013766516000032425,
-0.0652356818318367,
-0.17882375419139862,
-0.06113399937748909,
0.020582156255841255,
0.09155929833650589,
-0.04084502533078194,
0.09391363710165024,
-0.026796959340572357,
0.004645305685698986,
0.08013418316841125,
-0.01796312816441059,
-0.09550898522138596,
-0.10082390159368515,
-0.00909969862550497,
0.03703516349196434,
-0.04064198583364487,
-0.07605426013469696,
-0.09666287899017334,
-0.13658857345581055,
0.15877406299114227,
-0.05830254778265953,
-0.04235168546438217,
-0.1068197563290596,
0.05844207853078842,
0.052942413836717606,
-0.07899348437786102,
0.041699428111314774,
0.0046022674068808556,
0.07726394385099411,
0.017257291823625565,
-0.07134696841239929,
0.12145256996154785,
-0.07376263290643692,
-0.1742090880870819,
-0.06720238924026489,
0.11438346654176712,
0.021936217322945595,
0.04054776951670647,
-0.007344383280724287,
0.01236915122717619,
-0.022803237661719322,
-0.07962217926979065,
0.03161932900547981,
0.0019526710966601968,
0.06431519240140915,
0.02854548953473568,
-0.05537070333957672,
0.001658280030824244,
-0.057155005633831024,
-0.026210512965917587,
0.14412303268909454,
0.28888818621635437,
-0.08295685797929764,
0.016908500343561172,
0.058169834315776825,
-0.06689807772636414,
-0.2072044014930725,
0.028672806918621063,
0.02745681069791317,
-0.006629203446209431,
0.052799466997385025,
-0.1504259556531906,
0.10497435182332993,
0.09928493201732635,
-0.031025391072034836,
0.10259490460157394,
-0.29374393820762634,
-0.1377611607313156,
0.13706326484680176,
0.14606358110904694,
0.12527723610401154,
-0.1601434201002121,
-0.0426444336771965,
-0.04218650609254837,
-0.1054348573088646,
0.11504998058080673,
-0.1348186880350113,
0.107291080057621,
-0.010021120309829712,
0.045950524508953094,
0.004202555399388075,
-0.05058290809392929,
0.13730083405971527,
0.004724507220089436,
0.11888779699802399,
-0.06600680947303772,
0.007466739509254694,
0.060912419110536575,
-0.05935608595609665,
0.03075111284852028,
-0.12021870911121368,
0.04890831932425499,
-0.07222973555326462,
-0.021522412076592445,
-0.04639800637960434,
0.040495846420526505,
-0.03619338199496269,
-0.061351753771305084,
-0.04620622843503952,
0.021716110408306122,
0.04688999801874161,
-0.01030013058334589,
0.17415045201778412,
0.019475961104035378,
0.14379264414310455,
0.15695346891880035,
0.0693126991391182,
-0.07274787873029709,
-0.021798375993967056,
-0.016279613599181175,
-0.03600917384028435,
0.0653512179851532,
-0.1662880927324295,
0.04320202022790909,
0.12144879251718521,
0.010272623039782047,
0.14731676876544952,
0.06743146479129791,
-0.030750934034585953,
0.00586542347446084,
0.060652755200862885,
-0.16755343973636627,
-0.10317348688840866,
-0.011121110990643501,
-0.022520234808325768,
-0.11349447816610336,
0.06103353202342987,
0.1225145235657692,
-0.06553221493959427,
0.0028365820180624723,
-0.006149031687527895,
0.016495367512106895,
-0.03061574324965477,
0.188408762216568,
0.07328347861766815,
0.044130824506282806,
-0.08880586922168732,
0.10313757508993149,
0.060194339603185654,
-0.07862018048763275,
0.010694228112697601,
0.04599964991211891,
-0.08485861867666245,
-0.04810907319188118,
0.06055182218551636,
0.18969355523586273,
-0.027183033525943756,
-0.050699058920145035,
-0.1497320830821991,
-0.11459796130657196,
0.05883253738284111,
0.15336625277996063,
0.10169470310211182,
0.00912106316536665,
-0.03543134033679962,
0.0068871560506522655,
-0.10697313398122787,
0.12023723125457764,
0.05549609288573265,
0.08660361170768738,
-0.1517624706029892,
0.11390997469425201,
-0.0005332553992047906,
0.011905214749276638,
-0.026155643165111542,
0.044007688760757446,
-0.11188208311796188,
-0.010069887153804302,
-0.1412074714899063,
-0.005487140733748674,
-0.021098904311656952,
0.009606819599866867,
-0.0016334631945937872,
-0.05558263510465622,
-0.05382266640663147,
0.0165872760117054,
-0.10223732888698578,
-0.025551367551088333,
0.03542415425181389,
0.04823778197169304,
-0.12326515465974808,
-0.05341654270887375,
0.01830504648387432,
-0.07032789289951324,
0.06549231708049774,
0.02224108576774597,
0.01462445966899395,
0.04829104244709015,
-0.17762570083141327,
0.01983613893389702,
0.057756055146455765,
0.018319984897971153,
0.041578251868486404,
-0.08556288480758667,
-0.01887107640504837,
0.0051474603824317455,
0.03982694074511528,
0.01631723716855049,
0.0892069861292839,
-0.12760032713413239,
0.010985702276229858,
-0.03322078287601471,
-0.06444841623306274,
-0.05243319645524025,
0.029395557940006256,
0.08308985829353333,
0.016016820445656776,
0.215332493185997,
-0.09820207208395004,
0.01844814233481884,
-0.20750965178012848,
0.003971876576542854,
0.002736111404374242,
-0.12189524620771408,
-0.11513222754001617,
-0.04938247427344322,
0.05014324560761452,
-0.05967044085264206,
0.14138147234916687,
0.018795862793922424,
0.025177916511893272,
0.0389607772231102,
-0.02091044932603836,
0.04329782351851463,
0.020719312131404877,
0.21746519207954407,
0.027062881737947464,
-0.036503199487924576,
0.01944524422287941,
0.025750556960701942,
0.11552941799163818,
0.0825837254524231,
0.16918572783470154,
0.16783323884010315,
-0.04575946554541588,
0.09630053490400314,
0.03280067816376686,
-0.04715052619576454,
-0.13594625890254974,
0.06894367933273315,
-0.02593071572482586,
0.10879423469305038,
-0.015576260164380074,
0.1841539442539215,
0.0908350870013237,
-0.15373976528644562,
0.022130396217107773,
-0.04447329044342041,
-0.08311648666858673,
-0.11118071526288986,
-0.07133056968450546,
-0.097322016954422,
-0.15293431282043457,
-0.006561268121004105,
-0.11127962917089462,
0.014614882878959179,
0.10119283944368362,
0.0016647299053147435,
-0.019721413031220436,
0.15801654756069183,
-0.004913046956062317,
0.027977827936410904,
0.05977949500083923,
-0.005856530740857124,
-0.048820964992046356,
-0.06581347435712814,
-0.09455189853906631,
0.007006634958088398,
-0.0062814741395413876,
0.02078196592628956,
-0.04636525735259056,
-0.024145785719156265,
0.04004529491066933,
-0.011443573981523514,
-0.10937903076410294,
0.011449525132775307,
0.029701855033636093,
0.051340412348508835,
0.06144389510154724,
0.014230663888156414,
0.008610144257545471,
0.000908843707293272,
0.2246900200843811,
-0.07545411586761475,
-0.06827463954687119,
-0.09522698819637299,
0.2224937379360199,
0.031223749741911888,
0.006381338927894831,
0.016603123396635056,
-0.09439617395401001,
0.02722114883363247,
0.20731626451015472,
0.19005000591278076,
-0.0856798067688942,
-0.001503183040767908,
-0.02405187115073204,
-0.011081013828516006,
-0.02376588061451912,
0.09252552688121796,
0.11205791682004929,
0.004650635179132223,
-0.06886576861143112,
-0.048455361276865005,
-0.03746768832206726,
-0.006553171202540398,
-0.06351520121097565,
0.06892997771501541,
0.029045745730400085,
0.0007422540802508593,
-0.04252525418996811,
0.06081017851829529,
-0.026804398745298386,
-0.10273177921772003,
0.04176092892885208,
-0.1837586909532547,
-0.15061897039413452,
-0.017919406294822693,
0.11396563798189163,
-0.004851295612752438,
0.042152538895606995,
-0.0319334976375103,
0.006995183881372213,
0.07250195741653442,
-0.030733920633792877,
-0.0650906190276146,
-0.07098258286714554,
0.06426194310188293,
-0.1253982037305832,
0.22336925566196442,
-0.027693569660186768,
0.05006589740514755,
0.123532734811306,
0.0412566140294075,
-0.07428350299596786,
0.0855659618973732,
0.042000845074653625,
-0.05278504639863968,
0.039781354367733,
0.08848758786916733,
-0.038755375891923904,
0.1121215671300888,
0.06027621403336525,
-0.13248935341835022,
0.0034846989437937737,
-0.037641774863004684,
-0.09519516676664352,
-0.055335018783807755,
-0.03669809550046921,
-0.06169193983078003,
0.13045884668827057,
0.18099373579025269,
-0.04033561423420906,
0.008843999356031418,
-0.04824141040444374,
0.018942290917038918,
0.06574686616659164,
0.024731462821364403,
-0.030439600348472595,
-0.22261634469032288,
0.026952655985951424,
0.06366971880197525,
-0.003947112709283829,
-0.2710157036781311,
-0.08103446662425995,
-0.015503022819757462,
-0.05110010877251625,
-0.09626956284046173,
0.08799844980239868,
0.10621721297502518,
0.04578324779868126,
-0.06457311660051346,
-0.08579687029123306,
-0.0763295367360115,
0.1530027985572815,
-0.12597429752349854,
-0.08990342170000076
] |
null | null | null |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | null | LoneStriker/Wiedervereinigung-7b-dpo-GGUF | [
"gguf",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"region:us"
] | 2024-02-08T17:17:23+00:00 | [] | [
"de",
"en"
] | TAGS
#gguf #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#gguf #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
156,
148,
120,
4,
3
] | [
"passage: TAGS\n#gguf #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06840161234140396,
-0.0017888112924993038,
-0.0030518292915076017,
0.0388304628431797,
0.024843620136380196,
0.027147727087140083,
0.09619682282209396,
0.0874171331524849,
0.11155671626329422,
0.0542539544403553,
-0.03507915511727333,
-0.013619470410048962,
0.06772734224796295,
0.08502312749624252,
0.026791291311383247,
-0.2681794762611389,
0.09561140835285187,
-0.04525778442621231,
-0.025232266634702682,
0.031469009816646576,
0.1430734395980835,
-0.06246589124202728,
0.06688480824232101,
-0.006146128289401531,
-0.0500844269990921,
0.04531572759151459,
-0.05160384997725487,
-0.000232265010708943,
0.08475756645202637,
0.07735604047775269,
0.0441252738237381,
0.013860472477972507,
0.0013139918446540833,
-0.13803324103355408,
0.010634801350533962,
-0.013623183593153954,
-0.017985986545681953,
0.02841457910835743,
0.10188905894756317,
-0.008087546564638615,
0.12908077239990234,
-0.1365516632795334,
-0.00982335302978754,
0.054925840348005295,
-0.08953079581260681,
-0.10654260218143463,
-0.1938284933567047,
0.1756240576505661,
0.07733850926160812,
0.027085278183221817,
-0.048629194498062134,
0.05659037083387375,
-0.03808183595538139,
0.04227817803621292,
0.1705092340707779,
-0.20762620866298676,
-0.049266327172517776,
0.23333269357681274,
0.0005461449618451297,
-0.007300557568669319,
-0.06994989514350891,
0.04077264666557312,
0.04189057648181915,
0.009787443093955517,
-0.04205542057752609,
-0.05106431245803833,
0.2417343556880951,
-0.018990539014339447,
-0.1525419056415558,
-0.011635920032858849,
0.07343974709510803,
0.07428617030382156,
-0.07341983169317245,
-0.16962192952632904,
0.015645697712898254,
-0.0065629128366708755,
-0.04002932459115982,
-0.12758949398994446,
0.008198602125048637,
-0.016076084226369858,
0.09494868665933609,
-0.07637429237365723,
-0.053936682641506195,
0.03579752519726753,
-0.03440464660525322,
0.19957131147384644,
0.006943247746676207,
-0.0364774651825428,
0.03706680238246918,
0.07555364817380905,
-0.22371743619441986,
-0.09843448549509048,
-0.06838352233171463,
-0.09247118979692459,
-0.09215892851352692,
-0.054924916476011276,
-0.04837856441736221,
-0.07314795255661011,
-0.01751560904085636,
0.18222351372241974,
-0.03751533105969429,
0.0499316081404686,
0.07444304972887039,
-0.009718216024339199,
0.03925880789756775,
0.12960711121559143,
-0.08097010105848312,
-0.13039463758468628,
-0.02439415268599987,
0.002973714144900441,
0.04612143710255623,
0.03875245898962021,
-0.04190754517912865,
-0.053650278598070145,
0.0023283218033611774,
0.01770230010151863,
0.002027674810960889,
0.04521307349205017,
-0.07689335197210312,
-0.05825570225715637,
0.1900431513786316,
-0.11587149649858475,
0.015317718498408794,
-0.04085158929228783,
-0.08271851390600204,
0.11403845995664597,
0.002142189769074321,
0.013236995786428452,
-0.01768096163868904,
0.17041955888271332,
-0.03710823133587837,
-0.023710748180747032,
-0.037302445620298386,
-0.05981605499982834,
0.05273956432938576,
-0.02404119074344635,
-0.0416339673101902,
-0.09936192631721497,
-0.09640076011419296,
-0.08816414326429367,
0.0335099920630455,
-0.021210886538028717,
0.008597025647759438,
-0.007933327928185463,
-0.04441152140498161,
-0.024494096636772156,
0.028384748846292496,
-0.047643233090639114,
-0.0011925847502425313,
-0.03568118438124657,
-0.08044520020484924,
0.01668543741106987,
-0.11059300601482391,
0.030935117974877357,
-0.08627226203680038,
0.0373958982527256,
-0.3303810954093933,
0.09268131107091904,
-0.12310991436243057,
0.011590744368731976,
-0.14586637914180756,
-0.05776023119688034,
-0.014607766643166542,
0.0035792638082057238,
0.03211873024702072,
0.1417909562587738,
-0.13290925323963165,
-0.0789678543806076,
0.08475371450185776,
-0.12599648535251617,
-0.04133446514606476,
0.1300671547651291,
-0.015136588364839554,
0.03359086811542511,
0.1230473443865776,
0.11459854245185852,
0.20318780839443207,
-0.08640233427286148,
-0.11251287907361984,
-0.05428965762257576,
0.043083496391773224,
0.16670899093151093,
0.04927781596779823,
-0.03652097284793854,
0.012792776338756084,
0.048784807324409485,
-0.10010506212711334,
0.00003981325062341057,
0.016366735100746155,
-0.056102536618709564,
-0.028547482565045357,
0.0028197881765663624,
0.07283233106136322,
-0.030790505930781364,
-0.001084300922229886,
-0.03221023455262184,
-0.09255094826221466,
0.0514618381857872,
0.11896932870149612,
0.017088904976844788,
-0.006129765417426825,
-0.04336997866630554,
0.10137882083654404,
-0.01948629319667816,
0.017237713560461998,
-0.10447815805673599,
-0.14261774718761444,
0.01601540483534336,
-0.22129403054714203,
0.054493945091962814,
0.0818171426653862,
0.0838375985622406,
0.014272196218371391,
-0.08632466197013855,
-0.020988386124372482,
-0.061328452080488205,
-0.003967034164816141,
0.02728147618472576,
-0.1075519323348999,
-0.07382166385650635,
-0.05169977247714996,
0.1807737946510315,
-0.06999067217111588,
-0.004474112298339605,
0.05879734084010124,
0.20623378455638885,
0.03945375606417656,
-0.03183169290423393,
-0.04333674535155296,
0.03561781346797943,
0.03192317485809326,
-0.007314297370612621,
-0.006078056525439024,
-0.04463151469826698,
-0.14275383949279785,
0.05601172149181366,
-0.034948304295539856,
-0.04228996858000755,
0.07425958663225174,
0.07763238251209259,
-0.08604077249765396,
-0.06234272941946983,
-0.03424376621842384,
-0.03297135978937149,
-0.008990639820694923,
-0.10289739817380905,
0.08033552020788193,
0.04600188881158829,
0.0031668993178755045,
-0.04123180732131004,
-0.04893621802330017,
-0.003092672675848007,
0.016054129227995872,
-0.1078248843550682,
0.0805295929312706,
-0.07027620077133179,
-0.16854548454284668,
0.08521867543458939,
0.16619117558002472,
0.02855684608221054,
0.1411195695400238,
-0.01663399487733841,
-0.033001065254211426,
-0.08531580120325089,
0.032762739807367325,
-0.03033575788140297,
0.07077176868915558,
0.010432506911456585,
0.04705384746193886,
0.03455682843923569,
0.03147063031792641,
0.029617488384246826,
-0.036462508141994476,
0.040493324398994446,
0.007885719649493694,
-0.026744725182652473,
0.05494749918580055,
0.09379344433546066,
-0.004005890339612961,
0.0775795429944992,
0.05404417961835861,
-0.030621157959103584,
-0.01059874426573515,
-0.029847528785467148,
-0.07662083953619003,
0.09537645429372787,
-0.08770639449357986,
-0.11284710466861725,
-0.16593094170093536,
0.003319225972518325,
-0.17895303666591644,
-0.022466357797384262,
0.032577451318502426,
-0.052206944674253464,
-0.08680839836597443,
-0.054081030189991,
0.17820481956005096,
0.019742010161280632,
-0.05207791551947594,
-0.05703709274530411,
-0.05087970942258835,
0.008273288607597351,
-0.11960623413324356,
-0.0496126189827919,
-0.01899242401123047,
-0.014120257459580898,
-0.008374493569135666,
-0.009907256811857224,
0.0517691969871521,
0.03400042653083801,
0.024661051109433174,
-0.03619746118783951,
-0.04383928328752518,
0.20793527364730835,
-0.11589927971363068,
0.10326937586069107,
0.07705552875995636,
-0.02394883707165718,
0.059691864997148514,
0.15724070370197296,
0.05306728184223175,
-0.056635282933712006,
-0.0038286992348730564,
0.05797167494893074,
-0.02116609551012516,
-0.18086840212345123,
-0.1543416827917099,
-0.06449683010578156,
-0.024428104981780052,
0.0012033306993544102,
0.04585789144039154,
0.034679435193538666,
0.05752227455377579,
-0.10471518337726593,
-0.03291408345103264,
0.05969996750354767,
0.06053841859102249,
0.13129857182502747,
0.008958708494901657,
0.06077890843153,
-0.06131463125348091,
-0.05939200520515442,
0.10223813354969025,
-0.011268937960267067,
0.1548515260219574,
0.021775241941213608,
0.08050905168056488,
0.04164973273873329,
0.04495275020599365,
0.004773375112563372,
0.0709766075015068,
0.0275344867259264,
-0.03589010611176491,
-0.02264687791466713,
-0.10566751658916473,
0.03824448958039284,
0.08882484585046768,
0.07639138400554657,
0.010145734064280987,
-0.04122718796133995,
-0.05838623270392418,
0.09700897336006165,
0.2145635187625885,
0.04531486704945564,
-0.14198806881904602,
-0.1036786139011383,
0.046197496354579926,
0.001693897764198482,
0.030009644106030464,
-0.04655744880437851,
-0.009357464499771595,
-0.12088636308908463,
0.16102400422096252,
-0.03908431529998779,
0.062748983502388,
-0.03275253251194954,
0.0010969472350552678,
-0.004709696862846613,
0.0646546334028244,
0.0028901793994009495,
0.052895594388246536,
-0.17493727803230286,
0.10404045134782791,
0.007418571971356869,
0.049620918929576874,
-0.05013047903776169,
0.060389623045921326,
0.060659993439912796,
0.03242199495434761,
0.12456677854061127,
0.04426801949739456,
-0.05656754598021507,
-0.059626881033182144,
-0.05510077625513077,
-0.032016560435295105,
0.11391696333885193,
-0.09958772361278534,
0.08122432231903076,
-0.010713901370763779,
-0.01355021446943283,
-0.05999988317489624,
0.14424531161785126,
-0.16207104921340942,
-0.09823974221944809,
0.11107531189918518,
-0.041497521102428436,
0.049344029277563095,
-0.05613982677459717,
-0.05936241149902344,
-0.22692131996154785,
0.181588277220726,
-0.004003545269370079,
-0.0326298326253891,
-0.09952236711978912,
0.01417502574622631,
0.17320331931114197,
-0.10397206991910934,
0.04236747324466705,
0.011818446218967438,
0.08458010107278824,
-0.06539016962051392,
-0.10392622649669647,
0.017119476571679115,
-0.08514531701803207,
-0.10831049829721451,
0.042233988642692566,
0.13960778713226318,
0.07056258618831635,
0.022222649306058884,
0.022310800850391388,
0.039023708552122116,
0.07552062720060349,
-0.13744422793388367,
0.020096078515052795,
0.17061643302440643,
-0.034555573016405106,
0.15110036730766296,
-0.06946263462305069,
-0.1366567611694336,
-0.0645747184753418,
-0.0145940650254488,
0.043775711208581924,
0.2936514914035797,
-0.04839940741658211,
0.07461727410554886,
0.0825270563364029,
-0.03314992040395737,
-0.18228667974472046,
0.03560005873441696,
0.06827987730503082,
0.04022214189171791,
0.011418952606618404,
-0.08250870555639267,
0.10270102322101593,
0.13716429471969604,
-0.020948052406311035,
-0.011350592598319054,
-0.24887965619564056,
-0.11121756583452225,
0.04805891215801239,
-0.02069733291864395,
0.009748424403369427,
-0.058731596916913986,
-0.08168071508407593,
-0.04078337177634239,
-0.1808154284954071,
0.043770354241132736,
0.006688766181468964,
0.02532745525240898,
0.024722760543227196,
0.02942412532866001,
0.06275545060634613,
-0.01876659132540226,
0.2105501890182495,
0.09428294003009796,
0.047712862491607666,
-0.08455473929643631,
-0.05085933953523636,
0.026487428694963455,
-0.05029861629009247,
0.1595267653465271,
0.06061037629842758,
-0.00196132343262434,
-0.09412924945354462,
0.0016599632799625397,
-0.07553131133317947,
0.10083209723234177,
-0.06386395543813705,
-0.05701732635498047,
-0.09962472319602966,
0.1471213400363922,
0.02638273872435093,
-0.004370477050542831,
0.12057344615459442,
-0.03807448595762253,
0.0398310162127018,
0.07238969951868057,
0.05949443578720093,
0.015678543597459793,
-0.07740778475999832,
0.002297682221978903,
-0.03283010050654411,
0.06367343664169312,
0.015648530796170235,
0.04636754095554352,
0.13292866945266724,
-0.014244148507714272,
0.1502600759267807,
-0.011404857970774174,
-0.13799646496772766,
-0.0422905869781971,
0.1051933541893959,
-0.11967786401510239,
-0.2364330142736435,
-0.04458756372332573,
-0.03943411633372307,
-0.0049179014749825,
0.04030853509902954,
0.2107355296611786,
-0.062353603541851044,
-0.027217812836170197,
0.02592354081571102,
0.043288454413414,
-0.08820900321006775,
0.12484337389469147,
-0.021780861541628838,
0.014541259966790676,
-0.07590065896511078,
0.09937030076980591,
0.05768408253788948,
-0.21584051847457886,
-0.03150317072868347,
0.020220313221216202,
-0.06972553580999374,
-0.06029830873012543,
-0.07331515103578568,
0.1123185008764267,
-0.12647493183612823,
-0.019121482968330383,
-0.019532926380634308,
-0.1282310038805008,
0.015812573954463005,
0.021606123074889183,
0.07534259557723999,
0.04973883181810379,
0.03181879594922066,
0.0217901561409235,
0.011061257682740688,
0.040814414620399475,
0.12976741790771484,
0.10137387365102768,
-0.09492650628089905,
0.02660183422267437,
-0.0866955816745758,
0.017181580886244774,
-0.023337025195360184,
-0.0054610357619822025,
-0.11408933997154236,
-0.07531997561454773,
-0.14225172996520996,
-0.027895784005522728,
-0.0949355959892273,
0.002274530241265893,
-0.009548227302730083,
-0.013263814151287079,
0.023399828001856804,
-0.0032376274466514587,
-0.09483691304922104,
-0.04691332206130028,
-0.032208602875471115,
0.08426263928413391,
-0.1018306165933609,
0.00822862796485424,
0.04115734249353409,
-0.04635857790708542,
0.07905441522598267,
0.04148531332612038,
0.03290518373250961,
0.03525659814476967,
-0.014903362840414047,
-0.007509151939302683,
0.029024822637438774,
0.003667125478386879,
0.028296733275055885,
-0.14503978192806244,
-0.01197761669754982,
0.001313225831836462,
-0.0463910736143589,
0.02045964077115059,
-0.08159632235765457,
-0.1002078503370285,
-0.03126963973045349,
0.016943570226430893,
-0.05013694241642952,
-0.06916742026805878,
0.032034844160079956,
0.08368753641843796,
0.058873407542705536,
0.13978137075901031,
-0.03700299933552742,
0.03175528720021248,
-0.10256227850914001,
-0.030871769413352013,
-0.006160784978419542,
-0.06537114083766937,
0.030999554321169853,
-0.0422525592148304,
0.014495272189378738,
-0.012425215914845467,
0.08163832128047943,
-0.012993874959647655,
-0.04016747325658798,
0.042328570038080215,
-0.038857392966747284,
-0.023197470232844353,
0.025073828175663948,
0.08526187390089035,
0.028389565646648407,
-0.007557916920632124,
-0.10334204882383347,
0.0012589080724865198,
-0.05099719762802124,
-0.008943749591708183,
0.0930357277393341,
0.15324245393276215,
0.06707746535539627,
0.05850912258028984,
0.06615997850894928,
-0.0348985455930233,
-0.00036426997394300997,
0.014931169338524342,
0.01356563065201044,
0.07299178093671799,
-0.035501573234796524,
-0.004012246150523424,
0.06073828786611557,
-0.24714361131191254,
0.1567402333021164,
-0.013849245384335518,
0.003963680472224951,
-0.037547413259744644,
-0.140652135014534,
-0.057629913091659546,
-0.07498050481081009,
0.0057674008421599865,
-0.09328256547451019,
0.000979967531748116,
0.02212400548160076,
0.05972207337617874,
-0.02641250193119049,
0.15960755944252014,
-0.26378121972084045,
-0.08369841426610947,
0.13764065504074097,
0.006841121707111597,
-0.025919517502188683,
0.041076477617025375,
-0.018193626776337624,
-0.02810429222881794,
0.053398460149765015,
0.014392700046300888,
0.035647787153720856,
0.0385282002389431,
-0.03911355510354042,
-0.039136555045843124,
-0.05565799027681351,
-0.004775478504598141,
0.0007159403758123517,
0.008302449248731136,
0.09250792115926743,
0.021341122686862946,
-0.005673958454281092,
-0.03481858968734741,
0.14640982449054718,
-0.052672240883111954,
-0.06332338601350784,
-0.12540669739246368,
0.15395031869411469,
0.036194197833538055,
0.07027237862348557,
0.0035699396394193172,
-0.11412621289491653,
0.02314097248017788,
0.010996471159160137,
0.20751145482063293,
0.049539897590875626,
0.020889591425657272,
0.01681961677968502,
0.010770699940621853,
0.004453964531421661,
0.04339643940329552,
-0.013750729151070118,
0.22640033066272736,
-0.028032921254634857,
0.195931077003479,
-0.041006069630384445,
-0.04479644447565079,
0.0074816918931901455,
0.13494504988193512,
-0.013166988268494606,
-0.0262567438185215,
-0.05583420768380165,
0.13294793665409088,
-0.030791496858000755,
-0.19458989799022675,
0.014874827116727829,
-0.11037062108516693,
-0.14163967967033386,
-0.04084489494562149,
-0.01086729671806097,
0.09451998025178909,
0.124494768679142,
-0.010269269347190857,
-0.036028288304805756,
0.17765532433986664,
0.00209207646548748,
-0.03856346011161804,
-0.14158591628074646,
0.09252429008483887,
0.07367589324712753,
0.0981963500380516,
0.031007535755634308,
0.07018786668777466,
0.09927753359079361,
-0.017578229308128357,
-0.10750105232000351,
-0.00716734491288662,
0.0286480113863945,
-0.11748738586902618,
-0.006254499312490225,
0.15853746235370636,
-0.026517517864704132,
0.09956485033035278,
0.0520283542573452,
-0.0805673748254776,
0.04410067945718765,
0.11068209260702133,
-0.02493232861161232,
-0.08788687735795975,
0.17508861422538757,
-0.10910744220018387,
0.1390254944562912,
0.23879076540470123,
-0.015911605209112167,
0.03208133950829506,
-0.05106049031019211,
0.0694429874420166,
0.07210544496774673,
0.10063767433166504,
-0.002433470916002989,
-0.1457262486219406,
0.04079053923487663,
-0.02662983536720276,
0.03632882237434387,
-0.1207178384065628,
-0.10731413960456848,
0.0219006035476923,
0.022997351363301277,
0.0005509696784429252,
0.11006809771060944,
0.06640954315662384,
-0.012142610736191273,
-0.0036820643581449986,
-0.14573736488819122,
-0.03166817128658295,
0.057758916169404984,
-0.04269591346383095,
-0.03693113103508949
] |
null | null | diffusers |
# SDXL LoRA DreamBooth - linoyts/ad_huggy
<Gallery />
## Model description
### These are linoyts/ad_huggy LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
## Download model
### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke
- **LoRA**: download **[`ad_huggy.safetensors` here 💾](/linoyts/ad_huggy/blob/main/ad_huggy.safetensors)**.
- Place it on your `models/Lora` folder.
- On AUTOMATIC1111, load the LoRA by adding `<lora:ad_huggy:1>` to your prompt. On ComfyUI just [load it as a regular LoRA](https://comfyanonymous.github.io/ComfyUI_examples/lora/).
- *Embeddings*: download **[`ad_huggy_emb.safetensors` here 💾](/linoyts/ad_huggy/blob/main/ad_huggy_emb.safetensors)**.
- Place it on it on your `embeddings` folder
- Use it by adding `ad_huggy_emb` to your prompt. For example, `a ad_huggy_emb ad for lego with a emoji`
(you need both the LoRA and the embeddings as they were trained together for this LoRA)
## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
```py
from diffusers import AutoPipelineForText2Image
import torch
from huggingface_hub import hf_hub_download
from safetensors.torch import load_file
pipeline = AutoPipelineForText2Image.from_pretrained('stabilityai/stable-diffusion-xl-base-1.0', torch_dtype=torch.float16).to('cuda')
pipeline.load_lora_weights('linoyts/ad_huggy', weight_name='pytorch_lora_weights.safetensors')
embedding_path = hf_hub_download(repo_id='linoyts/ad_huggy', filename='ad_huggy_emb.safetensors', repo_type="model")
state_dict = load_file(embedding_path)
pipeline.load_textual_inversion(state_dict["clip_l"], token=["<s0>", "<s1>", "<s2>", "<s3>"], text_encoder=pipeline.text_encoder, tokenizer=pipeline.tokenizer)
pipeline.load_textual_inversion(state_dict["clip_g"], token=["<s0>", "<s1>", "<s2>", "<s3>"], text_encoder=pipeline.text_encoder_2, tokenizer=pipeline.tokenizer_2)
image = pipeline('a <s0><s1> ad for lego with a <s2><s3> emoji').images[0]
```
For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
## Trigger words
To trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:
to trigger concept `TOK` → use `<s0><s1>` in your prompt
to trigger concept `T2K` → use `<s2><s3>` in your prompt
## Details
All [Files & versions](/linoyts/ad_huggy/tree/main).
The weights were trained using [🧨 diffusers Advanced Dreambooth Training Script](https://github.com/huggingface/diffusers/blob/main/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py).
LoRA for the text encoder was enabled. False.
Pivotal tuning was enabled: True.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
| {"license": "openrail++", "tags": ["stable-diffusion-xl", "stable-diffusion-xl-diffusers", "text-to-image", "diffusers", "lora", "template:sd-lora"], "widget": [{"text": "a <s0><s1> ad for lego with a <s2><s3> emoji", "output": {"url": "image_0.png"}}, {"text": "a <s0><s1> ad for lego with a <s2><s3> emoji", "output": {"url": "image_1.png"}}, {"text": "a <s0><s1> ad for lego with a <s2><s3> emoji", "output": {"url": "image_2.png"}}, {"text": "a <s0><s1> ad for lego with a <s2><s3> emoji", "output": {"url": "image_3.png"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "a <s0><s1> ad for lego with a <s2><s3> emoji"} | text-to-image | linoyts/ad_huggy | [
"diffusers",
"stable-diffusion-xl",
"stable-diffusion-xl-diffusers",
"text-to-image",
"lora",
"template:sd-lora",
"base_model:stabilityai/stable-diffusion-xl-base-1.0",
"license:openrail++",
"region:us"
] | 2024-02-08T17:18:55+00:00 | [] | [] | TAGS
#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #region-us
|
# SDXL LoRA DreamBooth - linoyts/ad_huggy
<Gallery />
## Model description
### These are linoyts/ad_huggy LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.
## Download model
### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke
- LoRA: download 'ad_huggy.safetensors' here .
- Place it on your 'models/Lora' folder.
- On AUTOMATIC1111, load the LoRA by adding '<lora:ad_huggy:1>' to your prompt. On ComfyUI just load it as a regular LoRA.
- *Embeddings*: download 'ad_huggy_emb.safetensors' here .
- Place it on it on your 'embeddings' folder
- Use it by adding 'ad_huggy_emb' to your prompt. For example, 'a ad_huggy_emb ad for lego with a emoji'
(you need both the LoRA and the embeddings as they were trained together for this LoRA)
## Use it with the diffusers library
For more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers
## Trigger words
To trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:
to trigger concept 'TOK' → use '<s0><s1>' in your prompt
to trigger concept 'T2K' → use '<s2><s3>' in your prompt
## Details
All Files & versions.
The weights were trained using diffusers Advanced Dreambooth Training Script.
LoRA for the text encoder was enabled. False.
Pivotal tuning was enabled: True.
Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
| [
"# SDXL LoRA DreamBooth - linoyts/ad_huggy\n\n<Gallery />",
"## Model description",
"### These are linoyts/ad_huggy LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.",
"## Download model",
"### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke\n\n- LoRA: download 'ad_huggy.safetensors' here .\n - Place it on your 'models/Lora' folder.\n - On AUTOMATIC1111, load the LoRA by adding '<lora:ad_huggy:1>' to your prompt. On ComfyUI just load it as a regular LoRA.\n- *Embeddings*: download 'ad_huggy_emb.safetensors' here .\n - Place it on it on your 'embeddings' folder\n - Use it by adding 'ad_huggy_emb' to your prompt. For example, 'a ad_huggy_emb ad for lego with a emoji'\n (you need both the LoRA and the embeddings as they were trained together for this LoRA)",
"## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers",
"## Trigger words\n\nTo trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:\n\nto trigger concept 'TOK' → use '<s0><s1>' in your prompt \n\n\nto trigger concept 'T2K' → use '<s2><s3>' in your prompt",
"## Details\nAll Files & versions.\n\nThe weights were trained using diffusers Advanced Dreambooth Training Script.\n\nLoRA for the text encoder was enabled. False.\n\nPivotal tuning was enabled: True.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix."
] | [
"TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #region-us \n",
"# SDXL LoRA DreamBooth - linoyts/ad_huggy\n\n<Gallery />",
"## Model description",
"### These are linoyts/ad_huggy LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.",
"## Download model",
"### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke\n\n- LoRA: download 'ad_huggy.safetensors' here .\n - Place it on your 'models/Lora' folder.\n - On AUTOMATIC1111, load the LoRA by adding '<lora:ad_huggy:1>' to your prompt. On ComfyUI just load it as a regular LoRA.\n- *Embeddings*: download 'ad_huggy_emb.safetensors' here .\n - Place it on it on your 'embeddings' folder\n - Use it by adding 'ad_huggy_emb' to your prompt. For example, 'a ad_huggy_emb ad for lego with a emoji'\n (you need both the LoRA and the embeddings as they were trained together for this LoRA)",
"## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers",
"## Trigger words\n\nTo trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:\n\nto trigger concept 'TOK' → use '<s0><s1>' in your prompt \n\n\nto trigger concept 'T2K' → use '<s2><s3>' in your prompt",
"## Details\nAll Files & versions.\n\nThe weights were trained using diffusers Advanced Dreambooth Training Script.\n\nLoRA for the text encoder was enabled. False.\n\nPivotal tuning was enabled: True.\n\nSpecial VAE used for training: madebyollin/sdxl-vae-fp16-fix."
] | [
78,
23,
3,
37,
3,
202,
38,
76,
74
] | [
"passage: TAGS\n#diffusers #stable-diffusion-xl #stable-diffusion-xl-diffusers #text-to-image #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #license-openrail++ #region-us \n# SDXL LoRA DreamBooth - linoyts/ad_huggy\n\n<Gallery />## Model description### These are linoyts/ad_huggy LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0.## Download model### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke\n\n- LoRA: download 'ad_huggy.safetensors' here .\n - Place it on your 'models/Lora' folder.\n - On AUTOMATIC1111, load the LoRA by adding '<lora:ad_huggy:1>' to your prompt. On ComfyUI just load it as a regular LoRA.\n- *Embeddings*: download 'ad_huggy_emb.safetensors' here .\n - Place it on it on your 'embeddings' folder\n - Use it by adding 'ad_huggy_emb' to your prompt. For example, 'a ad_huggy_emb ad for lego with a emoji'\n (you need both the LoRA and the embeddings as they were trained together for this LoRA)## Use it with the diffusers library\n\n\n\nFor more details, including weighting, merging and fusing LoRAs, check the documentation on loading LoRAs in diffusers## Trigger words\n\nTo trigger image generation of trained concept(or concepts) replace each concept identifier in you prompt with the new inserted tokens:\n\nto trigger concept 'TOK' → use '<s0><s1>' in your prompt \n\n\nto trigger concept 'T2K' → use '<s2><s3>' in your prompt"
] | [
-0.008388958871364594,
0.05098827928304672,
-0.008621140383183956,
-0.0020394790917634964,
0.12760382890701294,
0.0355568490922451,
0.10871229320764542,
0.09044905006885529,
0.10360056161880493,
0.08923699706792831,
-0.022486036643385887,
0.12150032073259354,
0.07868421077728271,
0.14233216643333435,
0.026823656633496284,
-0.18206831812858582,
0.0077630216255784035,
-0.0766148567199707,
0.037268009036779404,
0.039503347128629684,
0.07829862087965012,
-0.04904525727033615,
0.12395378202199936,
-0.015162469819188118,
-0.07234439998865128,
0.02173534780740738,
-0.05089547112584114,
-0.010172574780881405,
0.034590043127536774,
0.04146037995815277,
0.0711791068315506,
0.007190489675849676,
0.06302487105131149,
-0.20753298699855804,
0.015887729823589325,
0.06836804002523422,
-0.0094569968059659,
0.04403550550341606,
0.10004017502069473,
-0.0430985763669014,
0.11126606166362762,
-0.12879805266857147,
0.06612216681241989,
0.0759430080652237,
-0.09580074995756149,
0.014215179719030857,
-0.11809446662664413,
0.02732805348932743,
0.1604384034872055,
0.04065684601664543,
0.0026694422122091055,
-0.03937496244907379,
0.009269775822758675,
0.031015412881970406,
0.21738018095493317,
-0.1211257055401802,
-0.008422273211181164,
0.0748506486415863,
0.038040995597839355,
-0.00038532231701537967,
-0.061440739780664444,
0.02334524691104889,
-0.030699484050273895,
-0.01875458098948002,
0.028361279517412186,
-0.06085031107068062,
-0.007978463545441628,
-0.04739995300769806,
-0.10433641076087952,
-0.005224193911999464,
0.16538745164871216,
-0.00022400720627047122,
-0.05654698610305786,
-0.14530712366104126,
-0.06707334518432617,
0.053417664021253586,
-0.03568464517593384,
-0.05335496738553047,
0.049837302416563034,
-0.00480792298913002,
0.1102360263466835,
-0.18133136630058289,
-0.06832170486450195,
-0.00200093281455338,
-0.004960773978382349,
0.06597874313592911,
0.007612503599375486,
-0.020060600712895393,
0.03603404015302658,
0.10703271627426147,
-0.01881175860762596,
-0.07717631757259369,
-0.03063741698861122,
-0.021433258429169655,
-0.14885194599628448,
-0.03563527762889862,
-0.027788864448666573,
-0.07068569958209991,
0.03990396857261658,
0.2306145578622818,
0.16694292426109314,
0.056920792907476425,
-0.09838604927062988,
0.05181160196661949,
0.0013261238345876336,
0.15693829953670502,
-0.009586341679096222,
-0.11223089694976807,
0.03829473629593849,
0.02709387056529522,
0.05664278566837311,
-0.02754714898765087,
-0.08427917212247849,
-0.034056827425956726,
-0.05052628740668297,
0.07841058075428009,
0.04856737330555916,
0.0547192357480526,
-0.07883105427026749,
-0.05450708419084549,
0.13373933732509613,
-0.1450454294681549,
0.09130983054637909,
-0.014212798327207565,
-0.030008336529135704,
0.09177432954311371,
0.12832596898078918,
0.000979574746452272,
-0.03105919063091278,
0.11729303747415543,
-0.05019373074173927,
0.05052143335342407,
-0.10214721411466599,
-0.042854875326156616,
0.022214531898498535,
-0.12421499937772751,
-0.022342104464769363,
-0.09140386432409286,
-0.16780810058116913,
-0.05158400535583496,
0.05241483822464943,
-0.046111393719911575,
-0.010998601093888283,
-0.0706368014216423,
-0.03386177122592926,
-0.024667374789714813,
0.028078626841306686,
0.039671458303928375,
0.016296347603201866,
0.021233871579170227,
-0.032234206795692444,
0.05246957391500473,
0.005848803091794252,
0.024528147652745247,
-0.019558021798729897,
0.06121312826871872,
-0.2496574968099594,
0.1626916229724884,
-0.053527865558862686,
0.015994247049093246,
-0.09918659925460815,
0.052994586527347565,
0.02922271378338337,
0.005678631365299225,
0.015174677595496178,
0.09613867104053497,
-0.20554910600185394,
-0.03795219957828522,
0.08405539393424988,
-0.07038582116365433,
-0.0279928520321846,
0.05578742176294327,
-0.0030116266570985317,
0.10885322093963623,
0.0671677216887474,
0.11983571946620941,
0.12052002549171448,
-0.2155068963766098,
-0.03914128988981247,
-0.054001301527023315,
-0.10700090229511261,
0.05001493915915489,
0.005490806419402361,
-0.0030334487091749907,
0.08278322219848633,
0.05241617560386658,
-0.07342393696308136,
0.026348961517214775,
0.049639541655778885,
-0.009746191091835499,
0.005179445259273052,
-0.02798689343035221,
-0.053494177758693695,
-0.0859917402267456,
-0.07630737125873566,
0.0469619557261467,
-0.07247032225131989,
0.029835479333996773,
0.09452234208583832,
0.023478791117668152,
0.06433402001857758,
-0.059616949409246445,
0.06742305308580399,
-0.015015031211078167,
0.018916955217719078,
-0.09262771904468536,
-0.11172585189342499,
0.027583295479416847,
0.03708643838763237,
0.09829345345497131,
-0.047784022986888885,
0.02700801193714142,
0.04882371425628662,
-0.0026684324257075787,
-0.029706407338380814,
0.027079612016677856,
-0.06718437373638153,
-0.03952683135867119,
-0.04861108958721161,
-0.07764138281345367,
-0.03135402873158455,
0.14417767524719238,
-0.01043914444744587,
0.05111498758196831,
-0.0261683389544487,
0.12276247888803482,
0.04577350988984108,
-0.06617185473442078,
-0.0033234080765396357,
-0.05379567667841911,
0.0006289434968493879,
-0.0378454215824604,
-0.012044243514537811,
0.01243703905493021,
-0.07022376358509064,
0.04889567941427231,
-0.15344667434692383,
-0.14474746584892273,
0.07910965383052826,
0.07260912656784058,
-0.05966711416840553,
-0.13838796317577362,
0.006553641054779291,
-0.006235803477466106,
-0.06118045002222061,
-0.1106037124991417,
0.1385449916124344,
0.08353908360004425,
0.05904576554894447,
-0.015883179381489754,
-0.03671596944332123,
-0.04715964198112488,
-0.010376516729593277,
-0.0486609973013401,
0.03146500140428543,
0.08973323553800583,
0.003071012906730175,
0.03275807946920395,
0.03005228005349636,
-0.01783578470349312,
0.01605216972529888,
0.05162316933274269,
-0.07229440659284592,
-0.049742456525564194,
-0.0013569581788033247,
0.0433037094771862,
-0.004445475991815329,
0.06277339160442352,
0.0599602647125721,
0.055825673043727875,
-0.03626495599746704,
0.014455003663897514,
-0.045751411467790604,
0.020195823162794113,
0.010122228413820267,
-0.01968240737915039,
0.10912898927927017,
0.06433811038732529,
0.0351560115814209,
0.03814991936087608,
-0.040482811629772186,
0.08446941524744034,
-0.03880548104643822,
-0.022783100605010986,
-0.05742660164833069,
0.009280761703848839,
-0.12349782884120941,
-0.19317539036273956,
-0.19109372794628143,
-0.09896877408027649,
-0.06638408452272415,
0.023233111947774887,
0.05122305080294609,
-0.0061772894114255905,
-0.07371727377176285,
-0.045184727758169174,
0.10989510267972946,
0.0690358430147171,
-0.07136686891317368,
-0.04254470393061638,
0.038219962269067764,
0.02973722293972969,
-0.07013632357120514,
-0.01818026416003704,
0.031363680958747864,
-0.06891236454248428,
0.02899305522441864,
0.08850720524787903,
0.09786820411682129,
0.01653449796140194,
0.04122067615389824,
0.016961151733994484,
0.03588255122303963,
0.1589617282152176,
-0.07386905699968338,
0.16402067244052887,
0.16322742402553558,
-0.04139312729239464,
0.09505774825811386,
0.14204250276088715,
-0.002039120765402913,
-0.06760101765394211,
0.005637960974127054,
0.07631782442331314,
-0.014959564432501793,
-0.1410628855228424,
-0.051046740263700485,
-0.027400536462664604,
-0.05337853357195854,
0.09411375224590302,
0.05417769029736519,
0.057763054966926575,
0.07200458645820618,
-0.04360738396644592,
-0.017335597425699234,
0.05693306028842926,
0.12028222531080246,
0.06651627272367477,
-0.040449272841215134,
0.02149595506489277,
-0.01719580590724945,
-0.03408605605363846,
0.04846308007836342,
0.04959796369075775,
0.14746099710464478,
-0.01722862385213375,
0.0741945281624794,
-0.003115422325208783,
-0.008724885061383247,
-0.044634636491537094,
0.06671641021966934,
0.005948462523519993,
0.04113604128360748,
-0.019688190892338753,
-0.060862962156534195,
-0.01990959607064724,
0.12070327252149582,
0.10409539192914963,
-0.04488341137766838,
0.04658525437116623,
0.03401680290699005,
0.09160400182008743,
0.12399082630872726,
-0.05261678248643875,
-0.191439688205719,
0.09718769788742065,
0.04186384752392769,
-0.01100841723382473,
-0.0630391538143158,
-0.0389895886182785,
0.04317416995763779,
-0.07837215811014175,
0.07363582402467728,
-0.020074328407645226,
0.07437968999147415,
-0.05773831531405449,
-0.04290398955345154,
0.02128070779144764,
0.2100924402475357,
0.015544363297522068,
0.027352312579751015,
-0.19298866391181946,
-0.04960789903998375,
0.02834196947515011,
0.025801757350564003,
-0.0459161102771759,
0.06288522481918335,
0.03155958652496338,
-0.0037462199106812477,
0.1201225221157074,
-0.012654697522521019,
-0.0700039267539978,
-0.08282061666250229,
-0.128550186753273,
-0.030090676620602608,
0.08011133223772049,
-0.12074185162782669,
0.08844242990016937,
-0.04183013737201691,
-0.07114417105913162,
-0.05327332392334938,
0.0312669463455677,
-0.10752987116575241,
-0.20767496526241302,
0.038206059485673904,
0.02460496686398983,
0.08921127766370773,
-0.012670913711190224,
0.04415881633758545,
-0.042562197893857956,
0.12441101670265198,
-0.030099239200353622,
-0.07763653248548508,
-0.0918530821800232,
-0.08110818266868591,
0.12692686915397644,
-0.0483134426176548,
0.006800375878810883,
0.0004707910702563822,
0.11104445159435272,
-0.10297922790050507,
-0.08440857380628586,
-0.0039504533633589745,
-0.058910902589559555,
-0.028957482427358627,
-0.036435093730688095,
0.1252201944589615,
0.03672696277499199,
0.0029385294765233994,
-0.005711948033422232,
0.029157795011997223,
0.04591565579175949,
-0.08121728152036667,
0.02037467062473297,
0.23854559659957886,
0.03120795451104641,
0.06399405747652054,
-0.140055313706398,
-0.043178312480449677,
-0.12557797133922577,
0.03729284927248955,
0.0976153016090393,
0.2375042587518692,
-0.03801827132701874,
0.10816369205713272,
-0.03842300549149513,
-0.1102772131562233,
-0.1348455548286438,
-0.034485310316085815,
0.03589194640517235,
-0.024397268891334534,
0.006376086268573999,
-0.1715545654296875,
0.08061347901821136,
0.0809304416179657,
-0.007015895564109087,
0.18147416412830353,
-0.2523373067378998,
-0.11418773233890533,
-0.0033839470706880093,
-0.019783122465014458,
-0.13922148942947388,
-0.10543642938137054,
-0.06230231374502182,
-0.07587350159883499,
0.0456555113196373,
0.09038420766592026,
0.09450937062501907,
0.06252852082252502,
0.01646622270345688,
0.09788200259208679,
0.08553461730480194,
-0.029757900163531303,
0.11446511000394821,
-0.013266244903206825,
0.051178161054849625,
-0.0534089095890522,
-0.06239093467593193,
0.02975514344871044,
-0.0889042392373085,
0.11401436477899551,
-0.06989817321300507,
-0.03024718165397644,
-0.018202148377895355,
-0.01290708128362894,
-0.018808940425515175,
0.08155430108308792,
-0.04203943908214569,
-0.007689942140132189,
-0.05547254905104637,
0.02460382878780365,
0.0773325189948082,
-0.005392476916313171,
-0.07635752111673355,
-0.06002272292971611,
-0.08834460377693176,
0.1355927586555481,
-0.006452007684856653,
0.08087461441755295,
-0.08103152364492416,
0.004613260272890329,
0.002136258175596595,
0.013555091805756092,
0.004693863447755575,
0.0342639721930027,
0.06245368346571922,
-0.03884464129805565,
0.13675910234451294,
-0.002066110959276557,
-0.0975823774933815,
-0.012503236532211304,
0.0665840357542038,
-0.08905868232250214,
-0.1023356094956398,
-0.02803577110171318,
0.05905649811029434,
-0.08180370926856995,
-0.10541683435440063,
0.11662084609270096,
0.018366917967796326,
-0.051133256405591965,
0.041249100118875504,
0.09164243936538696,
-0.032047707587480545,
0.014757429249584675,
-0.042540308088064194,
-0.007934867404401302,
-0.06007932871580124,
0.09950960427522659,
0.08117649704217911,
-0.0833432674407959,
0.00462412741035223,
0.13696952164173126,
-0.024209637194871902,
-0.05129070207476616,
-0.06910406053066254,
0.138264462351799,
-0.024156495928764343,
0.029666544869542122,
0.037857167422771454,
-0.035518672317266464,
0.016750363633036613,
0.04426195099949837,
0.003639448434114456,
-0.02740926295518875,
0.009155087172985077,
0.021612506359815598,
-0.06688365340232849,
0.07738392055034637,
0.05111559107899666,
0.06669984757900238,
-0.11537006497383118,
0.04777836799621582,
-0.01406325027346611,
0.0003734457422979176,
0.0005652242107316852,
0.0011940408730879426,
-0.05107022821903229,
-0.03844855725765228,
-0.04870833456516266,
0.0656391903758049,
-0.10067948698997498,
-0.0015535525744780898,
0.047470252960920334,
0.02273043803870678,
0.026979634538292885,
-0.018902329728007317,
-0.05800816789269447,
-0.1188020333647728,
-0.04422997683286667,
0.08274698257446289,
-0.1196863204240799,
-0.06641024351119995,
0.07357770204544067,
-0.1078723669052124,
0.0507328137755394,
0.04900572821497917,
-0.02620653808116913,
-0.0433780699968338,
-0.04550671577453613,
-0.030210282653570175,
0.05766608938574791,
0.027915168553590775,
0.04235079512000084,
-0.13638439774513245,
-0.00038817664608359337,
-0.06053714454174042,
-0.026424473151564598,
-0.0752510204911232,
0.08531761914491653,
-0.10531625896692276,
0.10187449306249619,
0.002975120674818754,
-0.08199131488800049,
-0.07876613736152649,
0.006409937050193548,
0.09803605079650879,
0.04366563633084297,
0.0943572148680687,
-0.05623616650700569,
0.10681744664907455,
-0.0787796899676323,
0.017641836777329445,
0.025108475238084793,
-0.001814071787521243,
0.06662370264530182,
-0.11361684650182724,
0.02805311232805252,
-0.008566481992602348,
0.03554137051105499,
-0.020453326404094696,
-0.009771902114152908,
0.0010647652670741081,
0.035993486642837524,
-0.07878022640943527,
0.01815205253660679,
-0.02098902128636837,
-0.011309471912682056,
0.014085470698773861,
-0.013892004266381264,
-0.007219961378723383,
0.04440758749842644,
-0.07551359385251999,
0.053183313459157944,
0.11037294566631317,
0.04666169732809067,
0.02717561274766922,
0.01973959244787693,
-0.08731583505868912,
-0.06435820460319519,
0.07828717678785324,
-0.0631629154086113,
-0.006068026181310415,
-0.08034326881170273,
0.14074431359767914,
0.06238994374871254,
-0.15108446776866913,
0.10006818175315857,
0.08208831399679184,
-0.02134021185338497,
-0.027017299085855484,
-0.19297175109386444,
-0.03998265415430069,
-0.052221618592739105,
0.001354519627057016,
-0.07004232704639435,
0.07227680832147598,
-0.044470224529504776,
-0.03933628275990486,
0.016605887562036514,
0.08844242990016937,
-0.11505013704299927,
-0.06684698909521103,
0.019385000690817833,
0.011409690603613853,
0.0016705524176359177,
0.034622352570295334,
-0.02779647335410118,
0.02028960920870304,
0.07893862575292587,
0.049816206097602844,
0.06373615562915802,
0.0913950577378273,
0.06795088201761246,
-0.09185504168272018,
-0.040154051035642624,
0.00910441018640995,
-0.004423313774168491,
-0.02457202784717083,
0.08487290143966675,
0.04966196417808533,
-0.04390290752053261,
-0.061497077345848083,
0.12269818037748337,
-0.06156712770462036,
-0.06896410882472992,
-0.1149337887763977,
0.05462678149342537,
-0.019857341423630714,
-0.0008444077684544027,
-0.029930900782346725,
-0.10070967674255371,
-0.06299793720245361,
0.20150119066238403,
0.12349975854158401,
-0.038967523723840714,
0.010201230645179749,
-0.007319250144064426,
0.02684376947581768,
-0.015310695394873619,
-0.006131859961897135,
0.031130217015743256,
0.08789379149675369,
-0.023160358890891075,
0.10390844196081161,
-0.0058818417601287365,
-0.03768076002597809,
-0.04055710881948471,
0.04712770879268646,
-0.12360312044620514,
-0.02256235107779503,
0.001420529093593359,
0.029632359743118286,
-0.11232651770114899,
-0.1881444901227951,
0.049251195043325424,
-0.011608364060521126,
-0.07120151072740555,
0.009048528037965298,
0.044469669461250305,
-0.0053708236664533615,
0.07122671604156494,
0.005603610072284937,
-0.05132272094488144,
0.1999959498643875,
-0.02715989015996456,
-0.09829127788543701,
-0.022630251944065094,
0.030737893655896187,
-0.1097598746418953,
0.15919426083564758,
0.011993317864835262,
-0.03083411604166031,
0.04188051447272301,
0.012289698235690594,
-0.11260586977005005,
0.003617451759055257,
0.016187092289328575,
-0.12532106041908264,
0.06132090091705322,
0.1568097621202469,
-0.034166622906923294,
0.057980261743068695,
0.045162755995988846,
-0.13027407228946686,
0.028806839138269424,
0.060179419815540314,
0.015838734805583954,
-0.052121829241514206,
0.06564659625291824,
-0.12402457743883133,
0.10871879756450653,
0.13111144304275513,
0.029635407030582428,
0.04877539724111557,
-0.04837747663259506,
-0.0007765802438370883,
0.052672602236270905,
0.07419048249721527,
-0.038830094039440155,
-0.040443822741508484,
-0.00847889669239521,
0.030060119926929474,
0.058730583637952805,
-0.09697021543979645,
-0.062002215534448624,
-0.04077679291367531,
-0.038772303611040115,
-0.04628792405128479,
0.09811414778232574,
0.09183192998170853,
0.009388324804604053,
0.01953716389834881,
-0.19026930630207062,
0.06324049830436707,
0.10477165877819061,
-0.1226898655295372,
-0.016583552584052086
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-large-lora-1.57M-squad-model3
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 54
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-large", "model-index": [{"name": "roberta-large-lora-1.57M-squad-model3", "results": []}]} | question-answering | varun-v-rao/roberta-large-lora-1.57M-squad-model3 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-large",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:23:38+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us
|
# roberta-large-lora-1.57M-squad-model3
This model is a fine-tuned version of roberta-large on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 54
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-large-lora-1.57M-squad-model3\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 54\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n",
"# roberta-large-lora-1.57M-squad-model3\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 54\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
70,
41,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n# roberta-large-lora-1.57M-squad-model3\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 54\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.08723487704992294,
0.11848623305559158,
-0.002330329967662692,
0.09665360301733017,
0.1480117291212082,
0.008159030228853226,
0.12183629721403122,
0.12074936181306839,
-0.0941850021481514,
0.06440624594688416,
0.07542012631893158,
0.04690675809979439,
0.024355873465538025,
0.10552429407835007,
-0.028213299810886383,
-0.21728159487247467,
-0.008100204169750214,
-0.019315555691719055,
-0.09385199844837189,
0.10033246129751205,
0.1062641590833664,
-0.10441606491804123,
0.08152083307504654,
-0.01794813573360443,
-0.15305334329605103,
0.05398956686258316,
-0.01791922003030777,
-0.03722402825951576,
0.09953777492046356,
0.03427852690219879,
0.09768839180469513,
0.0051507242023944855,
0.140668585896492,
-0.21658502519130707,
0.00871315784752369,
0.06885267049074173,
0.028365952894091606,
0.07459409534931183,
0.02645678073167801,
-0.016260528936982155,
0.08632846921682358,
-0.16725382208824158,
0.0981488898396492,
0.028692856431007385,
-0.0760069414973259,
-0.16955862939357758,
-0.09386542439460754,
0.05209092050790787,
0.08322618156671524,
0.1015634760260582,
0.003946835640817881,
0.15983565151691437,
-0.08208324015140533,
0.08630828559398651,
0.23515327274799347,
-0.2684452533721924,
-0.06645257025957108,
0.08176656067371368,
0.07822448015213013,
0.10199644416570663,
-0.12546691298484802,
-0.0028978779446333647,
0.05621766671538353,
0.009555247612297535,
0.0824074000120163,
-0.03932909667491913,
-0.09320280700922012,
0.01888042502105236,
-0.13292759656906128,
-0.0004583474656101316,
0.18232625722885132,
0.057063028216362,
-0.04076538234949112,
-0.07414773106575012,
-0.03849441185593605,
-0.04780465364456177,
-0.02533808723092079,
-0.035391438752412796,
0.05013807862997055,
-0.06642741709947586,
-0.09611935913562775,
-0.06867387890815735,
-0.08709733188152313,
-0.08874750137329102,
0.0043680137023329735,
0.10296203941106796,
0.04386656731367111,
0.010357538238167763,
-0.04326234385371208,
0.08042441308498383,
-0.031077643856406212,
-0.1031002476811409,
-0.03424321487545967,
0.005861229728907347,
-0.0838521346449852,
-0.05301862582564354,
-0.01875896006822586,
-0.03257352486252785,
0.03417516127228737,
0.1729176938533783,
-0.0542578399181366,
0.05117193982005119,
0.01904754340648651,
0.006439373828470707,
-0.025539472699165344,
0.14350903034210205,
-0.06769806146621704,
-0.06132764741778374,
0.02095918171107769,
0.08139073103666306,
0.02197854220867157,
0.006723546888679266,
-0.09068549424409866,
-0.026258559897542,
0.08373827487230301,
0.08581463992595673,
-0.033340632915496826,
0.032034534960985184,
0.004901669453829527,
-0.008215976879000664,
0.02131820097565651,
-0.1310933530330658,
0.038028065115213394,
-0.01506910752505064,
-0.06371065229177475,
-0.055199358612298965,
0.03517642989754677,
-0.00910269282758236,
0.009093335829675198,
0.053791046142578125,
-0.08660799264907837,
-0.010008091107010841,
-0.07994072884321213,
-0.0909738540649414,
0.01521262712776661,
-0.06478468328714371,
0.01423997338861227,
-0.084988072514534,
-0.1948385089635849,
-0.04715370759367943,
0.00977391842752695,
-0.061113834381103516,
-0.016906024888157845,
-0.03786537051200867,
-0.07153264433145523,
-0.015567747876048088,
0.002161978278309107,
0.13150441646575928,
-0.03779544681310654,
0.08660416305065155,
0.024978896602988243,
0.03249402344226837,
0.011691357009112835,
0.03201432153582573,
-0.09999798238277435,
0.02856188639998436,
-0.14051875472068787,
0.044169310480356216,
-0.10176753252744675,
0.06028388440608978,
-0.12649038434028625,
-0.09632738679647446,
-0.00026857820921577513,
-0.007443604990839958,
0.04992074519395828,
0.10984978824853897,
-0.1467106193304062,
-0.018122538924217224,
0.16224615275859833,
-0.07810654491186142,
-0.09251769632101059,
0.09843014180660248,
-0.05405588820576668,
0.05636819079518318,
0.07830217480659485,
0.1689717024564743,
0.11922147870063782,
-0.1298675686120987,
-0.012129820883274078,
0.0008877287618815899,
0.050753604620695114,
-0.016382666304707527,
0.06352370977401733,
-0.001394764636643231,
0.023987088352441788,
0.01013669278472662,
-0.08033031970262527,
0.0074962712824344635,
-0.08923488110303879,
-0.07756935060024261,
-0.04688841849565506,
-0.11183606088161469,
0.01173105277121067,
0.0461898073554039,
0.028642714023590088,
-0.08201907575130463,
-0.09965582937002182,
0.1055329442024231,
0.13573572039604187,
-0.05533888190984726,
0.0035610487684607506,
-0.06674632430076599,
0.04474201798439026,
-0.0687284916639328,
-0.03277777507901192,
-0.17908954620361328,
-0.12690679728984833,
0.01650252938270569,
-0.041130587458610535,
0.05419675633311272,
0.029642542824149132,
0.07734718173742294,
0.05740872025489807,
-0.06282462924718857,
-0.010924719274044037,
-0.05928923189640045,
0.003557767951861024,
-0.08916333317756653,
-0.19243431091308594,
-0.04355406388640404,
-0.03707696497440338,
0.13248707354068756,
-0.26030412316322327,
0.021341457962989807,
-0.03341861441731453,
0.10793420672416687,
0.03527791425585747,
-0.038394778966903687,
0.004687850829213858,
0.034122034907341,
-0.008558563888072968,
-0.09341869503259659,
0.023694708943367004,
-0.01653197593986988,
-0.0753956213593483,
-0.05513985827565193,
-0.14202652871608734,
0.07465559244155884,
0.06847428530454636,
0.07148012518882751,
-0.10938934981822968,
0.004916963633149862,
-0.0549774207174778,
-0.03419763222336769,
-0.07508955150842667,
-0.014269382692873478,
0.1210692897439003,
-0.0018772846087813377,
0.11745250970125198,
-0.06274837255477905,
-0.06597740948200226,
0.00020416387997101992,
-0.018783770501613617,
0.004218939691781998,
0.10808371007442474,
0.08145362138748169,
-0.14009921252727509,
0.09619587659835815,
0.07626135647296906,
-0.06436627358198166,
0.14564163982868195,
-0.05163373053073883,
-0.08818042278289795,
-0.035746246576309204,
0.02584950625896454,
0.0023365893866866827,
0.13697317242622375,
-0.05935881286859512,
-0.006850571371614933,
0.022225521504878998,
-0.005602737423032522,
0.016872592270374298,
-0.1569286584854126,
-0.033165354281663895,
0.023126574233174324,
-0.04511022940278053,
0.00930491741746664,
-0.0112262312322855,
0.02080562524497509,
0.09204436093568802,
0.012303276918828487,
-0.02348015084862709,
0.006428664084523916,
-0.008208684623241425,
-0.07754236459732056,
0.19533683359622955,
-0.07154954224824905,
-0.11887532472610474,
-0.13331133127212524,
0.027079379186034203,
-0.03921687230467796,
-0.029738500714302063,
0.024945417419075966,
-0.0883062481880188,
-0.054149314761161804,
-0.10122647881507874,
0.017414836212992668,
-0.029011063277721405,
-0.01911609247326851,
-0.0006714147748425603,
0.039272841066122055,
0.07951072603464127,
-0.1431923657655716,
0.025598982349038124,
-0.03329098969697952,
-0.12917345762252808,
-0.015129772946238518,
0.046244677156209946,
0.14519554376602173,
0.0958317443728447,
-0.02015262097120285,
0.00800979882478714,
-0.03534230589866638,
0.20629668235778809,
-0.07254815846681595,
0.011614320799708366,
0.11698288470506668,
0.027192117646336555,
0.0400136262178421,
0.13066232204437256,
0.03496024012565613,
-0.09383195638656616,
0.03787229210138321,
0.08500706404447556,
-0.0326409637928009,
-0.2556629180908203,
-0.029270043596625328,
-0.03261369466781616,
-0.059363506734371185,
0.06910660117864609,
0.06669556349515915,
0.023751376196742058,
0.04820908233523369,
-0.004537143278867006,
0.033408768475055695,
-0.011473706923425198,
0.08850546926259995,
0.09313555061817169,
0.028831373900175095,
0.10145688056945801,
-0.060502201318740845,
-0.053658075630664825,
0.055200763046741486,
0.017419055104255676,
0.31141602993011475,
-0.015817934647202492,
0.117644764482975,
0.05377558618783951,
0.12184903025627136,
-0.02532692439854145,
0.02641006000339985,
0.0076152426190674305,
-0.012036535888910294,
0.005784369073808193,
-0.05888405442237854,
0.013602430000901222,
0.0313357412815094,
-0.03493839129805565,
0.04819232225418091,
-0.06519418954849243,
0.03435451164841652,
0.0228397436439991,
0.21355943381786346,
0.03132062405347824,
-0.2811458110809326,
-0.07664798200130463,
0.024531759321689606,
-0.02627606689929962,
-0.039332225918769836,
0.019282842054963112,
0.13811852037906647,
-0.10685943067073822,
0.034108031541109085,
-0.049880847334861755,
0.08123023808002472,
-0.03536234796047211,
-0.011319435201585293,
0.018928181380033493,
0.12214367836713791,
-0.017325546592473984,
0.09910701960325241,
-0.18316435813903809,
0.23795545101165771,
0.025235824286937714,
0.1107262447476387,
-0.036957766860723495,
0.012404688633978367,
0.007215769495815039,
0.08381049335002899,
0.13301217555999756,
-0.01598561368882656,
-0.03734007850289345,
-0.18098533153533936,
-0.0650978535413742,
0.05001184716820717,
0.11090048402547836,
-0.029157985001802444,
0.10398721694946289,
-0.05722633749246597,
-0.008887311443686485,
0.054563187062740326,
-0.09907914698123932,
-0.15324686467647552,
-0.11943767964839935,
-0.01033758744597435,
0.001200188184157014,
-0.07470975071191788,
-0.0866493210196495,
-0.08410995453596115,
-0.029663054272532463,
0.1674615889787674,
0.0136113241314888,
-0.039453234523534775,
-0.1291828155517578,
0.09335559606552124,
0.1283380687236786,
-0.07754804193973541,
0.009001604281365871,
0.019750099629163742,
0.10613103210926056,
0.03638450801372528,
-0.0665670707821846,
0.05656984820961952,
-0.05704908445477486,
-0.15264742076396942,
-0.04490935057401657,
0.1257970929145813,
0.052411533892154694,
0.04492495208978653,
-0.00011867690045619383,
0.02750784158706665,
0.013142337091267109,
-0.09548404067754745,
0.0004150513559579849,
0.0725681483745575,
0.047485120594501495,
0.05749109387397766,
-0.08796650171279907,
0.018064917996525764,
-0.040228165686130524,
-0.02350793406367302,
0.13240565359592438,
0.21959686279296875,
-0.09180305898189545,
0.06732688844203949,
0.0820348858833313,
-0.08048562705516815,
-0.1916641891002655,
0.07113844901323318,
0.04862693324685097,
0.006117881741374731,
0.07790721207857132,
-0.15468789637088776,
0.12269757688045502,
0.1116609126329422,
-0.02017929032444954,
0.05937253683805466,
-0.3118264675140381,
-0.12770460546016693,
0.07922015339136124,
0.1308448165655136,
0.01701754331588745,
-0.1647561639547348,
-0.04278319701552391,
-0.022859767079353333,
-0.1350378692150116,
0.06827899813652039,
-0.11549696326255798,
0.10933253914117813,
-0.003688857425004244,
0.0990917757153511,
0.02023009955883026,
-0.04011531174182892,
0.13149215281009674,
0.020505806431174278,
0.10087887197732925,
-0.05621258541941643,
-0.022668495774269104,
0.13546526432037354,
-0.06198202073574066,
0.0711279809474945,
-0.05686686187982559,
0.07425040751695633,
-0.11494181305170059,
-0.036077339202165604,
-0.06949559599161148,
0.06641019880771637,
-0.054745156317949295,
-0.05768771097064018,
-0.07490568608045578,
0.06978710740804672,
0.0277000330388546,
-0.028589241206645966,
0.07820265740156174,
-0.00011021123646060005,
0.09290367364883423,
0.06335572898387909,
0.10028427839279175,
0.0371098667383194,
-0.0858723595738411,
0.008916991762816906,
-0.028845105320215225,
0.050707124173641205,
-0.1443084180355072,
0.024969812482595444,
0.11044222116470337,
0.06086161732673645,
0.1337449550628662,
0.024451330304145813,
-0.058095671236515045,
0.009405716322362423,
0.03777887672185898,
-0.08572280406951904,
-0.2013341188430786,
0.00760243134573102,
-0.06386350095272064,
-0.1542741358280182,
0.04076506942510605,
0.10657301545143127,
-0.06856527924537659,
-0.016505694016814232,
-0.026855014264583588,
0.033756595104932785,
-0.021659672260284424,
0.17712244391441345,
0.08095744252204895,
0.05947958678007126,
-0.07661045342683792,
0.09327719360589981,
0.07976149022579193,
-0.02176636829972267,
0.04155983030796051,
0.07896966487169266,
-0.06689182668924332,
-0.032129064202308655,
0.060706574469804764,
0.2014622837305069,
-0.06539367139339447,
-0.03319782763719559,
-0.10951779782772064,
-0.07737855613231659,
0.02924175001680851,
0.14212597906589508,
0.042541634291410446,
-0.024947911500930786,
-0.0028178261127322912,
0.033213917165994644,
-0.15907353162765503,
0.1158706322312355,
0.03968312218785286,
0.07614144682884216,
-0.14771762490272522,
0.07105731219053268,
0.002747295191511512,
0.031049983575940132,
-0.015308243222534657,
0.03781398758292198,
-0.10667777061462402,
-0.013452893123030663,
-0.13507114350795746,
-0.006237039342522621,
-0.014444579370319843,
0.01465445477515459,
-0.017014672979712486,
-0.0651397854089737,
-0.05836715176701546,
0.05248567461967468,
-0.0622730553150177,
-0.052770089358091354,
0.023064805194735527,
0.06417477130889893,
-0.16630768775939941,
-0.022373758256435394,
0.026948589831590652,
-0.0763041228055954,
0.05899057537317276,
0.028769856318831444,
0.039195407181978226,
0.03192158788442612,
-0.13256412744522095,
0.002972502028569579,
0.021847382187843323,
0.01905086077749729,
0.06366584450006485,
-0.09695152193307877,
-0.012489269487559795,
-0.02662658505141735,
0.07179485261440277,
0.021113233640789986,
0.013960635289549828,
-0.12056398391723633,
0.012976255267858505,
-0.09355766326189041,
-0.05201250687241554,
-0.05298813804984093,
0.0414389930665493,
0.08664804697036743,
0.03690682351589203,
0.14737862348556519,
-0.08606346696615219,
0.04482963681221008,
-0.21781620383262634,
-0.032076191157102585,
-0.005844990722835064,
-0.038164786994457245,
-0.07326668500900269,
-0.034118905663490295,
0.0688665434718132,
-0.05638796463608742,
0.12437904626131058,
0.004826494958251715,
0.08149024844169617,
0.0475492998957634,
-0.03434310480952263,
0.022437646985054016,
0.004996060393750668,
0.17351527512073517,
0.04257943481206894,
-0.019643059000372887,
0.07496637105941772,
-0.011091657914221287,
0.06597168743610382,
0.015957659110426903,
0.16713286936283112,
0.18984253704547882,
-0.03361862525343895,
0.038743820041418076,
0.08857288211584091,
-0.09548298269510269,
-0.09803926199674606,
0.10626181215047836,
-0.0014324303483590484,
0.0674455389380455,
-0.048563092947006226,
0.15762929618358612,
0.15986788272857666,
-0.15726003050804138,
0.03680425509810448,
-0.06017657741904259,
-0.10664806514978409,
-0.10648543387651443,
-0.03341173380613327,
-0.10494240373373032,
-0.1018424704670906,
0.03264811635017395,
-0.1210874542593956,
0.019754910841584206,
0.07814948260784149,
0.007139537949115038,
0.011549912393093109,
0.17409060895442963,
-0.006669220980256796,
0.0321534089744091,
0.029810503125190735,
0.00424826517701149,
-0.010458185337483883,
-0.04437968134880066,
-0.02743460237979889,
0.06990745663642883,
-0.0037690848112106323,
0.05486215278506279,
-0.03507940098643303,
0.0004688649205490947,
0.043214812874794006,
-0.02882491797208786,
-0.07329964637756348,
0.006670345086604357,
0.048397284001111984,
0.04262160509824753,
0.0406421534717083,
0.057703614234924316,
-0.010659290477633476,
-0.035594500601291656,
0.2875901758670807,
-0.0629430040717125,
-0.09575143456459045,
-0.11019080132246017,
0.16694404184818268,
0.026476042345166206,
-0.005412248428910971,
0.03910263627767563,
-0.13317067921161652,
0.02620534971356392,
0.18285296857357025,
0.17066162824630737,
-0.031360261142253876,
-0.007484883069992065,
-0.021655745804309845,
-0.013937423005700111,
-0.05079001188278198,
0.09028784185647964,
0.09890198707580566,
0.05033986270427704,
-0.05627596750855446,
-0.034929800778627396,
-0.007770000956952572,
-0.02261209487915039,
-0.0763300433754921,
0.07041894644498825,
0.029773183166980743,
0.018171384930610657,
-0.036779314279556274,
0.09027045965194702,
0.008521980606019497,
-0.22785498201847076,
0.05880472809076309,
-0.15314428508281708,
-0.1670481115579605,
-0.02727428637444973,
0.09460140764713287,
-0.022617405280470848,
0.041879624128341675,
-0.03205990791320801,
-0.0064866989850997925,
0.10391823947429657,
-0.00802893377840519,
-0.08870135992765427,
-0.1102595180273056,
0.08592838048934937,
-0.04911939427256584,
0.23521827161312103,
-0.01336450595408678,
0.04700358584523201,
0.11762304604053497,
-0.022247860208153725,
-0.14568914473056793,
0.06242163106799126,
0.07273459434509277,
-0.06755782663822174,
0.03911839798092842,
0.15542249381542206,
-0.048386700451374054,
0.11961272358894348,
0.039376262575387955,
-0.10983085632324219,
-0.0068592494353652,
-0.028044233098626137,
-0.026410996913909912,
-0.1050770953297615,
0.00368447694927454,
-0.07847966253757477,
0.15239903330802917,
0.18175670504570007,
-0.03940359130501747,
0.019142143428325653,
-0.07730574905872345,
0.037491340190172195,
0.0877743810415268,
0.07816887646913528,
-0.010281876660883427,
-0.18870747089385986,
0.0145945493131876,
0.009439647197723389,
0.030449895188212395,
-0.29840415716171265,
-0.0790230855345726,
0.034115344285964966,
-0.04020347073674202,
-0.04858698323369026,
0.11076182872056961,
0.11235661804676056,
0.04550366848707199,
-0.035591334104537964,
-0.16197136044502258,
-0.04917515814304352,
0.14919190108776093,
-0.16565340757369995,
-0.05177795886993408
] |
null | null | null | ### My-Pet-RABBIT-XZG Dreambooth model trained by saisindhu24 following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 218U1AO407
Sample pictures of this concept:
.jpg)
| {"license": "creativeml-openrail-m", "tags": ["NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion"]} | text-to-image | saisindhu24/my-pet-rabbit-xzg | [
"NxtWave-GenAI-Webinar",
"text-to-image",
"stable-diffusion",
"license:creativeml-openrail-m",
"region:us"
] | 2024-02-08T17:25:56+00:00 | [] | [] | TAGS
#NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #region-us
| ### My-Pet-RABBIT-XZG Dreambooth model trained by saisindhu24 following the "Build your own Gen AI model" session by NxtWave.
Project Submission Code: 218U1AO407
Sample pictures of this concept:
!0.jpg)
| [
"### My-Pet-RABBIT-XZG Dreambooth model trained by saisindhu24 following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 218U1AO407\n\nSample pictures of this concept:\n\n !0.jpg)"
] | [
"TAGS\n#NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #region-us \n",
"### My-Pet-RABBIT-XZG Dreambooth model trained by saisindhu24 following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 218U1AO407\n\nSample pictures of this concept:\n\n !0.jpg)"
] | [
43,
64
] | [
"passage: TAGS\n#NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #region-us \n### My-Pet-RABBIT-XZG Dreambooth model trained by saisindhu24 following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 218U1AO407\n\nSample pictures of this concept:\n\n !0.jpg)"
] | [
-0.08704815804958344,
0.1853010058403015,
-0.0011079086689278483,
0.035360101610422134,
0.0598352774977684,
0.012332645244896412,
0.22235170006752014,
0.0422673262655735,
0.06587671488523483,
0.029477907344698906,
0.14338578283786774,
0.026908904314041138,
0.044411398470401764,
0.19161687791347504,
0.05376255884766579,
-0.1302003264427185,
0.045132048428058624,
0.06737779825925827,
-0.030510591343045235,
0.02523539774119854,
0.029841968789696693,
-0.06183899939060211,
0.12696312367916107,
-0.014274266548454762,
-0.1970895230770111,
-0.014448966830968857,
-0.08602698892354965,
0.01688074693083763,
0.07506518065929413,
0.06226620823144913,
0.028153330087661743,
0.09136535972356796,
0.01074825320392847,
-0.0473208874464035,
0.04238560423254967,
-0.01600189320743084,
-0.0964093878865242,
0.09184227138757706,
0.07717128843069077,
0.02525024674832821,
0.21602924168109894,
0.18017658591270447,
-0.1555550992488861,
0.03681020438671112,
-0.08411980420351028,
0.018396558240056038,
0.03449534997344017,
0.14031392335891724,
0.06681233644485474,
0.05612368881702423,
-0.012419629842042923,
0.08054812252521515,
0.020274922251701355,
0.10557432472705841,
0.11389362812042236,
-0.18018433451652527,
-0.1101001501083374,
0.20490048825740814,
0.033205002546310425,
0.026497015729546547,
-0.03642534837126732,
0.1148686632514,
0.1181211844086647,
-0.015125218778848648,
-0.03701075166463852,
-0.03980481997132301,
0.026088088750839233,
-0.06492240726947784,
-0.0995611697435379,
0.02782427705824375,
0.2274790108203888,
0.0748768076300621,
-0.006016972009092569,
0.026912469416856766,
-0.095013827085495,
-0.04707726091146469,
-0.03785109147429466,
0.010156390257179737,
-0.045229118317365646,
0.023474056273698807,
-0.025558292865753174,
-0.1118200346827507,
-0.10404632985591888,
-0.06949123740196228,
-0.008858502842485905,
0.14818134903907776,
0.015464335680007935,
0.07729366421699524,
-0.12657475471496582,
0.074497751891613,
0.03399961441755295,
-0.09340187162160873,
0.04758043959736824,
-0.11363168805837631,
0.02930769883096218,
0.08418485522270203,
0.042350318282842636,
-0.029388712719082832,
0.12324173748493195,
-0.03339560702443123,
0.07730299234390259,
-0.03974442183971405,
0.020265309140086174,
0.11114006489515305,
0.02627437189221382,
-0.060512885451316833,
-0.11018595844507217,
-0.09560016542673111,
-0.013340306468307972,
-0.058952417224645615,
-0.03449074178934097,
-0.040322139859199524,
-0.145575150847435,
0.039193060249090195,
-0.07779635488986969,
0.012242751196026802,
0.024583563208580017,
0.033686451613903046,
-0.01743302121758461,
-0.06876299530267715,
0.1543785184621811,
0.09732359647750854,
-0.020578883588314056,
-0.022191625088453293,
-0.03614236041903496,
-0.03908326104283333,
0.02001187764108181,
-0.00615825317800045,
-0.01891632378101349,
-0.0936983972787857,
-0.10179978609085083,
-0.0361410416662693,
-0.07324172556400299,
-0.016881639137864113,
-0.012312856502830982,
-0.10962419211864471,
0.08320329338312149,
-0.1484098583459854,
-0.051013242453336716,
0.015021393075585365,
0.0756838396191597,
-0.04981143772602081,
-0.07117336988449097,
-0.0831335261464119,
-0.09369578957557678,
-0.03653644025325775,
-0.009249215014278889,
0.0508560910820961,
-0.007368946447968483,
0.03531772270798683,
-0.03037066012620926,
0.09200096130371094,
-0.2317655086517334,
0.01705174334347248,
-0.08900964260101318,
-0.007369850296527147,
0.12295938283205032,
-0.005470643285661936,
-0.034851353615522385,
0.08761058747768402,
-0.018632439896464348,
0.028077248483896255,
-0.005823296494781971,
0.02124025672674179,
0.02028832770884037,
0.1327599287033081,
-0.08228671550750732,
0.0031930124387145042,
0.1709684133529663,
-0.08312754333019257,
-0.17389804124832153,
0.10001687705516815,
0.038284990936517715,
0.12544991075992584,
0.03407750651240349,
0.17498286068439484,
0.0719825029373169,
-0.06169065833091736,
-0.05135060101747513,
-0.00555806327611208,
-0.10380008071660995,
-0.2000148594379425,
0.004351094365119934,
0.1545332968235016,
-0.1625470519065857,
0.0201861672103405,
-0.07674560695886612,
0.1386110931634903,
-0.10609778016805649,
-0.047060661017894745,
-0.00012012739171041176,
-0.12428144365549088,
-0.026850789785385132,
0.021837109699845314,
0.054983146488666534,
-0.014071350917220116,
0.035010725259780884,
-0.15719671547412872,
0.08019160479307175,
-0.011403762735426426,
-0.05382392555475235,
-0.11153596639633179,
0.053087688982486725,
-0.06560834497213364,
0.03840838745236397,
-0.008882641792297363,
-0.04917766898870468,
0.051622286438941956,
0.0374978631734848,
0.09059271216392517,
0.22687238454818726,
0.0281633660197258,
0.07059475034475327,
-0.003930957056581974,
-0.07199916243553162,
0.04119746387004852,
-0.02924746461212635,
-0.03269634768366814,
-0.15153199434280396,
0.07758817821741104,
-0.060701463371515274,
-0.031417571008205414,
-0.14340218901634216,
0.012000645510852337,
-0.0290245171636343,
0.07834967970848083,
0.025760943070054054,
0.0034637944772839546,
0.04785362631082535,
0.0014294624561443925,
-0.09886591881513596,
0.016045132651925087,
0.04305613040924072,
0.011335009709000587,
-0.13942283391952515,
0.1624523550271988,
-0.06112426146864891,
0.13299785554409027,
0.09782101958990097,
-0.043887872248888016,
0.008711705915629864,
0.04310737922787666,
-0.07229990512132645,
0.008284864947199821,
0.012460998259484768,
0.034137409180402756,
0.0034143365919589996,
-0.05003898963332176,
0.09135933965444565,
-0.04716799408197403,
0.03313516825437546,
0.06061231717467308,
-0.03898756951093674,
-0.014528726227581501,
0.11598630249500275,
0.1430879831314087,
-0.17884446680545807,
0.1495804637670517,
0.09588733315467834,
0.007349544670432806,
0.20783039927482605,
0.04857778921723366,
0.024342363700270653,
-0.04938311502337456,
0.04526399075984955,
-0.019960658624768257,
0.24533826112747192,
-0.11032330989837646,
0.024854760617017746,
0.022863177582621574,
-0.029246840626001358,
0.0813087448477745,
-0.13864043354988098,
-0.10214059799909592,
-0.059383977204561234,
-0.038293104618787766,
0.11230047047138214,
0.0776953250169754,
-0.15990255773067474,
0.10904770344495773,
-0.09789074212312698,
-0.18502405285835266,
0.02008277364075184,
-0.018202533945441246,
-0.03618273138999939,
0.06259801983833313,
-0.018526991829276085,
-0.227116197347641,
-0.10785342007875443,
-0.04677269980311394,
-0.10719788819551468,
0.008749142289161682,
0.029906537383794785,
-0.08717852085828781,
0.001322233467362821,
-0.050446998327970505,
-0.012107311747968197,
-0.12153251469135284,
-0.007517603691667318,
-0.008907133713364601,
0.03807727247476578,
-0.05772767961025238,
-0.020159052684903145,
0.023725807666778564,
-0.07036663591861725,
0.017187219113111496,
0.15428678691387177,
-0.05874950438737869,
0.18612024188041687,
0.09115716069936752,
0.01081429049372673,
-0.00009182203211821616,
-0.006824647542089224,
0.2386215776205063,
-0.08515787869691849,
0.07482621818780899,
0.07309336215257645,
0.10299762338399887,
0.062137361615896225,
0.16146929562091827,
0.039192307740449905,
-0.10747966170310974,
0.0359998382627964,
-0.06030305474996567,
-0.1166323572397232,
-0.13530969619750977,
-0.06093727424740791,
-0.05423756688833237,
0.16035887598991394,
0.011191549710929394,
0.11073919385671616,
0.16076989471912384,
0.17942500114440918,
-0.001279941527172923,
0.01006601843982935,
-0.11758462339639664,
0.05256560444831848,
-0.10797326266765594,
-0.0434933640062809,
0.024418724700808525,
-0.07180719822645187,
-0.08047451078891754,
0.11131729930639267,
0.09814267605543137,
0.20149290561676025,
0.09963255375623703,
0.06032951921224594,
0.07087548077106476,
0.14964716136455536,
0.09693121910095215,
0.05421239882707596,
0.005253448151051998,
-0.056833770126104355,
-0.06606114655733109,
-0.058977145701646805,
0.10019026696681976,
0.11819181591272354,
-0.024720191955566406,
-0.03621218353509903,
0.11621958017349243,
0.015795500949025154,
-0.00701625132933259,
0.034546054899692535,
0.11497843265533447,
-0.19888156652450562,
0.039539139717817307,
0.010410505346953869,
0.06883370876312256,
-0.07259846478700638,
0.0343523733317852,
0.20636726915836334,
-0.00026026717387139797,
0.005202343221753836,
-0.028124509379267693,
0.0942867174744606,
0.07224242389202118,
0.00640984158962965,
-0.07674987614154816,
0.0118370670825243,
0.010392026044428349,
0.05859214812517166,
-0.2069743424654007,
0.2149484008550644,
-0.030379226431250572,
0.0025395029224455357,
-0.02950553223490715,
-0.06252042204141617,
-0.005308138206601143,
0.138390451669693,
0.16249988973140717,
0.024935411289334297,
-0.04373888671398163,
-0.03926882520318031,
-0.0966123715043068,
0.060390815138816833,
0.03206758573651314,
-0.05474090576171875,
-0.0022999532520771027,
0.09332737326622009,
-0.013233793899416924,
-0.051190804690122604,
0.07141957432031631,
-0.13294070959091187,
-0.0457889698445797,
-0.002296170685440302,
0.2038312405347824,
0.08352143317461014,
0.02952359989285469,
0.02342287451028824,
-0.014780470170080662,
0.0683416947722435,
-0.09090273082256317,
-0.04515332728624344,
-0.06249057129025459,
-0.030529839918017387,
0.008266488090157509,
-0.054098013788461685,
-0.062037281692028046,
-0.09516659379005432,
0.012101919390261173,
-0.04160301387310028,
-0.1265200525522232,
0.03104054369032383,
-0.14301837980747223,
-0.088555246591568,
-0.11190120875835419,
-0.011577194556593895,
0.028828056529164314,
0.03559533879160881,
-0.013886184431612492,
-0.048816483467817307,
-0.09066171199083328,
-0.10866080969572067,
0.07393909990787506,
-0.01670568250119686,
-0.16876381635665894,
-0.019467324018478394,
-0.05014483630657196,
0.01289842464029789,
-0.08160390704870224,
-0.08925049751996994,
0.054894451051950455,
0.2861083447933197,
-0.07096029072999954,
0.03054196387529373,
0.23491284251213074,
-0.07731173932552338,
-0.256198525428772,
-0.1443495899438858,
-0.07242091745138168,
-0.006596957799047232,
0.005662695970386267,
-0.2362678050994873,
0.07470379769802094,
0.012748463079333305,
-0.055895667523145676,
0.22976598143577576,
-0.29384124279022217,
-0.031180400401353836,
0.045131243765354156,
0.13334965705871582,
0.30785098671913147,
-0.16947582364082336,
-0.023444313555955887,
-0.018615543842315674,
0.035205986350774765,
0.23151226341724396,
0.058741141110658646,
0.08353855460882187,
-0.056036874651908875,
0.06042104214429855,
0.0031201746314764023,
-0.0014633474638685584,
0.11402803659439087,
0.0016356230480596423,
0.07288841903209686,
-0.05366029962897301,
0.04161182790994644,
0.12933890521526337,
0.011934986338019371,
0.019454138353466988,
-0.11136392503976822,
0.012927849777042866,
-0.09839774668216705,
-0.0023235969711095095,
-0.06384991854429245,
0.0016600940143689513,
-0.018968217074871063,
-0.12369272112846375,
-0.1354912966489792,
0.05659394711256027,
-0.020870910957455635,
0.054464083164930344,
0.00408375496044755,
0.012160148471593857,
-0.04722977057099342,
0.11407572776079178,
-0.01332124974578619,
-0.0322238989174366,
0.05953516066074371,
-0.08454086631536484,
-0.056152597069740295,
0.14270637929439545,
-0.06896091252565384,
-0.06440053880214691,
0.11950033158063889,
-0.010541364550590515,
0.05990844592452049,
0.01183983776718378,
-0.0893048346042633,
0.0623859241604805,
0.13368265330791473,
-0.1301189512014389,
-0.2647034525871277,
-0.07869775593280792,
0.16244247555732727,
0.13008110225200653,
0.11029500514268875,
0.09841132164001465,
-0.10866234451532364,
0.012977750040590763,
-0.04088301211595535,
0.0033986347261816263,
0.0005735738668590784,
0.004079727455973625,
0.021335924044251442,
0.003814120078459382,
-0.0871664360165596,
0.009166872128844261,
0.012597544118762016,
-0.002970949048176408,
-0.014647900126874447,
0.05333178862929344,
-0.04489004239439964,
-0.10995341092348099,
0.03081848844885826,
0.15252159535884857,
-0.19638991355895996,
-0.06640930473804474,
-0.047338906675577164,
-0.03774712234735489,
0.022127822041511536,
-0.0016983331879600883,
0.029285527765750885,
-0.005597665440291166,
0.05712321773171425,
-0.010898126289248466,
-0.021135438233613968,
0.0031868901569396257,
-0.0750579908490181,
0.11566182225942612,
-0.17610280215740204,
-0.12053427845239639,
0.02396215684711933,
0.07496138662099838,
-0.08408733457326889,
-0.033146824687719345,
-0.10109636932611465,
0.022834541276097298,
-0.024756496772170067,
0.061466630548238754,
-0.10015248507261276,
-0.09082876890897751,
-0.03398524224758148,
-0.024415936321020126,
-0.07005280256271362,
0.002474366221576929,
-0.09669341892004013,
0.05549487844109535,
0.05677085742354393,
0.03210779279470444,
0.016841020435094833,
0.022852957248687744,
0.014885093085467815,
0.004359369166195393,
0.0826185941696167,
-0.05870051681995392,
-0.08075208216905594,
0.006987179163843393,
-0.18712931871414185,
0.03292899206280708,
0.05688703805208206,
0.02787056751549244,
0.004682025406509638,
0.1062982901930809,
-0.013830396346747875,
-0.018457436934113503,
0.03329175338149071,
-0.027022838592529297,
-0.05341852083802223,
-0.09967295080423355,
-0.06860048323869705,
-0.055634163320064545,
-0.02949560061097145,
-0.06859559565782547,
0.020537836477160454,
0.05605233833193779,
0.06511403620243073,
0.08064056932926178,
-0.04673425480723381,
0.05717954412102699,
-0.03881136700510979,
0.04793468862771988,
0.07169575244188309,
-0.06480159610509872,
-0.028075305745005608,
-0.10913240909576416,
-0.03384971246123314,
-0.01193357165902853,
0.1020069345831871,
0.02440609410405159,
-0.24575868248939514,
-0.06878297030925751,
-0.0743848904967308,
-0.08693158626556396,
-0.029475944116711617,
0.22447504103183746,
0.05280015617609024,
0.031061971560120583,
-0.10285027325153351,
0.06349951028823853,
0.03805927187204361,
0.19012632966041565,
0.01397002674639225,
0.004596271086484194,
0.06856939196586609,
0.06330512464046478,
0.07887532562017441,
-0.014727408066391945,
-0.027647841721773148,
0.03374466672539711,
-0.07974203675985336,
0.1366686373949051,
-0.04396775737404823,
0.1225706934928894,
0.13430021703243256,
0.03924637287855148,
0.02285510115325451,
0.09015711396932602,
-0.005086896475404501,
-0.031133243814110756,
-0.22967854142189026,
-0.07245177775621414,
-0.11066066473722458,
0.031795285642147064,
-0.0662924125790596,
-0.030588699504733086,
0.0004587012226693332,
0.039355549961328506,
-0.08057870715856552,
0.016604140400886536,
0.11741426587104797,
-0.035194605588912964,
0.07510898262262344,
-0.029707947745919228,
-0.04819406941533089,
-0.02338891290128231,
-0.0034207883290946484,
0.00214576069265604,
0.054705020040273666,
-0.0019660983234643936,
0.05298546329140663,
-0.04606349766254425,
0.028782658278942108,
-0.014589528553187847,
-0.05234483629465103,
-0.05010542646050453,
-0.05150458216667175,
0.01256981585174799,
0.07031230628490448,
-0.00002945248161267955,
-0.0276847705245018,
0.010452941060066223,
0.09545087069272995,
0.02733645588159561,
0.052278775721788406,
-0.0616869293153286,
0.10454364866018295,
-0.1368883103132248,
0.055291689932346344,
-0.04090648144483566,
0.032097768038511276,
-0.089079849421978,
0.22296267747879028,
0.19244511425495148,
-0.14582723379135132,
-0.0331404022872448,
-0.03066089004278183,
0.012008448131382465,
-0.08430591970682144,
0.12626197934150696,
-0.024881770834326744,
0.20863084495067596,
-0.05773089453577995,
0.017766306176781654,
-0.12388332188129425,
-0.04138588160276413,
-0.017879929393529892,
-0.07669641077518463,
0.07674262672662735,
-0.027031205594539642,
-0.14659693837165833,
0.08494164049625397,
-0.2510685324668884,
0.008264834992587566,
0.08265383541584015,
0.002821447793394327,
-0.05558259040117264,
-0.004388670437037945,
0.07739187031984329,
0.09417534619569778,
0.06997712701559067,
-0.13335835933685303,
0.045140959322452545,
0.02963542751967907,
-0.04479685798287392,
-0.14446307718753815,
0.037584658712148666,
0.023113256320357323,
-0.1850394755601883,
0.19788530468940735,
-0.021086998283863068,
-0.022339075803756714,
0.06292302161455154,
-0.09117438644170761,
-0.15900002419948578,
0.09226823598146439,
-0.020634230226278305,
-0.029218444600701332,
-0.013191865757107735,
0.11663388460874557,
0.027421973645687103,
-0.02346980758011341,
0.04799708351492882,
0.024283070117235184,
-0.048252757638692856,
0.07398660480976105,
0.05139981955289841,
-0.04374013468623161,
0.0695769190788269,
-0.055141232907772064,
0.0921609178185463,
0.010340199805796146,
-0.04918151721358299,
-0.053067028522491455,
-0.03731783851981163,
0.014693357981741428,
0.009142996743321419,
-0.014289048500359058,
0.0202066358178854,
-0.0651668831706047,
-0.07574629783630371,
0.05889279395341873,
0.04363785684108734,
-0.1901811808347702,
0.041659317910671234,
-0.1811765730381012,
0.03722834959626198,
-0.026837171986699104,
0.0329444482922554,
0.20987604558467865,
-0.005291904788464308,
0.01891886629164219,
-0.06380413472652435,
-0.0527394600212574,
-0.03128499910235405,
0.007495423778891563,
-0.14722274243831635
] |
null | null | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="Poliuszko/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Poliuszko/q-FrozenLake-v1-4x4-noSlippery | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-08T17:27:53+00:00 | [] | [] | TAGS
#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 FrozenLake-v1
This is a trained model of a Q-Learning agent playing FrozenLake-v1 .
## Usage
| [
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
"TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
40,
39
] | [
"passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
0.04578453302383423,
-0.08074592798948288,
-0.00430759321898222,
0.10720831900835037,
0.05034215748310089,
-0.040469273924827576,
0.11997015029191971,
0.018999949097633362,
0.20601962506771088,
-0.010012076236307621,
0.1455274522304535,
0.007022971753031015,
-0.006192410364747047,
0.1867983490228653,
0.04572829231619835,
-0.26324528455734253,
0.01831899583339691,
-0.09495259821414948,
-0.07281816750764847,
0.11870454251766205,
0.05470194295048714,
-0.01901467889547348,
-0.0007633853238075972,
0.056141503155231476,
-0.0673527717590332,
0.0007737681735306978,
0.031996939331293106,
-0.012976245954632759,
0.19804789125919342,
-0.02254498563706875,
0.06641989201307297,
0.054705578833818436,
0.0758768692612648,
-0.1998077929019928,
0.0358855277299881,
-0.04215473681688309,
-0.09439758956432343,
-0.03934839740395546,
-0.018780618906021118,
0.05878105387091637,
0.053356342017650604,
0.03858819976449013,
0.058354366570711136,
0.09384993463754654,
-0.0773480236530304,
0.04328357055783272,
0.04280758649110794,
0.024811049923300743,
0.04589218273758888,
-0.0237203948199749,
-0.027002155780792236,
0.08246652781963348,
-0.22182892262935638,
0.10318073630332947,
-0.010159241035580635,
-0.5270710587501526,
-0.00633762264624238,
0.24088262021541595,
0.11517096310853958,
0.05707438662648201,
-0.06903956830501556,
0.10566288232803345,
0.03913382440805435,
-0.007209456991404295,
0.03210983797907829,
0.02150118350982666,
0.12817370891571045,
0.06009242683649063,
-0.09581366181373596,
0.040699947625398636,
0.13722525537014008,
0.012822695076465607,
0.020306183025240898,
-0.08888901025056839,
0.0410032719373703,
-0.03461858257651329,
-0.007679527159780264,
-0.09758518636226654,
0.05478060990571976,
0.012466507963836193,
-0.0934976264834404,
-0.09247440844774246,
-0.04236573353409767,
-0.06708304584026337,
0.11252415925264359,
0.046419668942689896,
-0.0874939113855362,
0.03884070739150047,
-0.06760413944721222,
0.05918780341744423,
-0.16863860189914703,
0.02074250765144825,
-0.06627868115901947,
-0.09376336634159088,
-0.11799788475036621,
-0.01683047041296959,
-0.07946427166461945,
0.009092256426811218,
0.056664444506168365,
0.1447116881608963,
0.22076484560966492,
0.06690320372581482,
0.09728849679231644,
0.07456006109714508,
0.06531001627445221,
0.1538129299879074,
0.10918238013982773,
0.019075315445661545,
-0.015266558155417442,
0.0948706716299057,
-0.06445580720901489,
-0.1351388692855835,
-0.15579092502593994,
0.005488025024533272,
0.0983937531709671,
0.08871900290250778,
-0.044080477207899094,
-0.006702381651848555,
-0.024641724303364754,
0.08566431701183319,
-0.11314457654953003,
-0.024612564593553543,
-0.002267979085445404,
0.06882024556398392,
-0.024801667779684067,
0.020378148183226585,
-0.06242705136537552,
0.12715265154838562,
0.04222423583269119,
-0.059924717992544174,
-0.055308472365140915,
-0.03053177334368229,
-0.014276440255343914,
-0.027539284899830818,
0.02446848154067993,
-0.07659092545509338,
0.04767750948667526,
-0.16766095161437988,
-0.042871296405792236,
-0.04784649610519409,
0.025697942823171616,
-0.03907240927219391,
-0.13557587563991547,
-0.17699143290519714,
-0.048906855285167694,
-0.022438718006014824,
0.03549358621239662,
-0.038111843168735504,
0.006551501806825399,
-0.006318534724414349,
-0.1583600640296936,
0.09783563017845154,
0.09784027189016342,
-0.03643378987908363,
-0.02749447710812092,
0.056263517588377,
-0.07194498926401138,
0.1561182290315628,
-0.21054518222808838,
-0.054014235734939575,
-0.044764336198568344,
-0.06595750898122787,
0.19673264026641846,
0.012690845876932144,
-0.01202624011784792,
0.19873127341270447,
-0.29073721170425415,
-0.06078760325908661,
0.12533614039421082,
-0.07834373414516449,
-0.0936407670378685,
0.06941844522953033,
-0.04206686094403267,
0.023345354944467545,
0.046047765761613846,
0.36345911026000977,
-0.02069227211177349,
-0.16197136044502258,
-0.021782705560326576,
0.13971707224845886,
-0.1184760183095932,
0.059895481914281845,
0.04240793362259865,
0.12543781101703644,
-0.04250509291887283,
-0.018672896549105644,
-0.09023164212703705,
0.05999075248837471,
-0.05241934582591057,
-0.09016361832618713,
-0.03393383324146271,
-0.07645075023174286,
0.13294468820095062,
-0.0629684180021286,
0.05601520463824272,
-0.03255095332860947,
-0.07133250683546066,
-0.050324998795986176,
-0.016492370516061783,
0.04460815340280533,
0.05951254442334175,
-0.12794871628284454,
0.11029167473316193,
0.13025271892547607,
-0.0006193425506353378,
-0.07498852163553238,
-0.17872096598148346,
0.003240168560296297,
0.009576505981385708,
0.039837226271629333,
0.17141658067703247,
0.12209978699684143,
0.033295199275016785,
0.008770671673119068,
-0.06389404833316803,
-0.18276847898960114,
0.058129217475652695,
-0.056212130934000015,
-0.14230976998806,
-0.052409034222364426,
-0.0728459507226944,
0.017381802201271057,
-0.0859743058681488,
-0.017379917204380035,
0.021926190704107285,
0.006908397190272808,
0.02990424446761608,
-0.026645656675100327,
-0.049561817198991776,
0.021254703402519226,
0.06490101665258408,
-0.0037617047782987356,
0.12023693323135376,
0.008277264423668385,
-0.18308481574058533,
0.07930773496627808,
0.08478537946939468,
0.09196605533361435,
0.013250201940536499,
0.02685922384262085,
-0.021522263064980507,
-0.08061408251523972,
-0.054420311003923416,
0.02957955375313759,
0.11417073011398315,
0.1317172348499298,
0.2361993044614792,
0.08753683418035507,
0.04697408527135849,
-0.02164587564766407,
-0.016415923833847046,
0.002810494042932987,
-0.06318057328462601,
-0.029935607686638832,
0.10614971816539764,
0.05865858122706413,
-0.067733034491539,
-0.04576427489519119,
0.09590928256511688,
0.02732124738395214,
0.21205885708332062,
-0.03342745825648308,
0.01286078616976738,
-0.10957037657499313,
-0.06550975888967514,
-0.031982194632291794,
0.09201868623495102,
0.09498392790555954,
0.009755023755133152,
-0.022056059911847115,
-0.04259001836180687,
0.0012916827108711004,
-0.1334889680147171,
-0.10375088453292847,
0.026475343853235245,
0.013400445692241192,
-0.11206940561532974,
0.11674030870199203,
-0.11352457851171494,
0.039504457265138626,
0.06024791672825813,
-0.13837239146232605,
0.04428480193018913,
-0.029713207855820656,
-0.07886212319135666,
0.16866780817508698,
-0.11075661331415176,
-0.094340018928051,
-0.08831550180912018,
0.004082420375198126,
0.0075836325995624065,
-0.03922267258167267,
-0.009283260442316532,
-0.19952571392059326,
-0.005375816952437162,
-0.03544965013861656,
0.013616434298455715,
-0.06988783925771713,
-0.11287739872932434,
-0.010957922786474228,
0.07084179669618607,
-0.043388739228248596,
-0.07803605496883392,
0.007967432029545307,
-0.08923084288835526,
-0.10623309016227722,
0.028189711272716522,
0.019765101373195648,
-0.022883659228682518,
0.16152891516685486,
0.01816628873348236,
0.05626589432358742,
-0.03298520669341087,
0.30665266513824463,
-0.038163769990205765,
0.08371731638908386,
-0.02993497997522354,
-0.07433546334505081,
0.06130730360746384,
-0.022327827289700508,
0.06086638569831848,
-0.020221687853336334,
-0.02362890914082527,
0.0077952733263373375,
-0.08579335361719131,
-0.18365982174873352,
-0.05417544022202492,
0.03724347800016403,
0.195254847407341,
0.031118987128138542,
0.01910330168902874,
-0.0488768145442009,
-0.010547760874032974,
0.1665220558643341,
-0.10005921125411987,
0.04030545800924301,
-0.05366240441799164,
0.11506262421607971,
-0.08640182018280029,
0.06195629760622978,
0.020486772060394287,
0.04266135022044182,
-0.04877188801765442,
0.09486009180545807,
0.0826394334435463,
0.1121082529425621,
-0.02206910029053688,
0.046257395297288895,
0.019012698903679848,
0.07383184134960175,
0.11073657125234604,
0.0368414968252182,
-0.0729052945971489,
0.001982470043003559,
-0.006313489284366369,
-0.039427030831575394,
0.11933320760726929,
0.17963355779647827,
-0.11991413682699203,
-0.05106910318136215,
0.27167606353759766,
0.0031242913100868464,
0.19481229782104492,
-0.01315275114029646,
0.043591804802417755,
-0.04484925419092178,
0.04572054371237755,
-0.05338600277900696,
-0.04086209088563919,
0.2094656229019165,
0.08045925945043564,
-0.17165091633796692,
-0.08549032360315323,
-0.05912299454212189,
0.07081323862075806,
0.10728751868009567,
0.0013539529172703624,
-0.04156802222132683,
0.0004610282776411623,
0.0014198932331055403,
0.08339415490627289,
-0.14520122110843658,
0.11816094070672989,
-0.03172019124031067,
0.05612684786319733,
0.017555562779307365,
-0.045326150953769684,
0.04264266416430473,
0.07474290579557419,
0.26618310809135437,
0.0904107540845871,
-0.040318213403224945,
-0.0892091691493988,
-0.12260187417268753,
0.010461576282978058,
0.029102616012096405,
-0.03534553572535515,
0.0037547778338193893,
-0.020087555050849915,
0.0318896509706974,
0.008264793083071709,
0.016230624169111252,
-0.08987458795309067,
-0.03175399824976921,
-0.027736429125070572,
-0.023839212954044342,
0.10733365267515182,
-0.09495144337415695,
-0.1444292515516281,
-0.15713949501514435,
0.04191131144762039,
-0.0766405463218689,
-0.056593164801597595,
-0.054507751017808914,
-0.05239389091730118,
-0.0311186034232378,
-0.03773957118391991,
0.09099467098712921,
-0.0021037792321294546,
0.14807306230068207,
-0.1920108050107956,
-0.04220759496092796,
0.051812779158353806,
-0.07607918977737427,
-0.08729588985443115,
0.03410962224006653,
0.12136995792388916,
0.05116051807999611,
0.11504370719194412,
0.013609255664050579,
0.09567681699991226,
0.0045484392903745174,
-0.06713183224201202,
0.15302421152591705,
-0.14069625735282898,
-0.27875974774360657,
-0.03836318850517273,
0.016946332529187202,
0.1615200787782669,
-0.05613167956471443,
0.031766023486852646,
0.3335736393928528,
0.27782970666885376,
-0.1428707242012024,
0.25916144251823425,
0.019178593531250954,
0.004398873541504145,
-0.19130495190620422,
-0.10125631093978882,
0.025324683636426926,
0.04740457236766815,
0.12032642960548401,
-0.14564448595046997,
-0.010732659138739109,
-0.04543145373463631,
-0.025908485054969788,
0.10386138409376144,
-0.12300799041986465,
-0.07263197749853134,
0.07765276730060577,
0.039809420704841614,
0.1808302253484726,
0.03932500258088112,
0.0014799144119024277,
0.13626977801322937,
0.06612244248390198,
0.019124457612633705,
0.05216038227081299,
0.08028066903352737,
-0.018944554030895233,
0.14207926392555237,
0.05448179319500923,
-0.02551644667983055,
0.052681710571050644,
-0.0054580713622272015,
-0.03219012916088104,
0.015605825930833817,
-0.183198019862175,
-0.10147556662559509,
-0.0561356320977211,
-0.10798973590135574,
-0.04978342354297638,
0.056853994727134705,
-0.12395523488521576,
-0.007896827533841133,
-0.03841273859143257,
0.03718273714184761,
-0.07831971347332001,
-0.09360362589359283,
-0.036494381725788116,
0.1351792961359024,
0.07210618257522583,
0.04471297934651375,
0.035655103623867035,
-0.07390819489955902,
0.07097936421632767,
0.21671734750270844,
0.08159157633781433,
0.028919655829668045,
-0.19545674324035645,
-0.024042490869760513,
-0.0803457647562027,
0.06306298077106476,
-0.08856996893882751,
-0.016788700595498085,
0.11923003196716309,
0.08616556972265244,
0.05413002520799637,
0.09640096127986908,
-0.045083072036504745,
0.021686913445591927,
0.02684609219431877,
-0.15131035447120667,
-0.18501274287700653,
-0.08534606546163559,
-0.03519878163933754,
0.11561143398284912,
-0.06398691236972809,
0.10897188633680344,
-0.13615410029888153,
0.010051886551082134,
-0.006060056854039431,
0.02693452313542366,
-0.03596206381917,
-0.11251141875982285,
0.15348562598228455,
0.11999429017305374,
-0.06767056882381439,
0.03127254918217659,
-0.09527092427015305,
-0.04423454403877258,
0.12686803936958313,
-0.013623855076730251,
-0.0371493324637413,
-0.054547641426324844,
-0.03628576174378395,
0.15247689187526703,
-0.03436964750289917,
0.008244883269071579,
-0.041229065507650375,
-0.18217355012893677,
0.0798322781920433,
0.09045056998729706,
0.019827889278531075,
-0.031874191015958786,
-0.09797266125679016,
-0.010231015272438526,
-0.0011165260802954435,
0.11730700731277466,
-0.10696814209222794,
-0.10933240503072739,
-0.15144047141075134,
0.06713984161615372,
-0.0007159380475059152,
0.18502596020698547,
-0.06394898891448975,
-0.08904669433832169,
-0.12429379671812057,
0.02344517596065998,
-0.0027384376153349876,
-0.042264558374881744,
0.01618490368127823,
0.07992301136255264,
-0.04095321521162987,
0.02075677551329136,
-0.06651144474744797,
0.06372585147619247,
-0.11786920577287674,
0.09625071287155151,
0.01063506118953228,
0.016993753612041473,
-0.0417880080640316,
-0.01618220843374729,
0.039470795542001724,
-0.057925306260585785,
0.07921463251113892,
0.011758086271584034,
0.0010938759660348296,
0.10196787863969803,
-0.0034960443153977394,
0.06409632414579391,
-0.05372481048107147,
-0.023290161043405533,
0.06578411161899567,
-0.05874887853860855,
-0.03370826691389084,
-0.1573946475982666,
-0.0709633082151413,
0.020051732659339905,
-0.04775108024477959,
0.002077929675579071,
0.03673801198601723,
0.062159497290849686,
-0.06937079131603241,
-0.12125655263662338,
-0.043812792748212814,
-0.028638383373618126,
0.021301284432411194,
0.10829301923513412,
-0.07526551932096481,
0.1547859013080597,
-0.052787959575653076,
-0.00020603960729204118,
0.07437096536159515,
0.04048224538564682,
0.01393822580575943,
-0.10422444343566895,
-0.04698587954044342,
-0.11035211384296417,
0.1502903699874878,
-0.007902312092483044,
-0.03533121198415756,
0.03719403222203255,
-0.11946307867765427,
-0.1572723090648651,
0.03418220207095146,
0.10199101269245148,
0.0448341928422451,
0.025807438418269157,
0.027079269289970398,
-0.04042419046163559,
-0.021270349621772766,
-0.07034418731927872,
0.0882953479886055,
-0.12085357308387756,
-0.09669415652751923,
0.09555385261774063,
0.12178351730108261,
-0.0036850625183433294,
-0.07441367954015732,
0.11554073542356491,
-0.021787192672491074,
0.05525410920381546,
-0.02971339225769043,
0.10308072715997696,
0.0796005055308342,
-0.12273547053337097,
0.005693064536899328,
-0.036891788244247437,
-0.0741485133767128,
-0.12975730001926422,
0.019545545801520348,
-0.061916105449199677,
-0.13383042812347412,
0.12179028987884521,
-0.09376577287912369,
0.030037038028240204,
-0.10506992787122726,
0.021338803693652153,
0.01864001713693142,
0.061665527522563934,
-0.10988292098045349,
0.08575301617383957,
0.13424484431743622,
-0.043199893087148666,
-0.07184189558029175,
-0.12455986440181732,
-0.05022053420543671,
-0.04231856390833855,
-0.13957437872886658,
-0.11600435525178909,
0.0100301094353199,
-0.023418782278895378,
-0.05818291753530502,
0.0015462689334526658,
-0.03659068048000336,
0.008594646118581295,
0.021907730028033257,
0.04032021388411522,
-0.02693161368370056,
0.05134565755724907,
-0.057569269090890884,
-0.052510857582092285,
0.11489357799291611,
0.04113486409187317,
-0.03561042994260788,
-0.052359987050294876,
0.12997733056545258,
-0.11959461867809296,
0.07662346214056015,
-0.020313527435064316,
0.017129231244325638,
-0.06435854732990265,
0.17131924629211426,
0.11673715710639954,
-0.1367570012807846,
-0.005008010193705559,
-0.08210669457912445,
0.020409544929862022,
0.023555370047688484,
0.13693512976169586,
-0.03411718085408211,
-0.0012358218664303422,
-0.1580323874950409,
0.018575575202703476,
-0.18557456135749817,
-0.03716109320521355,
0.04671547934412956,
0.09917585551738739,
0.15293832123279572,
-0.0034432117827236652,
-0.1263325810432434,
0.10424192249774933,
-0.2118520885705948,
0.0907607227563858,
0.05121984705328941,
-0.11874113976955414,
-0.06765396893024445,
-0.06795281916856766,
0.1198519766330719,
0.009196433238685131,
0.2040700763463974,
-0.013615905307233334,
-0.09132910519838333,
-0.07060808688402176,
-0.01980910450220108,
-0.030524181202054024,
0.09714830666780472,
0.041414931416511536,
0.04653804749250412,
0.12821412086486816,
0.00368314771912992,
0.07533777505159378,
0.060310911387205124,
0.02759413793683052,
-0.012300663627684116,
0.04076618701219559,
0.08261215686798096,
-0.14588621258735657,
-0.1659701019525528,
0.1326720416545868,
0.025149408727884293,
0.11792458593845367,
0.03658788278698921,
-0.1549617499113083,
0.06687124073505402,
0.2523096203804016,
-0.11147607117891312,
0.02505038119852543,
0.12737524509429932,
-0.0366884209215641,
0.0672016367316246,
0.1144871786236763,
-0.02633814327418804,
-0.05217865854501724,
-0.011363590136170387,
0.10233135521411896,
0.028660254552960396,
-0.04646271467208862,
-0.02340836264193058,
-0.03373933956027031,
-0.019070526584982872,
-0.011738128960132599,
-0.0909019410610199,
-0.1543993502855301,
-0.10471053421497345,
-0.16619662940502167,
0.04399140924215317,
-0.04626438021659851,
0.13418889045715332,
0.09469578415155411,
-0.012723101302981377,
0.04568437114357948,
0.028575526550412178,
0.07275456190109253,
0.07916246354579926,
-0.02939477376639843,
-0.036159269511699677
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# SMIDS_3x_beit_large_Adamax_lr00001_fold5
This model is a fine-tuned version of [microsoft/beit-large-patch16-224](https://huggingface.co/microsoft/beit-large-patch16-224) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8314
- Accuracy: 0.905
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|
| 0.1537 | 1.0 | 450 | 0.2709 | 0.9067 |
| 0.1471 | 2.0 | 900 | 0.3716 | 0.8967 |
| 0.1235 | 3.0 | 1350 | 0.3515 | 0.91 |
| 0.0665 | 4.0 | 1800 | 0.4266 | 0.9117 |
| 0.0022 | 5.0 | 2250 | 0.5065 | 0.9033 |
| 0.0447 | 6.0 | 2700 | 0.5322 | 0.9133 |
| 0.0008 | 7.0 | 3150 | 0.6263 | 0.9083 |
| 0.02 | 8.0 | 3600 | 0.6755 | 0.9133 |
| 0.0006 | 9.0 | 4050 | 0.8080 | 0.905 |
| 0.0347 | 10.0 | 4500 | 0.9745 | 0.8917 |
| 0.0 | 11.0 | 4950 | 0.7945 | 0.9067 |
| 0.0087 | 12.0 | 5400 | 0.7652 | 0.9067 |
| 0.053 | 13.0 | 5850 | 0.7058 | 0.9117 |
| 0.0001 | 14.0 | 6300 | 0.6925 | 0.9117 |
| 0.0698 | 15.0 | 6750 | 0.8058 | 0.905 |
| 0.0098 | 16.0 | 7200 | 0.7520 | 0.9167 |
| 0.0074 | 17.0 | 7650 | 0.7782 | 0.905 |
| 0.0 | 18.0 | 8100 | 0.7799 | 0.9133 |
| 0.0003 | 19.0 | 8550 | 0.8447 | 0.9067 |
| 0.0 | 20.0 | 9000 | 0.8202 | 0.9117 |
| 0.0 | 21.0 | 9450 | 0.8251 | 0.9067 |
| 0.0004 | 22.0 | 9900 | 0.7880 | 0.9083 |
| 0.0356 | 23.0 | 10350 | 0.8288 | 0.905 |
| 0.0 | 24.0 | 10800 | 0.8379 | 0.9033 |
| 0.0 | 25.0 | 11250 | 0.7968 | 0.9083 |
| 0.0128 | 26.0 | 11700 | 0.8170 | 0.91 |
| 0.0 | 27.0 | 12150 | 0.8852 | 0.9033 |
| 0.0002 | 28.0 | 12600 | 0.8989 | 0.9067 |
| 0.0 | 29.0 | 13050 | 0.8791 | 0.905 |
| 0.0 | 30.0 | 13500 | 0.8663 | 0.9067 |
| 0.0 | 31.0 | 13950 | 0.8564 | 0.9083 |
| 0.0 | 32.0 | 14400 | 0.8462 | 0.9067 |
| 0.0 | 33.0 | 14850 | 0.8516 | 0.905 |
| 0.0007 | 34.0 | 15300 | 0.8682 | 0.9067 |
| 0.0 | 35.0 | 15750 | 0.8117 | 0.9067 |
| 0.0 | 36.0 | 16200 | 0.8616 | 0.9083 |
| 0.0 | 37.0 | 16650 | 0.8586 | 0.9083 |
| 0.0 | 38.0 | 17100 | 0.8326 | 0.9083 |
| 0.0 | 39.0 | 17550 | 0.8172 | 0.9083 |
| 0.0 | 40.0 | 18000 | 0.8481 | 0.9033 |
| 0.0 | 41.0 | 18450 | 0.8645 | 0.9067 |
| 0.0 | 42.0 | 18900 | 0.8596 | 0.9083 |
| 0.0 | 43.0 | 19350 | 0.8769 | 0.91 |
| 0.0 | 44.0 | 19800 | 0.8690 | 0.91 |
| 0.0001 | 45.0 | 20250 | 0.8430 | 0.905 |
| 0.0 | 46.0 | 20700 | 0.8411 | 0.905 |
| 0.0 | 47.0 | 21150 | 0.8399 | 0.9033 |
| 0.0 | 48.0 | 21600 | 0.8330 | 0.9067 |
| 0.0 | 49.0 | 22050 | 0.8319 | 0.905 |
| 0.0 | 50.0 | 22500 | 0.8314 | 0.905 |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1
- Datasets 2.12.0
- Tokenizers 0.13.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "metrics": ["accuracy"], "base_model": "microsoft/beit-large-patch16-224", "model-index": [{"name": "SMIDS_3x_beit_large_Adamax_lr00001_fold5", "results": [{"task": {"type": "image-classification", "name": "Image Classification"}, "dataset": {"name": "imagefolder", "type": "imagefolder", "config": "default", "split": "test", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.905, "name": "Accuracy"}]}]}]} | image-classification | onizukal/SMIDS_3x_beit_large_Adamax_lr00001_fold5 | [
"transformers",
"pytorch",
"beit",
"image-classification",
"generated_from_trainer",
"dataset:imagefolder",
"base_model:microsoft/beit-large-patch16-224",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:27:55+00:00 | [] | [] | TAGS
#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| SMIDS\_3x\_beit\_large\_Adamax\_lr00001\_fold5
==============================================
This model is a fine-tuned version of microsoft/beit-large-patch16-224 on the imagefolder dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8314
* Accuracy: 0.905
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* num\_epochs: 50
### Training results
### Framework versions
* Transformers 4.32.1
* Pytorch 2.0.1
* Datasets 2.12.0
* Tokenizers 0.13.2
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
"TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
81,
116,
4,
30
] | [
"passage: TAGS\n#transformers #pytorch #beit #image-classification #generated_from_trainer #dataset-imagefolder #base_model-microsoft/beit-large-patch16-224 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* num\\_epochs: 50### Training results### Framework versions\n\n\n* Transformers 4.32.1\n* Pytorch 2.0.1\n* Datasets 2.12.0\n* Tokenizers 0.13.2"
] | [
-0.1292150914669037,
0.17132072150707245,
-0.002415567170828581,
0.13183215260505676,
0.11657863855361938,
0.020753253251314163,
0.1335890144109726,
0.16620413959026337,
-0.08238927274942398,
0.04929587244987488,
0.13697229325771332,
0.1357421576976776,
0.04955337569117546,
0.20790311694145203,
-0.053285520523786545,
-0.26080378890037537,
0.0391765721142292,
0.03443576768040657,
-0.020672276616096497,
0.12494900077581406,
0.09484300017356873,
-0.1312379240989685,
0.11272566765546799,
0.025938162580132484,
-0.20840293169021606,
-0.033587437123060226,
-0.01026944164186716,
-0.06854863464832306,
0.10221196711063385,
0.001568986801430583,
0.0741027221083641,
0.037979885935783386,
0.08491890877485275,
-0.12677186727523804,
0.000941311358474195,
0.04326357692480087,
0.0062435888685286045,
0.1065368577837944,
0.062226198613643646,
-0.008521218784153461,
0.06926212459802628,
-0.07453521341085434,
0.06115834787487984,
0.008060229010879993,
-0.11478453874588013,
-0.2692618668079376,
-0.09817449003458023,
0.07377522438764572,
0.08109822124242783,
0.06491127610206604,
0.006432840134948492,
0.16222304105758667,
-0.015434488654136658,
0.1024109497666359,
0.23076069355010986,
-0.2713507413864136,
-0.054792311042547226,
0.022649891674518585,
0.0155020197853446,
0.06252340972423553,
-0.10333037376403809,
-0.01993185468018055,
0.019141921773552895,
0.042880840599536896,
0.14450453221797943,
-0.012332411482930183,
-0.03331032395362854,
-0.02637922763824463,
-0.11139829456806183,
-0.08930420875549316,
0.18604889512062073,
0.06140090152621269,
-0.04917457327246666,
-0.07841385900974274,
-0.07612120360136032,
-0.17419220507144928,
-0.03924720734357834,
0.008911197073757648,
0.046679239720106125,
-0.04711441695690155,
-0.10239296406507492,
-0.03511375933885574,
-0.07504668086767197,
-0.05196268856525421,
-0.026160720735788345,
0.1420334428548813,
0.03879573196172714,
0.05471520125865936,
-0.027205273509025574,
0.10149593651294708,
0.010796112939715385,
-0.1717151701450348,
-0.02661297097802162,
0.0005703883362002671,
0.010487399995326996,
-0.01821139082312584,
-0.029929913580417633,
-0.06737607717514038,
-0.003975129686295986,
0.15347014367580414,
-0.07002666592597961,
0.058850113302469254,
-0.0054583000019192696,
0.041531506925821304,
-0.049319881945848465,
0.1874888390302658,
-0.029916515573859215,
-0.016198426485061646,
0.019476165995001793,
0.08928463608026505,
0.0656052976846695,
-0.030047036707401276,
-0.12371734529733658,
0.021691862493753433,
0.13241209089756012,
0.006458523217588663,
-0.022870952263474464,
0.054544735699892044,
-0.0711979940533638,
-0.0584990456700325,
0.09274657070636749,
-0.09275025129318237,
0.035496871918439865,
-0.011692462489008904,
-0.08981472253799438,
-0.06787234544754028,
0.029122935608029366,
0.011931490153074265,
-0.009771439246833324,
0.06940538436174393,
-0.09093258529901505,
0.01846885494887829,
-0.06650768965482712,
-0.09852384030818939,
0.01388985849916935,
-0.11549968272447586,
0.010918805375695229,
-0.10079170018434525,
-0.19154705107212067,
0.0032797311432659626,
0.07527101784944534,
-0.06246669217944145,
-0.06951755285263062,
-0.033377837389707565,
-0.07729615271091461,
0.03790769353508949,
-0.01523390132933855,
0.07408059388399124,
-0.07056254893541336,
0.09071778506040573,
0.02892814762890339,
0.09002465009689331,
-0.052364569157361984,
0.048610031604766846,
-0.09854818880558014,
0.05158581584692001,
-0.19896768033504486,
0.0824570581316948,
-0.04529954120516777,
0.05730293318629265,
-0.10005063563585281,
-0.10804302245378494,
0.029095064848661423,
-0.0466112419962883,
0.07224688678979874,
0.09985066950321198,
-0.16068536043167114,
-0.05396431311964989,
0.14283035695552826,
-0.09281232208013535,
-0.14269256591796875,
0.09829698503017426,
-0.045770496129989624,
0.014614340849220753,
0.04329100251197815,
0.2130173146724701,
0.04901750758290291,
-0.08417420834302902,
-0.023242823779582977,
-0.02969830296933651,
0.03785223513841629,
-0.0668954998254776,
0.10032020509243011,
0.025215676054358482,
0.05325069651007652,
0.02284027636051178,
-0.029413679614663124,
0.04126512631773949,
-0.08672589063644409,
-0.09880872070789337,
-0.053216658532619476,
-0.0853687971830368,
0.03892384096980095,
0.05334646999835968,
0.0614997074007988,
-0.10279879719018936,
-0.09344549477100372,
0.0453280434012413,
0.09495674818754196,
-0.07567895948886871,
0.02865210548043251,
-0.08989366888999939,
0.10926083475351334,
-0.08635354787111282,
-0.02427433431148529,
-0.18316780030727386,
-0.041861772537231445,
0.04194685444235802,
-0.025394707918167114,
-0.007599220145493746,
-0.05216266214847565,
0.06521623581647873,
0.0848059430718422,
-0.05379978567361832,
-0.05897609516978264,
-0.05670713260769844,
0.002749721286818385,
-0.10883764177560806,
-0.17341645061969757,
-0.08353621512651443,
-0.03381705656647682,
0.14265403151512146,
-0.15880316495895386,
0.019960513338446617,
0.05115775763988495,
0.12808771431446075,
0.060330405831336975,
-0.044940851628780365,
-0.0009795452933758497,
0.02373526245355606,
-0.05278978496789932,
-0.09012233465909958,
0.059676408767700195,
0.0331520177423954,
-0.07579167187213898,
-0.016548609361052513,
-0.09850107133388519,
0.1460651308298111,
0.1280234009027481,
-0.010448831133544445,
-0.04986010119318962,
-0.011923554353415966,
-0.06967874616384506,
-0.030430803075432777,
-0.036602724343538284,
0.019139016047120094,
0.09450183063745499,
0.012393946759402752,
0.14818525314331055,
-0.09332848340272903,
-0.034156475216150284,
0.05024607852101326,
-0.028047295287251472,
-0.03259625658392906,
0.0731319710612297,
0.025664178654551506,
-0.14941470324993134,
0.14837577939033508,
0.14845694601535797,
-0.04714515432715416,
0.12564225494861603,
-0.03889495134353638,
-0.06329566240310669,
-0.04632000997662544,
-0.02844901941716671,
0.013190032914280891,
0.13346467912197113,
-0.076783187687397,
-0.004412572830915451,
0.05686868354678154,
0.017921162769198418,
-0.004722983110696077,
-0.1827412098646164,
0.003951311111450195,
0.0321657620370388,
-0.05121494084596634,
-0.011695281602442265,
-0.017026077955961227,
0.003609517589211464,
0.09151934087276459,
0.02040533348917961,
-0.06441836804151535,
0.05384209007024765,
0.012033452279865742,
-0.05366513133049011,
0.1677880585193634,
-0.07823625206947327,
-0.20364677906036377,
-0.12268579006195068,
-0.06752478331327438,
-0.10258819162845612,
0.012170074507594109,
0.06315170973539352,
-0.04569438472390175,
-0.050954580307006836,
-0.0997823104262352,
-0.037851084023714066,
0.021281057968735695,
0.026625970378518105,
0.05139283835887909,
-0.005415658466517925,
0.09185726940631866,
-0.09241294115781784,
-0.030897676944732666,
-0.01631389558315277,
0.009287231601774693,
0.06772445887327194,
0.019780615344643593,
0.1102219671010971,
0.07713042199611664,
-0.029881305992603302,
0.05137522891163826,
-0.013354548253118992,
0.2620471715927124,
-0.06917091459035873,
-0.002909549279138446,
0.1375615894794464,
-0.015162656083703041,
0.08283410966396332,
0.1273423582315445,
0.041794080287218094,
-0.09746479243040085,
-0.011291430331766605,
-0.0008301159832626581,
-0.049490246921777725,
-0.16143162548542023,
-0.04317644611001015,
-0.0434197373688221,
-0.010716320015490055,
0.1416788250207901,
0.03848205506801605,
0.024626927450299263,
0.07702240347862244,
0.015813151374459267,
0.057987019419670105,
-0.02077260985970497,
0.1017511859536171,
0.0805719867348671,
0.06816057115793228,
0.13305824995040894,
-0.036980245262384415,
-0.02092074789106846,
0.057033997029066086,
0.04002218693494797,
0.21362732350826263,
-0.02804172970354557,
0.15433214604854584,
0.026679744943976402,
0.1909136176109314,
0.019870078191161156,
0.07247955352067947,
-0.010095180943608284,
0.0028269465547055006,
-0.018500015139579773,
-0.04554403945803642,
-0.05979170650243759,
0.03185109794139862,
-0.016015755012631416,
0.05207211896777153,
-0.09269700944423676,
0.028567379340529442,
0.06037893891334534,
0.3028397262096405,
0.061388690024614334,
-0.41139692068099976,
-0.09273239970207214,
0.009406263940036297,
-0.002105827210471034,
-0.06053102761507034,
-0.011343861930072308,
0.09683393687009811,
-0.09968853741884232,
0.08300996571779251,
-0.09414921700954437,
0.08760150521993637,
-0.08863518387079239,
0.016419410705566406,
0.07728815078735352,
0.06722814589738846,
0.01766069419682026,
0.057678405195474625,
-0.22131015360355377,
0.2517315745353699,
0.02006395347416401,
0.04867706075310707,
-0.08515261113643646,
0.013813616707921028,
0.029918700456619263,
0.058915551751852036,
0.08619558066129684,
0.0083828279748559,
-0.09208258241415024,
-0.19043345749378204,
-0.12182265520095825,
-0.0015020827995613217,
0.06677291542291641,
-0.03118232637643814,
0.0942893773317337,
-0.01760665327310562,
-0.012930129654705524,
0.019664883613586426,
0.00020212549134157598,
-0.039232417941093445,
-0.09916181117296219,
0.019594477489590645,
0.03770963475108147,
-0.0040510352700948715,
-0.06473120301961899,
-0.1088499054312706,
-0.027749689295887947,
0.1611177921295166,
0.0489477813243866,
-0.07595206052064896,
-0.14163517951965332,
0.0831608697772026,
0.0844789668917656,
-0.08478974550962448,
0.046326830983161926,
-0.015740465372800827,
0.14427345991134644,
0.02813553437590599,
-0.08791226893663406,
0.10567717254161835,
-0.05589807406067848,
-0.18345315754413605,
-0.035460758954286575,
0.09823724627494812,
0.006449915003031492,
0.047238387167453766,
0.0029976284131407738,
0.05834325775504112,
-0.03208146244287491,
-0.05784951522946358,
0.06896662712097168,
-0.0034485149662941694,
0.1075923964381218,
-0.0061480943113565445,
-0.0032397336326539516,
0.02182089537382126,
-0.04197082296013832,
-0.0014782516518607736,
0.1645156890153885,
0.23995232582092285,
-0.10496784001588821,
0.055536478757858276,
0.030249565839767456,
-0.03645236790180206,
-0.18277540802955627,
0.009984065778553486,
0.08414819091558456,
0.0021475672256201506,
0.040169790387153625,
-0.1663118302822113,
0.05386544391512871,
0.10983236879110336,
-0.04191310703754425,
0.07995743304491043,
-0.2803034782409668,
-0.1190505102276802,
0.08906996995210648,
0.13602600991725922,
0.06884066760540009,
-0.13274545967578888,
-0.045290667563676834,
-0.039063699543476105,
-0.16666166484355927,
0.1351267695426941,
-0.04754851385951042,
0.11997194588184357,
-0.040666740387678146,
0.06989686191082001,
0.015085658058524132,
-0.05448267608880997,
0.14587333798408508,
0.00877679605036974,
0.0857420563697815,
-0.07118549197912216,
0.0021252231672406197,
0.10074540972709656,
-0.0982399731874466,
0.07668103277683258,
-0.08308075368404388,
0.06399426609277725,
-0.11283876746892929,
-0.007322354707866907,
-0.07328318059444427,
0.015542288310825825,
-0.012007588520646095,
-0.043488435447216034,
-0.04113076627254486,
0.03472091257572174,
0.06403200328350067,
-0.015996064990758896,
0.20271754264831543,
0.0629286915063858,
0.08313194662332535,
0.17939580976963043,
0.04974674805998802,
-0.096995510160923,
-0.09814400225877762,
-0.04502987116575241,
-0.028452320024371147,
0.06312472373247147,
-0.13321243226528168,
0.05335186421871185,
0.1209464818239212,
0.008661448024213314,
0.12983813881874084,
0.054849762469530106,
-0.0316605418920517,
0.033173978328704834,
0.06366948038339615,
-0.16513317823410034,
-0.08843576163053513,
-0.011303714476525784,
0.01758752204477787,
-0.12545546889305115,
0.0447046272456646,
0.12079240381717682,
-0.057224519550800323,
-0.015418118797242641,
-0.0026640621945261955,
0.03586487099528313,
-0.00886022113263607,
0.16030296683311462,
0.05005719140172005,
0.05675157532095909,
-0.11541767418384552,
0.1181424930691719,
0.06067226454615593,
-0.0710521712899208,
0.031696248799562454,
0.05698402598500252,
-0.10586927086114883,
-0.022646361961960793,
0.03662630170583725,
0.14154238998889923,
-0.06414706259965897,
-0.04990902543067932,
-0.13196614384651184,
-0.0909038558602333,
0.07024894654750824,
0.0724560096859932,
0.09284354001283646,
0.016252439469099045,
-0.031063025817275047,
-0.014114780351519585,
-0.10623957961797714,
0.10545456409454346,
0.04753988981246948,
0.09451808035373688,
-0.17563696205615997,
0.06374634802341461,
0.0007657874375581741,
0.07206296175718307,
-0.024532334879040718,
0.005616967566311359,
-0.09020458161830902,
-0.0008940583793446422,
-0.10660925507545471,
0.025940274819731712,
-0.04968960955739021,
0.0027822551783174276,
-0.020955873653292656,
-0.058104176074266434,
-0.06385789811611176,
0.02704726532101631,
-0.11796805262565613,
-0.05728267878293991,
0.01832517236471176,
0.029680335894227028,
-0.11609132587909698,
-0.04758497327566147,
0.014494677074253559,
-0.09034118801355362,
0.09993617236614227,
0.05929066613316536,
-0.006737631745636463,
0.0029803363140672445,
0.011042662896215916,
-0.02363271825015545,
0.06827948242425919,
0.006517379079014063,
0.07795335352420807,
-0.11366859823465347,
-0.018052512779831886,
0.017967568710446358,
-0.002112566027790308,
0.011524608358740807,
0.15499049425125122,
-0.12699781358242035,
-0.0033930845092982054,
-0.022802060469985008,
-0.06095515564084053,
-0.06754840165376663,
0.06765563786029816,
0.10613249987363815,
0.0214694757014513,
0.2064255326986313,
-0.054858945310115814,
0.01148067507892847,
-0.21229742467403412,
-0.011367390863597393,
0.0014767643297091126,
-0.1394193321466446,
-0.10240225493907928,
-0.03432944789528847,
0.0646229088306427,
-0.07021024078130722,
0.1212792620062828,
0.036924295127391815,
0.015180133283138275,
0.028698688372969627,
0.025451842695474625,
-0.009322993457317352,
0.01828060857951641,
0.16467928886413574,
0.014544252306222916,
-0.030929861590266228,
0.12307319045066833,
0.026831358671188354,
0.0918813943862915,
0.11550118029117584,
0.17162561416625977,
0.1226300448179245,
0.042329173535108566,
0.09527058154344559,
0.05073356628417969,
-0.032373297959566116,
-0.2198440134525299,
0.04109371080994606,
-0.043747998774051666,
0.14987531304359436,
-0.0034218686632812023,
0.15886609256267548,
0.08696271479129791,
-0.1824999451637268,
0.04266338422894478,
-0.02988567017018795,
-0.08202743530273438,
-0.08238054066896439,
-0.1163601353764534,
-0.10495591163635254,
-0.15148837864398956,
0.0012598474277183414,
-0.10238117724657059,
0.02373862825334072,
0.11528778076171875,
-0.010980993509292603,
-0.00952758826315403,
0.1250862330198288,
-0.01644187793135643,
0.019042596220970154,
0.04508042708039284,
0.007425562012940645,
-0.05218745768070221,
-0.04613304138183594,
-0.08413935452699661,
0.015972480177879333,
0.0363130047917366,
0.05680973082780838,
-0.03208919242024422,
-0.008708061650395393,
0.03847881406545639,
-0.008026620373129845,
-0.12142552435398102,
0.013289375230669975,
0.007551861461251974,
0.04767835885286331,
-0.004989264067262411,
0.007813788950443268,
0.026865217834711075,
-0.01780105195939541,
0.195222407579422,
-0.06977689266204834,
-0.02860948257148266,
-0.12041912227869034,
0.17737813293933868,
0.00569287920370698,
-0.048185933381319046,
0.05394943431019783,
-0.09105358272790909,
-0.02213868498802185,
0.15108588337898254,
0.18787547945976257,
-0.06683575361967087,
-0.017941389232873917,
-0.014669668860733509,
-0.01477136928588152,
-0.01832989603281021,
0.10442051291465759,
0.09986825287342072,
-0.004740583244711161,
-0.07264549285173416,
-0.024389909580349922,
-0.06369390338659286,
-0.032235804945230484,
-0.04127946496009827,
0.07026855647563934,
-0.001124961650930345,
0.005972458980977535,
-0.07571399211883545,
0.03954308480024338,
-0.020357538014650345,
-0.06112333759665489,
0.07204564660787582,
-0.21083933115005493,
-0.1802441030740738,
0.0017737408634275198,
0.07683850824832916,
0.0021866720635443926,
0.04613208398222923,
-0.012570524588227272,
0.018509654328227043,
0.07427240163087845,
-0.02333001233637333,
-0.08794470131397247,
-0.09525144845247269,
0.1020299568772316,
-0.13951729238033295,
0.24700812995433807,
-0.03552914783358574,
0.0377071388065815,
0.1201176866889,
0.03583609312772751,
-0.13580889999866486,
0.03513867408037186,
0.03722600266337395,
-0.02918340638279915,
0.0181744247674942,
0.14616045355796814,
-0.03901152312755585,
0.07440102845430374,
0.04275068640708923,
-0.10678882896900177,
-0.04424819350242615,
-0.04619530588388443,
-0.015570126473903656,
-0.02712010033428669,
-0.05963090807199478,
-0.04089967906475067,
0.12949442863464355,
0.17410574853420258,
-0.04094170406460762,
-0.021948745474219322,
-0.06438223272562027,
0.035308949649333954,
0.08067496865987778,
-0.026465818285942078,
-0.04482371732592583,
-0.2364819198846817,
0.0028874515555799007,
0.050913918763399124,
-0.008316555991768837,
-0.19871793687343597,
-0.10607530176639557,
-0.00044736277777701616,
-0.05943094193935394,
-0.08227076381444931,
0.09325046092271805,
0.06211918964982033,
0.03563893958926201,
-0.06190048158168793,
0.02738066203892231,
-0.07750356942415237,
0.14178979396820068,
-0.14600589871406555,
-0.07656177133321762
] |
null | null | sentence-transformers |
For more details please refer to our github repo: https://github.com/FlagOpen/FlagEmbedding
# BGE-M3
In this project, we introduce BGE-M3, which is distinguished for its versatility in Multi-Functionality, Multi-Linguality, and Multi-Granularity.
- Multi-Functionality: It can simultaneously perform the three common retrieval functionalities of embedding model: dense retrieval, multi-vector retrieval, and sparse retrieval.
- Multi-Linguality: It can support more than 100 working languages.
- Multi-Granularity: It is able to process inputs of different granularities, spanning from short sentences to long documents of up to 8192 tokens.
**Some suggestions for retrieval pipeline in RAG:**
We recommend to use following pipeline: hybrid retrieval + re-ranking.
- Hybrid retrieval leverages the strengths of various methods, offering higher accuracy and stronger generalization capabilities.
A classic example: using both embedding retrieval and the BM25 algorithm.
Now, you can try to use BGE-M3, which supports both embedding and sparse retrieval.
This allows you to obtain token weights (similar to the BM25) without any additional cost when generate dense embeddings.
- As cross-encoder models, re-ranker demonstrates higher accuracy than bi-encoder embedding model.
Utilizing the re-ranking model (e.g., [bge-reranker](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/reranker), [cohere-reranker](https://txt.cohere.com/rerank/)) after retrieval can further filter the selected text.
## News:
- 2/1/2024: **Thanks for the excellent tool from Vespa.** You can easily use multiple modes of BGE-M3 following this [notebook](https://github.com/vespa-engine/pyvespa/blob/master/docs/sphinx/source/examples/mother-of-all-embedding-models-cloud.ipynb)
## Model Specs
| Model Name | Dimension | Sequence Length |
|:----:|:---:|:---:|
| [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3) | 1024 | 8192 |
| [BAAI/bge-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5) | 1024 | 512 |
| [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 768 | 512 |
| [BAAI/bge-small-en-v1.5](https://huggingface.co/BAAI/bge-small-en-v1.5) | 384 | 512 |
## FAQ
**1. Introduction for different retrieval methods**
- Dense retrieval: map the text into a single embedding, e.g., [DPR](https://arxiv.org/abs/2004.04906), [BGE-v1.5](https://github.com/FlagOpen/FlagEmbedding)
- Sparse retrieval (lexical matching): a vector of size equal to the vocabulary, with the majority of positions set to zero, calculating a weight only for tokens present in the text. e.g., BM25, [unicoil](https://arxiv.org/pdf/2106.14807.pdf), and [splade](https://arxiv.org/abs/2107.05720)
- Multi-vector retrieval: use multiple vectors to represent a text, e.g., [ColBERT](https://arxiv.org/abs/2004.12832).
**2. Comparison with BGE-v1.5 and other monolingual models**
BGE-M3 is a multilingual model, and its ability in monolingual embedding retrieval may not surpass models specifically designed for single languages.
However, we still recommend trying BGE-M3 because of its versatility (support for multiple languages and long texts).
Moreover, it can simultaneously generate multiple representations, and using them together can enhance accuracy and generalization,
unlike most existing models that can only perform dense retrieval.
In the open-source community, there are many excellent models (e.g., jina-embedding, colbert, e5, etc),
and users can choose a model that suits their specific needs based on practical considerations,
such as whether to require multilingual or cross-language support, and whether to process long texts.
**3. How to use BGE-M3 in other projects?**
For embedding retrieval, you can employ the BGE-M3 model using the same approach as BGE.
The only difference is that the BGE-M3 model no longer requires adding instructions to the queries.
For sparse retrieval methods, most open-source libraries currently do not support direct utilization of the BGE-M3 model.
Contributions from the community are welcome.
**4. How to fine-tune bge-M3 model?**
You can follow the common in this [example](https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/finetune)
to fine-tune the dense embedding.
Our code and data for unified fine-tuning (dense, sparse, and multi-vectors) will be released.
## Usage
Install:
```
git clone https://github.com/FlagOpen/FlagEmbedding.git
cd FlagEmbedding
pip install -e .
```
or:
```
pip install -U FlagEmbedding
```
### Generate Embedding for text
- Dense Embedding
```python
from FlagEmbedding import BGEM3FlagModel
model = BGEM3FlagModel('BAAI/bge-m3',
use_fp16=True) # Setting use_fp16 to True speeds up computation with a slight performance degradation
sentences_1 = ["What is BGE M3?", "Defination of BM25"]
sentences_2 = ["BGE M3 is an embedding model supporting dense retrieval, lexical matching and multi-vector interaction.",
"BM25 is a bag-of-words retrieval function that ranks a set of documents based on the query terms appearing in each document"]
embeddings_1 = model.encode(sentences_1,
batch_size=12,
max_length=8192, # If you don't need such a long length, you can set a smaller value to speed up the encoding process.
)['dense_vecs']
embeddings_2 = model.encode(sentences_2)['dense_vecs']
similarity = embeddings_1 @ embeddings_2.T
print(similarity)
# [[0.6265, 0.3477], [0.3499, 0.678 ]]
```
You also can use sentence-transformers and huggingface transformers to generate dense embeddings.
Refer to [baai_general_embedding](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/baai_general_embedding#usage) for details.
- Sparse Embedding (Lexical Weight)
```python
from FlagEmbedding import BGEM3FlagModel
model = BGEM3FlagModel('BAAI/bge-m3', use_fp16=True) # Setting use_fp16 to True speeds up computation with a slight performance degradation
sentences_1 = ["What is BGE M3?", "Defination of BM25"]
sentences_2 = ["BGE M3 is an embedding model supporting dense retrieval, lexical matching and multi-vector interaction.",
"BM25 is a bag-of-words retrieval function that ranks a set of documents based on the query terms appearing in each document"]
output_1 = model.encode(sentences_1, return_dense=True, return_sparse=True, return_colbert_vecs=False)
output_2 = model.encode(sentences_2, return_dense=True, return_sparse=True, return_colbert_vecs=False)
# you can see the weight for each token:
print(model.convert_id_to_token(output_1['lexical_weights']))
# [{'What': 0.08356, 'is': 0.0814, 'B': 0.1296, 'GE': 0.252, 'M': 0.1702, '3': 0.2695, '?': 0.04092},
# {'De': 0.05005, 'fin': 0.1368, 'ation': 0.04498, 'of': 0.0633, 'BM': 0.2515, '25': 0.3335}]
# compute the scores via lexical mathcing
lexical_scores = model.compute_lexical_matching_score(output_1['lexical_weights'][0], output_2['lexical_weights'][0])
print(lexical_scores)
# 0.19554901123046875
print(model.compute_lexical_matching_score(output_1['lexical_weights'][0], output_1['lexical_weights'][1]))
# 0.0
```
- Multi-Vector (ColBERT)
```python
from FlagEmbedding import BGEM3FlagModel
model = BGEM3FlagModel('BAAI/bge-m3', use_fp16=True)
sentences_1 = ["What is BGE M3?", "Defination of BM25"]
sentences_2 = ["BGE M3 is an embedding model supporting dense retrieval, lexical matching and multi-vector interaction.",
"BM25 is a bag-of-words retrieval function that ranks a set of documents based on the query terms appearing in each document"]
output_1 = model.encode(sentences_1, return_dense=True, return_sparse=True, return_colbert_vecs=True)
output_2 = model.encode(sentences_2, return_dense=True, return_sparse=True, return_colbert_vecs=True)
print(model.colbert_score(output_1['colbert_vecs'][0], output_2['colbert_vecs'][0]))
print(model.colbert_score(output_1['colbert_vecs'][0], output_2['colbert_vecs'][1]))
# 0.7797
# 0.4620
```
### Compute score for text pairs
Input a list of text pairs, you can get the scores computed by different methods.
```python
from FlagEmbedding import BGEM3FlagModel
model = BGEM3FlagModel('BAAI/bge-m3', use_fp16=True)
sentences_1 = ["What is BGE M3?", "Defination of BM25"]
sentences_2 = ["BGE M3 is an embedding model supporting dense retrieval, lexical matching and multi-vector interaction.",
"BM25 is a bag-of-words retrieval function that ranks a set of documents based on the query terms appearing in each document"]
sentence_pairs = [[i,j] for i in sentences_1 for j in sentences_2]
print(model.compute_score(sentence_pairs,
max_passage_length=128, # a smaller max length leads to a lower latency
weights_for_different_modes=[0.4, 0.2, 0.4])) # weights_for_different_modes(w) is used to do weighted sum: w[0]*dense_score + w[1]*sparse_score + w[2]*colbert_score
# {
# 'colbert': [0.7796499729156494, 0.4621465802192688, 0.4523794651031494, 0.7898575067520142],
# 'sparse': [0.195556640625, 0.00879669189453125, 0.0, 0.1802978515625],
# 'dense': [0.6259765625, 0.347412109375, 0.349853515625, 0.67822265625],
# 'sparse+dense': [0.482503205537796, 0.23454029858112335, 0.2332356721162796, 0.5122477412223816],
# 'colbert+sparse+dense': [0.6013619303703308, 0.3255828022956848, 0.32089319825172424, 0.6232916116714478]
# }
```
## Evaluation
- Multilingual (Miracl dataset)

- Cross-lingual (MKQA dataset)

- Long Document Retrieval
- MLDR:

Please note that MLDR is a document retrieval dataset we constructed via LLM,
covering 13 languages, including test set, validation set, and training set.
We utilized the training set from MLDR to enhance the model's long document retrieval capabilities.
Therefore, comparing baseline with `Dense w.o.long`(fine-tuning without long document dataset) is more equitable.
Additionally, this long document retrieval dataset will be open-sourced to address the current lack of open-source multilingual long text retrieval datasets.
We believe that this data will be helpful for the open-source community in training document retrieval models.
- NarritiveQA:

## Training
- Self-knowledge Distillation: combining multiple outputs from different
retrieval modes as reward signal to enhance the performance of single mode(especially for sparse retrieval and multi-vec(colbert) retrival)
- Efficient Batching: Improve the efficiency when fine-tuning on long text.
The small-batch strategy is simple but effective, which also can used to fine-tune large embedding model.
- MCLS: A simple method to improve the performance on long text without fine-tuning.
If you have no enough resource to fine-tuning model with long text, the method is useful.
Refer to our [report](https://github.com/FlagOpen/FlagEmbedding/blob/master/FlagEmbedding/BGE_M3/BGE_M3.pdf) for more details.
**The fine-tuning codes and datasets will be open-sourced in the near future.**
## Models
We release two versions:
- BAAI/bge-m3-unsupervised: the model after contrastive learning in a large-scale dataset
- BAAI/bge-m3: the final model fine-tuned from BAAI/bge-m3-unsupervised
## Acknowledgement
Thanks the authors of open-sourced datasets, including Miracl, MKQA, NarritiveQA, etc.
## Citation
If you find this repository useful, please consider giving a star :star: and citation
```
``` | {"license": "mit", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity"], "pipeline_tag": "sentence-similarity"} | sentence-similarity | Enno-Ai/bge-m3 | [
"sentence-transformers",
"safetensors",
"xlm-roberta",
"feature-extraction",
"sentence-similarity",
"arxiv:2004.04906",
"arxiv:2106.14807",
"arxiv:2107.05720",
"arxiv:2004.12832",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:32:22+00:00 | [
"2004.04906",
"2106.14807",
"2107.05720",
"2004.12832"
] | [] | TAGS
#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #arxiv-2004.04906 #arxiv-2106.14807 #arxiv-2107.05720 #arxiv-2004.12832 #license-mit #endpoints_compatible #region-us
| For more details please refer to our github repo: URL
BGE-M3
======
In this project, we introduce BGE-M3, which is distinguished for its versatility in Multi-Functionality, Multi-Linguality, and Multi-Granularity.
* Multi-Functionality: It can simultaneously perform the three common retrieval functionalities of embedding model: dense retrieval, multi-vector retrieval, and sparse retrieval.
* Multi-Linguality: It can support more than 100 working languages.
* Multi-Granularity: It is able to process inputs of different granularities, spanning from short sentences to long documents of up to 8192 tokens.
Some suggestions for retrieval pipeline in RAG:
We recommend to use following pipeline: hybrid retrieval + re-ranking.
* Hybrid retrieval leverages the strengths of various methods, offering higher accuracy and stronger generalization capabilities.
A classic example: using both embedding retrieval and the BM25 algorithm.
Now, you can try to use BGE-M3, which supports both embedding and sparse retrieval.
This allows you to obtain token weights (similar to the BM25) without any additional cost when generate dense embeddings.
* As cross-encoder models, re-ranker demonstrates higher accuracy than bi-encoder embedding model.
Utilizing the re-ranking model (e.g., bge-reranker, cohere-reranker) after retrieval can further filter the selected text.
News:
-----
* 2/1/2024: Thanks for the excellent tool from Vespa. You can easily use multiple modes of BGE-M3 following this notebook
Model Specs
-----------
FAQ
---
1. Introduction for different retrieval methods
* Dense retrieval: map the text into a single embedding, e.g., DPR, BGE-v1.5
* Sparse retrieval (lexical matching): a vector of size equal to the vocabulary, with the majority of positions set to zero, calculating a weight only for tokens present in the text. e.g., BM25, unicoil, and splade
* Multi-vector retrieval: use multiple vectors to represent a text, e.g., ColBERT.
2. Comparison with BGE-v1.5 and other monolingual models
BGE-M3 is a multilingual model, and its ability in monolingual embedding retrieval may not surpass models specifically designed for single languages.
However, we still recommend trying BGE-M3 because of its versatility (support for multiple languages and long texts).
Moreover, it can simultaneously generate multiple representations, and using them together can enhance accuracy and generalization,
unlike most existing models that can only perform dense retrieval.
In the open-source community, there are many excellent models (e.g., jina-embedding, colbert, e5, etc),
and users can choose a model that suits their specific needs based on practical considerations,
such as whether to require multilingual or cross-language support, and whether to process long texts.
3. How to use BGE-M3 in other projects?
For embedding retrieval, you can employ the BGE-M3 model using the same approach as BGE.
The only difference is that the BGE-M3 model no longer requires adding instructions to the queries.
For sparse retrieval methods, most open-source libraries currently do not support direct utilization of the BGE-M3 model.
Contributions from the community are welcome.
4. How to fine-tune bge-M3 model?
You can follow the common in this example
to fine-tune the dense embedding.
Our code and data for unified fine-tuning (dense, sparse, and multi-vectors) will be released.
Usage
-----
Install:
or:
### Generate Embedding for text
* Dense Embedding
You also can use sentence-transformers and huggingface transformers to generate dense embeddings.
Refer to baai\_general\_embedding for details.
* Sparse Embedding (Lexical Weight)
* Multi-Vector (ColBERT)
### Compute score for text pairs
Input a list of text pairs, you can get the scores computed by different methods.
Evaluation
----------
* Multilingual (Miracl dataset)
!avatar
* Cross-lingual (MKQA dataset)
!avatar
* Long Document Retrieval
+ MLDR:
!avatar
Please note that MLDR is a document retrieval dataset we constructed via LLM,
covering 13 languages, including test set, validation set, and training set.
We utilized the training set from MLDR to enhance the model's long document retrieval capabilities.
Therefore, comparing baseline with 'Dense w.o.long'(fine-tuning without long document dataset) is more equitable.
Additionally, this long document retrieval dataset will be open-sourced to address the current lack of open-source multilingual long text retrieval datasets.
We believe that this data will be helpful for the open-source community in training document retrieval models.
+ NarritiveQA:
!avatar
Training
--------
* Self-knowledge Distillation: combining multiple outputs from different
retrieval modes as reward signal to enhance the performance of single mode(especially for sparse retrieval and multi-vec(colbert) retrival)
* Efficient Batching: Improve the efficiency when fine-tuning on long text.
The small-batch strategy is simple but effective, which also can used to fine-tune large embedding model.
* MCLS: A simple method to improve the performance on long text without fine-tuning.
If you have no enough resource to fine-tuning model with long text, the method is useful.
Refer to our report for more details.
The fine-tuning codes and datasets will be open-sourced in the near future.
Models
------
We release two versions:
* BAAI/bge-m3-unsupervised: the model after contrastive learning in a large-scale dataset
* BAAI/bge-m3: the final model fine-tuned from BAAI/bge-m3-unsupervised
Acknowledgement
---------------
Thanks the authors of open-sourced datasets, including Miracl, MKQA, NarritiveQA, etc.
If you find this repository useful, please consider giving a star :star: and citation
| [
"### Generate Embedding for text\n\n\n* Dense Embedding\n\n\nYou also can use sentence-transformers and huggingface transformers to generate dense embeddings.\nRefer to baai\\_general\\_embedding for details.\n\n\n* Sparse Embedding (Lexical Weight)\n* Multi-Vector (ColBERT)",
"### Compute score for text pairs\n\n\nInput a list of text pairs, you can get the scores computed by different methods.\n\n\nEvaluation\n----------\n\n\n* Multilingual (Miracl dataset)\n\n\n!avatar\n\n\n* Cross-lingual (MKQA dataset)\n\n\n!avatar\n\n\n* Long Document Retrieval\n\t+ MLDR: \n\t\n\t!avatar\n\tPlease note that MLDR is a document retrieval dataset we constructed via LLM,\n\tcovering 13 languages, including test set, validation set, and training set.\n\tWe utilized the training set from MLDR to enhance the model's long document retrieval capabilities.\n\tTherefore, comparing baseline with 'Dense w.o.long'(fine-tuning without long document dataset) is more equitable.\n\tAdditionally, this long document retrieval dataset will be open-sourced to address the current lack of open-source multilingual long text retrieval datasets.\n\tWe believe that this data will be helpful for the open-source community in training document retrieval models.\n\t+ NarritiveQA: \n\t\n\t!avatar\n\n\nTraining\n--------\n\n\n* Self-knowledge Distillation: combining multiple outputs from different\nretrieval modes as reward signal to enhance the performance of single mode(especially for sparse retrieval and multi-vec(colbert) retrival)\n* Efficient Batching: Improve the efficiency when fine-tuning on long text.\nThe small-batch strategy is simple but effective, which also can used to fine-tune large embedding model.\n* MCLS: A simple method to improve the performance on long text without fine-tuning.\nIf you have no enough resource to fine-tuning model with long text, the method is useful.\n\n\nRefer to our report for more details.\n\n\nThe fine-tuning codes and datasets will be open-sourced in the near future.\n\n\nModels\n------\n\n\nWe release two versions:\n\n\n* BAAI/bge-m3-unsupervised: the model after contrastive learning in a large-scale dataset\n* BAAI/bge-m3: the final model fine-tuned from BAAI/bge-m3-unsupervised\n\n\nAcknowledgement\n---------------\n\n\nThanks the authors of open-sourced datasets, including Miracl, MKQA, NarritiveQA, etc.\n\n\nIf you find this repository useful, please consider giving a star :star: and citation"
] | [
"TAGS\n#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #arxiv-2004.04906 #arxiv-2106.14807 #arxiv-2107.05720 #arxiv-2004.12832 #license-mit #endpoints_compatible #region-us \n",
"### Generate Embedding for text\n\n\n* Dense Embedding\n\n\nYou also can use sentence-transformers and huggingface transformers to generate dense embeddings.\nRefer to baai\\_general\\_embedding for details.\n\n\n* Sparse Embedding (Lexical Weight)\n* Multi-Vector (ColBERT)",
"### Compute score for text pairs\n\n\nInput a list of text pairs, you can get the scores computed by different methods.\n\n\nEvaluation\n----------\n\n\n* Multilingual (Miracl dataset)\n\n\n!avatar\n\n\n* Cross-lingual (MKQA dataset)\n\n\n!avatar\n\n\n* Long Document Retrieval\n\t+ MLDR: \n\t\n\t!avatar\n\tPlease note that MLDR is a document retrieval dataset we constructed via LLM,\n\tcovering 13 languages, including test set, validation set, and training set.\n\tWe utilized the training set from MLDR to enhance the model's long document retrieval capabilities.\n\tTherefore, comparing baseline with 'Dense w.o.long'(fine-tuning without long document dataset) is more equitable.\n\tAdditionally, this long document retrieval dataset will be open-sourced to address the current lack of open-source multilingual long text retrieval datasets.\n\tWe believe that this data will be helpful for the open-source community in training document retrieval models.\n\t+ NarritiveQA: \n\t\n\t!avatar\n\n\nTraining\n--------\n\n\n* Self-knowledge Distillation: combining multiple outputs from different\nretrieval modes as reward signal to enhance the performance of single mode(especially for sparse retrieval and multi-vec(colbert) retrival)\n* Efficient Batching: Improve the efficiency when fine-tuning on long text.\nThe small-batch strategy is simple but effective, which also can used to fine-tune large embedding model.\n* MCLS: A simple method to improve the performance on long text without fine-tuning.\nIf you have no enough resource to fine-tuning model with long text, the method is useful.\n\n\nRefer to our report for more details.\n\n\nThe fine-tuning codes and datasets will be open-sourced in the near future.\n\n\nModels\n------\n\n\nWe release two versions:\n\n\n* BAAI/bge-m3-unsupervised: the model after contrastive learning in a large-scale dataset\n* BAAI/bge-m3: the final model fine-tuned from BAAI/bge-m3-unsupervised\n\n\nAcknowledgement\n---------------\n\n\nThanks the authors of open-sourced datasets, including Miracl, MKQA, NarritiveQA, etc.\n\n\nIf you find this repository useful, please consider giving a star :star: and citation"
] | [
82,
74,
537
] | [
"passage: TAGS\n#sentence-transformers #safetensors #xlm-roberta #feature-extraction #sentence-similarity #arxiv-2004.04906 #arxiv-2106.14807 #arxiv-2107.05720 #arxiv-2004.12832 #license-mit #endpoints_compatible #region-us \n### Generate Embedding for text\n\n\n* Dense Embedding\n\n\nYou also can use sentence-transformers and huggingface transformers to generate dense embeddings.\nRefer to baai\\_general\\_embedding for details.\n\n\n* Sparse Embedding (Lexical Weight)\n* Multi-Vector (ColBERT)"
] | [
-0.015302288345992565,
-0.12127773463726044,
-0.004402799066156149,
0.03694140538573265,
0.09887946397066116,
0.05450037866830826,
0.12041968107223511,
0.08347742259502411,
-0.05488051846623421,
0.10220537334680557,
0.1520014852285385,
0.10382281243801117,
-0.0020213108509778976,
0.03403595834970474,
-0.03663105517625809,
-0.22993457317352295,
0.012917560525238514,
0.10010568052530289,
0.03129306435585022,
0.023388879373669624,
0.10304318368434906,
-0.07168123871088028,
0.11963491141796112,
-0.017296545207500458,
-0.1537988930940628,
0.0774555429816246,
-0.006167071871459484,
-0.029808636754751205,
0.09424160420894623,
0.09643533825874329,
0.08189351111650467,
-0.0020407182164490223,
-0.026809850707650185,
-0.24652712047100067,
0.011666369624435902,
0.00008430904563283548,
-0.11003535985946655,
0.06600989401340485,
0.028227541595697403,
-0.08370664715766907,
0.08088923990726471,
-0.17285844683647156,
0.003503235522657633,
0.04629344493150711,
-0.07393712550401688,
-0.13674084842205048,
0.01733039878308773,
-0.045743536204099655,
0.050176143646240234,
-0.017087945714592934,
-0.050148047506809235,
0.011867912486195564,
-0.011970866471529007,
0.08305096626281738,
0.22632794082164764,
-0.36234134435653687,
-0.021662911400198936,
0.06466101109981537,
0.07443846762180328,
0.09714210778474808,
-0.07106523960828781,
0.056406278163194656,
0.09584777057170868,
0.034169752150774,
0.11167863756418228,
-0.04896853119134903,
0.05046946555376053,
0.039722271263599396,
-0.11452827602624893,
0.1031564474105835,
0.11846687644720078,
0.019179729744791985,
0.008356177248060703,
-0.1147414892911911,
-0.0982903391122818,
-0.004958157893270254,
-0.05461294576525688,
0.016105905175209045,
0.04069352522492409,
0.03059421479701996,
-0.0037325755693018436,
-0.025979960337281227,
-0.10369575023651123,
-0.027783062309026718,
-0.13871194422245026,
0.04822540283203125,
0.020600667223334312,
-0.036529622972011566,
-0.01901637203991413,
-0.01839752495288849,
-0.1598609983921051,
-0.1353682577610016,
0.036394257098436356,
-0.052389055490493774,
0.04802914336323738,
0.062483519315719604,
-0.03435438126325607,
-0.10717175900936127,
0.09835737198591232,
0.023259857669472694,
0.0766313299536705,
0.08689908683300018,
-0.07790861278772354,
0.09117919951677322,
-0.07366540282964706,
0.11620469391345978,
-0.03956330195069313,
-0.09691338986158371,
0.023351257666945457,
-0.034283947199583054,
0.09070835262537003,
-0.023856788873672485,
-0.18940940499305725,
-0.008830470964312553,
-0.0034401719458401203,
0.02670641802251339,
-0.07086767256259918,
0.04866701364517212,
-0.08085550367832184,
0.05559638887643814,
-0.02146768383681774,
-0.05163775011897087,
0.04476453736424446,
-0.048941586166620255,
0.10523940622806549,
0.028837472200393677,
0.04341420531272888,
0.01576637104153633,
-0.01646038517355919,
0.07218270003795624,
-0.11278349161148071,
0.041043300181627274,
-0.047996215522289276,
-0.1538926512002945,
0.015251673758029938,
-0.11044763028621674,
0.011021208018064499,
-0.11281571537256241,
-0.009734532795846462,
0.03452330455183983,
-0.004415729083120823,
-0.03937528282403946,
0.10116580873727798,
-0.1048063188791275,
-0.006012539844959974,
0.05515178292989731,
-0.01118846982717514,
-0.06739906966686249,
-0.008816120214760303,
0.03450148180127144,
0.06896545737981796,
0.07825630903244019,
-0.14600685238838196,
0.02762085571885109,
-0.08434339612722397,
0.02552870288491249,
-0.10467026382684708,
0.08604929596185684,
-0.030904611572623253,
0.042862433940172195,
-0.06817074865102768,
-0.01335667259991169,
-0.045564573258161545,
0.05332580953836441,
0.050430066883563995,
0.08609034866094589,
-0.23719243705272675,
-0.07903364300727844,
0.2003832310438156,
-0.05613061785697937,
-0.1703454703092575,
0.18773989379405975,
-0.032974325120449066,
-0.017454514279961586,
0.1003006100654602,
0.20825786888599396,
-0.09263075888156891,
-0.04337639734148979,
0.026762638241052628,
0.0749758929014206,
-0.002627238165587187,
0.0816602036356926,
0.09830605983734131,
-0.009166165255010128,
-0.07496694475412369,
0.017453370615839958,
0.017095986753702164,
0.037413615733385086,
-0.01869046315550804,
-0.006726394407451153,
-0.02236391231417656,
0.002711871638894081,
0.09565896540880203,
-0.021807963028550148,
-0.014299744740128517,
-0.10288472473621368,
-0.11059371381998062,
0.04265762120485306,
0.014177478849887848,
-0.13166311383247375,
0.052307404577732086,
-0.026723414659500122,
0.0504768006503582,
-0.16023117303848267,
-0.008880517445504665,
-0.09480924904346466,
-0.02957732416689396,
0.016983598470687866,
0.17426809668540955,
0.10376788675785065,
0.025137851014733315,
0.043968942016363144,
0.09059842675924301,
-0.07181435078382492,
-0.005681317299604416,
0.11369398981332779,
0.0067739286459982395,
-0.09806326776742935,
-0.11411678791046143,
0.0232053492218256,
-0.08427007496356964,
0.001986198825761676,
-0.0918002501130104,
0.036414001137018204,
-0.01401278842240572,
0.01986752077937126,
0.04552038758993149,
0.023160038515925407,
-0.04770198091864586,
0.013946217484772205,
-0.07261568307876587,
0.025126084685325623,
0.07448750734329224,
-0.05120227485895157,
-0.2213248759508133,
0.1304241120815277,
-0.2175101935863495,
0.14365479350090027,
0.1248554214835167,
-0.06464780867099762,
0.04172218218445778,
-0.2123090773820877,
0.009422230534255505,
0.042985714972019196,
0.020630531013011932,
-0.10457383096218109,
0.20222635567188263,
-0.029916882514953613,
0.10468482971191406,
-0.12748245894908905,
-0.00486591923981905,
-0.02789447084069252,
-0.026333943009376526,
-0.06580602377653122,
0.14964559674263,
-0.021796466782689095,
-0.1367272138595581,
0.10616648197174072,
0.22721734642982483,
-0.04157336801290512,
0.08522610366344452,
0.021341990679502487,
-0.04509441927075386,
-0.02174699865281582,
0.019805297255516052,
0.027513742446899414,
0.018221400678157806,
-0.13100743293762207,
-0.028376689180731773,
0.0214646365493536,
-0.007260962389409542,
0.06423518806695938,
-0.08597299456596375,
0.0007556000491604209,
0.04477068781852722,
-0.02308041602373123,
0.014644250273704529,
0.03323108330368996,
-0.027471911162137985,
0.07941611111164093,
-0.04408887401223183,
0.013842674903571606,
0.015493394806981087,
0.007796184625476599,
-0.09450697898864746,
0.16973501443862915,
-0.17408303916454315,
-0.1492345631122589,
-0.18100309371948242,
-0.12691630423069,
-0.07377754151821136,
0.04496560990810394,
0.13392984867095947,
-0.13627780973911285,
-0.039204515516757965,
-0.10221832990646362,
0.16453580558300018,
-0.0013692259090021253,
0.017513427883386612,
-0.05184178426861763,
-0.003072053659707308,
-0.03881283476948738,
-0.15890219807624817,
-0.050981324166059494,
-0.04937630146741867,
0.02802141383290291,
0.03667842224240303,
-0.06244348734617233,
0.1061447486281395,
0.10663087666034698,
0.029506761580705643,
0.009401921182870865,
-0.07304792106151581,
0.10161156952381134,
0.013510306365787983,
-0.006115994416177273,
0.13376715779304504,
-0.0324939601123333,
0.08421893417835236,
0.11394686996936798,
0.04927322268486023,
-0.0888681560754776,
0.02033229172229767,
0.01605101488530636,
-0.02447398379445076,
-0.1360182911157608,
-0.08888275921344757,
-0.1166769489645958,
0.02954862080514431,
0.04511961713433266,
0.011317253112792969,
0.008500693365931511,
0.07339823246002197,
-0.024978747591376305,
0.06518159061670303,
0.07306992262601852,
0.10972334444522858,
0.1159919872879982,
-0.006092987954616547,
0.10321186482906342,
-0.028676483780145645,
-0.14215759932994843,
0.044274069368839264,
-0.05229794979095459,
0.13728806376457214,
0.05643660947680473,
0.14875584840774536,
0.05360657721757889,
-0.019441314041614532,
0.01771029829978943,
0.11752685159444809,
-0.1323091834783554,
-0.027479171752929688,
-0.07942467927932739,
-0.09752703458070755,
0.04166964814066887,
0.0794864147901535,
0.029200782999396324,
-0.0006633948651142418,
-0.05167193338274956,
0.008893744088709354,
0.1393081694841385,
-0.05935041233897209,
0.08357802033424377,
-0.29403066635131836,
0.044754981994628906,
0.1157391220331192,
0.016792524605989456,
-0.08019917458295822,
0.013033824972808361,
0.12457749992609024,
0.01982053555548191,
0.11048950254917145,
-0.037615906447172165,
0.07819054275751114,
0.10579685866832733,
0.05204655975103378,
-0.17445588111877441,
0.130466029047966,
-0.006909996271133423,
0.07058722525835037,
-0.28293341398239136,
0.1516517549753189,
0.07533299177885056,
-0.03201042488217354,
-0.06909526139497757,
0.02426941692829132,
0.060087837278842926,
0.029829824343323708,
0.1423766314983368,
-0.06655299663543701,
0.03559982031583786,
-0.026348724961280823,
-0.028479835018515587,
0.07231874018907547,
0.141896590590477,
-0.07920265197753906,
0.040302522480487823,
-0.01703738607466221,
-0.022976338863372803,
0.05423346906900406,
0.12431405484676361,
-0.09706536680459976,
-0.19673749804496765,
-0.025480765849351883,
0.08910486847162247,
-0.03397911787033081,
0.010728336870670319,
-0.024263616651296616,
-0.04488658905029297,
0.11209895461797714,
-0.14774835109710693,
-0.057000625878572464,
-0.06739100813865662,
0.00607362249866128,
0.09632064402103424,
-0.08592704683542252,
0.010824499651789665,
-0.0340343676507473,
0.1652938723564148,
-0.13841941952705383,
-0.12423662841320038,
0.12708547711372375,
-0.04367917776107788,
0.004303146619349718,
-0.06662113964557648,
0.14751394093036652,
-0.04679189249873161,
0.0687347799539566,
0.020804213359951973,
0.03517443314194679,
-0.041880495846271515,
-0.07750310003757477,
-0.020175812765955925,
0.06650297343730927,
0.13705798983573914,
0.053837500512599945,
-0.23554372787475586,
-0.01907818764448166,
-0.05836181715130806,
0.04829103499650955,
0.25734466314315796,
0.2301822006702423,
-0.061071015894412994,
0.03934427350759506,
0.11220848560333252,
-0.07379887253046036,
-0.23404379189014435,
-0.07488147914409637,
-0.02835497260093689,
0.04929531365633011,
0.045465923845767975,
-0.013876447454094887,
0.1188807263970375,
0.0636511743068695,
0.05009535700082779,
0.007820804603397846,
-0.16443249583244324,
-0.14768941700458527,
0.08090347051620483,
0.032714128494262695,
0.17395801842212677,
-0.14596548676490784,
-0.058364760130643845,
-0.15016603469848633,
-0.038290537893772125,
0.04992041364312172,
-0.10589156299829483,
0.10652581602334976,
0.027632366865873337,
-0.07823801040649414,
0.0007243742002174258,
0.002623017178848386,
0.14605776965618134,
0.0541694238781929,
0.08562533557415009,
-0.019375350326299667,
-0.12459471076726913,
-0.00021981968893669546,
-0.038636792451143265,
0.09758833050727844,
-0.29261574149131775,
-0.03994721919298172,
-0.09248299896717072,
-0.021830396726727486,
0.010110455565154552,
0.02073568105697632,
0.006614902056753635,
-0.06931109726428986,
-0.08114403486251831,
-0.050503332167863846,
0.04335319250822067,
-0.00539602292701602,
0.21233098208904266,
-0.0991852730512619,
0.07813863456249237,
0.05556156486272812,
0.1361694633960724,
-0.2077561914920807,
-0.029840335249900818,
0.020643850788474083,
-0.04700479283928871,
0.06194344162940979,
-0.16559401154518127,
0.0881313681602478,
0.08901660144329071,
-0.05598185211420059,
0.14282271265983582,
0.10184906423091888,
0.013203968293964863,
0.018108714371919632,
0.11537288129329681,
-0.05601691082119942,
-0.06393995136022568,
-0.045686013996601105,
-0.00005413982944446616,
-0.04412863403558731,
0.05201559513807297,
0.08640943467617035,
-0.010741738602519035,
-0.008883470669388771,
0.00213888892903924,
0.059190403670072556,
-0.08956669270992279,
0.057735614478588104,
0.04425010830163956,
0.026469171047210693,
-0.0956103652715683,
0.04190903529524803,
0.003256307216361165,
-0.10962975025177002,
0.025402357801795006,
0.09026612341403961,
-0.03671419247984886,
-0.059660397469997406,
0.058858729898929596,
0.29895660281181335,
-0.03278406709432602,
-0.0031645006965845823,
-0.042823366820812225,
-0.13157464563846588,
0.007927530445158482,
0.23086802661418915,
0.0870819166302681,
0.02255757339298725,
-0.04616907984018326,
-0.014704577624797821,
-0.05572551488876343,
0.06961873918771744,
0.06754626333713531,
0.09977442771196365,
-0.07305624336004257,
0.04926697164773941,
-0.0008146621403284371,
0.04695802927017212,
-0.06892106682062149,
0.07485798001289368,
-0.07810740172863007,
-0.0553780198097229,
-0.06271611154079437,
0.05475183576345444,
-0.09223981201648712,
-0.015694204717874527,
0.03579268604516983,
0.015923386439681053,
-0.039341818541288376,
-0.011449101381003857,
-0.03117641806602478,
-0.01001034677028656,
-0.01180058903992176,
0.08333690464496613,
-0.04786834865808487,
-0.06018553301692009,
0.030731460079550743,
-0.06994646787643433,
0.06278149783611298,
-0.045947618782520294,
-0.03459595888853073,
0.027371317148208618,
-0.20895397663116455,
-0.03954999893903732,
0.11511977761983871,
0.0022842218168079853,
-0.004419288598001003,
-0.04674191027879715,
0.026158148422837257,
-0.020490726456046104,
0.02033805474638939,
-0.018666036427021027,
0.15508608520030975,
-0.06260983645915985,
0.03450516611337662,
-0.1736539900302887,
-0.12250576168298721,
-0.04444960877299309,
-0.04143287241458893,
0.017004696652293205,
0.04209741950035095,
0.11831174045801163,
-0.09788227081298828,
0.03207239508628845,
-0.11579185724258423,
0.03913820534944534,
-0.015946071594953537,
-0.14199641346931458,
-0.013232509605586529,
-0.11758915334939957,
0.009192204102873802,
-0.01374692190438509,
0.1740047186613083,
-0.02902987226843834,
-0.059684641659259796,
0.0166329238563776,
0.04933425784111023,
0.07902093231678009,
0.0010662944987416267,
0.1591518372297287,
-0.001437843544408679,
-0.04952007532119751,
-0.16605640947818756,
0.07519194483757019,
0.13722731173038483,
0.07753538340330124,
0.02741684950888157,
0.08458449691534042,
0.003702176734805107,
0.13729166984558105,
0.07677335292100906,
0.017563089728355408,
-0.013785590417683125,
0.04326152801513672,
0.06580288708209991,
0.005668623838573694,
0.029919862747192383,
0.1572478711605072,
0.1966233253479004,
-0.0726773589849472,
0.014824261888861656,
0.05788656324148178,
-0.06804123520851135,
-0.10423710197210312,
-0.018835389986634254,
-0.08278171718120575,
-0.12020887434482574,
-0.015877598896622658,
-0.11667384952306747,
-0.056171514093875885,
0.04965033382177353,
0.0030002244748175144,
0.04598810523748398,
0.16694015264511108,
-0.07971572875976562,
-0.08063128590583801,
0.062026284635066986,
-0.005010633263736963,
0.04137270525097847,
0.08892728388309479,
-0.01945391483604908,
0.03500055521726608,
-0.056825149804353714,
0.032984472811222076,
0.0404425710439682,
0.057240355759859085,
0.02508166804909706,
-0.11286886036396027,
-0.10011044889688492,
-0.04118197783827782,
0.0340571254491806,
0.02925679087638855,
0.1947265863418579,
0.028840353712439537,
-0.10031052678823471,
0.008174493908882141,
0.019521862268447876,
-0.05265814810991287,
0.023166105151176453,
-0.071596659719944,
0.09554510563611984,
0.05511864274740219,
0.11859528720378876,
-0.08282436430454254,
-0.06443138420581818,
-0.017168531194329262,
0.18404953181743622,
0.1801094263792038,
-0.03973730280995369,
-0.005781153216958046,
0.04450967162847519,
0.03245951980352402,
0.03894100338220596,
0.01657208241522312,
0.035062238574028015,
0.22767429053783417,
-0.07987827807664871,
-0.03162839636206627,
-0.04675731062889099,
0.03937652334570885,
-0.12043644487857819,
0.023197179660201073,
0.023972641676664352,
-0.0456603467464447,
0.056439679116010666,
0.06360745429992676,
-0.024246666580438614,
0.042016275227069855,
0.04165751487016678,
-0.16928914189338684,
-0.021220430731773376,
-0.005583515390753746,
0.12672321498394012,
0.01173172052949667,
0.10023049265146255,
-0.016241377219557762,
-0.05710212513804436,
0.011381861753761768,
0.007336446549743414,
-0.16744907200336456,
0.05386202782392502,
-0.026687821373343468,
-0.031838394701480865,
0.09810464084148407,
-0.01950344629585743,
0.03640200197696686,
0.12197693437337875,
0.042609237134456635,
0.040794868022203445,
0.11014819145202637,
0.00044021979556418955,
0.03952200338244438,
0.09967818111181259,
-0.01380455493927002,
-0.03987820819020271,
0.022221988067030907,
0.06155537813901901,
-0.16322246193885803,
0.07875055074691772,
0.024140525609254837,
-0.12016090005636215,
-0.010685337707400322,
0.045524414628744125,
-0.11537699401378632,
0.0608285591006279,
0.05727078393101692,
0.016336804255843163,
-0.08004416525363922,
-0.014215951785445213,
-0.014751891605556011,
0.03239745646715164,
0.09587845951318741,
0.006652926094830036,
-0.12037336826324463,
0.020507410168647766,
0.11144702136516571,
-0.013806651346385479,
-0.3710376024246216,
-0.07592280209064484,
-0.05494071915745735,
0.04091927781701088,
-0.14327140152454376,
0.03970072418451309,
0.19789648056030273,
-0.037934206426143646,
-0.046997010707855225,
-0.3856077790260315,
0.06526483595371246,
0.07588254660367966,
-0.08017081022262573,
-0.08591030538082123
] |
null | null | transformers |
<br>

<br>
Tess, short for Tesoro (Treasure in Italian), is a general purpose Large Language Model series. Tess-72B-v1.5b was trained on the Qwen-72B base.
# Prompt Format:
```
SYSTEM: <ANY SYSTEM CONTEXT>
USER:
ASSISTANT:
``` | {"license": "other", "license_name": "qwen-72b-licence", "license_link": "https://huggingface.co/Qwen/Qwen-72B/blob/main/LICENSE"} | text-generation | migtissera/Tess-72B-v1.5b | [
"transformers",
"safetensors",
"llama",
"text-generation",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:34:38+00:00 | [] | [] | TAGS
#transformers #safetensors #llama #text-generation #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
<br>
!Tesoro
<br>
Tess, short for Tesoro (Treasure in Italian), is a general purpose Large Language Model series. Tess-72B-v1.5b was trained on the Qwen-72B base.
# Prompt Format:
| [
"# Prompt Format:"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Prompt Format:"
] | [
52,
6
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #license-other #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Prompt Format:"
] | [
-0.011230595409870148,
0.009728879667818546,
-0.005859217140823603,
-0.011872829869389534,
0.13850584626197815,
-0.0014071150217205286,
0.19027571380138397,
0.10097196698188782,
-0.04979602247476578,
-0.020273063331842422,
0.11999130249023438,
0.18056467175483704,
-0.015663573518395424,
0.11136814206838608,
-0.1184680238366127,
-0.20867086946964264,
0.0705202966928482,
0.00809504184871912,
0.08602900803089142,
0.08904710412025452,
0.11213148385286331,
-0.05304219573736191,
0.08734852075576782,
-0.0427827388048172,
-0.08511706441640854,
0.0438048280775547,
0.052062418311834335,
-0.12507659196853638,
0.07752614468336105,
0.09322108328342438,
0.09386591613292694,
0.07408663630485535,
-0.004626085516065359,
-0.20419816672801971,
0.018742607906460762,
-0.011174402199685574,
-0.08587568253278732,
0.006605461239814758,
0.06593527644872665,
-0.05897798761725426,
0.08846415579319,
0.004944521468132734,
-0.018882805481553078,
0.06466129422187805,
-0.10738222301006317,
-0.04911492392420769,
-0.04197341576218605,
-0.011097663082182407,
0.11042089760303497,
0.08079608529806137,
-0.00017626614135224372,
0.11962153762578964,
-0.02990906499326229,
0.09131637960672379,
0.0876397117972374,
-0.307723730802536,
0.013106954284012318,
0.09963063895702362,
0.0810266062617302,
0.0669260025024414,
-0.04522309824824333,
0.08574073016643524,
0.06802158802747726,
-0.029641050845384598,
0.02762787602841854,
-0.06270596385002136,
-0.07840806990861893,
0.025785576552152634,
-0.06490705162286758,
-0.033035580068826675,
0.24535052478313446,
-0.050522446632385254,
0.006616826634854078,
-0.07958005368709564,
-0.051837507635354996,
-0.006124316249042749,
-0.04203476384282112,
0.067174032330513,
-0.000005956438599241665,
0.08664010465145111,
0.050204355269670486,
-0.03152322769165039,
-0.15515759587287903,
-0.02515527978539467,
-0.19263620674610138,
0.15358741581439972,
0.00863807462155819,
0.03094109334051609,
-0.14588606357574463,
0.042581845074892044,
-0.022976400330662727,
-0.0946086123585701,
-0.021655021235346794,
-0.0755881741642952,
0.0822882205247879,
-0.03125045448541641,
-0.06277678161859512,
-0.11920250952243805,
0.14552347362041473,
0.16322433948516846,
0.026665769517421722,
0.023712649941444397,
-0.12333846837282181,
0.08215209096670151,
-0.0060946098528802395,
0.054526183754205704,
0.05251728743314743,
0.0026293836999684572,
0.0751541256904602,
-0.08394976705312729,
0.0893404632806778,
-0.027634961530566216,
-0.17638002336025238,
-0.009019384160637856,
0.0013793298276141286,
0.13634832203388214,
0.000529722950886935,
0.10170867294073105,
-0.02844710648059845,
0.05295119807124138,
0.11092213541269302,
-0.09248976409435272,
-0.013455154374241829,
-0.02895260974764824,
0.07350392639636993,
0.04511698707938194,
0.015026845969259739,
0.021038852632045746,
-0.042713288217782974,
0.07580512017011642,
-0.05058886110782623,
-0.014464174397289753,
-0.053881000727415085,
-0.062461886554956436,
0.052326034754514694,
-0.08159215748310089,
0.031453583389520645,
-0.16611216962337494,
-0.21917767822742462,
0.025396086275577545,
0.003137284656986594,
-0.010582582093775272,
-0.0011697603622451425,
-0.030024278908967972,
-0.025441600009799004,
0.033554546535015106,
-0.08018714934587479,
-0.04237774759531021,
-0.09291253238916397,
0.1010625809431076,
-0.02625402621924877,
0.04520146921277046,
-0.188853919506073,
0.04240311309695244,
-0.10787472873926163,
-0.0048163351602852345,
-0.05433540791273117,
0.025732241570949554,
-0.07082465291023254,
0.1576930433511734,
-0.010748016647994518,
0.014379767701029778,
-0.05153276398777962,
0.07720371335744858,
-0.051340680569410324,
0.1669580191373825,
-0.1517992913722992,
-0.04189193621277809,
0.22989217936992645,
-0.15381744503974915,
-0.187440887093544,
0.07875935733318329,
0.012949526309967041,
0.06445389986038208,
0.09429863840341568,
0.16389378905296326,
0.05018449202179909,
-0.10638891905546188,
0.040453433990478516,
0.10179533809423447,
-0.0753815546631813,
-0.12886714935302734,
0.010182864964008331,
-0.032067347317934036,
-0.11988534778356552,
0.05595243722200394,
0.054574377834796906,
0.057566266506910324,
-0.005611748900264502,
-0.047865934669971466,
-0.06133764609694481,
-0.026357699185609818,
-0.03393348678946495,
-0.0287166815251112,
0.05277490243315697,
-0.09556654840707779,
-0.002371495822444558,
0.04912681132555008,
-0.020435286685824394,
0.01226859912276268,
0.018337378278374672,
-0.07950891554355621,
0.0613693930208683,
-0.006269086617976427,
0.042842693626880646,
-0.12360318750143051,
-0.11479947715997696,
-0.025211604312062263,
0.09047112613916397,
0.02429935149848461,
0.019750652834773064,
0.03468196466565132,
-0.001550982124172151,
-0.037933964282274246,
0.005597901064902544,
0.21086366474628448,
0.02248769998550415,
-0.052435923367738724,
-0.07973003387451172,
0.10579945892095566,
-0.05242419242858887,
-0.008814187720417976,
-0.09406975656747818,
0.030696045607328415,
0.060472920536994934,
0.07072199881076813,
0.014986153692007065,
0.06772185117006302,
-0.0005683779017999768,
0.017471294850111008,
-0.10342472046613693,
0.016948891803622246,
0.06851296126842499,
0.016283556818962097,
-0.10498926788568497,
0.21297919750213623,
-0.2235386222600937,
0.2712683081626892,
0.217091903090477,
-0.15056699514389038,
0.016323581337928772,
-0.13563649356365204,
0.009032662026584148,
0.02384147234261036,
-0.02304299920797348,
-0.026391075924038887,
-0.06071239337325096,
-0.022208133712410927,
0.1754799485206604,
-0.07539781928062439,
-0.01844455674290657,
-0.005674710962921381,
-0.06424371898174286,
-0.05401167273521423,
0.0501604788005352,
0.14648376405239105,
-0.14764782786369324,
0.17113307118415833,
0.24849355220794678,
0.020699050277471542,
0.1793595552444458,
-0.04803755134344101,
-0.011507845483720303,
0.050224367529153824,
0.046320218592882156,
0.029511671513319016,
-0.050548527389764786,
-0.054554153233766556,
-0.002176472684368491,
0.07370831817388535,
0.023502284660935402,
0.04966426640748978,
-0.14227955043315887,
-0.06192408874630928,
-0.0033864672295749187,
-0.05196771398186684,
-0.02039084956049919,
0.09252176433801651,
0.02433866821229458,
0.10541243106126785,
-0.08246752619743347,
-0.016503246501088142,
0.1067500039935112,
0.00036323454696685076,
-0.07921642810106277,
0.16810515522956848,
-0.12470674514770508,
-0.24259819090366364,
-0.22421039640903473,
-0.1712326556444168,
-0.05619212985038757,
0.0720207691192627,
0.14771462976932526,
-0.041029077023267746,
-0.08084529638290405,
-0.08572524040937424,
0.017084132879972458,
0.008296975865960121,
0.024386921897530556,
-0.09100169688463211,
0.08248630911111832,
-0.018122157081961632,
-0.10829252004623413,
-0.058050062507390976,
0.03448985144495964,
-0.04032720997929573,
0.13439640402793884,
-0.08897080272436142,
0.11098228394985199,
0.07773694396018982,
0.014641252346336842,
0.009900129400193691,
-0.06757640838623047,
0.12976080179214478,
-0.052642568945884705,
-0.005984837654978037,
0.20048627257347107,
-0.030733611434698105,
0.05727292224764824,
0.17378856241703033,
0.017761969938874245,
-0.1119663268327713,
0.059212639927864075,
-0.02587246522307396,
-0.0943274274468422,
-0.20982758700847626,
-0.10102568566799164,
-0.0912313163280487,
0.04262140765786171,
0.03203180059790611,
0.07879512757062912,
0.12958306074142456,
0.08535943180322647,
-0.051044438034296036,
0.03342677652835846,
0.07618563622236252,
0.10730252414941788,
0.2735042870044708,
-0.0012243870878592134,
0.1412700116634369,
-0.11995819956064224,
-0.10784589499235153,
0.09871463477611542,
0.0293703805655241,
0.15021979808807373,
0.13232439756393433,
0.062211502343416214,
0.05580836907029152,
0.006681852508336306,
0.12473894655704498,
0.13578800857067108,
0.050685640424489975,
-0.033952172845602036,
-0.02288590371608734,
-0.0482449047267437,
-0.0049053640104830265,
0.06675806641578674,
-0.09025739878416061,
-0.13901522755622864,
-0.023714642971754074,
-0.08186337351799011,
0.06963435560464859,
0.07332522422075272,
-0.007279854733496904,
-0.21411322057247162,
0.05524946004152298,
0.1161351129412651,
-0.003336492460221052,
-0.10448984801769257,
0.10665199905633926,
-0.01092730276286602,
-0.012579883448779583,
0.13724693655967712,
-0.018934719264507294,
0.11561207473278046,
0.004778416361659765,
0.06771999597549438,
-0.08718886226415634,
-0.05423153564333916,
0.000991365173831582,
0.12270155549049377,
-0.29557862877845764,
0.21381144225597382,
0.02584897354245186,
0.006600636523216963,
-0.06980449706315994,
0.0030566847417503595,
0.025887256488204002,
0.2148788720369339,
0.13700813055038452,
-0.03857250139117241,
-0.14299242198467255,
-0.09565745294094086,
-0.027844440191984177,
0.018420225009322166,
0.12926067411899567,
0.010141201317310333,
0.0399414598941803,
-0.05269626900553703,
-0.0161657202988863,
0.0022575485054403543,
-0.045592475682497025,
-0.047184430062770844,
-0.19655705988407135,
0.026418738067150116,
0.1249714195728302,
0.12395386397838593,
-0.05089331790804863,
0.0030041616410017014,
-0.15375593304634094,
0.12843729555606842,
-0.09093771874904633,
-0.07659506797790527,
-0.12515388429164886,
-0.13955490291118622,
0.027814723551273346,
-0.022954966872930527,
0.05805334076285362,
-0.06188292056322098,
0.06085588037967682,
-0.06592448055744171,
-0.16372273862361908,
0.1019846573472023,
-0.10169263184070587,
-0.03343276306986809,
-0.054387178272008896,
0.1270492672920227,
-0.12298092246055603,
-0.03031737357378006,
0.042904146015644073,
0.03258517012000084,
-0.05428963527083397,
-0.1228342354297638,
-0.02908068709075451,
0.034373365342617035,
0.041818443685770035,
-0.0049539427272975445,
-0.1545449048280716,
-0.11817555874586105,
0.020539339631795883,
-0.0811554566025734,
0.20194785296916962,
0.2680841386318207,
-0.025160573422908783,
0.12831217050552368,
0.13549809157848358,
-0.11917131394147873,
-0.32758793234825134,
-0.12407363951206207,
-0.15239983797073364,
-0.05095592513680458,
0.00882591214030981,
-0.10395659506320953,
0.07762642204761505,
0.06510517746210098,
-0.041921716183423996,
0.12094627320766449,
-0.1548520028591156,
-0.11835837364196777,
0.15272565186023712,
0.026967110112309456,
0.2664203643798828,
-0.20665369927883148,
-0.11639323085546494,
-0.13038921356201172,
-0.0933978259563446,
0.1499910205602646,
-0.0665140151977539,
0.08534207940101624,
0.009631140157580376,
0.018574580550193787,
0.010057426989078522,
-0.043879199773073196,
0.1105385422706604,
-0.05368558317422867,
0.08035364747047424,
-0.1299164891242981,
0.05808880925178528,
0.08768007904291153,
-0.0062548210844397545,
0.09141276031732559,
-0.178704634308815,
0.0016284341691061854,
-0.07529835402965546,
-0.07074474543333054,
-0.01551639474928379,
0.08231621235609055,
0.012958918698132038,
-0.042646992951631546,
-0.008281978778541088,
-0.0725179985165596,
-0.008844875730574131,
-0.04081033542752266,
0.23907312750816345,
-0.09410812705755234,
0.13040636479854584,
0.1468389481306076,
0.14807210862636566,
-0.1218821331858635,
0.11312868446111679,
-0.043329935520887375,
-0.09934332221746445,
0.05388219282031059,
-0.10115030407905579,
0.05222824588418007,
0.0697028785943985,
-0.057885147631168365,
0.09145873039960861,
0.07165422290563583,
0.051767297089099884,
-0.004323496017605066,
0.18380512297153473,
-0.1723783165216446,
-0.055289141833782196,
-0.045920539647340775,
0.10567928850650787,
0.05289298668503761,
0.04461351037025452,
0.13449308276176453,
-0.0007475948659703135,
0.016100766137242317,
0.0007324452162720263,
0.029087385162711143,
0.01093771681189537,
0.03018307313323021,
0.022647961974143982,
0.0072632841765880585,
-0.12650814652442932,
0.06913184374570847,
0.020991569384932518,
-0.13081851601600647,
-0.00265874364413321,
0.1005372628569603,
-0.13803063333034515,
-0.12511609494686127,
-0.002295474288985133,
0.11976693570613861,
-0.17252464592456818,
-0.0841871052980423,
-0.061444491147994995,
-0.16946035623550415,
0.05109841749072075,
0.23239894211292267,
0.060094211250543594,
0.05577747896313667,
0.02647061087191105,
-0.04239752143621445,
-0.022248517721891403,
0.04281039908528328,
-0.05609063804149628,
0.06367908418178558,
-0.10254205018281937,
-0.004346790723502636,
-0.04577472805976868,
0.020679185166954994,
-0.08175413310527802,
0.004373271018266678,
-0.1345316469669342,
0.006237950176000595,
-0.17351068556308746,
0.03277764841914177,
-0.07635585963726044,
-0.015740303322672844,
0.02272094041109085,
-0.023710813373327255,
-0.04567459225654602,
-0.06103118881583214,
-0.10514330863952637,
-0.0014663106994703412,
-0.03781585395336151,
0.08518895506858826,
-0.10872376710176468,
-0.04307400435209274,
0.05272804573178291,
-0.03165611997246742,
0.07062987238168716,
0.04747313633561134,
-0.08171312510967255,
0.09370120614767075,
-0.2015981376171112,
-0.04510153830051422,
0.13515105843544006,
0.026574518531560898,
0.007883847691118717,
0.07953620702028275,
0.01890578120946884,
0.14500905573368073,
0.00773583073168993,
0.049986131489276886,
-0.05690780654549599,
-0.13399584591388702,
0.03625623136758804,
-0.02646530419588089,
-0.1093776673078537,
-0.011807448230683804,
-0.07835014909505844,
0.09522731602191925,
-0.02784382365643978,
0.20265868306159973,
-0.08990230411291122,
0.006856369320303202,
-0.02646126039326191,
0.03588639199733734,
-0.013975897803902626,
-0.1772764027118683,
-0.13786084949970245,
-0.09223205596208572,
-0.01300746388733387,
0.01413044985383749,
0.3105114996433258,
0.0356825590133667,
-0.021229756996035576,
0.07884224504232407,
0.07427646964788437,
0.05357539653778076,
0.009452782571315765,
0.3088417053222656,
0.09997312724590302,
0.004432865884155035,
-0.12129399925470352,
0.002349408343434334,
0.035522524267435074,
-0.12077357620000839,
0.05727066099643707,
0.09332437813282013,
-0.10440989583730698,
0.09622621536254883,
0.048750679939985275,
0.015185982920229435,
-0.03793200105428696,
-0.0945933610200882,
-0.027227912098169327,
0.07297004014253616,
0.009094971232116222,
0.10387582331895828,
0.18241587281227112,
-0.03192252293229103,
0.006461684592068195,
-0.025010745972394943,
-0.020247507840394974,
-0.16335377097129822,
-0.13482029736042023,
-0.10908669233322144,
-0.0993284285068512,
0.03728736937046051,
-0.06503691524267197,
0.07587417960166931,
0.04341675341129303,
0.06514830142259598,
-0.01709168218076229,
0.10952291637659073,
-0.04379665479063988,
-0.05948502942919731,
0.025360452011227608,
-0.059517815709114075,
0.035336531698703766,
-0.03537719324231148,
-0.06099284440279007,
-0.02906166948378086,
-0.06585084646940231,
-0.061115410178899765,
0.07760155946016312,
0.0301591195166111,
0.0398273840546608,
-0.16050957143306732,
-0.06133757531642914,
-0.02413959614932537,
0.0724308118224144,
-0.005776391364634037,
0.11292558163404465,
0.010924669913947582,
-0.030279817059636116,
0.08098281174898148,
0.1791081577539444,
-0.047708164900541306,
-0.18830224871635437,
-0.012126044370234013,
0.18869157135486603,
0.04895121976733208,
0.11678896099328995,
-0.05389630049467087,
-0.012669963762164116,
-0.002739167073741555,
0.3189498484134674,
0.23185373842716217,
-0.008961835876107216,
0.031145842745900154,
-0.05014354735612869,
0.037917375564575195,
0.09471171349287033,
0.09567858278751373,
0.07439254224300385,
0.20330707728862762,
-0.01842869073152542,
-0.019972940906882286,
-0.022472156211733818,
0.03963448479771614,
-0.15274758636951447,
0.09462541341781616,
-0.00884321704506874,
-0.07109043002128601,
-0.019290277734398842,
0.11111433058977127,
-0.1262006312608719,
0.10743539780378342,
-0.055962104350328445,
-0.06619032472372055,
-0.0055690971203148365,
-0.006067302078008652,
0.17370517551898956,
-0.014357276260852814,
-0.011063736863434315,
-0.03105381689965725,
-0.05988373979926109,
0.020967844873666763,
-0.011092768982052803,
-0.20756275951862335,
0.0097003523260355,
0.003118488471955061,
0.008441489189863205,
0.05121782049536705,
0.016646597534418106,
0.029383962973952293,
0.0634746178984642,
0.0030316112097352743,
-0.05439230054616928,
0.19196172058582306,
0.001409298973158002,
-0.07849456369876862,
0.08907874673604965,
-0.03589985892176628,
-0.011932888068258762,
-0.008216619491577148,
0.027493737637996674,
-0.06670785695314407,
0.0692308321595192,
-0.009348316118121147,
-0.1220974400639534,
-0.02113187685608864,
0.038745127618312836,
-0.08591125160455704,
0.06819646805524826,
0.04177335649728775,
0.0038918093778192997,
-0.003309905296191573,
-0.04228295385837555,
0.04971704259514809,
0.011059688404202461,
-0.1607132852077484,
-0.027725324034690857,
-0.09222875535488129,
-0.07640449702739716,
0.15832896530628204,
0.0336291678249836,
-0.2695191502571106,
0.014002724550664425,
-0.1347365826368332,
0.05643495172262192,
-0.19540663063526154,
0.06738363206386566,
0.20934461057186127,
0.0032795541919767857,
-0.028614817187190056,
-0.16454988718032837,
0.06895467638969421,
0.04787858575582504,
-0.06242532283067703,
-0.07411511242389679
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-3.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:40:27+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
194,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06399306654930115,
-0.029307762160897255,
-0.004331388510763645,
0.03184039518237114,
0.018488328903913498,
0.03748767077922821,
0.10618676990270615,
0.07814439386129379,
0.07063937187194824,
0.07519446313381195,
-0.04871003329753876,
-0.02137892320752144,
0.07951480895280838,
0.10956592857837677,
0.007890605367720127,
-0.23595072329044342,
0.08250588178634644,
-0.03575211763381958,
-0.026412149891257286,
0.03893791511654854,
0.14763003587722778,
-0.05930296331644058,
0.05646873265504837,
-0.0034751244820654392,
-0.05727600306272507,
0.05131329223513603,
-0.026073675602674484,
-0.007598201744258404,
0.09508240222930908,
0.0628398060798645,
0.041950274258852005,
0.0221012644469738,
-0.018701795488595963,
-0.15748009085655212,
0.004922133404761553,
-0.017628159373998642,
-0.04106835648417473,
0.027759606018662453,
0.11616206914186478,
-0.045659661293029785,
0.11880303174257278,
-0.12680159509181976,
-0.018632367253303528,
0.08243369311094284,
-0.12379294633865356,
-0.10155918449163437,
-0.20841088891029358,
0.15503807365894318,
0.07979700714349747,
0.008151463232934475,
-0.04288718104362488,
0.07205401360988617,
-0.054492317140102386,
0.0433170422911644,
0.08753809332847595,
-0.2521143853664398,
-0.04418989270925522,
0.17082107067108154,
0.007189796306192875,
-0.04941372200846672,
-0.05787082388997078,
0.06996709108352661,
0.029280396178364754,
0.0067648859694600105,
-0.04790947958827019,
-0.04312778264284134,
0.2201330065727234,
-0.03288035839796066,
-0.1199050173163414,
-0.024701880291104317,
0.08322399854660034,
0.08842694014310837,
-0.04718442261219025,
-0.15207047760486603,
-0.011922257021069527,
-0.022255953401327133,
-0.04938272759318352,
-0.09886841475963593,
0.003574386006221175,
-0.02095438726246357,
0.07563887536525726,
-0.05726063251495361,
-0.04117260500788689,
0.0583573654294014,
-0.05984218046069145,
0.1629648655653,
-0.0014757480239495635,
-0.03302643075585365,
0.05737544223666191,
0.03662708401679993,
-0.2148275226354599,
-0.08567781001329422,
-0.04239440709352493,
-0.054734885692596436,
-0.1342768371105194,
-0.0431794635951519,
-0.04418531805276871,
-0.08217091858386993,
-0.007381834089756012,
0.23022499680519104,
-0.02492726407945156,
0.02804037742316723,
0.041754867881536484,
0.010314702056348324,
0.04419277235865593,
0.1185997873544693,
-0.05733427032828331,
-0.1965552270412445,
-0.02631615847349167,
-0.01892445981502533,
0.06395958364009857,
0.013402312994003296,
-0.02684073895215988,
-0.01519184559583664,
0.02383972890675068,
0.02890627458691597,
0.02422332763671875,
0.03168574720621109,
-0.07110860198736191,
-0.03037358447909355,
0.14692901074886322,
-0.10123810917139053,
0.04444355517625809,
-0.03552069887518883,
-0.03803017735481262,
0.07610528916120529,
-0.030228903517127037,
0.023446066305041313,
-0.01123465783894062,
0.14718079566955566,
-0.04342922940850258,
-0.02282293513417244,
-0.04051238298416138,
-0.05655231699347496,
0.056725140661001205,
-0.08175330609083176,
-0.06928885728120804,
-0.0921410471200943,
-0.07629581540822983,
-0.0781061053276062,
0.019695017486810684,
-0.04772234335541725,
-0.00727304769679904,
-0.00963415578007698,
-0.03471999987959862,
0.004572734236717224,
0.03700930252671242,
-0.03850525617599487,
-0.006236410234123468,
-0.05285993963479996,
-0.02465738169848919,
0.016344578936696053,
-0.07584261149168015,
0.020161736756563187,
-0.09869375079870224,
0.055850550532341,
-0.3218844532966614,
0.09342993050813675,
-0.11400728672742844,
0.009930809028446674,
-0.1314808875322342,
-0.0335058830678463,
0.016112782061100006,
0.005322059616446495,
0.013146123848855495,
0.11860170215368271,
-0.15911765396595,
-0.09210728108882904,
0.09710568189620972,
-0.13270564377307892,
-0.06123318895697594,
0.13335616886615753,
0.006097014062106609,
-0.007668835576623678,
0.0786266103386879,
0.1425439566373825,
0.14123733341693878,
-0.04838160052895546,
-0.13677071034908295,
-0.06624309718608856,
0.027206536382436752,
0.1544196456670761,
0.06648306548595428,
-0.0522027313709259,
-0.0182489063590765,
0.03598593920469284,
-0.031956933438777924,
-0.0026491705793887377,
-0.016834797337651253,
-0.05705594643950462,
-0.023978661745786667,
-0.002377238357439637,
0.096376933157444,
-0.015163181349635124,
0.002383065177127719,
-0.07838787883520126,
-0.09363235533237457,
0.039118777960538864,
0.0882512778043747,
0.041436195373535156,
0.01435884553939104,
-0.06161881238222122,
0.09994251281023026,
-0.03056234121322632,
0.022030595690011978,
-0.10238077491521835,
-0.15781597793102264,
0.03258318454027176,
-0.15215061604976654,
0.06616183370351791,
0.046265859156847,
0.06192070245742798,
0.027611956000328064,
-0.11285637319087982,
-0.06351099908351898,
-0.02237211912870407,
0.016440147534012794,
0.007992693223059177,
-0.12434909492731094,
-0.08924297988414764,
-0.04348140209913254,
0.22069425880908966,
-0.054984692484140396,
-0.007727441843599081,
0.06026673689484596,
0.2515117824077606,
0.018960142508149147,
-0.014410633593797684,
-0.05082327499985695,
0.07166947424411774,
0.00031130402931012213,
0.0197899267077446,
0.01994451880455017,
-0.04336968809366226,
-0.14637522399425507,
0.095527283847332,
-0.055198993533849716,
-0.02979397401213646,
0.06157594546675682,
0.054885584861040115,
-0.09651610255241394,
-0.011951263062655926,
-0.05258362367749214,
-0.029914764687418938,
0.006216008216142654,
-0.12069366872310638,
0.042220812290906906,
0.04080578312277794,
0.03252416476607323,
-0.0428742878139019,
-0.048637744039297104,
0.01096632331609726,
0.00038561708061024547,
-0.12152905762195587,
0.10360410064458847,
-0.08400224894285202,
-0.21976810693740845,
0.10374713689088821,
0.15290415287017822,
0.07434318959712982,
0.12372945249080658,
-0.013827105984091759,
-0.022371480241417885,
-0.1197407990694046,
0.0015785678988322616,
0.021130502223968506,
0.042604196816682816,
0.04024510458111763,
0.05408281460404396,
0.03447001054883003,
0.043456509709358215,
0.02863817848265171,
-0.02880154922604561,
0.04750415310263634,
-0.009588376618921757,
0.0040976181626319885,
0.12772449851036072,
0.10417608171701431,
0.02852170541882515,
0.08533679693937302,
0.03910043090581894,
-0.0404503159224987,
0.01288803294301033,
-0.02749737910926342,
-0.06698718667030334,
0.10686532407999039,
-0.10741190612316132,
-0.16609297692775726,
-0.17008981108665466,
-0.051623594015836716,
-0.18006007373332977,
-0.024816229939460754,
0.03015192784368992,
-0.03652867674827576,
-0.07785869389772415,
-0.07390360534191132,
0.13494938611984253,
0.03926362842321396,
-0.05240222439169884,
-0.05028913915157318,
-0.03344966843724251,
0.042253971099853516,
-0.10105009377002716,
-0.04346547648310661,
0.0018156386213377118,
-0.050230301916599274,
0.021803520619869232,
0.027263345196843147,
0.0939897820353508,
0.060186538845300674,
-0.00003084815762122162,
-0.02964363619685173,
-0.040875863283872604,
0.17850545048713684,
-0.1293039619922638,
0.11540871858596802,
0.04773532599210739,
-0.0027224498335272074,
0.06698734313249588,
0.16732339560985565,
0.05765581503510475,
-0.04658689722418785,
-0.011350571177899837,
0.06666777282953262,
-0.030819516628980637,
-0.19200317561626434,
-0.1769828051328659,
-0.06770068407058716,
0.007487337104976177,
-0.027668775990605354,
0.04515873268246651,
0.06839024275541306,
-0.0003817266842816025,
-0.09345950931310654,
-0.07025164365768433,
0.036232683807611465,
0.04945642128586769,
0.17080222070217133,
0.009348781779408455,
0.05902401730418205,
-0.06639035046100616,
-0.06039080023765564,
0.09840981662273407,
0.00460409838706255,
0.08121812343597412,
0.06772522628307343,
0.06868960708379745,
0.051509808748960495,
0.012179503217339516,
0.0003098213055636734,
0.04621569812297821,
0.015963001176714897,
-0.02683101035654545,
-0.011713831685483456,
-0.08623592555522919,
0.07211727648973465,
0.08716612309217453,
0.10278671234846115,
0.03722165524959564,
-0.051564089953899384,
-0.06990828365087509,
0.12818388640880585,
0.1740635186433792,
0.11096327751874924,
-0.12066809087991714,
-0.10149893164634705,
0.059367019683122635,
-0.012059872969985008,
0.020104901865124702,
-0.03295712172985077,
-0.051309432834386826,
-0.13940629363059998,
0.20302264392375946,
-0.018019717186689377,
0.06347235292196274,
0.0055109113454818726,
-0.0001949289726326242,
-0.02158956229686737,
0.06696530431509018,
0.017838791012763977,
0.050336673855781555,
-0.141648069024086,
0.14686071872711182,
0.010985979810357094,
0.021909072995185852,
-0.01216028444468975,
0.05184603109955788,
0.01702127605676651,
0.05610823258757591,
0.10946010798215866,
0.04561372473835945,
-0.09489621967077255,
-0.08675345033407211,
-0.04737729951739311,
-0.05841971933841705,
0.13532322645187378,
-0.0978664755821228,
0.08815576881170273,
-0.005218944977968931,
-0.04548734799027443,
-0.058141354471445084,
0.16146430373191833,
-0.15934163331985474,
-0.10354094952344894,
0.10525287687778473,
-0.04850301891565323,
0.028323976323008537,
-0.07308647781610489,
-0.04719609022140503,
-0.24660754203796387,
0.2158661186695099,
0.015384898521006107,
0.00384410098195076,
-0.09645523875951767,
-0.0031059239991009235,
0.16681231558322906,
-0.09507244825363159,
0.038081102073192596,
-0.0030803470872342587,
0.08820296823978424,
-0.09653045237064362,
-0.08165580779314041,
0.04465007409453392,
-0.07744460552930832,
-0.14185139536857605,
0.011764489114284515,
0.19781433045864105,
0.022616581991314888,
0.037971191108226776,
0.025954479351639748,
0.09160146117210388,
0.08607495576143265,
-0.10742948949337006,
0.08422011882066727,
0.0879930630326271,
-0.030707523226737976,
0.1495898962020874,
-0.033895764499902725,
-0.15872009098529816,
-0.05146879330277443,
-0.0411202572286129,
0.10311200469732285,
0.34257373213768005,
-0.06467219442129135,
0.08678370714187622,
0.08080018311738968,
-0.042148686945438385,
-0.19632753729820251,
0.007789987605065107,
0.01917961612343788,
0.02523370087146759,
0.00445806048810482,
-0.05062577873468399,
0.10844220966100693,
0.13072700798511505,
-0.020657971501350403,
-0.04788164049386978,
-0.20526863634586334,
-0.1290767788887024,
0.011098041199147701,
-0.06513512134552002,
-0.012189746834337711,
-0.06006240472197533,
-0.10408233106136322,
-0.07066939771175385,
-0.16428698599338531,
0.05405878648161888,
0.04084285721182823,
0.048617810010910034,
0.0022039911709725857,
-0.022370649501681328,
0.03627573698759079,
-0.018924014642834663,
0.18879440426826477,
0.06920293718576431,
0.02023874782025814,
-0.07239631563425064,
-0.04903245344758034,
0.11341241002082825,
-0.037153460085392,
0.09947021305561066,
0.04320525750517845,
-0.005356981884688139,
-0.11361989378929138,
0.012292295694351196,
-0.09132620692253113,
0.11476991325616837,
-0.09256471693515778,
-0.030837757512927055,
-0.08548001945018768,
0.140853151679039,
0.0238761268556118,
-0.015673771500587463,
0.12679731845855713,
-0.032148636877536774,
0.1091376468539238,
0.13080862164497375,
0.0701589584350586,
0.0031420262530446053,
-0.07472140341997147,
0.0012759605888277292,
-0.06170511618256569,
0.019098622724413872,
0.05546579137444496,
0.04380559176206589,
0.13516387343406677,
-0.01975567452609539,
0.14712172746658325,
-0.018929120153188705,
-0.13689684867858887,
-0.023719700053334236,
0.11753503233194351,
-0.11223208159208298,
-0.2821040451526642,
-0.04018678143620491,
-0.034599367529153824,
0.012139244005084038,
0.061747223138809204,
0.19457398355007172,
-0.03780088946223259,
-0.023348631337285042,
0.021921996027231216,
0.04527831822633743,
-0.09146711230278015,
0.08037439733743668,
-0.021092766895890236,
0.042854998260736465,
-0.06821004301309586,
0.10744444280862808,
0.07745382934808731,
-0.22681817412376404,
-0.025234894827008247,
0.09869363903999329,
-0.06257352977991104,
-0.07561001926660538,
-0.09376278519630432,
0.1380273848772049,
-0.1314297765493393,
-0.0158812515437603,
-0.020428206771612167,
-0.1451670229434967,
0.04510067030787468,
0.09574063867330551,
0.04472881183028221,
0.011867307126522064,
0.035759247839450836,
0.00018209819972980767,
0.01861182041466236,
0.03738055005669594,
0.09328993409872055,
0.10640279203653336,
-0.042188696563243866,
0.03816690295934677,
-0.08412777632474899,
0.015412991866469383,
-0.021022463217377663,
-0.01296116691082716,
-0.09465239942073822,
-0.07739552855491638,
-0.10236255824565887,
-0.06780892610549927,
-0.08637113869190216,
-0.02516632154583931,
-0.0233630258589983,
-0.023995107039809227,
0.004357992671430111,
-0.027227098122239113,
-0.07138573378324509,
-0.050982195883989334,
-0.0155563373118639,
0.08955743908882141,
-0.1038399413228035,
-0.0027917600236833096,
0.06826390326023102,
-0.051174331456422806,
0.0997040867805481,
0.024214884266257286,
0.0514301136136055,
0.03840871900320053,
-0.03028056025505066,
-0.03847609460353851,
0.01905941218137741,
0.006461322773247957,
0.05086956173181534,
-0.14872215688228607,
-0.012150841765105724,
-0.0032515712082386017,
-0.051708951592445374,
0.050017524510622025,
-0.0625491663813591,
-0.10236836224794388,
0.008505094796419144,
0.016627702862024307,
-0.09351813793182373,
-0.07604184001684189,
0.027614127844572067,
0.0729801207780838,
0.035474032163619995,
0.11152558028697968,
-0.058112598955631256,
0.05574757605791092,
-0.1357552707195282,
-0.01836516335606575,
-0.02554010972380638,
-0.1030651405453682,
0.016653461381793022,
-0.011252963915467262,
0.022870756685733795,
-0.0031690942123532295,
0.11289511620998383,
-0.026996877044439316,
-0.025828121230006218,
0.06756260991096497,
-0.028583889827132225,
-0.0780753642320633,
0.0507938489317894,
0.07868042588233948,
0.04812269285321236,
-0.0161933284252882,
-0.08120898902416229,
-0.0014988429611548781,
-0.05973583459854126,
-0.007319761905819178,
0.1114073172211647,
0.14946618676185608,
0.11061113327741623,
0.05770944431424141,
0.08207276463508606,
-0.04891999810934067,
-0.028829647228121758,
-0.009984364733099937,
-0.029658254235982895,
0.08592110872268677,
-0.02771167829632759,
0.046906858682632446,
0.07054735720157623,
-0.23260782659053802,
0.1542171984910965,
-0.04036210849881172,
0.003158458275720477,
-0.019460761919617653,
-0.11383481323719025,
-0.07883796095848083,
-0.042559150606393814,
0.010454006493091583,
-0.08244454860687256,
0.009923603385686874,
0.00909425225108862,
0.03983308747410774,
-0.018626000732183456,
0.14684614539146423,
-0.22303219139575958,
-0.020058225840330124,
0.1173686534166336,
0.01384452823549509,
-0.01986311748623848,
0.04652818292379379,
-0.01833965815603733,
-0.05564456060528755,
0.0526103749871254,
0.003485421882942319,
0.042003098875284195,
0.01768621988594532,
-0.009287383407354355,
-0.030709892511367798,
-0.0902811586856842,
0.015258834697306156,
-0.013983458280563354,
0.019041143357753754,
0.026359856128692627,
0.02583683282136917,
0.016558820381760597,
-0.0304630808532238,
0.1203962042927742,
-0.04186068847775459,
-0.08964694291353226,
-0.09655430167913437,
0.18649248778820038,
0.015295648016035557,
0.07767751067876816,
0.014468489214777946,
-0.11231113970279694,
0.003929391037672758,
0.011538079008460045,
0.2675095796585083,
0.06229402497410774,
0.0249533299356699,
0.022313563153147697,
0.0121700968593359,
0.03887152299284935,
0.019862765446305275,
-0.021693918853998184,
0.17263448238372803,
-0.035285405814647675,
0.2010856717824936,
-0.05225873738527298,
-0.046340879052877426,
0.0050239525735378265,
0.1740422248840332,
-0.01995527185499668,
-0.01296838466078043,
-0.015919938683509827,
0.11686369776725769,
-0.02941286377608776,
-0.15430203080177307,
0.030166465789079666,
-0.1680508852005005,
-0.1555458903312683,
-0.04883193224668503,
0.03437260910868645,
0.0537441223859787,
0.11792732775211334,
0.015498639084398746,
-0.03355927765369415,
0.11172260344028473,
0.013434622436761856,
-0.01853151060640812,
-0.12593184411525726,
0.06253454834222794,
-0.005848423577845097,
0.10684798657894135,
0.020717212930321693,
0.04494085907936096,
0.12986114621162415,
-0.024106910452246666,
-0.10693443566560745,
0.014985187910497189,
0.045302849262952805,
-0.08386611938476562,
0.0049399384297430515,
0.15897472202777863,
-0.014517411589622498,
0.11524591594934464,
0.05899004638195038,
-0.13243594765663147,
0.07087180018424988,
0.0894564762711525,
-0.041566379368305206,
-0.06131380423903465,
0.1714923232793808,
-0.08582644164562225,
0.13566255569458008,
0.23618710041046143,
-0.026338893920183182,
0.025297190994024277,
-0.03987700119614601,
0.044342536479234695,
0.07475463300943375,
0.12010806798934937,
-0.017106499522924423,
-0.16066502034664154,
0.06056363880634308,
0.023605575785040855,
0.04031899571418762,
-0.14200104773044586,
-0.11799366027116776,
-0.015118377283215523,
0.02434436045587063,
-0.030886957421898842,
0.12695416808128357,
0.08972851932048798,
-0.0224088616669178,
0.001209444715641439,
-0.1888996660709381,
-0.008432535454630852,
0.07094074040651321,
-0.045595064759254456,
-0.016666574403643608
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# test_TwoWayLoss_25K_bs64_P4_N1
This model is a fine-tuned version of [allenai/scibert_scivocab_uncased](https://huggingface.co/allenai/scibert_scivocab_uncased) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 8.8668
- Accuracy: 0.5430
- Precision: 0.0114
- Recall: 0.5359
- F1: 0.0224
- Hamming: 0.4570
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Hamming |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:-------:|
| 8.9097 | 0.0 | 5 | 8.8974 | 0.5287 | 0.0107 | 0.5189 | 0.0210 | 0.4713 |
| 8.1387 | 0.0 | 10 | 8.8668 | 0.5430 | 0.0114 | 0.5359 | 0.0224 | 0.4570 |
### Framework versions
- Transformers 4.35.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.7.1
- Tokenizers 0.14.1
| {"tags": ["generated_from_trainer"], "metrics": ["accuracy", "precision", "recall", "f1"], "base_model": "allenai/scibert_scivocab_uncased", "model-index": [{"name": "test_TwoWayLoss_25K_bs64_P4_N1", "results": []}]} | text-classification | bdpc/test_TwoWayLoss_25K_bs64_P4_N1 | [
"transformers",
"pytorch",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:allenai/scibert_scivocab_uncased",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:41:32+00:00 | [] | [] | TAGS
#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us
| test\_TwoWayLoss\_25K\_bs64\_P4\_N1
===================================
This model is a fine-tuned version of allenai/scibert\_scivocab\_uncased on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 8.8668
* Accuracy: 0.5430
* Precision: 0.0114
* Recall: 0.5359
* F1: 0.0224
* Hamming: 0.4570
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_ratio: 0.1
* training\_steps: 10
### Training results
### Framework versions
* Transformers 4.35.0.dev0
* Pytorch 2.0.1+cu118
* Datasets 2.7.1
* Tokenizers 0.14.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
"TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
61,
115,
4,
36
] | [
"passage: TAGS\n#transformers #pytorch #bert #text-classification #generated_from_trainer #base_model-allenai/scibert_scivocab_uncased #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.1\n* training\\_steps: 10### Training results### Framework versions\n\n\n* Transformers 4.35.0.dev0\n* Pytorch 2.0.1+cu118\n* Datasets 2.7.1\n* Tokenizers 0.14.1"
] | [
-0.11901397258043289,
0.05818003788590431,
-0.002576048718765378,
0.1277925819158554,
0.17159773409366608,
0.03451113402843475,
0.12229453772306442,
0.1298956274986267,
-0.06671148538589478,
0.033512748777866364,
0.12027176469564438,
0.14603783190250397,
0.020510073751211166,
0.126299649477005,
-0.05034932121634483,
-0.29416623711586,
-0.00404456676915288,
0.017384430393576622,
-0.0642859935760498,
0.13318662345409393,
0.08560918271541595,
-0.13605624437332153,
0.09114351868629456,
-0.015392021276056767,
-0.159823939204216,
-0.0037332659121602774,
0.013471226207911968,
-0.04724876955151558,
0.15224698185920715,
0.014589891768991947,
0.12311960011720657,
0.022003665566444397,
0.11630265414714813,
-0.1794971525669098,
0.00951378047466278,
0.04702853411436081,
0.024641675874590874,
0.09206708520650864,
0.059074025601148605,
-0.004342362750321627,
0.11590910702943802,
-0.08297522366046906,
0.06359444558620453,
0.022770538926124573,
-0.1352519392967224,
-0.1828848421573639,
-0.07456278055906296,
0.033562902361154556,
0.07364761829376221,
0.08986175805330276,
-0.007848646491765976,
0.10108339786529541,
-0.09455756843090057,
0.11143719404935837,
0.22979314625263214,
-0.259819895029068,
-0.07738423347473145,
0.005848895292729139,
0.009438644163310528,
0.09249238669872284,
-0.11150462180376053,
-0.013401368632912636,
0.032099317759275436,
0.03735756501555443,
0.12356588244438171,
-0.026762839406728745,
-0.12812402844429016,
0.025542449206113815,
-0.13912871479988098,
-0.028142064809799194,
0.0989532321691513,
0.022599712014198303,
-0.041416529566049576,
-0.03637605905532837,
-0.07243399322032928,
-0.17004109919071198,
-0.038593802601099014,
-0.013557426631450653,
0.030867302790284157,
-0.04645764082670212,
-0.07389318943023682,
-0.029522769153118134,
-0.08763272315263748,
-0.09792207926511765,
-0.0317096970975399,
0.1754024177789688,
0.04006700962781906,
0.011795608326792717,
-0.005651060026139021,
0.13273045420646667,
0.01592053286731243,
-0.1539955884218216,
0.010227619670331478,
0.011534040793776512,
-0.04718701168894768,
-0.046790141612291336,
-0.054531071335077286,
-0.01920982636511326,
-0.01937848888337612,
0.13068316876888275,
-0.0224196445196867,
0.04813467711210251,
0.036371175199747086,
0.019081860780715942,
-0.10121357440948486,
0.20113731920719147,
-0.05527842044830322,
-0.01925429329276085,
-0.01563994772732258,
0.09469735622406006,
0.0031972702126950026,
-0.034310970455408096,
-0.10510993003845215,
-0.009774315170943737,
0.10418610274791718,
0.025440795347094536,
-0.07965856045484543,
0.06936469674110413,
-0.04465416446328163,
-0.037810344249010086,
0.013018951751291752,
-0.09988680481910706,
0.0360497385263443,
0.02196982316672802,
-0.10035677254199982,
-0.012892549857497215,
0.026745913550257683,
0.007516754325479269,
-0.015335002914071083,
0.13310319185256958,
-0.09553587436676025,
0.03754452243447304,
-0.09161999076604843,
-0.12142348289489746,
-0.006635898258537054,
-0.10080308467149734,
0.004844044800847769,
-0.0797857791185379,
-0.18677537143230438,
-0.026562189683318138,
0.043503765016794205,
-0.04893375560641289,
-0.033419907093048096,
-0.06850971281528473,
-0.06500028818845749,
0.024173462763428688,
-0.00661003403365612,
0.11994551867246628,
-0.057083480060100555,
0.12049513310194016,
0.03969322890043259,
0.06851155310869217,
-0.021050171926617622,
0.058683522045612335,
-0.10637225210666656,
0.009545463137328625,
-0.18570315837860107,
0.06510823965072632,
-0.056876182556152344,
0.04891849309206009,
-0.08016486465930939,
-0.10640599578619003,
0.02265378274023533,
0.0010017750319093466,
0.09074576944112778,
0.14159701764583588,
-0.17861756682395935,
-0.07789342105388641,
0.14320816099643707,
-0.07025504857301712,
-0.09437485039234161,
0.12230890244245529,
-0.06581716239452362,
0.03938906267285347,
0.06416157633066177,
0.16026924550533295,
0.08882725238800049,
-0.07350697368383408,
0.021781641989946365,
-0.008974314667284489,
0.08519653230905533,
-0.00022044969955459237,
0.07018810510635376,
0.014109667390584946,
0.002545085269957781,
0.026437684893608093,
-0.0651792585849762,
0.043953098356723785,
-0.10880604386329651,
-0.09554602950811386,
-0.0324493907392025,
-0.1022239476442337,
0.0916934609413147,
0.07505540549755096,
0.07579222321510315,
-0.08850742131471634,
-0.089222252368927,
0.055619221180677414,
0.0985446646809578,
-0.06437988579273224,
0.019499313086271286,
-0.060649748891592026,
0.056890618056058884,
-0.01900000497698784,
-0.02205251157283783,
-0.18833237886428833,
-0.044436659663915634,
0.019157113507390022,
0.047872450202703476,
0.02186409756541252,
0.00045015805517323315,
0.08708575367927551,
0.0866568312048912,
-0.06160221993923187,
-0.0361439511179924,
-0.045959919691085815,
-0.007862553000450134,
-0.13928194344043732,
-0.20520615577697754,
-0.06745630502700806,
-0.01925692893564701,
0.13769719004631042,
-0.2183861881494522,
0.037079013884067535,
-0.007348588202148676,
0.08332866430282593,
0.024990936741232872,
-0.024930676445364952,
-0.034517690539360046,
0.07566022127866745,
-0.03521628677845001,
-0.054699141532182693,
0.06778186559677124,
-0.013281183317303658,
-0.08445002138614655,
-0.05159826949238777,
-0.12069648504257202,
0.13833706080913544,
0.10510299354791641,
-0.09664520621299744,
-0.09384512156248093,
0.00017051029135473073,
-0.05287089943885803,
-0.030761366710066795,
-0.04911035671830177,
0.023375196382403374,
0.17236484587192535,
0.008742145262658596,
0.14939284324645996,
-0.05581888183951378,
-0.03788982704281807,
0.019421620294451714,
-0.011583386920392513,
0.04256878420710564,
0.13375967741012573,
0.09930118918418884,
-0.0776350200176239,
0.12650702893733978,
0.12022749334573746,
-0.08333223313093185,
0.14708787202835083,
-0.024369217455387115,
-0.08075268566608429,
-0.012210254557430744,
-0.02505146712064743,
0.0002824731054715812,
0.10511519014835358,
-0.12431631237268448,
-0.0215995404869318,
0.011175156570971012,
0.014787323772907257,
-0.0028115040622651577,
-0.20958121120929718,
-0.036434486508369446,
0.045592378824949265,
-0.05043448507785797,
-0.03260205686092377,
-0.011865167878568172,
0.007554288487881422,
0.11771690845489502,
0.010509821586310863,
-0.09166694432497025,
0.013673322275280952,
-0.0037684461567550898,
-0.06935665011405945,
0.20832806825637817,
-0.06984768062829971,
-0.11672678589820862,
-0.12158824503421783,
-0.05929696187376976,
-0.0586533397436142,
0.0280260369181633,
0.03781340643763542,
-0.09433113783597946,
-0.006740112788975239,
-0.05962882936000824,
0.029822850599884987,
0.007949607446789742,
0.03813331574201584,
-0.025276321917772293,
-0.019040174782276154,
0.056528396904468536,
-0.09242632985115051,
-0.005058479029685259,
-0.05822993814945221,
-0.07987936586141586,
0.03479737788438797,
0.032254450023174286,
0.11672104895114899,
0.15515868365764618,
-0.01878446154296398,
0.014477611519396305,
-0.02562427893280983,
0.22950634360313416,
-0.07527721673250198,
-0.008824112825095654,
0.14120011031627655,
-0.030566204339265823,
0.04989225044846535,
0.12992459535598755,
0.06580356508493423,
-0.0791751816868782,
0.008396334014832973,
0.04307877644896507,
-0.03539864346385002,
-0.2109011560678482,
-0.03431623801589012,
-0.05362283065915108,
0.009744088165462017,
0.10231099277734756,
0.013568169437348843,
0.030129902064800262,
0.07117430120706558,
0.015666477382183075,
0.04315505921840668,
-0.02757888287305832,
0.059478260576725006,
0.09366433322429657,
0.041944701224565506,
0.13579276204109192,
-0.024375123903155327,
-0.06819424033164978,
0.037948522716760635,
-0.022085273638367653,
0.20014189183712006,
-0.025523032993078232,
0.11680585891008377,
0.02222176268696785,
0.15596207976341248,
-0.01410122774541378,
0.08551263064146042,
0.00762378191575408,
-0.038557764142751694,
-0.009901408106088638,
-0.04264175891876221,
-0.059738889336586,
0.024259664118289948,
-0.05837314948439598,
0.06663483381271362,
-0.15758764743804932,
0.013186830095946789,
0.055136557668447495,
0.2756836712360382,
0.0657530203461647,
-0.324420690536499,
-0.11121397465467453,
0.004656370263546705,
-0.0358431302011013,
-0.0288983266800642,
0.010279613547027111,
0.12801319360733032,
-0.1060156524181366,
0.023086318746209145,
-0.07185039669275284,
0.09030719846487045,
-0.059915751218795776,
0.05608891695737839,
0.0810890644788742,
0.09270965307950974,
-0.009921351447701454,
0.07352972775697708,
-0.27440086007118225,
0.27836936712265015,
0.011724126525223255,
0.06460896134376526,
-0.07201258093118668,
-0.008825221098959446,
0.05213971436023712,
0.08747752010822296,
0.061413757503032684,
-0.015369323082268238,
-0.03311979025602341,
-0.23811845481395721,
-0.06900801509618759,
0.02216516248881817,
0.1168346032500267,
-0.07069330662488937,
0.11355267465114594,
-0.04348219558596611,
-0.001260775257833302,
0.05706687644124031,
-0.0360967181622982,
-0.05026037618517876,
-0.08130277693271637,
0.004692509304732084,
0.0006710363086313009,
-0.0056432634592056274,
-0.0522150956094265,
-0.12059471756219864,
-0.07378477603197098,
0.15097208321094513,
-0.03317046910524368,
-0.03501072898507118,
-0.13393531739711761,
0.08239786326885223,
0.09744047373533249,
-0.09147252142429352,
0.02531256154179573,
0.015379579737782478,
0.07419422268867493,
0.03589802235364914,
-0.057002075016498566,
0.11208871006965637,
-0.06780438125133514,
-0.20542557537555695,
-0.06215536594390869,
0.11192924529314041,
0.04735414683818817,
0.07036179304122925,
-0.020548027008771896,
0.020865444093942642,
-0.022519182413816452,
-0.08443796634674072,
0.028605297207832336,
-0.011877520009875298,
0.04282940551638603,
0.04024451971054077,
-0.06162833794951439,
0.04856330528855324,
-0.07127387076616287,
-0.006145413964986801,
0.1625979244709015,
0.24637360870838165,
-0.09921432286500931,
0.02788374572992325,
0.034775931388139725,
-0.0656699389219284,
-0.20158319175243378,
0.036423228681087494,
0.07201619446277618,
0.021215995773673058,
0.05048254504799843,
-0.20201563835144043,
0.09048860520124435,
0.09247171133756638,
-0.010620692744851112,
0.08988986164331436,
-0.310491144657135,
-0.12542957067489624,
0.11341096460819244,
0.12561342120170593,
0.1108081042766571,
-0.14193251729011536,
-0.02108059450984001,
-0.005432373844087124,
-0.08898790180683136,
0.08648601174354553,
-0.07302761822938919,
0.1282854825258255,
-0.026651719585061073,
0.08701416105031967,
0.03420214354991913,
-0.04500488564372063,
0.11857745051383972,
0.015701593831181526,
0.09402630478143692,
-0.050300292670726776,
-0.05290866270661354,
0.019195053726434708,
-0.050733648240566254,
0.008490592241287231,
-0.07844416797161102,
0.04326234757900238,
-0.11332516372203827,
-0.01768225058913231,
-0.09427167475223541,
0.022101793438196182,
-0.037489935755729675,
-0.06919790804386139,
-0.014210204593837261,
0.03682906925678253,
0.0452902615070343,
-0.003283283207565546,
0.12508025765419006,
-0.01746389828622341,
0.15768669545650482,
0.10516431927680969,
0.0875639095902443,
-0.0227096825838089,
-0.03194094076752663,
-0.0052976650185883045,
-0.004041146021336317,
0.05932541564106941,
-0.12370879203081131,
0.03571442887187004,
0.14751116931438446,
0.03673496097326279,
0.13183650374412537,
0.08534897863864899,
-0.00836755521595478,
0.00935033243149519,
0.051651839166879654,
-0.17602138221263885,
-0.06006268411874771,
-0.021389655768871307,
-0.06267506629228592,
-0.13471774756908417,
0.05380474030971527,
0.11202312260866165,
-0.0660976991057396,
-0.013245919719338417,
-0.011564429849386215,
0.011793426238000393,
-0.041937701404094696,
0.2176782488822937,
0.060630086809396744,
0.062402766197919846,
-0.10295094549655914,
0.06215422600507736,
0.050319839268922806,
-0.058207303285598755,
0.003895225003361702,
0.09990619868040085,
-0.08766604214906693,
-0.037679675966501236,
0.05678989738225937,
0.1470041126012802,
-0.06388426572084427,
-0.016999077051877975,
-0.15979807078838348,
-0.11466078460216522,
0.06842485070228577,
0.17307837307453156,
0.10142982006072998,
0.021465353667736053,
-0.07063306123018265,
0.028022920712828636,
-0.12008003890514374,
0.09886689484119415,
0.058014143258333206,
0.07358678430318832,
-0.14977742731571198,
0.1904129981994629,
-0.020299771800637245,
0.054479263722896576,
-0.02782033383846283,
0.014952929690480232,
-0.11292684078216553,
0.01348118670284748,
-0.12491442263126373,
-0.04283739626407623,
-0.009110449813306332,
0.002026754431426525,
-0.009667567908763885,
-0.07269591838121414,
-0.04411114752292633,
0.0006685203406959772,
-0.11673595011234283,
-0.02537863701581955,
0.022737909108400345,
0.037733759731054306,
-0.12518583238124847,
-0.0402621254324913,
0.03482877090573311,
-0.08498319238424301,
0.08136511594057083,
0.059110239148139954,
0.018400680273771286,
0.05162622034549713,
-0.1328829973936081,
0.009403408505022526,
0.033759020268917084,
0.006414457224309444,
0.06114351376891136,
-0.06425705552101135,
-0.002251851372420788,
-0.03882884234189987,
0.06158752739429474,
0.020769556984305382,
0.10363008081912994,
-0.12195582687854767,
0.029798191040754318,
-0.006369945127516985,
-0.07678867131471634,
-0.06485657393932343,
0.05917947366833687,
0.06739173084497452,
0.0283164344727993,
0.1685207635164261,
-0.08948566019535065,
0.05891484394669533,
-0.21968673169612885,
-0.001253497670404613,
-0.007980077527463436,
-0.11817620694637299,
-0.1272432655096054,
-0.0810258612036705,
0.08541562408208847,
-0.05129242688417435,
0.08580049127340317,
0.019304202869534492,
0.10081743448972702,
0.024991974234580994,
-0.031711556017398834,
0.03155599161982536,
0.03141847625374794,
0.17715874314308167,
0.04634115844964981,
-0.04880451038479805,
0.07107756286859512,
0.049985405057668686,
0.09009335190057755,
0.14697232842445374,
0.21901912987232208,
0.12634581327438354,
-0.014868056401610374,
0.08794914186000824,
0.04522809758782387,
-0.07032813131809235,
-0.16993020474910736,
-0.003409351222217083,
-0.04117885231971741,
0.08751802891492844,
-0.04521610587835312,
0.21402950584888458,
0.05324692651629448,
-0.16769829392433167,
0.04709490388631821,
-0.04904929921030998,
-0.10169430077075958,
-0.12233421206474304,
-0.020101699978113174,
-0.07345074415206909,
-0.1422247737646103,
-0.003045957302674651,
-0.10814426094293594,
0.0366266667842865,
0.10503026843070984,
0.010557522065937519,
-0.012487920001149178,
0.1692514270544052,
0.020734649151563644,
0.03901068493723869,
0.07398701459169388,
0.010335315018892288,
-0.008188899606466293,
-0.08353403955698013,
-0.08800452947616577,
-0.006714549846947193,
-0.014311039820313454,
0.02907426655292511,
-0.07704468071460724,
-0.06452365964651108,
0.028548449277877808,
-0.008489912375807762,
-0.11360584944486618,
0.017295027151703835,
0.009815744124352932,
0.07060473412275314,
0.05808265879750252,
0.00941329263150692,
0.010206412523984909,
-0.02238629199564457,
0.25125652551651,
-0.0960116758942604,
-0.049817103892564774,
-0.11013137549161911,
0.29036957025527954,
0.03676379472017288,
0.0009454129030928016,
0.023617547005414963,
-0.0763344019651413,
-0.005779159255325794,
0.23034073412418365,
0.1833788901567459,
-0.11508965492248535,
-0.005871182307600975,
0.004006389994174242,
-0.005603387951850891,
-0.005525580607354641,
0.10405922681093216,
0.09686476737260818,
0.04188643768429756,
-0.09830514341592789,
-0.029429392889142036,
-0.025304166600108147,
-0.03369209170341492,
-0.03233819082379341,
0.07901940494775772,
0.03476811945438385,
0.01910577528178692,
-0.059262629598379135,
0.04909488186240196,
-0.07310646027326584,
-0.10365349799394608,
0.06296191364526749,
-0.2146100252866745,
-0.17133377492427826,
-0.03237602859735489,
0.0671398937702179,
0.023249730467796326,
0.07686960697174072,
-0.025128716602921486,
-0.012129192240536213,
0.08507080376148224,
-0.012151479721069336,
-0.06048182398080826,
-0.10727165639400482,
0.11010511964559555,
-0.09671607613563538,
0.2037721872329712,
-0.05013314262032509,
0.04296911135315895,
0.11505188047885895,
0.04981324449181557,
-0.06970255821943283,
0.039352573454380035,
0.059884145855903625,
-0.09763424843549728,
0.019823962822556496,
0.1353226900100708,
-0.04499897360801697,
0.07689474523067474,
0.03739776462316513,
-0.15700021386146545,
0.020831217989325523,
-0.07056008279323578,
-0.07273272424936295,
-0.0268925242125988,
-0.04564611241221428,
-0.03274153172969818,
0.13747341930866241,
0.25035497546195984,
-0.031473394483327866,
0.010550284758210182,
-0.07453306019306183,
0.0051280297338962555,
0.057716067880392075,
0.05326291173696518,
-0.06679362803697586,
-0.23404237627983093,
0.003170912154018879,
0.06216126307845116,
-0.0139358164742589,
-0.2350955605506897,
-0.0933397188782692,
0.010343954898416996,
-0.0533423013985157,
-0.08519208431243896,
0.09380955994129181,
0.06191926449537277,
0.04345443844795227,
-0.04334281012415886,
-0.06560803204774857,
-0.06860619783401489,
0.16565631330013275,
-0.15820372104644775,
-0.08228906244039536
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-4.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:41:54+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
194,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06399306654930115,
-0.029307762160897255,
-0.004331388510763645,
0.03184039518237114,
0.018488328903913498,
0.03748767077922821,
0.10618676990270615,
0.07814439386129379,
0.07063937187194824,
0.07519446313381195,
-0.04871003329753876,
-0.02137892320752144,
0.07951480895280838,
0.10956592857837677,
0.007890605367720127,
-0.23595072329044342,
0.08250588178634644,
-0.03575211763381958,
-0.026412149891257286,
0.03893791511654854,
0.14763003587722778,
-0.05930296331644058,
0.05646873265504837,
-0.0034751244820654392,
-0.05727600306272507,
0.05131329223513603,
-0.026073675602674484,
-0.007598201744258404,
0.09508240222930908,
0.0628398060798645,
0.041950274258852005,
0.0221012644469738,
-0.018701795488595963,
-0.15748009085655212,
0.004922133404761553,
-0.017628159373998642,
-0.04106835648417473,
0.027759606018662453,
0.11616206914186478,
-0.045659661293029785,
0.11880303174257278,
-0.12680159509181976,
-0.018632367253303528,
0.08243369311094284,
-0.12379294633865356,
-0.10155918449163437,
-0.20841088891029358,
0.15503807365894318,
0.07979700714349747,
0.008151463232934475,
-0.04288718104362488,
0.07205401360988617,
-0.054492317140102386,
0.0433170422911644,
0.08753809332847595,
-0.2521143853664398,
-0.04418989270925522,
0.17082107067108154,
0.007189796306192875,
-0.04941372200846672,
-0.05787082388997078,
0.06996709108352661,
0.029280396178364754,
0.0067648859694600105,
-0.04790947958827019,
-0.04312778264284134,
0.2201330065727234,
-0.03288035839796066,
-0.1199050173163414,
-0.024701880291104317,
0.08322399854660034,
0.08842694014310837,
-0.04718442261219025,
-0.15207047760486603,
-0.011922257021069527,
-0.022255953401327133,
-0.04938272759318352,
-0.09886841475963593,
0.003574386006221175,
-0.02095438726246357,
0.07563887536525726,
-0.05726063251495361,
-0.04117260500788689,
0.0583573654294014,
-0.05984218046069145,
0.1629648655653,
-0.0014757480239495635,
-0.03302643075585365,
0.05737544223666191,
0.03662708401679993,
-0.2148275226354599,
-0.08567781001329422,
-0.04239440709352493,
-0.054734885692596436,
-0.1342768371105194,
-0.0431794635951519,
-0.04418531805276871,
-0.08217091858386993,
-0.007381834089756012,
0.23022499680519104,
-0.02492726407945156,
0.02804037742316723,
0.041754867881536484,
0.010314702056348324,
0.04419277235865593,
0.1185997873544693,
-0.05733427032828331,
-0.1965552270412445,
-0.02631615847349167,
-0.01892445981502533,
0.06395958364009857,
0.013402312994003296,
-0.02684073895215988,
-0.01519184559583664,
0.02383972890675068,
0.02890627458691597,
0.02422332763671875,
0.03168574720621109,
-0.07110860198736191,
-0.03037358447909355,
0.14692901074886322,
-0.10123810917139053,
0.04444355517625809,
-0.03552069887518883,
-0.03803017735481262,
0.07610528916120529,
-0.030228903517127037,
0.023446066305041313,
-0.01123465783894062,
0.14718079566955566,
-0.04342922940850258,
-0.02282293513417244,
-0.04051238298416138,
-0.05655231699347496,
0.056725140661001205,
-0.08175330609083176,
-0.06928885728120804,
-0.0921410471200943,
-0.07629581540822983,
-0.0781061053276062,
0.019695017486810684,
-0.04772234335541725,
-0.00727304769679904,
-0.00963415578007698,
-0.03471999987959862,
0.004572734236717224,
0.03700930252671242,
-0.03850525617599487,
-0.006236410234123468,
-0.05285993963479996,
-0.02465738169848919,
0.016344578936696053,
-0.07584261149168015,
0.020161736756563187,
-0.09869375079870224,
0.055850550532341,
-0.3218844532966614,
0.09342993050813675,
-0.11400728672742844,
0.009930809028446674,
-0.1314808875322342,
-0.0335058830678463,
0.016112782061100006,
0.005322059616446495,
0.013146123848855495,
0.11860170215368271,
-0.15911765396595,
-0.09210728108882904,
0.09710568189620972,
-0.13270564377307892,
-0.06123318895697594,
0.13335616886615753,
0.006097014062106609,
-0.007668835576623678,
0.0786266103386879,
0.1425439566373825,
0.14123733341693878,
-0.04838160052895546,
-0.13677071034908295,
-0.06624309718608856,
0.027206536382436752,
0.1544196456670761,
0.06648306548595428,
-0.0522027313709259,
-0.0182489063590765,
0.03598593920469284,
-0.031956933438777924,
-0.0026491705793887377,
-0.016834797337651253,
-0.05705594643950462,
-0.023978661745786667,
-0.002377238357439637,
0.096376933157444,
-0.015163181349635124,
0.002383065177127719,
-0.07838787883520126,
-0.09363235533237457,
0.039118777960538864,
0.0882512778043747,
0.041436195373535156,
0.01435884553939104,
-0.06161881238222122,
0.09994251281023026,
-0.03056234121322632,
0.022030595690011978,
-0.10238077491521835,
-0.15781597793102264,
0.03258318454027176,
-0.15215061604976654,
0.06616183370351791,
0.046265859156847,
0.06192070245742798,
0.027611956000328064,
-0.11285637319087982,
-0.06351099908351898,
-0.02237211912870407,
0.016440147534012794,
0.007992693223059177,
-0.12434909492731094,
-0.08924297988414764,
-0.04348140209913254,
0.22069425880908966,
-0.054984692484140396,
-0.007727441843599081,
0.06026673689484596,
0.2515117824077606,
0.018960142508149147,
-0.014410633593797684,
-0.05082327499985695,
0.07166947424411774,
0.00031130402931012213,
0.0197899267077446,
0.01994451880455017,
-0.04336968809366226,
-0.14637522399425507,
0.095527283847332,
-0.055198993533849716,
-0.02979397401213646,
0.06157594546675682,
0.054885584861040115,
-0.09651610255241394,
-0.011951263062655926,
-0.05258362367749214,
-0.029914764687418938,
0.006216008216142654,
-0.12069366872310638,
0.042220812290906906,
0.04080578312277794,
0.03252416476607323,
-0.0428742878139019,
-0.048637744039297104,
0.01096632331609726,
0.00038561708061024547,
-0.12152905762195587,
0.10360410064458847,
-0.08400224894285202,
-0.21976810693740845,
0.10374713689088821,
0.15290415287017822,
0.07434318959712982,
0.12372945249080658,
-0.013827105984091759,
-0.022371480241417885,
-0.1197407990694046,
0.0015785678988322616,
0.021130502223968506,
0.042604196816682816,
0.04024510458111763,
0.05408281460404396,
0.03447001054883003,
0.043456509709358215,
0.02863817848265171,
-0.02880154922604561,
0.04750415310263634,
-0.009588376618921757,
0.0040976181626319885,
0.12772449851036072,
0.10417608171701431,
0.02852170541882515,
0.08533679693937302,
0.03910043090581894,
-0.0404503159224987,
0.01288803294301033,
-0.02749737910926342,
-0.06698718667030334,
0.10686532407999039,
-0.10741190612316132,
-0.16609297692775726,
-0.17008981108665466,
-0.051623594015836716,
-0.18006007373332977,
-0.024816229939460754,
0.03015192784368992,
-0.03652867674827576,
-0.07785869389772415,
-0.07390360534191132,
0.13494938611984253,
0.03926362842321396,
-0.05240222439169884,
-0.05028913915157318,
-0.03344966843724251,
0.042253971099853516,
-0.10105009377002716,
-0.04346547648310661,
0.0018156386213377118,
-0.050230301916599274,
0.021803520619869232,
0.027263345196843147,
0.0939897820353508,
0.060186538845300674,
-0.00003084815762122162,
-0.02964363619685173,
-0.040875863283872604,
0.17850545048713684,
-0.1293039619922638,
0.11540871858596802,
0.04773532599210739,
-0.0027224498335272074,
0.06698734313249588,
0.16732339560985565,
0.05765581503510475,
-0.04658689722418785,
-0.011350571177899837,
0.06666777282953262,
-0.030819516628980637,
-0.19200317561626434,
-0.1769828051328659,
-0.06770068407058716,
0.007487337104976177,
-0.027668775990605354,
0.04515873268246651,
0.06839024275541306,
-0.0003817266842816025,
-0.09345950931310654,
-0.07025164365768433,
0.036232683807611465,
0.04945642128586769,
0.17080222070217133,
0.009348781779408455,
0.05902401730418205,
-0.06639035046100616,
-0.06039080023765564,
0.09840981662273407,
0.00460409838706255,
0.08121812343597412,
0.06772522628307343,
0.06868960708379745,
0.051509808748960495,
0.012179503217339516,
0.0003098213055636734,
0.04621569812297821,
0.015963001176714897,
-0.02683101035654545,
-0.011713831685483456,
-0.08623592555522919,
0.07211727648973465,
0.08716612309217453,
0.10278671234846115,
0.03722165524959564,
-0.051564089953899384,
-0.06990828365087509,
0.12818388640880585,
0.1740635186433792,
0.11096327751874924,
-0.12066809087991714,
-0.10149893164634705,
0.059367019683122635,
-0.012059872969985008,
0.020104901865124702,
-0.03295712172985077,
-0.051309432834386826,
-0.13940629363059998,
0.20302264392375946,
-0.018019717186689377,
0.06347235292196274,
0.0055109113454818726,
-0.0001949289726326242,
-0.02158956229686737,
0.06696530431509018,
0.017838791012763977,
0.050336673855781555,
-0.141648069024086,
0.14686071872711182,
0.010985979810357094,
0.021909072995185852,
-0.01216028444468975,
0.05184603109955788,
0.01702127605676651,
0.05610823258757591,
0.10946010798215866,
0.04561372473835945,
-0.09489621967077255,
-0.08675345033407211,
-0.04737729951739311,
-0.05841971933841705,
0.13532322645187378,
-0.0978664755821228,
0.08815576881170273,
-0.005218944977968931,
-0.04548734799027443,
-0.058141354471445084,
0.16146430373191833,
-0.15934163331985474,
-0.10354094952344894,
0.10525287687778473,
-0.04850301891565323,
0.028323976323008537,
-0.07308647781610489,
-0.04719609022140503,
-0.24660754203796387,
0.2158661186695099,
0.015384898521006107,
0.00384410098195076,
-0.09645523875951767,
-0.0031059239991009235,
0.16681231558322906,
-0.09507244825363159,
0.038081102073192596,
-0.0030803470872342587,
0.08820296823978424,
-0.09653045237064362,
-0.08165580779314041,
0.04465007409453392,
-0.07744460552930832,
-0.14185139536857605,
0.011764489114284515,
0.19781433045864105,
0.022616581991314888,
0.037971191108226776,
0.025954479351639748,
0.09160146117210388,
0.08607495576143265,
-0.10742948949337006,
0.08422011882066727,
0.0879930630326271,
-0.030707523226737976,
0.1495898962020874,
-0.033895764499902725,
-0.15872009098529816,
-0.05146879330277443,
-0.0411202572286129,
0.10311200469732285,
0.34257373213768005,
-0.06467219442129135,
0.08678370714187622,
0.08080018311738968,
-0.042148686945438385,
-0.19632753729820251,
0.007789987605065107,
0.01917961612343788,
0.02523370087146759,
0.00445806048810482,
-0.05062577873468399,
0.10844220966100693,
0.13072700798511505,
-0.020657971501350403,
-0.04788164049386978,
-0.20526863634586334,
-0.1290767788887024,
0.011098041199147701,
-0.06513512134552002,
-0.012189746834337711,
-0.06006240472197533,
-0.10408233106136322,
-0.07066939771175385,
-0.16428698599338531,
0.05405878648161888,
0.04084285721182823,
0.048617810010910034,
0.0022039911709725857,
-0.022370649501681328,
0.03627573698759079,
-0.018924014642834663,
0.18879440426826477,
0.06920293718576431,
0.02023874782025814,
-0.07239631563425064,
-0.04903245344758034,
0.11341241002082825,
-0.037153460085392,
0.09947021305561066,
0.04320525750517845,
-0.005356981884688139,
-0.11361989378929138,
0.012292295694351196,
-0.09132620692253113,
0.11476991325616837,
-0.09256471693515778,
-0.030837757512927055,
-0.08548001945018768,
0.140853151679039,
0.0238761268556118,
-0.015673771500587463,
0.12679731845855713,
-0.032148636877536774,
0.1091376468539238,
0.13080862164497375,
0.0701589584350586,
0.0031420262530446053,
-0.07472140341997147,
0.0012759605888277292,
-0.06170511618256569,
0.019098622724413872,
0.05546579137444496,
0.04380559176206589,
0.13516387343406677,
-0.01975567452609539,
0.14712172746658325,
-0.018929120153188705,
-0.13689684867858887,
-0.023719700053334236,
0.11753503233194351,
-0.11223208159208298,
-0.2821040451526642,
-0.04018678143620491,
-0.034599367529153824,
0.012139244005084038,
0.061747223138809204,
0.19457398355007172,
-0.03780088946223259,
-0.023348631337285042,
0.021921996027231216,
0.04527831822633743,
-0.09146711230278015,
0.08037439733743668,
-0.021092766895890236,
0.042854998260736465,
-0.06821004301309586,
0.10744444280862808,
0.07745382934808731,
-0.22681817412376404,
-0.025234894827008247,
0.09869363903999329,
-0.06257352977991104,
-0.07561001926660538,
-0.09376278519630432,
0.1380273848772049,
-0.1314297765493393,
-0.0158812515437603,
-0.020428206771612167,
-0.1451670229434967,
0.04510067030787468,
0.09574063867330551,
0.04472881183028221,
0.011867307126522064,
0.035759247839450836,
0.00018209819972980767,
0.01861182041466236,
0.03738055005669594,
0.09328993409872055,
0.10640279203653336,
-0.042188696563243866,
0.03816690295934677,
-0.08412777632474899,
0.015412991866469383,
-0.021022463217377663,
-0.01296116691082716,
-0.09465239942073822,
-0.07739552855491638,
-0.10236255824565887,
-0.06780892610549927,
-0.08637113869190216,
-0.02516632154583931,
-0.0233630258589983,
-0.023995107039809227,
0.004357992671430111,
-0.027227098122239113,
-0.07138573378324509,
-0.050982195883989334,
-0.0155563373118639,
0.08955743908882141,
-0.1038399413228035,
-0.0027917600236833096,
0.06826390326023102,
-0.051174331456422806,
0.0997040867805481,
0.024214884266257286,
0.0514301136136055,
0.03840871900320053,
-0.03028056025505066,
-0.03847609460353851,
0.01905941218137741,
0.006461322773247957,
0.05086956173181534,
-0.14872215688228607,
-0.012150841765105724,
-0.0032515712082386017,
-0.051708951592445374,
0.050017524510622025,
-0.0625491663813591,
-0.10236836224794388,
0.008505094796419144,
0.016627702862024307,
-0.09351813793182373,
-0.07604184001684189,
0.027614127844572067,
0.0729801207780838,
0.035474032163619995,
0.11152558028697968,
-0.058112598955631256,
0.05574757605791092,
-0.1357552707195282,
-0.01836516335606575,
-0.02554010972380638,
-0.1030651405453682,
0.016653461381793022,
-0.011252963915467262,
0.022870756685733795,
-0.0031690942123532295,
0.11289511620998383,
-0.026996877044439316,
-0.025828121230006218,
0.06756260991096497,
-0.028583889827132225,
-0.0780753642320633,
0.0507938489317894,
0.07868042588233948,
0.04812269285321236,
-0.0161933284252882,
-0.08120898902416229,
-0.0014988429611548781,
-0.05973583459854126,
-0.007319761905819178,
0.1114073172211647,
0.14946618676185608,
0.11061113327741623,
0.05770944431424141,
0.08207276463508606,
-0.04891999810934067,
-0.028829647228121758,
-0.009984364733099937,
-0.029658254235982895,
0.08592110872268677,
-0.02771167829632759,
0.046906858682632446,
0.07054735720157623,
-0.23260782659053802,
0.1542171984910965,
-0.04036210849881172,
0.003158458275720477,
-0.019460761919617653,
-0.11383481323719025,
-0.07883796095848083,
-0.042559150606393814,
0.010454006493091583,
-0.08244454860687256,
0.009923603385686874,
0.00909425225108862,
0.03983308747410774,
-0.018626000732183456,
0.14684614539146423,
-0.22303219139575958,
-0.020058225840330124,
0.1173686534166336,
0.01384452823549509,
-0.01986311748623848,
0.04652818292379379,
-0.01833965815603733,
-0.05564456060528755,
0.0526103749871254,
0.003485421882942319,
0.042003098875284195,
0.01768621988594532,
-0.009287383407354355,
-0.030709892511367798,
-0.0902811586856842,
0.015258834697306156,
-0.013983458280563354,
0.019041143357753754,
0.026359856128692627,
0.02583683282136917,
0.016558820381760597,
-0.0304630808532238,
0.1203962042927742,
-0.04186068847775459,
-0.08964694291353226,
-0.09655430167913437,
0.18649248778820038,
0.015295648016035557,
0.07767751067876816,
0.014468489214777946,
-0.11231113970279694,
0.003929391037672758,
0.011538079008460045,
0.2675095796585083,
0.06229402497410774,
0.0249533299356699,
0.022313563153147697,
0.0121700968593359,
0.03887152299284935,
0.019862765446305275,
-0.021693918853998184,
0.17263448238372803,
-0.035285405814647675,
0.2010856717824936,
-0.05225873738527298,
-0.046340879052877426,
0.0050239525735378265,
0.1740422248840332,
-0.01995527185499668,
-0.01296838466078043,
-0.015919938683509827,
0.11686369776725769,
-0.02941286377608776,
-0.15430203080177307,
0.030166465789079666,
-0.1680508852005005,
-0.1555458903312683,
-0.04883193224668503,
0.03437260910868645,
0.0537441223859787,
0.11792732775211334,
0.015498639084398746,
-0.03355927765369415,
0.11172260344028473,
0.013434622436761856,
-0.01853151060640812,
-0.12593184411525726,
0.06253454834222794,
-0.005848423577845097,
0.10684798657894135,
0.020717212930321693,
0.04494085907936096,
0.12986114621162415,
-0.024106910452246666,
-0.10693443566560745,
0.014985187910497189,
0.045302849262952805,
-0.08386611938476562,
0.0049399384297430515,
0.15897472202777863,
-0.014517411589622498,
0.11524591594934464,
0.05899004638195038,
-0.13243594765663147,
0.07087180018424988,
0.0894564762711525,
-0.041566379368305206,
-0.06131380423903465,
0.1714923232793808,
-0.08582644164562225,
0.13566255569458008,
0.23618710041046143,
-0.026338893920183182,
0.025297190994024277,
-0.03987700119614601,
0.044342536479234695,
0.07475463300943375,
0.12010806798934937,
-0.017106499522924423,
-0.16066502034664154,
0.06056363880634308,
0.023605575785040855,
0.04031899571418762,
-0.14200104773044586,
-0.11799366027116776,
-0.015118377283215523,
0.02434436045587063,
-0.030886957421898842,
0.12695416808128357,
0.08972851932048798,
-0.0224088616669178,
0.001209444715641439,
-0.1888996660709381,
-0.008432535454630852,
0.07094074040651321,
-0.045595064759254456,
-0.016666574403643608
] |
null | null | sentence-transformers |
# celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
<!--- Describe your model here -->
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
```
pip install -U sentence-transformers
```
Then you can use the model like this:
```python
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp')
embeddings = model.encode(sentences)
print(embeddings)
```
## Evaluation Results
<!--- Describe how your model was evaluated -->
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp)
## Training
The model was trained with the parameters:
**DataLoader**:
`torch.utils.data.dataloader.DataLoader` of length 794 with parameters:
```
{'batch_size': 4, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
```
**Loss**:
`sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss` with parameters:
```
{'scale': 20.0, 'similarity_fct': 'cos_sim'}
```
**DataLoader**:
`torch.utils.data.dataloader.DataLoader` of length 989 with parameters:
```
{'batch_size': 4, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
```
**Loss**:
`sentence_transformers.losses.OnlineContrastiveLoss.OnlineContrastiveLoss`
Parameters of the fit()-Method:
```
{
"epochs": 1,
"evaluation_steps": 0,
"evaluator": "NoneType",
"max_grad_norm": 1,
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
"optimizer_params": {
"lr": 2.43e-06
},
"scheduler": "WarmupLinear",
"steps_per_epoch": null,
"warmup_steps": 99.0,
"weight_decay": 0.1
}
```
## Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': True, 'pooling_mode_mean_sqrt_len_tokens': True, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False})
(2): Dense({'in_features': 3072, 'out_features': 768, 'bias': True, 'activation_function': 'torch.nn.modules.activation.Tanh'})
(3): Normalize()
)
```
## Citing & Authors
<!--- Describe where people can find more information --> | {"library_name": "sentence-transformers", "tags": ["sentence-transformers", "feature-extraction", "sentence-similarity"], "pipeline_tag": "sentence-similarity"} | sentence-similarity | celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp | [
"sentence-transformers",
"pytorch",
"tflite",
"safetensors",
"mpnet",
"feature-extraction",
"sentence-similarity",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:42:12+00:00 | [] | [] | TAGS
#sentence-transformers #pytorch #tflite #safetensors #mpnet #feature-extraction #sentence-similarity #endpoints_compatible #region-us
|
# celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp
This is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have sentence-transformers installed:
Then you can use the model like this:
## Evaluation Results
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL
## Training
The model was trained with the parameters:
DataLoader:
'URL.dataloader.DataLoader' of length 794 with parameters:
Loss:
'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:
DataLoader:
'URL.dataloader.DataLoader' of length 989 with parameters:
Loss:
'sentence_transformers.losses.OnlineContrastiveLoss.OnlineContrastiveLoss'
Parameters of the fit()-Method:
## Full Model Architecture
## Citing & Authors
| [
"# celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 794 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 989 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.OnlineContrastiveLoss.OnlineContrastiveLoss' \n\nParameters of the fit()-Method:",
"## Full Model Architecture",
"## Citing & Authors"
] | [
"TAGS\n#sentence-transformers #pytorch #tflite #safetensors #mpnet #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n",
"# celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.",
"## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:",
"## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL",
"## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 794 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 989 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.OnlineContrastiveLoss.OnlineContrastiveLoss' \n\nParameters of the fit()-Method:",
"## Full Model Architecture",
"## Citing & Authors"
] | [
49,
74,
38,
29,
136,
5,
6
] | [
"passage: TAGS\n#sentence-transformers #pytorch #tflite #safetensors #mpnet #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n# celik-muhammed/multi-qa-mpnet-base-cos-v1-finetuned-dtc-zoomcamp\n\nThis is a sentence-transformers model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.## Usage (Sentence-Transformers)\n\nUsing this model becomes easy when you have sentence-transformers installed:\n\n\n\nThen you can use the model like this:## Evaluation Results\n\n\n\nFor an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: URL## Training\nThe model was trained with the parameters:\n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 794 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.MultipleNegativesRankingLoss.MultipleNegativesRankingLoss' with parameters:\n \n\nDataLoader:\n\n'URL.dataloader.DataLoader' of length 989 with parameters:\n\n\nLoss:\n\n'sentence_transformers.losses.OnlineContrastiveLoss.OnlineContrastiveLoss' \n\nParameters of the fit()-Method:## Full Model Architecture## Citing & Authors"
] | [
-0.05125037580728531,
0.20253150165081024,
-0.007422384340316057,
0.053562115877866745,
0.07779410481452942,
0.046034324914216995,
0.08760224282741547,
0.11339737474918365,
-0.04566231369972229,
0.09642405807971954,
0.020030032843351364,
0.12055186182260513,
0.013104146346449852,
0.0023012515157461166,
-0.02399730123579502,
-0.22971229255199432,
0.02032032236456871,
-0.05261306092143059,
0.020170515403151512,
0.06465499103069305,
0.11078551411628723,
-0.08171363919973373,
0.05736469104886055,
-0.014610196463763714,
-0.0023217280395329,
0.010602216236293316,
0.005869620013982058,
-0.04480837285518646,
0.06145088002085686,
0.0360189750790596,
0.0539294071495533,
0.028304899111390114,
0.03941350057721138,
-0.1989029049873352,
0.016905073076486588,
0.07148891687393188,
-0.0037231401074677706,
0.058223657310009,
0.0010068521369248629,
0.003823285922408104,
0.0876985415816307,
-0.11423685401678085,
0.06437987089157104,
0.05122056230902672,
-0.062022965401411057,
-0.06578326970338821,
-0.07140114158391953,
0.037911973893642426,
0.0725981667637825,
0.10069909691810608,
-0.017509981989860535,
0.16949114203453064,
-0.0656704232096672,
0.08922997117042542,
0.116847462952137,
-0.2666207551956177,
-0.029109839349985123,
0.08461786061525345,
0.0194346122443676,
0.04259302094578743,
-0.12100865691900253,
-0.0006435082759708166,
-0.039291344583034515,
0.020708642899990082,
0.04322965443134308,
-0.05581243708729744,
-0.022680792957544327,
0.005473637022078037,
-0.07411610335111618,
0.012607534416019917,
0.18345873057842255,
0.04174532741308212,
-0.02831895649433136,
-0.11657099425792694,
-0.07389798760414124,
0.07261961698532104,
-0.05331028997898102,
0.00016922758368309587,
0.03612620756030083,
0.04338548332452774,
-0.03850734978914261,
-0.08248566091060638,
-0.09597164392471313,
-0.00873311422765255,
-0.05626267194747925,
-0.04349198192358017,
0.011486124247312546,
-0.02008131891489029,
0.032182034105062485,
0.04205818846821785,
-0.050137415528297424,
-0.09711548686027527,
-0.02641233056783676,
-0.04319113865494728,
-0.09007782489061356,
-0.009115193970501423,
-0.034363940358161926,
-0.11166397482156754,
0.05482633784413338,
0.046040184795856476,
0.08501803129911423,
0.021982666105031967,
-0.03341757878661156,
0.05370113626122475,
-0.027794629335403442,
0.1235010027885437,
-0.04996982216835022,
-0.13475866615772247,
0.03180655092000961,
0.030289173126220703,
0.04420895129442215,
0.00017077449592761695,
-0.059017110615968704,
-0.034254491329193115,
0.05581608787178993,
0.06527765840291977,
0.05297612398862839,
0.06575611978769302,
-0.024286139756441116,
-0.056738562881946564,
0.07146915048360825,
-0.10844489932060242,
0.02746001072227955,
0.029751135036349297,
-0.06891150027513504,
0.06984109431505203,
0.09155590832233429,
-0.015060419216752052,
-0.1060604378581047,
-0.023212771862745285,
-0.08443976193666458,
-0.03038944862782955,
-0.034804392606019974,
-0.16837407648563385,
0.010761933401226997,
-0.0013655686052516103,
-0.05874878913164139,
-0.10683304816484451,
-0.1916140913963318,
-0.05398954451084137,
0.040862131863832474,
-0.0004620069812517613,
-0.030166223645210266,
-0.09550004452466965,
-0.025059040635824203,
0.0007421763730235398,
-0.006273491308093071,
-0.09232684969902039,
-0.010548382066190243,
0.036889079958200455,
-0.11878418177366257,
0.07016931474208832,
0.01796366646885872,
0.04428589716553688,
-0.06345328688621521,
0.00036109070060774684,
-0.1431211233139038,
0.15574300289154053,
-0.053463973104953766,
0.09841565787792206,
-0.11585987359285355,
-0.026076918467879295,
0.08811993896961212,
-0.006552261300384998,
-0.0015110610984265804,
0.12147185951471329,
-0.20806743204593658,
-0.008595870807766914,
0.11956866830587387,
-0.03448624536395073,
-0.09798631817102432,
0.09492479264736176,
-0.05357646942138672,
0.11908192932605743,
0.1720183938741684,
0.10791879147291183,
0.19838613271713257,
-0.08625814318656921,
-0.03362121060490608,
0.030812902376055717,
-0.018716299906373024,
0.04058932885527611,
0.026584316045045853,
-0.032336317002773285,
0.10783018916845322,
-0.0074420650489628315,
0.003995784558355808,
0.028388185426592827,
0.0038545276038348675,
-0.055182747542858124,
0.006797545123845339,
-0.08216489106416702,
0.04112978279590607,
-0.03510347753763199,
0.002833649516105652,
-0.004399196244776249,
-0.07177402824163437,
0.08628195524215698,
0.09828977286815643,
-0.07213758677244186,
0.033604446798563004,
-0.06841056793928146,
-0.028648145496845245,
-0.03126562014222145,
0.004042743239551783,
-0.13967783749103546,
-0.19633261859416962,
0.0030596849974244833,
0.027695465832948685,
0.057378366589546204,
0.024661190807819366,
0.03374959155917168,
0.01588883250951767,
-0.020735297352075577,
0.003163169603794813,
0.04538378119468689,
0.016438346356153488,
-0.07384473830461502,
-0.09587015211582184,
-0.009219272062182426,
0.0026226232293993235,
0.07229450345039368,
-0.1141546368598938,
0.034473516047000885,
0.003016892122104764,
0.05436905100941658,
0.03581780567765236,
-0.01988856866955757,
0.024665649980306625,
-0.05787893384695053,
0.021239642053842545,
-0.07126650214195251,
0.03784926235675812,
0.042609479278326035,
-0.10735885798931122,
0.08647648245096207,
-0.14085668325424194,
-0.09336736053228378,
0.06596197932958603,
0.0762430727481842,
-0.058997973799705505,
-0.016718244180083275,
-0.016599737107753754,
0.00002682697777345311,
-0.047312699258327484,
-0.06753953546285629,
0.10826777666807175,
0.08218491077423096,
0.13293366134166718,
-0.07919613271951675,
-0.024978969246149063,
-0.03280346095561981,
-0.023504648357629776,
0.008733885362744331,
0.08058763295412064,
-0.05816837400197983,
-0.18217876553535461,
0.043222323060035706,
0.057495567947626114,
-0.06042462959885597,
0.09952287375926971,
-0.013302285224199295,
-0.06290014088153839,
-0.04241475835442543,
0.036334335803985596,
0.03627666458487511,
0.008935890160501003,
-0.03096124529838562,
0.030032075941562653,
0.05782607942819595,
0.02424447052180767,
-0.0005607487983070314,
-0.0704946219921112,
0.03567543625831604,
0.03480838984251022,
-0.02451413683593273,
0.025064455345273018,
0.004364845342934132,
0.03790600225329399,
0.0746937096118927,
0.018029751256108284,
-0.022748557850718498,
-0.0042187818326056,
-0.0464821420609951,
-0.0899668037891388,
0.15766137838363647,
-0.14996488392353058,
-0.1832309067249298,
-0.11743016541004181,
-0.01341862790286541,
-0.08164671808481216,
-0.0252677109092474,
0.052901748567819595,
-0.031191745772957802,
-0.09153957664966583,
-0.09883735328912735,
0.017571186646819115,
0.06358126550912857,
-0.026906508952379227,
-0.021434038877487183,
0.0608363077044487,
0.008257122710347176,
-0.12649090588092804,
-0.0025870841927826405,
-0.019638601690530777,
-0.07325142621994019,
0.0020815953612327576,
-0.05550893768668175,
0.037125591188669205,
0.09629352390766144,
0.0024396071676164865,
0.002574592363089323,
0.01019508671015501,
0.217134490609169,
-0.05954708904027939,
0.13374526798725128,
0.14066235721111298,
0.03682717680931091,
0.08221853524446487,
0.11040041595697403,
0.025512447580695152,
-0.04054349660873413,
0.05938825383782387,
0.07535450905561447,
0.002405855106189847,
-0.18826237320899963,
-0.11352341622114182,
-0.08855520188808441,
-0.02317110449075699,
0.09469053149223328,
0.03550860658288002,
0.01987254060804844,
0.0778312236070633,
-0.029215261340141296,
0.017384754493832588,
0.0586252361536026,
0.11635029315948486,
0.16726191341876984,
-0.004009745549410582,
0.07554352283477783,
-0.07254297286272049,
-0.03460068628191948,
0.08253621309995651,
0.03558758273720741,
0.16986270248889923,
-0.04479813948273659,
0.14970089495182037,
0.06043713167309761,
0.0778137594461441,
-0.016701102256774902,
0.10399489104747772,
-0.05593950301408768,
0.026114199310541153,
-0.01517887506633997,
-0.0871942862868309,
-0.015813518315553665,
0.04259772226214409,
0.07432319223880768,
-0.02674000710248947,
-0.02847384102642536,
0.06399901956319809,
0.09540817886590958,
0.1576710045337677,
0.08612234890460968,
-0.20776601135730743,
-0.05535255745053291,
0.020660899579524994,
-0.0630781501531601,
-0.06888636201620102,
0.012007968500256538,
0.07849263399839401,
-0.11584123969078064,
0.05928511917591095,
-0.012064961716532707,
0.10577594488859177,
-0.080084428191185,
-0.0015998957678675652,
-0.03770630806684494,
0.03913705050945282,
-0.018857741728425026,
0.08037197589874268,
-0.11916135996580124,
0.06907594203948975,
0.05196931213140488,
0.05202901363372803,
-0.037415795028209686,
0.015019557438790798,
0.08903315663337708,
0.05958741158246994,
0.14015378057956696,
-0.015364350751042366,
0.07820795476436615,
0.0014149390626698732,
-0.12541499733924866,
0.011303813196718693,
0.0349043570458889,
-0.03680606931447983,
0.1216132789850235,
-0.03952605277299881,
-0.038732271641492844,
0.010385550558567047,
0.009630183689296246,
-0.05891572684049606,
-0.18659155070781708,
-0.01454855129122734,
0.13361424207687378,
0.01081633660942316,
-0.02535964921116829,
-0.021001243963837624,
0.009287411347031593,
0.22666259109973907,
-0.03631790354847908,
-0.11022475361824036,
-0.11896374076604843,
-0.04310597851872444,
0.15139438211917877,
-0.08095291256904602,
0.012625906616449356,
-0.0230731014162302,
0.13289913535118103,
-0.0381665974855423,
-0.06585243344306946,
0.03230762109160423,
-0.054601140320301056,
-0.0658625140786171,
-0.016752850264310837,
0.103477843105793,
0.017629120498895645,
0.03843221813440323,
0.05459261313080788,
0.02756570838391781,
-0.05028385668992996,
-0.12360991537570953,
-0.11960206925868988,
0.0941847413778305,
0.0316949300467968,
0.07076836377382278,
-0.13997235894203186,
-0.048686202615499496,
-0.06801871210336685,
0.05442144721746445,
0.22747282683849335,
0.1718812733888626,
-0.0901825800538063,
0.0875406414270401,
0.11660460382699966,
-0.054311592131853104,
-0.21818111836910248,
-0.05387360230088234,
0.017826052382588387,
0.012832244858145714,
-0.0006386165623553097,
-0.10562966018915176,
0.08602543920278549,
0.051316775381565094,
-0.00030900872661732137,
-0.05551019683480263,
-0.254757821559906,
-0.11134985834360123,
0.09920329600572586,
0.01760333776473999,
-0.028922030702233315,
-0.11395702511072159,
-0.03200999274849892,
-0.07081454247236252,
-0.05468238890171051,
0.14128796756267548,
-0.12103988975286484,
0.11396284401416779,
0.02674713358283043,
0.0777212604880333,
0.056174036115407944,
-0.020769726485013962,
0.09141682088375092,
0.06833605468273163,
0.02631823532283306,
-0.03513181954622269,
0.0003595559683162719,
0.14607568085193634,
-0.09574950486421585,
0.1717047095298767,
-0.016794566065073013,
0.06858223676681519,
-0.1254180371761322,
-0.008525043725967407,
-0.07665177434682846,
-0.006352853029966354,
-0.03884662687778473,
-0.04610201716423035,
-0.0430137999355793,
0.06036911532282829,
0.17713136970996857,
-0.016445526853203773,
0.045345958322286606,
-0.027092237025499344,
0.009423209354281425,
0.1608041524887085,
0.06886180490255356,
0.0043032048270106316,
-0.17227067053318024,
0.03948577120900154,
-0.013570763170719147,
0.03441986069083214,
-0.10961342602968216,
0.06746009737253189,
0.07772853970527649,
-0.0025796587578952312,
0.10800641775131226,
0.024397222325205803,
-0.056991349905729294,
-0.017580589279532433,
0.025276385247707367,
-0.08466414362192154,
-0.11515051871538162,
-0.02921963483095169,
-0.019337795674800873,
-0.13200053572654724,
-0.03917849808931351,
0.1402691751718521,
-0.04031830281019211,
0.018429968506097794,
0.015940017998218536,
0.024360116571187973,
-0.0038695326074957848,
0.10080257803201675,
0.05430271103978157,
0.03043128177523613,
-0.07858633249998093,
0.10881046950817108,
0.0981387123465538,
-0.07401149719953537,
0.010935172438621521,
0.13864700496196747,
-0.11681187152862549,
-0.06305471062660217,
-0.030518658459186554,
0.09035714715719223,
-0.11036872118711472,
0.011307090520858765,
-0.06245531886816025,
-0.04613903537392616,
0.02646685391664505,
0.08020450919866562,
0.01959051750600338,
0.06258872896432877,
-0.08255556970834732,
-0.058910541236400604,
-0.11242792755365372,
0.10006683319807053,
0.0665280744433403,
0.0043266951106488705,
-0.05230662226676941,
0.09168552607297897,
-0.036307401955127716,
0.017952850088477135,
-0.00948554091155529,
-0.027594784274697304,
-0.03821529448032379,
0.003375439206138253,
-0.07189972698688507,
0.038269806653261185,
-0.09192701429128647,
0.010349063202738762,
0.002241012640297413,
0.04868635907769203,
-0.01725112274289131,
0.020481668412685394,
-0.046318419277668,
-0.07930769771337509,
-0.04783976450562477,
0.06135604903101921,
-0.14157472550868988,
-0.03860032185912132,
0.0014029062585905194,
-0.08174227923154831,
0.08197267353534698,
-0.03164709731936455,
-0.034562401473522186,
-0.00028132187435403466,
-0.03927324712276459,
-0.07527898997068405,
0.028119944036006927,
0.06240291893482208,
0.06979291141033173,
-0.06953106820583344,
0.007748579140752554,
-0.017289308831095695,
-0.013068669475615025,
0.001390043762512505,
0.046842388808727264,
-0.08990922570228577,
0.028632929548621178,
-0.04889357462525368,
0.01723279245197773,
-0.09317901730537415,
0.010666497983038425,
0.025707343593239784,
0.027666259557008743,
0.1366271823644638,
-0.05683235079050064,
0.07595393061637878,
-0.11393817514181137,
-0.022133110091090202,
0.010473870672285557,
-0.038244664669036865,
0.056467629969120026,
-0.048402346670627594,
0.06469495594501495,
-0.054868802428245544,
0.03927222639322281,
-0.038886211812496185,
0.017014672979712486,
0.04897768795490265,
0.008075334131717682,
-0.02365606278181076,
-0.016583450138568878,
0.07972129434347153,
0.044854786247015,
-0.017626117914915085,
0.01351007167249918,
0.006698156241327524,
0.041737571358680725,
0.015110389329493046,
0.07415305078029633,
0.0904664695262909,
0.0977846011519432,
0.07584735751152039,
0.032137636095285416,
-0.0065412926487624645,
-0.06669600307941437,
0.056103579699993134,
-0.04350929707288742,
0.060312408953905106,
-0.04873322695493698,
-0.009304715320467949,
0.1264428198337555,
-0.13684402406215668,
0.09757380187511444,
0.028239812701940536,
-0.09005201607942581,
-0.09391754120588303,
-0.051393840461969376,
-0.08823449164628983,
-0.032682858407497406,
-0.02395346574485302,
-0.13582473993301392,
0.024268580600619316,
0.0014446661807596684,
0.024052713066339493,
0.01834745705127716,
0.1203489601612091,
-0.12675078213214874,
-0.09651730954647064,
0.08238539099693298,
-0.016034871339797974,
0.04042103514075279,
0.08359920233488083,
0.03170151636004448,
0.003433442208915949,
0.01744021661579609,
0.05931423231959343,
0.08596202731132507,
0.09072752296924591,
0.04638217017054558,
-0.08009887486696243,
-0.0711556002497673,
-0.016012145206332207,
-0.0021180463954806328,
0.015561114996671677,
0.08689839392900467,
0.06216488406062126,
-0.05871458724141121,
-0.01707036979496479,
0.1932884156703949,
-0.08932337164878845,
-0.14171476662158966,
-0.18240240216255188,
0.10464713722467422,
0.051114946603775024,
-0.007999400608241558,
0.0076899356208741665,
-0.10459613054990768,
-0.01279715821146965,
0.17296569049358368,
0.18679676949977875,
-0.10525629669427872,
0.017027705907821655,
0.023007098585367203,
-0.000774757529143244,
0.010229947976768017,
0.07267575711011887,
0.06829831749200821,
0.20767943561077118,
-0.049210838973522186,
0.051545649766922,
-0.03057042695581913,
-0.03250160813331604,
-0.11065836250782013,
0.10774233937263489,
0.015552501194179058,
0.02343878708779812,
-0.029522132128477097,
0.10770190507173538,
-0.04870707914233208,
-0.09068647027015686,
0.004579952452331781,
-0.10683882981538773,
-0.12397699803113937,
-0.04101479798555374,
0.007905181497335434,
0.006747276056557894,
0.0685613825917244,
0.014444425702095032,
-0.014722014777362347,
0.13192950189113617,
-0.041195098310709,
-0.1124616414308548,
-0.044580571353435516,
0.06897701323032379,
0.008335938677191734,
0.15386490523815155,
0.008130541071295738,
-0.007797651924192905,
0.11519928276538849,
-0.047636695206165314,
-0.08260306715965271,
0.11205373704433441,
0.07751331478357315,
-0.08254493027925491,
0.09728313982486725,
0.05837195739150047,
-0.02986966446042061,
0.14685216546058655,
0.06397312134504318,
-0.11797399073839188,
0.01375780999660492,
0.005666985176503658,
-0.027400894090533257,
-0.09635713696479797,
0.01949954964220524,
-0.05463704839348793,
0.11808423697948456,
0.16369009017944336,
-0.03396870940923691,
0.016485171392560005,
-0.020609434694051743,
-0.03324286639690399,
0.029313858598470688,
0.006885528564453125,
-0.027671359479427338,
-0.07335327565670013,
-0.0025397555436939,
-0.0436885692179203,
0.06171588599681854,
-0.2498425841331482,
-0.06967368721961975,
-0.010928110219538212,
-0.03327591344714165,
-0.05695710331201553,
0.13903409242630005,
0.03439677879214287,
0.004882477689534426,
-0.020335517823696136,
-0.10379881411790848,
0.007587481290102005,
0.08529263734817505,
-0.09779208898544312,
-0.10450685024261475
] |
null | null | diffusers | # fcdsf
<Gallery />
## Model description
fgsdg

## Trigger words
You should use `dsfds` to trigger the image generation.
## Download model
[Download](/zz001/llll/tree/main) them in the Files & versions tab.
| {"license": "bigscience-openrail-m", "tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "wewe", "parameters": {"negative_prompt": "eretr"}, "output": {"url": "images/test.png"}}], "base_model": "InstantX/InstantID", "instance_prompt": "dsfds"} | text-to-image | zz001/llll | [
"diffusers",
"text-to-image",
"stable-diffusion",
"lora",
"template:sd-lora",
"base_model:InstantX/InstantID",
"license:bigscience-openrail-m",
"region:us"
] | 2024-02-08T17:42:44+00:00 | [] | [] | TAGS
#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-InstantX/InstantID #license-bigscience-openrail-m #region-us
| # fcdsf
<Gallery />
## Model description
fgsdg
!URL
## Trigger words
You should use 'dsfds' to trigger the image generation.
## Download model
Download them in the Files & versions tab.
| [
"# fcdsf\n\n<Gallery />",
"## Model description \n\nfgsdg\n!URL",
"## Trigger words\n\nYou should use 'dsfds' to trigger the image generation.",
"## Download model\n\n\nDownload them in the Files & versions tab."
] | [
"TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-InstantX/InstantID #license-bigscience-openrail-m #region-us \n",
"# fcdsf\n\n<Gallery />",
"## Model description \n\nfgsdg\n!URL",
"## Trigger words\n\nYou should use 'dsfds' to trigger the image generation.",
"## Download model\n\n\nDownload them in the Files & versions tab."
] | [
58,
10,
9,
18,
14
] | [
"passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-InstantX/InstantID #license-bigscience-openrail-m #region-us \n# fcdsf\n\n<Gallery />## Model description \n\nfgsdg\n!URL## Trigger words\n\nYou should use 'dsfds' to trigger the image generation.## Download model\n\n\nDownload them in the Files & versions tab."
] | [
-0.11219175159931183,
-0.013810201548039913,
-0.0040895589627325535,
0.048869337886571884,
0.10376913845539093,
0.03246120363473892,
0.16407015919685364,
0.043192919343709946,
0.1509876847267151,
0.04881671816110611,
0.1316148340702057,
0.061275456100702286,
0.020650293678045273,
0.23678487539291382,
-0.07125825434923172,
-0.22438770532608032,
0.05853695049881935,
-0.002553991274908185,
-0.11910584568977356,
-0.010545254684984684,
0.09266766160726547,
-0.06236831843852997,
0.14134915173053741,
-0.06473095715045929,
-0.15080755949020386,
-0.02371499128639698,
-0.0016543398378416896,
-0.040553413331508636,
0.032030194997787476,
0.08595074713230133,
-0.07511837780475616,
0.1366802304983139,
0.08817245811223984,
-0.04790607839822769,
0.0597541481256485,
-0.013149299658834934,
-0.0381072536110878,
0.058956079185009,
-0.0004028374678455293,
0.09113947302103043,
0.17531804740428925,
-0.0076414053328335285,
-0.038660913705825806,
-0.011089840903878212,
-0.05645294114947319,
-0.2324870079755783,
0.014394134283065796,
0.05981064587831497,
0.05341736972332001,
-0.02939451113343239,
0.0492996908724308,
-0.10202667862176895,
-0.004747712519019842,
0.0710642859339714,
0.09997521340847015,
-0.0877876803278923,
-0.03488006442785263,
0.2634302079677582,
0.04568307101726532,
0.12724071741104126,
-0.052496813237667084,
0.09656970202922821,
0.042925164103507996,
-0.03354619815945625,
0.013845075853168964,
-0.030534911900758743,
0.1300986111164093,
-0.04169337451457977,
-0.06537332385778427,
0.06948433071374893,
0.33351027965545654,
0.03422459214925766,
-0.039733804762363434,
-0.14576128125190735,
-0.08840502798557281,
0.1470816731452942,
-0.14365209639072418,
-0.07532811909914017,
0.06214994564652443,
0.03306064009666443,
0.06457280367612839,
-0.13622891902923584,
-0.06695988029241562,
-0.05277814716100693,
-0.06155415251851082,
0.1310448944568634,
-0.05331483855843544,
0.08975249528884888,
-0.05831212177872658,
0.03387998044490814,
-0.1412312388420105,
-0.16484078764915466,
-0.052136730402708054,
-0.16942501068115234,
0.10636545717716217,
0.06990735232830048,
-0.010052756406366825,
-0.03170796483755112,
0.13316521048545837,
0.08279146254062653,
0.07098066061735153,
-0.024104345589876175,
-0.010095756500959396,
0.10550206899642944,
0.031523697078228,
-0.019441308453679085,
-0.10488186776638031,
-0.008254697546362877,
0.08310751616954803,
0.034309178590774536,
0.03418195992708206,
-0.08048063516616821,
-0.1570851057767868,
-0.02342287078499794,
-0.16120624542236328,
-0.050619855523109436,
0.028163475915789604,
0.020873481407761574,
0.018611406907439232,
-0.05903162062168121,
0.21420620381832123,
0.01683899573981762,
-0.04257259517908096,
0.01638374850153923,
-0.05506424233317375,
0.1859571784734726,
0.07697789371013641,
0.004235551692545414,
0.16247068345546722,
0.10019471496343613,
-0.0704081580042839,
-0.04246688261628151,
-0.04903049021959305,
-0.013959732837975025,
-0.003952267579734325,
-0.042824238538742065,
0.0719626396894455,
-0.13525666296482086,
-0.23322468996047974,
0.01368328183889389,
0.05752553790807724,
-0.0655725747346878,
0.02540351264178753,
0.02052656002342701,
-0.08574438095092773,
0.05985395982861519,
0.01308355014771223,
-0.019008155912160873,
-0.05327865481376648,
0.09235725551843643,
-0.031397078186273575,
0.170715793967247,
-0.2774212062358856,
0.026116637513041496,
-0.048929326236248016,
0.030942179262638092,
-0.2298758625984192,
0.10049674659967422,
-0.12770672142505646,
0.06569044291973114,
0.01559280976653099,
-0.023966040462255478,
-0.0955338180065155,
0.05595054849982262,
0.009866503067314625,
0.24433310329914093,
-0.18802006542682648,
-0.04113287106156349,
0.0773729532957077,
-0.14504402875900269,
-0.09824329614639282,
0.0497555285692215,
-0.00979546271264553,
0.12428505718708038,
0.03354766592383385,
0.19665753841400146,
0.0584741048514843,
-0.17278464138507843,
0.09121984243392944,
0.16098327934741974,
-0.053476013243198395,
-0.034258633852005005,
0.0792485773563385,
0.0811735987663269,
0.0879461020231247,
0.09697438031435013,
-0.1219882071018219,
0.09181277453899384,
-0.05522884428501129,
-0.006401911377906799,
-0.009733668528497219,
-0.05056779459118843,
0.02487579546868801,
0.03446033596992493,
0.046208009123802185,
0.006879622116684914,
0.04632620885968208,
-0.006325963418930769,
0.059191253036260605,
0.0007151550380513072,
-0.056313656270504,
-0.03420534357428551,
0.12892456352710724,
-0.06775206327438354,
0.057508379220962524,
-0.03689740598201752,
-0.12289206683635712,
0.004635229241102934,
0.15997126698493958,
0.019249869510531425,
0.09395930171012878,
0.08650307357311249,
0.028439030051231384,
-0.024189630523324013,
0.055790454149246216,
0.032761912792921066,
-0.0499802827835083,
-0.004193323198705912,
-0.13732948899269104,
0.03515937179327011,
-0.08078278601169586,
0.07176246494054794,
-0.26197198033332825,
0.03131120279431343,
0.13257138431072235,
0.01412581279873848,
0.07504409551620483,
-0.02179671823978424,
-0.017430761829018593,
-0.08581749349832535,
-0.03727801516652107,
-0.0061340234242379665,
0.0638352558016777,
0.0007853748393245041,
-0.0562722347676754,
0.10639780014753342,
-0.01633928343653679,
0.2693837285041809,
0.1357211470603943,
0.0026955639477819204,
-0.024326208978891373,
-0.12033332139253616,
-0.008954491466283798,
0.0565362386405468,
0.049294471740722656,
-0.09390616416931152,
-0.10622332990169525,
-0.026708893477916718,
0.04392861947417259,
-0.03278012573719025,
0.05054810643196106,
0.06179462745785713,
-0.010023059323430061,
-0.05645092576742172,
0.05341720208525658,
0.1872262805700302,
-0.037164222449064255,
0.14241091907024384,
0.15057986974716187,
0.013790837489068508,
0.15760241448879242,
-0.059818871319293976,
-0.06577138602733612,
0.031485073268413544,
0.02302052453160286,
-0.03265108913183212,
0.22633326053619385,
-0.013664311729371548,
-0.008990069851279259,
0.08791720122098923,
-0.07813794165849686,
-0.008258369751274586,
-0.059485308825969696,
-0.09573130309581757,
0.03200095519423485,
-0.03855157643556595,
0.06716844439506531,
0.10576889663934708,
-0.0890154168009758,
0.11528690904378891,
-0.03974200040102005,
-0.10110599547624588,
0.03731268271803856,
-0.017321396619081497,
-0.056000351905822754,
0.10715049505233765,
-0.019439872354269028,
-0.13795672357082367,
-0.14142978191375732,
-0.019474904984235764,
-0.10604534298181534,
0.04631289094686508,
0.013530278578400612,
-0.04657885059714317,
-0.05083363875746727,
-0.019899612292647362,
-0.005757458042353392,
0.02569449320435524,
-0.04736833646893501,
0.02736371010541916,
-0.022224172949790955,
-0.07940039783716202,
-0.11520952731370926,
-0.03759603202342987,
-0.05235539376735687,
0.07990433275699615,
0.10263921320438385,
0.0018644973170012236,
0.12138314545154572,
0.04443350434303284,
0.06407169997692108,
0.015056408010423183,
0.03874557837843895,
0.24944579601287842,
-0.10317590087652206,
0.08355394750833511,
0.25707361102104187,
0.02462194859981537,
0.022527256980538368,
0.16030101478099823,
0.07588022947311401,
-0.1477510631084442,
0.016440169885754585,
-0.07373692095279694,
-0.1429019272327423,
-0.02169763296842575,
-0.08340350538492203,
-0.10225103050470352,
0.04675263911485672,
-0.05684226006269455,
0.04642026126384735,
0.09486677497625351,
0.16691266000270844,
0.007362920790910721,
0.018893593922257423,
0.06034240126609802,
0.015629302710294724,
0.10547707974910736,
-0.02095123939216137,
0.09241343289613724,
-0.04479977488517761,
-0.04098637029528618,
0.148548424243927,
-0.06886085867881775,
0.11248528212308884,
0.019612746313214302,
0.004736880771815777,
0.019279813393950462,
-0.04842907562851906,
0.09021831303834915,
0.0899139866232872,
-0.054702453315258026,
-0.07286283373832703,
-0.06138230860233307,
-0.13107575476169586,
0.03572864085435867,
0.11944364756345749,
-0.07554282993078232,
-0.09509609639644623,
0.0115659199655056,
0.014066268689930439,
-0.016917187720537186,
0.020213723182678223,
0.017972595989704132,
-0.2817154824733734,
-0.008939570747315884,
0.059519074857234955,
0.1487097144126892,
-0.0672890692949295,
0.02469281665980816,
0.1527627408504486,
-0.047211095690727234,
0.02640911191701889,
0.02869994193315506,
0.05518074333667755,
0.019246593117713928,
-0.006078486330807209,
0.020648876205086708,
0.025052694603800774,
0.02667776681482792,
-0.026713846251368523,
-0.2165721207857132,
0.03691646084189415,
0.006544168572872877,
-0.062148842960596085,
-0.02357036992907524,
0.019405361264944077,
0.0806269720196724,
0.17165759205818176,
0.1683603972196579,
0.005003298167139292,
0.013059496879577637,
-0.10600131005048752,
-0.09851148724555969,
0.007658340502530336,
0.09731565415859222,
-0.11702035367488861,
-0.04705654829740524,
0.050823427736759186,
0.0011675728019326925,
0.01085561141371727,
0.18384379148483276,
-0.08118124306201935,
-0.11370403319597244,
0.04144470766186714,
0.043277889490127563,
0.05017812177538872,
-0.021306157112121582,
0.001745707355439663,
-0.14302831888198853,
0.02396337129175663,
0.015914658084511757,
-0.1286846548318863,
-0.15109942853450775,
-0.023762330412864685,
0.002370159374549985,
-0.007363641634583473,
0.0774226039648056,
-0.021584106609225273,
0.05831499397754669,
-0.0335310660302639,
-0.2156745195388794,
0.09843920916318893,
-0.10847026109695435,
-0.06073674187064171,
-0.07991503924131393,
0.08793178200721741,
-0.019026126712560654,
-0.009837961755692959,
0.007960361428558826,
0.024392442777752876,
0.030740391463041306,
-0.11472055315971375,
0.06955544650554657,
0.09827612340450287,
-0.06048460677266121,
0.0171064306050539,
-0.08995542675256729,
0.03384923189878464,
0.08310503512620926,
0.032567739486694336,
0.05512700974941254,
0.30511000752449036,
-0.0910668596625328,
0.09043654054403305,
0.08824316412210464,
0.008719969540834427,
-0.3179195523262024,
0.04593261331319809,
-0.005577330477535725,
-0.009871047921478748,
0.013120496645569801,
-0.18858110904693604,
0.18493720889091492,
0.12318597733974457,
-0.06701914221048355,
0.34699222445487976,
-0.2359028160572052,
-0.10487741231918335,
0.017802363261580467,
0.07398757338523865,
0.23017892241477966,
-0.25155872106552124,
-0.0535244457423687,
-0.06710357218980789,
-0.20531602203845978,
0.16691383719444275,
-0.022370455786585808,
-0.007573007605969906,
-0.013810546137392521,
-0.0351940393447876,
0.03109908290207386,
-0.018648840487003326,
0.19547396898269653,
-0.04680759087204933,
0.1190771833062172,
-0.09848807007074356,
0.026474300771951675,
0.12320057302713394,
0.04172128811478615,
0.10926251113414764,
-0.029597699642181396,
0.0073945652693510056,
-0.14009717106819153,
-0.009635020978748798,
-0.04277098923921585,
0.05432911217212677,
-0.017316428944468498,
-0.1008039265871048,
-0.078281931579113,
0.006393310148268938,
-0.04860745370388031,
0.012310666963458061,
0.02189020998775959,
-0.06484829634428024,
0.021181384101510048,
0.16735254228115082,
-0.0009157651220448315,
-0.03867333382368088,
-0.03779057785868645,
-0.058547697961330414,
-0.044537849724292755,
0.10343633592128754,
-0.16207964718341827,
-0.030141755938529968,
0.10266754031181335,
0.05879128724336624,
0.05906257778406143,
0.0601731576025486,
0.05826098844408989,
0.06321591883897781,
0.12061617523431778,
-0.11298365890979767,
-0.17089249193668365,
-0.09735694527626038,
-0.05656903609633446,
0.08690164238214493,
0.009030969813466072,
0.10243993252515793,
-0.03205612301826477,
0.05278971046209335,
0.001980400877073407,
0.02235925756394863,
-0.03558384254574776,
0.03269890695810318,
0.119635671377182,
-0.009299438446760178,
-0.11974183470010757,
-0.013489007018506527,
-0.02990279532968998,
-0.02266782335937023,
-0.07884310930967331,
0.048971597105264664,
-0.08003703504800797,
-0.05123087763786316,
-0.06243326514959335,
0.05966835096478462,
-0.09065461903810501,
0.031785789877176285,
0.013113383203744888,
-0.1290709674358368,
-0.018025344237685204,
0.05505525320768356,
0.06145111843943596,
-0.06486516445875168,
0.03686016425490379,
0.07480965554714203,
0.012147479690611362,
0.053877245634794235,
-0.08178076893091202,
0.12401414662599564,
-0.183451846241951,
-0.12234146893024445,
-0.11277974396944046,
0.027017733082175255,
-0.13660933077335358,
0.004719817545264959,
-0.11582548916339874,
-0.02206086739897728,
-0.03956965357065201,
0.04039321094751358,
-0.058015376329422,
-0.0391322523355484,
-0.04173239320516586,
-0.040976203978061676,
0.009455294348299503,
-0.03153475001454353,
-0.08323682099580765,
-0.0070708682760596275,
0.011022202670574188,
0.020959826186299324,
-0.06503982096910477,
-0.07084520906209946,
-0.005234777927398682,
-0.07723215222358704,
0.11965179443359375,
0.07946843653917313,
-0.03551698103547096,
-0.057041049003601074,
-0.19363319873809814,
-0.005714720580726862,
0.1232849657535553,
0.08421323448419571,
-0.03654301166534424,
0.1432362198829651,
0.046927545219659805,
-0.03929935768246651,
-0.043282121419906616,
-0.0010348039213567972,
-0.015157323330640793,
-0.09511929005384445,
0.05528079345822334,
-0.10288834571838379,
0.019762134179472923,
-0.02243378385901451,
0.006126471795141697,
0.1582428216934204,
0.1296808421611786,
0.053694598376750946,
-0.048295311629772186,
0.010580637492239475,
-0.08673184365034103,
0.0007811617688275874,
0.04863766208291054,
-0.0991913452744484,
0.01638670079410076,
-0.08858619630336761,
0.01846483163535595,
-0.016855264082551003,
0.207858607172966,
0.033309582620859146,
-0.16021692752838135,
-0.008592711761593819,
-0.013226631097495556,
0.22431477904319763,
0.036652032285928726,
0.24237388372421265,
0.12866730988025665,
0.0686381459236145,
-0.10398421436548233,
0.140180304646492,
0.13774316012859344,
-0.03757069259881973,
0.007642779499292374,
0.10782214999198914,
-0.11774346977472305,
0.03601870685815811,
0.09401631355285645,
-0.01838318258523941,
0.057839419692754745,
-0.006582579575479031,
-0.017142686992883682,
0.0442298948764801,
-0.09323916584253311,
0.028515534475445747,
0.12867438793182373,
-0.07161515206098557,
0.04593445360660553,
0.046156276017427444,
-0.0068208943121135235,
-0.09063395857810974,
-0.11854435503482819,
-0.053804799914360046,
-0.1920260339975357,
0.03781392425298691,
-0.07239298522472382,
-0.02205936424434185,
0.047304827719926834,
0.04767894372344017,
-0.05657046288251877,
0.08059068024158478,
-0.07613296806812286,
-0.04946263134479523,
0.15306439995765686,
-0.046099547296762466,
-0.13869774341583252,
-0.09344114363193512,
-0.06137197092175484,
0.04669057950377464,
0.017653213813900948,
-0.023964177817106247,
0.039411306381225586,
-0.009627653285861015,
-0.0370730385184288,
-0.03443846479058266,
-0.04384420067071915,
-0.035976968705654144,
0.003468782640993595,
-0.046609967947006226,
0.05849044770002365,
0.0014876116765663028,
-0.009488780051469803,
0.003324696561321616,
0.10489656776189804,
-0.024774888530373573,
0.002552237594500184,
-0.04545534402132034,
-0.009088652208447456,
-0.11855422705411911,
0.149689719080925,
-0.09260216355323792,
-0.0784507542848587,
-0.06768792867660522,
0.12170428782701492,
0.20180152356624603,
-0.23830245435237885,
0.03328174725174904,
0.04190843179821968,
0.008921346627175808,
-0.037082310765981674,
0.12127195298671722,
0.028872204944491386,
0.32282185554504395,
-0.030027011409401894,
0.013061060570180416,
-0.04826844111084938,
-0.023271294310688972,
-0.08078894019126892,
0.009066312573850155,
0.08605712652206421,
-0.07036899775266647,
-0.07563349604606628,
0.07834035158157349,
-0.1981177031993866,
0.003424219088628888,
0.03092656284570694,
-0.0939374566078186,
-0.010268579237163067,
-0.09488548338413239,
-0.16360996663570404,
0.04564669355750084,
-0.03465176746249199,
-0.1130329817533493,
0.006905883084982634,
-0.06997187435626984,
-0.0258072130382061,
-0.10758759081363678,
0.039433885365724564,
0.023932939395308495,
-0.10218460857868195,
0.09686598926782608,
-0.05127086490392685,
0.047813404351472855,
-0.004328431095927954,
0.002059931168332696,
-0.04614792764186859,
0.004373989067971706,
0.013598755933344364,
-0.10561958700418472,
-0.0663791298866272,
0.06720733642578125,
-0.007893230766057968,
0.0011025724234059453,
0.0450160838663578,
-0.1270085722208023,
-0.0034479813184589148,
0.15740759670734406,
-0.09750876575708389,
-0.06278156489133835,
0.04047424718737602,
-0.14476454257965088,
0.049515292048454285,
0.023030726239085197,
0.022371532395482063,
-0.05272124707698822,
-0.03472520038485527,
0.07847268134355545,
0.10480236262083054,
-0.03523015230894089,
0.004485681187361479,
-0.050574418157339096,
-0.05512990057468414,
0.02541191130876541,
0.029620986431837082,
-0.1486896276473999,
-0.03153663128614426,
-0.14871947467327118,
0.08960813283920288,
0.010719960555434227,
0.012222325429320335,
0.21391133964061737,
-0.032717157155275345,
-0.027295086532831192,
-0.16602931916713715,
0.013713913038372993,
0.06172528117895126,
-0.07534159719944,
-0.032131899148225784
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-5.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:43:44+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
194,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06399306654930115,
-0.029307762160897255,
-0.004331388510763645,
0.03184039518237114,
0.018488328903913498,
0.03748767077922821,
0.10618676990270615,
0.07814439386129379,
0.07063937187194824,
0.07519446313381195,
-0.04871003329753876,
-0.02137892320752144,
0.07951480895280838,
0.10956592857837677,
0.007890605367720127,
-0.23595072329044342,
0.08250588178634644,
-0.03575211763381958,
-0.026412149891257286,
0.03893791511654854,
0.14763003587722778,
-0.05930296331644058,
0.05646873265504837,
-0.0034751244820654392,
-0.05727600306272507,
0.05131329223513603,
-0.026073675602674484,
-0.007598201744258404,
0.09508240222930908,
0.0628398060798645,
0.041950274258852005,
0.0221012644469738,
-0.018701795488595963,
-0.15748009085655212,
0.004922133404761553,
-0.017628159373998642,
-0.04106835648417473,
0.027759606018662453,
0.11616206914186478,
-0.045659661293029785,
0.11880303174257278,
-0.12680159509181976,
-0.018632367253303528,
0.08243369311094284,
-0.12379294633865356,
-0.10155918449163437,
-0.20841088891029358,
0.15503807365894318,
0.07979700714349747,
0.008151463232934475,
-0.04288718104362488,
0.07205401360988617,
-0.054492317140102386,
0.0433170422911644,
0.08753809332847595,
-0.2521143853664398,
-0.04418989270925522,
0.17082107067108154,
0.007189796306192875,
-0.04941372200846672,
-0.05787082388997078,
0.06996709108352661,
0.029280396178364754,
0.0067648859694600105,
-0.04790947958827019,
-0.04312778264284134,
0.2201330065727234,
-0.03288035839796066,
-0.1199050173163414,
-0.024701880291104317,
0.08322399854660034,
0.08842694014310837,
-0.04718442261219025,
-0.15207047760486603,
-0.011922257021069527,
-0.022255953401327133,
-0.04938272759318352,
-0.09886841475963593,
0.003574386006221175,
-0.02095438726246357,
0.07563887536525726,
-0.05726063251495361,
-0.04117260500788689,
0.0583573654294014,
-0.05984218046069145,
0.1629648655653,
-0.0014757480239495635,
-0.03302643075585365,
0.05737544223666191,
0.03662708401679993,
-0.2148275226354599,
-0.08567781001329422,
-0.04239440709352493,
-0.054734885692596436,
-0.1342768371105194,
-0.0431794635951519,
-0.04418531805276871,
-0.08217091858386993,
-0.007381834089756012,
0.23022499680519104,
-0.02492726407945156,
0.02804037742316723,
0.041754867881536484,
0.010314702056348324,
0.04419277235865593,
0.1185997873544693,
-0.05733427032828331,
-0.1965552270412445,
-0.02631615847349167,
-0.01892445981502533,
0.06395958364009857,
0.013402312994003296,
-0.02684073895215988,
-0.01519184559583664,
0.02383972890675068,
0.02890627458691597,
0.02422332763671875,
0.03168574720621109,
-0.07110860198736191,
-0.03037358447909355,
0.14692901074886322,
-0.10123810917139053,
0.04444355517625809,
-0.03552069887518883,
-0.03803017735481262,
0.07610528916120529,
-0.030228903517127037,
0.023446066305041313,
-0.01123465783894062,
0.14718079566955566,
-0.04342922940850258,
-0.02282293513417244,
-0.04051238298416138,
-0.05655231699347496,
0.056725140661001205,
-0.08175330609083176,
-0.06928885728120804,
-0.0921410471200943,
-0.07629581540822983,
-0.0781061053276062,
0.019695017486810684,
-0.04772234335541725,
-0.00727304769679904,
-0.00963415578007698,
-0.03471999987959862,
0.004572734236717224,
0.03700930252671242,
-0.03850525617599487,
-0.006236410234123468,
-0.05285993963479996,
-0.02465738169848919,
0.016344578936696053,
-0.07584261149168015,
0.020161736756563187,
-0.09869375079870224,
0.055850550532341,
-0.3218844532966614,
0.09342993050813675,
-0.11400728672742844,
0.009930809028446674,
-0.1314808875322342,
-0.0335058830678463,
0.016112782061100006,
0.005322059616446495,
0.013146123848855495,
0.11860170215368271,
-0.15911765396595,
-0.09210728108882904,
0.09710568189620972,
-0.13270564377307892,
-0.06123318895697594,
0.13335616886615753,
0.006097014062106609,
-0.007668835576623678,
0.0786266103386879,
0.1425439566373825,
0.14123733341693878,
-0.04838160052895546,
-0.13677071034908295,
-0.06624309718608856,
0.027206536382436752,
0.1544196456670761,
0.06648306548595428,
-0.0522027313709259,
-0.0182489063590765,
0.03598593920469284,
-0.031956933438777924,
-0.0026491705793887377,
-0.016834797337651253,
-0.05705594643950462,
-0.023978661745786667,
-0.002377238357439637,
0.096376933157444,
-0.015163181349635124,
0.002383065177127719,
-0.07838787883520126,
-0.09363235533237457,
0.039118777960538864,
0.0882512778043747,
0.041436195373535156,
0.01435884553939104,
-0.06161881238222122,
0.09994251281023026,
-0.03056234121322632,
0.022030595690011978,
-0.10238077491521835,
-0.15781597793102264,
0.03258318454027176,
-0.15215061604976654,
0.06616183370351791,
0.046265859156847,
0.06192070245742798,
0.027611956000328064,
-0.11285637319087982,
-0.06351099908351898,
-0.02237211912870407,
0.016440147534012794,
0.007992693223059177,
-0.12434909492731094,
-0.08924297988414764,
-0.04348140209913254,
0.22069425880908966,
-0.054984692484140396,
-0.007727441843599081,
0.06026673689484596,
0.2515117824077606,
0.018960142508149147,
-0.014410633593797684,
-0.05082327499985695,
0.07166947424411774,
0.00031130402931012213,
0.0197899267077446,
0.01994451880455017,
-0.04336968809366226,
-0.14637522399425507,
0.095527283847332,
-0.055198993533849716,
-0.02979397401213646,
0.06157594546675682,
0.054885584861040115,
-0.09651610255241394,
-0.011951263062655926,
-0.05258362367749214,
-0.029914764687418938,
0.006216008216142654,
-0.12069366872310638,
0.042220812290906906,
0.04080578312277794,
0.03252416476607323,
-0.0428742878139019,
-0.048637744039297104,
0.01096632331609726,
0.00038561708061024547,
-0.12152905762195587,
0.10360410064458847,
-0.08400224894285202,
-0.21976810693740845,
0.10374713689088821,
0.15290415287017822,
0.07434318959712982,
0.12372945249080658,
-0.013827105984091759,
-0.022371480241417885,
-0.1197407990694046,
0.0015785678988322616,
0.021130502223968506,
0.042604196816682816,
0.04024510458111763,
0.05408281460404396,
0.03447001054883003,
0.043456509709358215,
0.02863817848265171,
-0.02880154922604561,
0.04750415310263634,
-0.009588376618921757,
0.0040976181626319885,
0.12772449851036072,
0.10417608171701431,
0.02852170541882515,
0.08533679693937302,
0.03910043090581894,
-0.0404503159224987,
0.01288803294301033,
-0.02749737910926342,
-0.06698718667030334,
0.10686532407999039,
-0.10741190612316132,
-0.16609297692775726,
-0.17008981108665466,
-0.051623594015836716,
-0.18006007373332977,
-0.024816229939460754,
0.03015192784368992,
-0.03652867674827576,
-0.07785869389772415,
-0.07390360534191132,
0.13494938611984253,
0.03926362842321396,
-0.05240222439169884,
-0.05028913915157318,
-0.03344966843724251,
0.042253971099853516,
-0.10105009377002716,
-0.04346547648310661,
0.0018156386213377118,
-0.050230301916599274,
0.021803520619869232,
0.027263345196843147,
0.0939897820353508,
0.060186538845300674,
-0.00003084815762122162,
-0.02964363619685173,
-0.040875863283872604,
0.17850545048713684,
-0.1293039619922638,
0.11540871858596802,
0.04773532599210739,
-0.0027224498335272074,
0.06698734313249588,
0.16732339560985565,
0.05765581503510475,
-0.04658689722418785,
-0.011350571177899837,
0.06666777282953262,
-0.030819516628980637,
-0.19200317561626434,
-0.1769828051328659,
-0.06770068407058716,
0.007487337104976177,
-0.027668775990605354,
0.04515873268246651,
0.06839024275541306,
-0.0003817266842816025,
-0.09345950931310654,
-0.07025164365768433,
0.036232683807611465,
0.04945642128586769,
0.17080222070217133,
0.009348781779408455,
0.05902401730418205,
-0.06639035046100616,
-0.06039080023765564,
0.09840981662273407,
0.00460409838706255,
0.08121812343597412,
0.06772522628307343,
0.06868960708379745,
0.051509808748960495,
0.012179503217339516,
0.0003098213055636734,
0.04621569812297821,
0.015963001176714897,
-0.02683101035654545,
-0.011713831685483456,
-0.08623592555522919,
0.07211727648973465,
0.08716612309217453,
0.10278671234846115,
0.03722165524959564,
-0.051564089953899384,
-0.06990828365087509,
0.12818388640880585,
0.1740635186433792,
0.11096327751874924,
-0.12066809087991714,
-0.10149893164634705,
0.059367019683122635,
-0.012059872969985008,
0.020104901865124702,
-0.03295712172985077,
-0.051309432834386826,
-0.13940629363059998,
0.20302264392375946,
-0.018019717186689377,
0.06347235292196274,
0.0055109113454818726,
-0.0001949289726326242,
-0.02158956229686737,
0.06696530431509018,
0.017838791012763977,
0.050336673855781555,
-0.141648069024086,
0.14686071872711182,
0.010985979810357094,
0.021909072995185852,
-0.01216028444468975,
0.05184603109955788,
0.01702127605676651,
0.05610823258757591,
0.10946010798215866,
0.04561372473835945,
-0.09489621967077255,
-0.08675345033407211,
-0.04737729951739311,
-0.05841971933841705,
0.13532322645187378,
-0.0978664755821228,
0.08815576881170273,
-0.005218944977968931,
-0.04548734799027443,
-0.058141354471445084,
0.16146430373191833,
-0.15934163331985474,
-0.10354094952344894,
0.10525287687778473,
-0.04850301891565323,
0.028323976323008537,
-0.07308647781610489,
-0.04719609022140503,
-0.24660754203796387,
0.2158661186695099,
0.015384898521006107,
0.00384410098195076,
-0.09645523875951767,
-0.0031059239991009235,
0.16681231558322906,
-0.09507244825363159,
0.038081102073192596,
-0.0030803470872342587,
0.08820296823978424,
-0.09653045237064362,
-0.08165580779314041,
0.04465007409453392,
-0.07744460552930832,
-0.14185139536857605,
0.011764489114284515,
0.19781433045864105,
0.022616581991314888,
0.037971191108226776,
0.025954479351639748,
0.09160146117210388,
0.08607495576143265,
-0.10742948949337006,
0.08422011882066727,
0.0879930630326271,
-0.030707523226737976,
0.1495898962020874,
-0.033895764499902725,
-0.15872009098529816,
-0.05146879330277443,
-0.0411202572286129,
0.10311200469732285,
0.34257373213768005,
-0.06467219442129135,
0.08678370714187622,
0.08080018311738968,
-0.042148686945438385,
-0.19632753729820251,
0.007789987605065107,
0.01917961612343788,
0.02523370087146759,
0.00445806048810482,
-0.05062577873468399,
0.10844220966100693,
0.13072700798511505,
-0.020657971501350403,
-0.04788164049386978,
-0.20526863634586334,
-0.1290767788887024,
0.011098041199147701,
-0.06513512134552002,
-0.012189746834337711,
-0.06006240472197533,
-0.10408233106136322,
-0.07066939771175385,
-0.16428698599338531,
0.05405878648161888,
0.04084285721182823,
0.048617810010910034,
0.0022039911709725857,
-0.022370649501681328,
0.03627573698759079,
-0.018924014642834663,
0.18879440426826477,
0.06920293718576431,
0.02023874782025814,
-0.07239631563425064,
-0.04903245344758034,
0.11341241002082825,
-0.037153460085392,
0.09947021305561066,
0.04320525750517845,
-0.005356981884688139,
-0.11361989378929138,
0.012292295694351196,
-0.09132620692253113,
0.11476991325616837,
-0.09256471693515778,
-0.030837757512927055,
-0.08548001945018768,
0.140853151679039,
0.0238761268556118,
-0.015673771500587463,
0.12679731845855713,
-0.032148636877536774,
0.1091376468539238,
0.13080862164497375,
0.0701589584350586,
0.0031420262530446053,
-0.07472140341997147,
0.0012759605888277292,
-0.06170511618256569,
0.019098622724413872,
0.05546579137444496,
0.04380559176206589,
0.13516387343406677,
-0.01975567452609539,
0.14712172746658325,
-0.018929120153188705,
-0.13689684867858887,
-0.023719700053334236,
0.11753503233194351,
-0.11223208159208298,
-0.2821040451526642,
-0.04018678143620491,
-0.034599367529153824,
0.012139244005084038,
0.061747223138809204,
0.19457398355007172,
-0.03780088946223259,
-0.023348631337285042,
0.021921996027231216,
0.04527831822633743,
-0.09146711230278015,
0.08037439733743668,
-0.021092766895890236,
0.042854998260736465,
-0.06821004301309586,
0.10744444280862808,
0.07745382934808731,
-0.22681817412376404,
-0.025234894827008247,
0.09869363903999329,
-0.06257352977991104,
-0.07561001926660538,
-0.09376278519630432,
0.1380273848772049,
-0.1314297765493393,
-0.0158812515437603,
-0.020428206771612167,
-0.1451670229434967,
0.04510067030787468,
0.09574063867330551,
0.04472881183028221,
0.011867307126522064,
0.035759247839450836,
0.00018209819972980767,
0.01861182041466236,
0.03738055005669594,
0.09328993409872055,
0.10640279203653336,
-0.042188696563243866,
0.03816690295934677,
-0.08412777632474899,
0.015412991866469383,
-0.021022463217377663,
-0.01296116691082716,
-0.09465239942073822,
-0.07739552855491638,
-0.10236255824565887,
-0.06780892610549927,
-0.08637113869190216,
-0.02516632154583931,
-0.0233630258589983,
-0.023995107039809227,
0.004357992671430111,
-0.027227098122239113,
-0.07138573378324509,
-0.050982195883989334,
-0.0155563373118639,
0.08955743908882141,
-0.1038399413228035,
-0.0027917600236833096,
0.06826390326023102,
-0.051174331456422806,
0.0997040867805481,
0.024214884266257286,
0.0514301136136055,
0.03840871900320053,
-0.03028056025505066,
-0.03847609460353851,
0.01905941218137741,
0.006461322773247957,
0.05086956173181534,
-0.14872215688228607,
-0.012150841765105724,
-0.0032515712082386017,
-0.051708951592445374,
0.050017524510622025,
-0.0625491663813591,
-0.10236836224794388,
0.008505094796419144,
0.016627702862024307,
-0.09351813793182373,
-0.07604184001684189,
0.027614127844572067,
0.0729801207780838,
0.035474032163619995,
0.11152558028697968,
-0.058112598955631256,
0.05574757605791092,
-0.1357552707195282,
-0.01836516335606575,
-0.02554010972380638,
-0.1030651405453682,
0.016653461381793022,
-0.011252963915467262,
0.022870756685733795,
-0.0031690942123532295,
0.11289511620998383,
-0.026996877044439316,
-0.025828121230006218,
0.06756260991096497,
-0.028583889827132225,
-0.0780753642320633,
0.0507938489317894,
0.07868042588233948,
0.04812269285321236,
-0.0161933284252882,
-0.08120898902416229,
-0.0014988429611548781,
-0.05973583459854126,
-0.007319761905819178,
0.1114073172211647,
0.14946618676185608,
0.11061113327741623,
0.05770944431424141,
0.08207276463508606,
-0.04891999810934067,
-0.028829647228121758,
-0.009984364733099937,
-0.029658254235982895,
0.08592110872268677,
-0.02771167829632759,
0.046906858682632446,
0.07054735720157623,
-0.23260782659053802,
0.1542171984910965,
-0.04036210849881172,
0.003158458275720477,
-0.019460761919617653,
-0.11383481323719025,
-0.07883796095848083,
-0.042559150606393814,
0.010454006493091583,
-0.08244454860687256,
0.009923603385686874,
0.00909425225108862,
0.03983308747410774,
-0.018626000732183456,
0.14684614539146423,
-0.22303219139575958,
-0.020058225840330124,
0.1173686534166336,
0.01384452823549509,
-0.01986311748623848,
0.04652818292379379,
-0.01833965815603733,
-0.05564456060528755,
0.0526103749871254,
0.003485421882942319,
0.042003098875284195,
0.01768621988594532,
-0.009287383407354355,
-0.030709892511367798,
-0.0902811586856842,
0.015258834697306156,
-0.013983458280563354,
0.019041143357753754,
0.026359856128692627,
0.02583683282136917,
0.016558820381760597,
-0.0304630808532238,
0.1203962042927742,
-0.04186068847775459,
-0.08964694291353226,
-0.09655430167913437,
0.18649248778820038,
0.015295648016035557,
0.07767751067876816,
0.014468489214777946,
-0.11231113970279694,
0.003929391037672758,
0.011538079008460045,
0.2675095796585083,
0.06229402497410774,
0.0249533299356699,
0.022313563153147697,
0.0121700968593359,
0.03887152299284935,
0.019862765446305275,
-0.021693918853998184,
0.17263448238372803,
-0.035285405814647675,
0.2010856717824936,
-0.05225873738527298,
-0.046340879052877426,
0.0050239525735378265,
0.1740422248840332,
-0.01995527185499668,
-0.01296838466078043,
-0.015919938683509827,
0.11686369776725769,
-0.02941286377608776,
-0.15430203080177307,
0.030166465789079666,
-0.1680508852005005,
-0.1555458903312683,
-0.04883193224668503,
0.03437260910868645,
0.0537441223859787,
0.11792732775211334,
0.015498639084398746,
-0.03355927765369415,
0.11172260344028473,
0.013434622436761856,
-0.01853151060640812,
-0.12593184411525726,
0.06253454834222794,
-0.005848423577845097,
0.10684798657894135,
0.020717212930321693,
0.04494085907936096,
0.12986114621162415,
-0.024106910452246666,
-0.10693443566560745,
0.014985187910497189,
0.045302849262952805,
-0.08386611938476562,
0.0049399384297430515,
0.15897472202777863,
-0.014517411589622498,
0.11524591594934464,
0.05899004638195038,
-0.13243594765663147,
0.07087180018424988,
0.0894564762711525,
-0.041566379368305206,
-0.06131380423903465,
0.1714923232793808,
-0.08582644164562225,
0.13566255569458008,
0.23618710041046143,
-0.026338893920183182,
0.025297190994024277,
-0.03987700119614601,
0.044342536479234695,
0.07475463300943375,
0.12010806798934937,
-0.017106499522924423,
-0.16066502034664154,
0.06056363880634308,
0.023605575785040855,
0.04031899571418762,
-0.14200104773044586,
-0.11799366027116776,
-0.015118377283215523,
0.02434436045587063,
-0.030886957421898842,
0.12695416808128357,
0.08972851932048798,
-0.0224088616669178,
0.001209444715641439,
-0.1888996660709381,
-0.008432535454630852,
0.07094074040651321,
-0.045595064759254456,
-0.016666574403643608
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | tavalenzuelag/mistral-7b-e2e-mod-5 | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:44:48+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-6.0bpw-h6-exl2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:45:54+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
194,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06399306654930115,
-0.029307762160897255,
-0.004331388510763645,
0.03184039518237114,
0.018488328903913498,
0.03748767077922821,
0.10618676990270615,
0.07814439386129379,
0.07063937187194824,
0.07519446313381195,
-0.04871003329753876,
-0.02137892320752144,
0.07951480895280838,
0.10956592857837677,
0.007890605367720127,
-0.23595072329044342,
0.08250588178634644,
-0.03575211763381958,
-0.026412149891257286,
0.03893791511654854,
0.14763003587722778,
-0.05930296331644058,
0.05646873265504837,
-0.0034751244820654392,
-0.05727600306272507,
0.05131329223513603,
-0.026073675602674484,
-0.007598201744258404,
0.09508240222930908,
0.0628398060798645,
0.041950274258852005,
0.0221012644469738,
-0.018701795488595963,
-0.15748009085655212,
0.004922133404761553,
-0.017628159373998642,
-0.04106835648417473,
0.027759606018662453,
0.11616206914186478,
-0.045659661293029785,
0.11880303174257278,
-0.12680159509181976,
-0.018632367253303528,
0.08243369311094284,
-0.12379294633865356,
-0.10155918449163437,
-0.20841088891029358,
0.15503807365894318,
0.07979700714349747,
0.008151463232934475,
-0.04288718104362488,
0.07205401360988617,
-0.054492317140102386,
0.0433170422911644,
0.08753809332847595,
-0.2521143853664398,
-0.04418989270925522,
0.17082107067108154,
0.007189796306192875,
-0.04941372200846672,
-0.05787082388997078,
0.06996709108352661,
0.029280396178364754,
0.0067648859694600105,
-0.04790947958827019,
-0.04312778264284134,
0.2201330065727234,
-0.03288035839796066,
-0.1199050173163414,
-0.024701880291104317,
0.08322399854660034,
0.08842694014310837,
-0.04718442261219025,
-0.15207047760486603,
-0.011922257021069527,
-0.022255953401327133,
-0.04938272759318352,
-0.09886841475963593,
0.003574386006221175,
-0.02095438726246357,
0.07563887536525726,
-0.05726063251495361,
-0.04117260500788689,
0.0583573654294014,
-0.05984218046069145,
0.1629648655653,
-0.0014757480239495635,
-0.03302643075585365,
0.05737544223666191,
0.03662708401679993,
-0.2148275226354599,
-0.08567781001329422,
-0.04239440709352493,
-0.054734885692596436,
-0.1342768371105194,
-0.0431794635951519,
-0.04418531805276871,
-0.08217091858386993,
-0.007381834089756012,
0.23022499680519104,
-0.02492726407945156,
0.02804037742316723,
0.041754867881536484,
0.010314702056348324,
0.04419277235865593,
0.1185997873544693,
-0.05733427032828331,
-0.1965552270412445,
-0.02631615847349167,
-0.01892445981502533,
0.06395958364009857,
0.013402312994003296,
-0.02684073895215988,
-0.01519184559583664,
0.02383972890675068,
0.02890627458691597,
0.02422332763671875,
0.03168574720621109,
-0.07110860198736191,
-0.03037358447909355,
0.14692901074886322,
-0.10123810917139053,
0.04444355517625809,
-0.03552069887518883,
-0.03803017735481262,
0.07610528916120529,
-0.030228903517127037,
0.023446066305041313,
-0.01123465783894062,
0.14718079566955566,
-0.04342922940850258,
-0.02282293513417244,
-0.04051238298416138,
-0.05655231699347496,
0.056725140661001205,
-0.08175330609083176,
-0.06928885728120804,
-0.0921410471200943,
-0.07629581540822983,
-0.0781061053276062,
0.019695017486810684,
-0.04772234335541725,
-0.00727304769679904,
-0.00963415578007698,
-0.03471999987959862,
0.004572734236717224,
0.03700930252671242,
-0.03850525617599487,
-0.006236410234123468,
-0.05285993963479996,
-0.02465738169848919,
0.016344578936696053,
-0.07584261149168015,
0.020161736756563187,
-0.09869375079870224,
0.055850550532341,
-0.3218844532966614,
0.09342993050813675,
-0.11400728672742844,
0.009930809028446674,
-0.1314808875322342,
-0.0335058830678463,
0.016112782061100006,
0.005322059616446495,
0.013146123848855495,
0.11860170215368271,
-0.15911765396595,
-0.09210728108882904,
0.09710568189620972,
-0.13270564377307892,
-0.06123318895697594,
0.13335616886615753,
0.006097014062106609,
-0.007668835576623678,
0.0786266103386879,
0.1425439566373825,
0.14123733341693878,
-0.04838160052895546,
-0.13677071034908295,
-0.06624309718608856,
0.027206536382436752,
0.1544196456670761,
0.06648306548595428,
-0.0522027313709259,
-0.0182489063590765,
0.03598593920469284,
-0.031956933438777924,
-0.0026491705793887377,
-0.016834797337651253,
-0.05705594643950462,
-0.023978661745786667,
-0.002377238357439637,
0.096376933157444,
-0.015163181349635124,
0.002383065177127719,
-0.07838787883520126,
-0.09363235533237457,
0.039118777960538864,
0.0882512778043747,
0.041436195373535156,
0.01435884553939104,
-0.06161881238222122,
0.09994251281023026,
-0.03056234121322632,
0.022030595690011978,
-0.10238077491521835,
-0.15781597793102264,
0.03258318454027176,
-0.15215061604976654,
0.06616183370351791,
0.046265859156847,
0.06192070245742798,
0.027611956000328064,
-0.11285637319087982,
-0.06351099908351898,
-0.02237211912870407,
0.016440147534012794,
0.007992693223059177,
-0.12434909492731094,
-0.08924297988414764,
-0.04348140209913254,
0.22069425880908966,
-0.054984692484140396,
-0.007727441843599081,
0.06026673689484596,
0.2515117824077606,
0.018960142508149147,
-0.014410633593797684,
-0.05082327499985695,
0.07166947424411774,
0.00031130402931012213,
0.0197899267077446,
0.01994451880455017,
-0.04336968809366226,
-0.14637522399425507,
0.095527283847332,
-0.055198993533849716,
-0.02979397401213646,
0.06157594546675682,
0.054885584861040115,
-0.09651610255241394,
-0.011951263062655926,
-0.05258362367749214,
-0.029914764687418938,
0.006216008216142654,
-0.12069366872310638,
0.042220812290906906,
0.04080578312277794,
0.03252416476607323,
-0.0428742878139019,
-0.048637744039297104,
0.01096632331609726,
0.00038561708061024547,
-0.12152905762195587,
0.10360410064458847,
-0.08400224894285202,
-0.21976810693740845,
0.10374713689088821,
0.15290415287017822,
0.07434318959712982,
0.12372945249080658,
-0.013827105984091759,
-0.022371480241417885,
-0.1197407990694046,
0.0015785678988322616,
0.021130502223968506,
0.042604196816682816,
0.04024510458111763,
0.05408281460404396,
0.03447001054883003,
0.043456509709358215,
0.02863817848265171,
-0.02880154922604561,
0.04750415310263634,
-0.009588376618921757,
0.0040976181626319885,
0.12772449851036072,
0.10417608171701431,
0.02852170541882515,
0.08533679693937302,
0.03910043090581894,
-0.0404503159224987,
0.01288803294301033,
-0.02749737910926342,
-0.06698718667030334,
0.10686532407999039,
-0.10741190612316132,
-0.16609297692775726,
-0.17008981108665466,
-0.051623594015836716,
-0.18006007373332977,
-0.024816229939460754,
0.03015192784368992,
-0.03652867674827576,
-0.07785869389772415,
-0.07390360534191132,
0.13494938611984253,
0.03926362842321396,
-0.05240222439169884,
-0.05028913915157318,
-0.03344966843724251,
0.042253971099853516,
-0.10105009377002716,
-0.04346547648310661,
0.0018156386213377118,
-0.050230301916599274,
0.021803520619869232,
0.027263345196843147,
0.0939897820353508,
0.060186538845300674,
-0.00003084815762122162,
-0.02964363619685173,
-0.040875863283872604,
0.17850545048713684,
-0.1293039619922638,
0.11540871858596802,
0.04773532599210739,
-0.0027224498335272074,
0.06698734313249588,
0.16732339560985565,
0.05765581503510475,
-0.04658689722418785,
-0.011350571177899837,
0.06666777282953262,
-0.030819516628980637,
-0.19200317561626434,
-0.1769828051328659,
-0.06770068407058716,
0.007487337104976177,
-0.027668775990605354,
0.04515873268246651,
0.06839024275541306,
-0.0003817266842816025,
-0.09345950931310654,
-0.07025164365768433,
0.036232683807611465,
0.04945642128586769,
0.17080222070217133,
0.009348781779408455,
0.05902401730418205,
-0.06639035046100616,
-0.06039080023765564,
0.09840981662273407,
0.00460409838706255,
0.08121812343597412,
0.06772522628307343,
0.06868960708379745,
0.051509808748960495,
0.012179503217339516,
0.0003098213055636734,
0.04621569812297821,
0.015963001176714897,
-0.02683101035654545,
-0.011713831685483456,
-0.08623592555522919,
0.07211727648973465,
0.08716612309217453,
0.10278671234846115,
0.03722165524959564,
-0.051564089953899384,
-0.06990828365087509,
0.12818388640880585,
0.1740635186433792,
0.11096327751874924,
-0.12066809087991714,
-0.10149893164634705,
0.059367019683122635,
-0.012059872969985008,
0.020104901865124702,
-0.03295712172985077,
-0.051309432834386826,
-0.13940629363059998,
0.20302264392375946,
-0.018019717186689377,
0.06347235292196274,
0.0055109113454818726,
-0.0001949289726326242,
-0.02158956229686737,
0.06696530431509018,
0.017838791012763977,
0.050336673855781555,
-0.141648069024086,
0.14686071872711182,
0.010985979810357094,
0.021909072995185852,
-0.01216028444468975,
0.05184603109955788,
0.01702127605676651,
0.05610823258757591,
0.10946010798215866,
0.04561372473835945,
-0.09489621967077255,
-0.08675345033407211,
-0.04737729951739311,
-0.05841971933841705,
0.13532322645187378,
-0.0978664755821228,
0.08815576881170273,
-0.005218944977968931,
-0.04548734799027443,
-0.058141354471445084,
0.16146430373191833,
-0.15934163331985474,
-0.10354094952344894,
0.10525287687778473,
-0.04850301891565323,
0.028323976323008537,
-0.07308647781610489,
-0.04719609022140503,
-0.24660754203796387,
0.2158661186695099,
0.015384898521006107,
0.00384410098195076,
-0.09645523875951767,
-0.0031059239991009235,
0.16681231558322906,
-0.09507244825363159,
0.038081102073192596,
-0.0030803470872342587,
0.08820296823978424,
-0.09653045237064362,
-0.08165580779314041,
0.04465007409453392,
-0.07744460552930832,
-0.14185139536857605,
0.011764489114284515,
0.19781433045864105,
0.022616581991314888,
0.037971191108226776,
0.025954479351639748,
0.09160146117210388,
0.08607495576143265,
-0.10742948949337006,
0.08422011882066727,
0.0879930630326271,
-0.030707523226737976,
0.1495898962020874,
-0.033895764499902725,
-0.15872009098529816,
-0.05146879330277443,
-0.0411202572286129,
0.10311200469732285,
0.34257373213768005,
-0.06467219442129135,
0.08678370714187622,
0.08080018311738968,
-0.042148686945438385,
-0.19632753729820251,
0.007789987605065107,
0.01917961612343788,
0.02523370087146759,
0.00445806048810482,
-0.05062577873468399,
0.10844220966100693,
0.13072700798511505,
-0.020657971501350403,
-0.04788164049386978,
-0.20526863634586334,
-0.1290767788887024,
0.011098041199147701,
-0.06513512134552002,
-0.012189746834337711,
-0.06006240472197533,
-0.10408233106136322,
-0.07066939771175385,
-0.16428698599338531,
0.05405878648161888,
0.04084285721182823,
0.048617810010910034,
0.0022039911709725857,
-0.022370649501681328,
0.03627573698759079,
-0.018924014642834663,
0.18879440426826477,
0.06920293718576431,
0.02023874782025814,
-0.07239631563425064,
-0.04903245344758034,
0.11341241002082825,
-0.037153460085392,
0.09947021305561066,
0.04320525750517845,
-0.005356981884688139,
-0.11361989378929138,
0.012292295694351196,
-0.09132620692253113,
0.11476991325616837,
-0.09256471693515778,
-0.030837757512927055,
-0.08548001945018768,
0.140853151679039,
0.0238761268556118,
-0.015673771500587463,
0.12679731845855713,
-0.032148636877536774,
0.1091376468539238,
0.13080862164497375,
0.0701589584350586,
0.0031420262530446053,
-0.07472140341997147,
0.0012759605888277292,
-0.06170511618256569,
0.019098622724413872,
0.05546579137444496,
0.04380559176206589,
0.13516387343406677,
-0.01975567452609539,
0.14712172746658325,
-0.018929120153188705,
-0.13689684867858887,
-0.023719700053334236,
0.11753503233194351,
-0.11223208159208298,
-0.2821040451526642,
-0.04018678143620491,
-0.034599367529153824,
0.012139244005084038,
0.061747223138809204,
0.19457398355007172,
-0.03780088946223259,
-0.023348631337285042,
0.021921996027231216,
0.04527831822633743,
-0.09146711230278015,
0.08037439733743668,
-0.021092766895890236,
0.042854998260736465,
-0.06821004301309586,
0.10744444280862808,
0.07745382934808731,
-0.22681817412376404,
-0.025234894827008247,
0.09869363903999329,
-0.06257352977991104,
-0.07561001926660538,
-0.09376278519630432,
0.1380273848772049,
-0.1314297765493393,
-0.0158812515437603,
-0.020428206771612167,
-0.1451670229434967,
0.04510067030787468,
0.09574063867330551,
0.04472881183028221,
0.011867307126522064,
0.035759247839450836,
0.00018209819972980767,
0.01861182041466236,
0.03738055005669594,
0.09328993409872055,
0.10640279203653336,
-0.042188696563243866,
0.03816690295934677,
-0.08412777632474899,
0.015412991866469383,
-0.021022463217377663,
-0.01296116691082716,
-0.09465239942073822,
-0.07739552855491638,
-0.10236255824565887,
-0.06780892610549927,
-0.08637113869190216,
-0.02516632154583931,
-0.0233630258589983,
-0.023995107039809227,
0.004357992671430111,
-0.027227098122239113,
-0.07138573378324509,
-0.050982195883989334,
-0.0155563373118639,
0.08955743908882141,
-0.1038399413228035,
-0.0027917600236833096,
0.06826390326023102,
-0.051174331456422806,
0.0997040867805481,
0.024214884266257286,
0.0514301136136055,
0.03840871900320053,
-0.03028056025505066,
-0.03847609460353851,
0.01905941218137741,
0.006461322773247957,
0.05086956173181534,
-0.14872215688228607,
-0.012150841765105724,
-0.0032515712082386017,
-0.051708951592445374,
0.050017524510622025,
-0.0625491663813591,
-0.10236836224794388,
0.008505094796419144,
0.016627702862024307,
-0.09351813793182373,
-0.07604184001684189,
0.027614127844572067,
0.0729801207780838,
0.035474032163619995,
0.11152558028697968,
-0.058112598955631256,
0.05574757605791092,
-0.1357552707195282,
-0.01836516335606575,
-0.02554010972380638,
-0.1030651405453682,
0.016653461381793022,
-0.011252963915467262,
0.022870756685733795,
-0.0031690942123532295,
0.11289511620998383,
-0.026996877044439316,
-0.025828121230006218,
0.06756260991096497,
-0.028583889827132225,
-0.0780753642320633,
0.0507938489317894,
0.07868042588233948,
0.04812269285321236,
-0.0161933284252882,
-0.08120898902416229,
-0.0014988429611548781,
-0.05973583459854126,
-0.007319761905819178,
0.1114073172211647,
0.14946618676185608,
0.11061113327741623,
0.05770944431424141,
0.08207276463508606,
-0.04891999810934067,
-0.028829647228121758,
-0.009984364733099937,
-0.029658254235982895,
0.08592110872268677,
-0.02771167829632759,
0.046906858682632446,
0.07054735720157623,
-0.23260782659053802,
0.1542171984910965,
-0.04036210849881172,
0.003158458275720477,
-0.019460761919617653,
-0.11383481323719025,
-0.07883796095848083,
-0.042559150606393814,
0.010454006493091583,
-0.08244454860687256,
0.009923603385686874,
0.00909425225108862,
0.03983308747410774,
-0.018626000732183456,
0.14684614539146423,
-0.22303219139575958,
-0.020058225840330124,
0.1173686534166336,
0.01384452823549509,
-0.01986311748623848,
0.04652818292379379,
-0.01833965815603733,
-0.05564456060528755,
0.0526103749871254,
0.003485421882942319,
0.042003098875284195,
0.01768621988594532,
-0.009287383407354355,
-0.030709892511367798,
-0.0902811586856842,
0.015258834697306156,
-0.013983458280563354,
0.019041143357753754,
0.026359856128692627,
0.02583683282136917,
0.016558820381760597,
-0.0304630808532238,
0.1203962042927742,
-0.04186068847775459,
-0.08964694291353226,
-0.09655430167913437,
0.18649248778820038,
0.015295648016035557,
0.07767751067876816,
0.014468489214777946,
-0.11231113970279694,
0.003929391037672758,
0.011538079008460045,
0.2675095796585083,
0.06229402497410774,
0.0249533299356699,
0.022313563153147697,
0.0121700968593359,
0.03887152299284935,
0.019862765446305275,
-0.021693918853998184,
0.17263448238372803,
-0.035285405814647675,
0.2010856717824936,
-0.05225873738527298,
-0.046340879052877426,
0.0050239525735378265,
0.1740422248840332,
-0.01995527185499668,
-0.01296838466078043,
-0.015919938683509827,
0.11686369776725769,
-0.02941286377608776,
-0.15430203080177307,
0.030166465789079666,
-0.1680508852005005,
-0.1555458903312683,
-0.04883193224668503,
0.03437260910868645,
0.0537441223859787,
0.11792732775211334,
0.015498639084398746,
-0.03355927765369415,
0.11172260344028473,
0.013434622436761856,
-0.01853151060640812,
-0.12593184411525726,
0.06253454834222794,
-0.005848423577845097,
0.10684798657894135,
0.020717212930321693,
0.04494085907936096,
0.12986114621162415,
-0.024106910452246666,
-0.10693443566560745,
0.014985187910497189,
0.045302849262952805,
-0.08386611938476562,
0.0049399384297430515,
0.15897472202777863,
-0.014517411589622498,
0.11524591594934464,
0.05899004638195038,
-0.13243594765663147,
0.07087180018424988,
0.0894564762711525,
-0.041566379368305206,
-0.06131380423903465,
0.1714923232793808,
-0.08582644164562225,
0.13566255569458008,
0.23618710041046143,
-0.026338893920183182,
0.025297190994024277,
-0.03987700119614601,
0.044342536479234695,
0.07475463300943375,
0.12010806798934937,
-0.017106499522924423,
-0.16066502034664154,
0.06056363880634308,
0.023605575785040855,
0.04031899571418762,
-0.14200104773044586,
-0.11799366027116776,
-0.015118377283215523,
0.02434436045587063,
-0.030886957421898842,
0.12695416808128357,
0.08972851932048798,
-0.0224088616669178,
0.001209444715641439,
-0.1888996660709381,
-0.008432535454630852,
0.07094074040651321,
-0.045595064759254456,
-0.016666574403643608
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# output-7b-26k-lora
This model is a fine-tuned version of [deepseek-ai/deepseek-coder-7b-instruct-v1.5](https://huggingface.co/deepseek-ai/deepseek-coder-7b-instruct-v1.5) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 2
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "other", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "deepseek-ai/deepseek-coder-7b-instruct-v1.5", "model-index": [{"name": "output-7b-26k-lora", "results": []}]} | null | zzz99/output-7b-26k-lora | [
"peft",
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:deepseek-ai/deepseek-coder-7b-instruct-v1.5",
"license:other",
"region:us"
] | 2024-02-08T17:46:36+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #generated_from_trainer #base_model-deepseek-ai/deepseek-coder-7b-instruct-v1.5 #license-other #region-us
|
# output-7b-26k-lora
This model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 4
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 16
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_steps: 10
- num_epochs: 2
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# output-7b-26k-lora\n\nThis model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 4\n- eval_batch_size: 1\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 64\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 2",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-deepseek-ai/deepseek-coder-7b-instruct-v1.5 #license-other #region-us \n",
"# output-7b-26k-lora\n\nThis model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 4\n- eval_batch_size: 1\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 64\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 2",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
55,
46,
6,
12,
8,
3,
129,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-deepseek-ai/deepseek-coder-7b-instruct-v1.5 #license-other #region-us \n# output-7b-26k-lora\n\nThis model is a fine-tuned version of deepseek-ai/deepseek-coder-7b-instruct-v1.5 on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 4\n- eval_batch_size: 1\n- seed: 42\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 64\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 2### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.08614028245210648,
0.08098314702510834,
-0.003283097641542554,
0.08254305273294449,
0.12209969758987427,
0.022924095392227173,
0.11968836188316345,
0.12029042840003967,
-0.07489021122455597,
0.0968521386384964,
0.0350957028567791,
0.05529683083295822,
0.05858607217669487,
0.14050590991973877,
-0.04510785639286041,
-0.22825901210308075,
0.013398556038737297,
-0.021985825151205063,
-0.09267715364694595,
0.10612037032842636,
0.10306701809167862,
-0.1099899485707283,
0.057669270783662796,
0.020021354779601097,
-0.13969403505325317,
0.021785853430628777,
-0.025897929444909096,
-0.040247708559036255,
0.09664888679981232,
-0.00069021200761199,
0.13693787157535553,
0.001165088266134262,
0.14140929281711578,
-0.20298714935779572,
0.003421869594603777,
0.11148522049188614,
0.03686715289950371,
0.10159236192703247,
0.06314510852098465,
0.02325616404414177,
0.012988334521651268,
-0.1238524466753006,
0.10959777981042862,
0.008064260706305504,
-0.09402329474687576,
-0.1925833821296692,
-0.09645739197731018,
0.07571743428707123,
0.1239277720451355,
0.08236108720302582,
0.0136871924623847,
0.16934847831726074,
-0.06465951353311539,
0.060423072427511215,
0.22067782282829285,
-0.2917259931564331,
-0.07054103910923004,
0.0419071763753891,
0.033136170357465744,
0.07492212951183319,
-0.10282547771930695,
-0.03540099784731865,
0.039411891251802444,
0.016553781926631927,
0.07476378977298737,
0.016920888796448708,
-0.05321347340941429,
-0.0335056409239769,
-0.12299919873476028,
-0.019162554293870926,
0.08566348254680634,
0.04034634307026863,
-0.03046710230410099,
-0.10870519280433655,
-0.0839860737323761,
-0.16628475487232208,
-0.028257355093955994,
-0.06619161367416382,
0.042266808450222015,
-0.037367600947618484,
-0.013691222295165062,
-0.027838287875056267,
-0.07910230755805969,
-0.06837150454521179,
0.009797637350857258,
0.13944032788276672,
0.0493309460580349,
0.01777048222720623,
-0.023478345945477486,
0.11630924046039581,
0.011847524903714657,
-0.14831556379795074,
0.008428952656686306,
-0.001522752339951694,
-0.07700630277395248,
-0.06674028187990189,
-0.04210132732987404,
-0.01592552475631237,
-0.01993161253631115,
0.17128421366214752,
-0.07694108784198761,
0.07886204868555069,
-0.01772216148674488,
0.0124895004555583,
-0.048998840153217316,
0.1409725546836853,
-0.04796083644032478,
0.005225380416959524,
0.018904419615864754,
0.12754805386066437,
0.05021107569336891,
-0.011712318286299706,
-0.08603863418102264,
-0.016390567645430565,
0.07254776358604431,
0.04660506919026375,
-0.050658147782087326,
0.0035072513855993748,
-0.057324301451444626,
-0.042563073337078094,
0.07412932068109512,
-0.1353379487991333,
0.05876503884792328,
0.019578032195568085,
-0.0736149400472641,
-0.042575910687446594,
0.014940180815756321,
0.04154500365257263,
-0.02214079722762108,
0.11342230439186096,
-0.058329083025455475,
0.013691543601453304,
-0.09501726925373077,
-0.06555972248315811,
0.013379848562180996,
-0.05393213406205177,
-0.014415976591408253,
-0.06108316779136658,
-0.159471333026886,
-0.0365060456097126,
0.04655204713344574,
-0.07451014220714569,
-0.04565782472491264,
-0.0025922113563865423,
-0.08237046003341675,
0.005986865144222975,
-0.004789876751601696,
0.11341126263141632,
-0.039975326508283615,
0.07253411412239075,
-0.005261390004307032,
0.027716370299458504,
0.03713257610797882,
0.020718401297926903,
-0.08552902191877365,
0.027989406138658524,
-0.17636793851852417,
0.03812168166041374,
-0.0736059695482254,
0.007037541829049587,
-0.1048457995057106,
-0.10062903165817261,
-0.040565524250268936,
-0.013655168935656548,
0.07223504036664963,
0.08152057975530624,
-0.17273515462875366,
-0.023723306134343147,
0.13757392764091492,
-0.09820312261581421,
-0.09170994162559509,
0.08992918580770493,
-0.040749840438365936,
0.02091379277408123,
0.029382267966866493,
0.14535145461559296,
0.10323547571897507,
-0.14951109886169434,
-0.004027163609862328,
-0.01461684051901102,
0.07730206847190857,
0.032450392842292786,
0.06706668436527252,
-0.010555868037045002,
0.09646883606910706,
-0.01801571622490883,
-0.03755859285593033,
-0.022878825664520264,
-0.06659425795078278,
-0.08224226534366608,
-0.05190407857298851,
-0.06589117646217346,
0.038800716400146484,
0.0350365936756134,
0.0479884073138237,
-0.07214376330375671,
-0.13066603243350983,
0.09611436724662781,
0.1169915497303009,
-0.04648828133940697,
0.021847300231456757,
-0.07948583364486694,
0.03996163606643677,
-0.019376195967197418,
-0.04725026711821556,
-0.19515460729599,
-0.10111821442842484,
0.04853801056742668,
-0.08554209768772125,
0.00396727817133069,
0.027236804366111755,
0.07905590534210205,
0.08454364538192749,
-0.044679056853055954,
-0.035665981471538544,
-0.11079999804496765,
-0.00716862129047513,
-0.09235115349292755,
-0.19151318073272705,
-0.06523209065198898,
-0.021824825555086136,
0.162618950009346,
-0.2010728269815445,
0.014296941459178925,
-0.022035516798496246,
0.1863906979560852,
0.0447111651301384,
-0.06333433836698532,
-0.022759506478905678,
0.05089307203888893,
0.015812581405043602,
-0.08815640956163406,
0.025621697306632996,
-0.008660080842673779,
-0.06034207344055176,
-0.0669412910938263,
-0.16874954104423523,
0.10719301551580429,
0.08383787423372269,
0.027809692546725273,
-0.07818692922592163,
0.005224599968641996,
-0.06452695280313492,
-0.03828093409538269,
-0.06095314025878906,
0.005008417647331953,
0.19435901939868927,
0.02563202567398548,
0.12267297506332397,
-0.09197311103343964,
-0.08672341704368591,
-0.005604657810181379,
-0.003006385639309883,
0.016754966229200363,
0.052057795226573944,
0.07323653995990753,
-0.11072909086942673,
0.09205081313848495,
0.1275029480457306,
-0.04101094603538513,
0.11734867095947266,
-0.07383863627910614,
-0.08512347936630249,
-0.02296956442296505,
0.02204514481127262,
-0.004317250102758408,
0.1163339614868164,
-0.016721554100513458,
0.04383384436368942,
0.02523432858288288,
0.01324415672570467,
0.02887308970093727,
-0.19950003921985626,
-0.0027224349323660135,
0.02693518064916134,
-0.030521780252456665,
0.014560804702341557,
-0.03435300663113594,
0.062243420630693436,
0.086301289498806,
0.02061646617949009,
-0.0080857640132308,
0.007172466721385717,
-0.01928592659533024,
-0.07546685636043549,
0.1645844727754593,
-0.1137256920337677,
-0.12412073463201523,
-0.1276654154062271,
0.06454028189182281,
-0.023097053170204163,
-0.03487076237797737,
-0.00988179724663496,
-0.08649972826242447,
-0.056206993758678436,
-0.09958367794752121,
-0.025089461356401443,
0.007140897214412689,
-0.017264604568481445,
0.11499954015016556,
0.02188151143491268,
0.11206842213869095,
-0.13013701140880585,
0.015753624960780144,
-0.03238344192504883,
-0.07752124965190887,
-0.018969863653182983,
0.09645885974168777,
0.051713693886995316,
0.13251641392707825,
-0.01450743805617094,
0.014169913716614246,
-0.023015059530735016,
0.21630743145942688,
-0.1015615239739418,
0.0018409984186291695,
0.1551712453365326,
-0.04125286638736725,
0.05646173283457756,
0.1348995417356491,
0.05764133855700493,
-0.09309489279985428,
0.016440104693174362,
0.0757843554019928,
-0.00232772552408278,
-0.2175448089838028,
-0.03817394748330116,
-0.00894457008689642,
-0.08473797887563705,
0.08607590943574905,
0.04284609854221344,
0.002510561142116785,
0.021238230168819427,
-0.03327756002545357,
0.023369915783405304,
0.010307593271136284,
0.07355330884456635,
0.07867250591516495,
0.05823551118373871,
0.09874295443296432,
-0.024669015780091286,
-0.039226140826940536,
0.04436623305082321,
0.00682451669126749,
0.184485524892807,
-0.020267337560653687,
0.09528864920139313,
0.0173796396702528,
0.1378052681684494,
-0.007103152573108673,
0.04458146542310715,
0.005395484622567892,
-0.026009568944573402,
0.012869895435869694,
-0.0684487372636795,
-0.02673233300447464,
0.026028664782643318,
-0.04823915287852287,
0.06806759536266327,
-0.03963087126612663,
0.05152108520269394,
0.025951500982046127,
0.25259318947792053,
0.044369764626026154,
-0.2695312201976776,
-0.0663161501288414,
0.012007830664515495,
-0.03480001166462898,
-0.060270413756370544,
0.006563376169651747,
0.10964859277009964,
-0.10725949704647064,
0.052119556814432144,
-0.0809050053358078,
0.07320256531238556,
-0.03719867020845413,
0.013559810817241669,
0.10008841007947922,
0.12650185823440552,
0.013164425268769264,
0.050608955323696136,
-0.15661205351352692,
0.1937166452407837,
0.021489961072802544,
0.11599556356668472,
-0.058036625385284424,
0.05508200451731682,
0.00042395657510496676,
0.034467052668333054,
0.05981418490409851,
-0.017149416729807854,
-0.020105578005313873,
-0.18004873394966125,
-0.11569993942975998,
0.02415836974978447,
0.13887237012386322,
-0.05805046111345291,
0.08356595784425735,
-0.043748192489147186,
-0.005659583956003189,
0.046199019998311996,
-0.07329071313142776,
-0.13829784095287323,
-0.10323072969913483,
0.030207350850105286,
-0.02334849163889885,
-0.04530337452888489,
-0.0966368094086647,
-0.10483895242214203,
-0.06392387300729752,
0.1454606056213379,
-0.03265862539410591,
-0.038530170917510986,
-0.13043789565563202,
0.08082591742277145,
0.1261356920003891,
-0.0424443781375885,
0.033159058541059494,
0.02752521261572838,
0.12848886847496033,
0.03441443666815758,
-0.07941550761461258,
0.08401957899332047,
-0.06881815940141678,
-0.21308255195617676,
-0.048239387571811676,
0.15723916888237,
0.025000249966979027,
0.04477326199412346,
-0.014944848604500294,
0.040959082543849945,
0.03249500319361687,
-0.07843899726867676,
0.02099102921783924,
0.06293614208698273,
0.06322897970676422,
0.030722130089998245,
-0.07448957860469818,
0.0423014834523201,
-0.021711770445108414,
-0.00029490224551409483,
0.09121384471654892,
0.24155525863170624,
-0.07992663979530334,
0.06576144695281982,
0.019023030996322632,
-0.05234675481915474,
-0.1605723649263382,
0.047863446176052094,
0.14187458157539368,
0.018597405403852463,
0.062431223690509796,
-0.17016537487506866,
0.12792029976844788,
0.11659921705722809,
-0.040962059050798416,
0.044603489339351654,
-0.32224148511886597,
-0.14575757086277008,
0.05685020983219147,
0.104176826775074,
-0.01101667433977127,
-0.11901398003101349,
-0.03947945311665535,
-0.018456872552633286,
-0.13767439126968384,
0.104782834649086,
-0.09900060296058655,
0.08049323409795761,
0.0045913089998066425,
0.0399005301296711,
0.024727633222937584,
-0.034068576991558075,
0.13779835402965546,
-0.020022129639983177,
0.10542502254247665,
-0.04612375795841217,
0.03616580739617348,
0.06006492301821709,
-0.05491029471158981,
0.010037112981081009,
0.003972949925810099,
0.05043928325176239,
-0.11994786560535431,
-0.002390650799497962,
-0.060338929295539856,
0.02502569742500782,
-0.0651865303516388,
-0.04986095055937767,
-0.04194759950041771,
0.05458182096481323,
0.039197709411382675,
-0.03809233382344246,
0.11491981893777847,
0.04516839608550072,
0.13924232125282288,
0.11809235066175461,
0.08847009390592575,
0.018886076286435127,
-0.08190644532442093,
-0.0004363581247162074,
-0.013532436452805996,
0.06354209035634995,
-0.10576450824737549,
0.03772750496864319,
0.11240578442811966,
0.02565278857946396,
0.09765632450580597,
0.03221983462572098,
-0.07092013955116272,
-0.012829873710870743,
0.04229491576552391,
-0.1063002273440361,
-0.12337682396173477,
0.009989969432353973,
0.01967892050743103,
-0.12037152051925659,
0.014675803482532501,
0.12383728474378586,
-0.036976318806409836,
-0.019537923857569695,
0.012761305086314678,
0.01787574030458927,
-0.010890125297009945,
0.15567176043987274,
0.03876660019159317,
0.07275541871786118,
-0.07512664794921875,
0.11508435755968094,
0.06634900718927383,
-0.033397141844034195,
0.048074644058942795,
0.07270841300487518,
-0.07771912217140198,
0.0016081632347777486,
0.08252523839473724,
0.1439196616411209,
0.03017369657754898,
-0.03797965496778488,
-0.11160121858119965,
-0.09798227250576019,
0.039563268423080444,
0.12951454520225525,
0.04466438293457031,
-0.01579052396118641,
-0.018901100382208824,
0.040659792721271515,
-0.13489416241645813,
0.10389523208141327,
0.019816739484667778,
0.07760225236415863,
-0.15754711627960205,
0.09447076171636581,
0.011128026060760021,
-0.007980612106621265,
-0.014272552914917469,
0.047388363629579544,
-0.08404449373483658,
-0.02513270266354084,
-0.09884050488471985,
-0.02808910422027111,
0.002746689599007368,
0.00017729002865962684,
-0.007185469381511211,
-0.03957592695951462,
-0.045489612966775894,
0.04024018347263336,
-0.07254290580749512,
-0.056189198046922684,
-0.00014299582107923925,
0.021942172199487686,
-0.15831421315670013,
0.0019379016011953354,
0.03273483365774155,
-0.09963573515415192,
0.08256315439939499,
0.047360196709632874,
0.05207614600658417,
0.038446273654699326,
-0.11415885388851166,
-0.010135749354958534,
0.041247621178627014,
0.03456306457519531,
0.053555771708488464,
-0.08085731416940689,
-0.008980127051472664,
-0.04017551615834236,
0.01492486335337162,
0.011845726519823074,
0.0805625170469284,
-0.12801198661327362,
-0.007915690541267395,
-0.040941715240478516,
-0.05090327188372612,
-0.045775383710861206,
0.02386646904051304,
0.0934823527932167,
0.01192823052406311,
0.152825266122818,
-0.07869643718004227,
0.05683138966560364,
-0.22829657793045044,
-0.036475393921136856,
0.006654495373368263,
-0.00569529365748167,
-0.08104995638132095,
-0.023796234279870987,
0.0804659053683281,
-0.058554377406835556,
0.1012558713555336,
-0.007258610799908638,
0.07994726300239563,
0.05507579818367958,
-0.047303151339292526,
-0.01930195651948452,
0.0463881641626358,
0.1429498940706253,
0.0638372078537941,
-0.004914695397019386,
0.09221258014440536,
-0.04402418062090874,
0.06039964407682419,
-0.004143711179494858,
0.1834283173084259,
0.1856648176908493,
-0.024555925279855728,
0.04526621103286743,
0.08485755324363708,
-0.12358272075653076,
-0.12388475239276886,
0.13317063450813293,
-0.06169629842042923,
0.09062673896551132,
-0.05618595704436302,
0.1349785327911377,
0.08915574103593826,
-0.16536764800548553,
0.021115481853485107,
-0.05729217454791069,
-0.0863608717918396,
-0.13073214888572693,
-0.01635848544538021,
-0.08878657221794128,
-0.13204547762870789,
0.029198158532381058,
-0.11990823596715927,
0.02798350341618061,
0.10495124012231827,
0.01144582498818636,
0.012485850602388382,
0.1936526894569397,
-0.018226712942123413,
0.03157000243663788,
0.03707962855696678,
0.03762480616569519,
0.0146798649802804,
-0.04835374653339386,
-0.08889465779066086,
0.042870718985795975,
0.010275235399603844,
0.07665689289569855,
-0.0559377521276474,
-0.021572979167103767,
0.021994492039084435,
0.0029994617216289043,
-0.0905129611492157,
0.05141547694802284,
-0.007402791175991297,
0.020433494821190834,
0.03307333216071129,
0.03598378598690033,
0.037385065108537674,
-0.06534447520971298,
0.27385348081588745,
-0.0716160461306572,
-0.07361255586147308,
-0.13272738456726074,
0.17519785463809967,
0.02740306966006756,
0.0031196512281894684,
0.07022994011640549,
-0.12228818237781525,
-0.039539217948913574,
0.1002817153930664,
0.13040883839130402,
-0.09582424908876419,
-0.002275798236951232,
0.022614408284425735,
-0.01434698048979044,
-0.08100677281618118,
0.11607390642166138,
0.12215408682823181,
0.004411947913467884,
-0.053008779883384705,
-0.02392967790365219,
0.0057034059427678585,
-0.032972682267427444,
-0.08227110654115677,
0.06512881815433502,
0.017444372177124023,
0.009741383604705334,
-0.034590959548950195,
0.06757459789514542,
0.044312335550785065,
-0.16837933659553528,
0.07121849060058594,
-0.22145698964595795,
-0.20384781062602997,
-0.006229436956346035,
0.117470383644104,
-0.045370250940322876,
0.0512520857155323,
0.006896679755300283,
-0.03775838389992714,
0.11544936150312424,
-0.019127635285258293,
-0.010374755598604679,
-0.09150731563568115,
0.05303631350398064,
-0.13146638870239258,
0.22348442673683167,
-0.000003957301942136837,
0.07768532633781433,
0.09122028201818466,
0.005091962404549122,
-0.09889533370733261,
0.028343548998236656,
0.07409542798995972,
-0.10734100639820099,
-0.022090477868914604,
0.1688891053199768,
-0.042992882430553436,
0.09733818471431732,
0.06665732711553574,
-0.18456721305847168,
-0.020777055993676186,
-0.058310411870479584,
-0.05431383475661278,
-0.07084754854440689,
-0.008271733298897743,
-0.06944264471530914,
0.150910884141922,
0.20585502684116364,
-0.026903223246335983,
0.02152220346033573,
-0.043750084936618805,
0.0407634899020195,
0.054011303931474686,
0.09876930713653564,
-0.02245117537677288,
-0.23696020245552063,
0.03963650390505791,
0.027116868644952774,
0.02472032606601715,
-0.19087903201580048,
-0.09421120584011078,
0.05582846701145172,
-0.05999605357646942,
-0.09126754850149155,
0.08579890429973602,
0.04452516511082649,
0.036983247846364975,
-0.029140695929527283,
-0.1113438829779625,
-0.03841441497206688,
0.14122925698757172,
-0.16070657968521118,
-0.04059425741434097
] |
null | null | null |
# RWKV-6 World
Use rwkv pip package 0.8.24+ for RWKV-6 inference: https://pypi.org/project/rwkv/ (pipeline = PIPELINE(model, "rwkv_vocab_v20230424") for rwkv-world models)
Trained on same world-2 dataset as https://huggingface.co/BlinkDL/rwkv-5-world.
| {"language": ["en", "zh", "fr", "es", "de", "pt", "ru", "it", "ja", "ko", "vi", "ar"], "license": "apache-2.0", "tags": ["pytorch", "text-generation", "causal-lm", "rwkv"], "datasets": ["cerebras/SlimPajama-627B", "EleutherAI/pile", "bigcode/starcoderdata", "oscar-corpus/OSCAR-2301"]} | text-generation | BlinkDL/rwkv-6-world | [
"pytorch",
"text-generation",
"causal-lm",
"rwkv",
"en",
"zh",
"fr",
"es",
"de",
"pt",
"ru",
"it",
"ja",
"ko",
"vi",
"ar",
"dataset:cerebras/SlimPajama-627B",
"dataset:EleutherAI/pile",
"dataset:bigcode/starcoderdata",
"dataset:oscar-corpus/OSCAR-2301",
"license:apache-2.0",
"has_space",
"region:us"
] | 2024-02-08T17:47:54+00:00 | [] | [
"en",
"zh",
"fr",
"es",
"de",
"pt",
"ru",
"it",
"ja",
"ko",
"vi",
"ar"
] | TAGS
#pytorch #text-generation #causal-lm #rwkv #en #zh #fr #es #de #pt #ru #it #ja #ko #vi #ar #dataset-cerebras/SlimPajama-627B #dataset-EleutherAI/pile #dataset-bigcode/starcoderdata #dataset-oscar-corpus/OSCAR-2301 #license-apache-2.0 #has_space #region-us
|
# RWKV-6 World
Use rwkv pip package 0.8.24+ for RWKV-6 inference: URL (pipeline = PIPELINE(model, "rwkv_vocab_v20230424") for rwkv-world models)
Trained on same world-2 dataset as URL
| [
"# RWKV-6 World\n\nUse rwkv pip package 0.8.24+ for RWKV-6 inference: URL (pipeline = PIPELINE(model, \"rwkv_vocab_v20230424\") for rwkv-world models)\n\nTrained on same world-2 dataset as URL"
] | [
"TAGS\n#pytorch #text-generation #causal-lm #rwkv #en #zh #fr #es #de #pt #ru #it #ja #ko #vi #ar #dataset-cerebras/SlimPajama-627B #dataset-EleutherAI/pile #dataset-bigcode/starcoderdata #dataset-oscar-corpus/OSCAR-2301 #license-apache-2.0 #has_space #region-us \n",
"# RWKV-6 World\n\nUse rwkv pip package 0.8.24+ for RWKV-6 inference: URL (pipeline = PIPELINE(model, \"rwkv_vocab_v20230424\") for rwkv-world models)\n\nTrained on same world-2 dataset as URL"
] | [
111,
69
] | [
"passage: TAGS\n#pytorch #text-generation #causal-lm #rwkv #en #zh #fr #es #de #pt #ru #it #ja #ko #vi #ar #dataset-cerebras/SlimPajama-627B #dataset-EleutherAI/pile #dataset-bigcode/starcoderdata #dataset-oscar-corpus/OSCAR-2301 #license-apache-2.0 #has_space #region-us \n# RWKV-6 World\n\nUse rwkv pip package 0.8.24+ for RWKV-6 inference: URL (pipeline = PIPELINE(model, \"rwkv_vocab_v20230424\") for rwkv-world models)\n\nTrained on same world-2 dataset as URL"
] | [
-0.10198764503002167,
0.134396493434906,
-0.0039993892423808575,
0.12186887115240097,
0.04101612791419029,
0.0358707495033741,
0.0438925139605999,
0.09894493222236633,
0.09481742233037949,
0.05652429908514023,
0.17796245217323303,
0.05362986400723457,
0.06405408680438995,
0.08354879170656204,
0.01968694105744362,
-0.17931616306304932,
0.015256350859999657,
-0.013928848318755627,
0.004437324590981007,
0.1274155080318451,
0.05558359622955322,
-0.07153119891881943,
0.11704416573047638,
-0.022560475394129753,
-0.05059131234884262,
0.03347720205783844,
-0.02023632451891899,
-0.04264632984995842,
0.056917618960142136,
-0.0015245153335854411,
0.08625653386116028,
0.04394572600722313,
0.1006535068154335,
-0.07303279638290405,
0.014540844596922398,
0.010598777793347836,
-0.05878305435180664,
0.07793852686882019,
0.04278099164366722,
-0.07655008882284164,
0.0869414210319519,
-0.09223198890686035,
-0.03380700200796127,
-0.0007117612403817475,
-0.06894167512655258,
-0.18299102783203125,
-0.05983312055468559,
0.10567200928926468,
-0.07426778972148895,
0.09136748313903809,
-0.0001249987108167261,
0.19966144859790802,
-0.1412222683429718,
0.06430073827505112,
0.22936658561229706,
-0.24761494994163513,
-0.01370574627071619,
0.16142310202121735,
-0.006397494114935398,
0.08197153359651566,
-0.058652110397815704,
-0.022531820461153984,
0.04437001794576645,
0.054635342210531235,
-0.034824613481760025,
-0.01783115416765213,
-0.10742942988872528,
0.04795685410499573,
-0.1197013109922409,
-0.03567172586917877,
0.31506916880607605,
0.04369660094380379,
0.03355100750923157,
0.07035606354475021,
-0.09948211908340454,
-0.18986330926418304,
0.02343570627272129,
0.0715581476688385,
-0.030894014984369278,
-0.05937228351831436,
0.025080014020204544,
0.06720080226659775,
-0.10749101638793945,
-0.10641302913427353,
-0.10791550576686859,
0.205582395195961,
0.03228910639882088,
0.13646116852760315,
-0.13311246037483215,
0.12826408445835114,
-0.011591983959078789,
-0.1668221801519394,
-0.014230352826416492,
-0.12659548223018646,
0.03567340970039368,
0.10800473392009735,
0.0037957464810460806,
0.16306249797344208,
0.07870513200759888,
0.046187624335289,
-0.0017160644056275487,
0.024695754051208496,
0.12543624639511108,
0.09906427562236786,
0.019624123349785805,
0.06675787270069122,
-0.20032744109630585,
-0.020954685285687447,
0.019245881587266922,
-0.004210036247968674,
0.033803511410951614,
-0.01504488755017519,
-0.05228250101208687,
-0.10113687068223953,
-0.029792770743370056,
0.008649853989481926,
-0.0020315847359597683,
0.03574221208691597,
-0.07602871209383011,
-0.030819328501820564,
0.0828046053647995,
-0.06353409588336945,
-0.01714874431490898,
0.06658150255680084,
-0.13557009398937225,
0.12304764240980148,
0.0376000702381134,
-0.012609203346073627,
0.004108978435397148,
0.022907759994268417,
-0.08213596045970917,
0.023551782593131065,
-0.041634153574705124,
-0.05851851403713226,
0.0535094179213047,
-0.11169231683015823,
0.0593450628221035,
-0.13531604409217834,
-0.12741674482822418,
-0.0355832614004612,
-0.0017904830165207386,
-0.07327552139759064,
0.026522118598222733,
-0.03458676114678383,
-0.09253080189228058,
0.022977231070399284,
0.0014074015198275447,
0.05941058695316315,
-0.08780219405889511,
0.09109801054000854,
-0.0778837576508522,
0.11909175664186478,
-0.12491215020418167,
0.02346498891711235,
-0.09456456452608109,
-0.018536841496825218,
-0.0464283712208271,
0.04643603041768074,
-0.12594962120056152,
-0.03482339158654213,
-0.09541402757167816,
-0.11676955223083496,
-0.0691244825720787,
-0.0007651217165403068,
0.05595941096544266,
0.15222424268722534,
-0.2030002772808075,
-0.005023923236876726,
0.15783476829528809,
-0.04206681624054909,
-0.1043100357055664,
0.06482728570699692,
-0.014820071868598461,
0.02273719757795334,
0.004806024022400379,
0.1810208559036255,
0.06621819734573364,
-0.1041051521897316,
-0.10430944710969925,
0.07037776708602905,
-0.03217686712741852,
-0.09574596583843231,
0.13023221492767334,
0.07474348694086075,
-0.06405381113290787,
0.04990773648023605,
-0.04449307546019554,
-0.007453117519617081,
-0.0562308244407177,
-0.06965326517820358,
-0.03379136323928833,
-0.05645817890763283,
0.017552493140101433,
-0.011183192022144794,
0.03513599559664726,
-0.031187651678919792,
-0.04291491210460663,
-0.1352790892124176,
0.17739714682102203,
0.042008813470602036,
-0.012437681667506695,
-0.06754713505506516,
0.09898090362548828,
-0.09865079075098038,
-0.016502391546964645,
-0.11396782100200653,
0.06077006459236145,
-0.006642011925578117,
0.12050173431634903,
0.008981132879853249,
0.08310414105653763,
0.07648226618766785,
-0.009371057152748108,
0.03696294128894806,
-0.09177346527576447,
-0.01909487694501877,
-0.0035926271229982376,
-0.028813224285840988,
-0.1457129716873169,
-0.010388309136033058,
-0.04979325830936432,
-0.0018421780550852418,
-0.08508631587028503,
-0.020285282284021378,
0.04829202592372894,
0.09662789106369019,
0.002083786064758897,
0.04899520054459572,
0.04583042114973068,
-0.009008507244288921,
-0.008175312541425228,
-0.008144467137753963,
0.0016237683594226837,
-0.06869238615036011,
-0.08642225712537766,
0.03234328329563141,
0.007656577974557877,
0.1604517102241516,
0.11239208281040192,
-0.00482011865824461,
0.01147102564573288,
0.07847071439027786,
-0.006923474837094545,
0.015004975721240044,
-0.027350308373570442,
0.049614496529102325,
0.07359734922647476,
0.009058821015059948,
0.03131909668445587,
-0.0981542319059372,
-0.015388715080916882,
0.04531314969062805,
-0.09334121644496918,
0.008022719994187355,
0.22729100286960602,
0.04758216068148613,
-0.11182104051113129,
0.14087845385074615,
0.1301371306180954,
-0.053983211517333984,
0.15624190866947174,
-0.005075694061815739,
-0.051446229219436646,
-0.04486927390098572,
0.00470083300024271,
-0.05819518491625786,
0.1591119021177292,
-0.04398001730442047,
0.0568627268075943,
0.09800701588392258,
0.024140484631061554,
0.016000520437955856,
-0.11777946352958679,
-0.10622411221265793,
-0.05735583230853081,
-0.020455043762922287,
-0.1426493227481842,
0.09182403236627579,
0.013506430201232433,
0.14790330827236176,
-0.04347274452447891,
-0.08915796130895615,
0.014167693443596363,
-0.03343481570482254,
-0.03274254873394966,
0.20269182324409485,
-0.08931022137403488,
-0.14727982878684998,
-0.09492474794387817,
0.005397099535912275,
-0.07750414311885834,
-0.02998954802751541,
0.030188102275133133,
-0.1593453288078308,
0.004766549449414015,
-0.013778076507151127,
0.04539891332387924,
-0.04087282344698906,
-0.005728765856474638,
-0.04373962804675102,
0.02305169217288494,
-0.0045679728500545025,
-0.09451045840978622,
-0.02923988550901413,
-0.07691571116447449,
-0.0006180745549499989,
0.1608995795249939,
-0.028990792110562325,
0.06551709026098251,
0.07170535624027252,
0.0049129207618534565,
0.03295312076807022,
0.04417358338832855,
0.18734391033649445,
-0.025289833545684814,
0.009289765730500221,
0.16534753143787384,
0.030912775546312332,
0.05358238145709038,
0.10764264315366745,
0.06362602114677429,
-0.03155270218849182,
-0.01108530256897211,
-0.06685091555118561,
-0.07200314849615097,
-0.22502723336219788,
-0.1351807713508606,
-0.04682653024792671,
0.03707096725702286,
0.015199568122625351,
0.03734404966235161,
-0.14619886875152588,
0.11790183186531067,
-0.0030238910112529993,
-0.01302123349159956,
-0.09592288732528687,
0.011429395526647568,
-0.04458649083971977,
-0.0024964173790067434,
0.07693085819482803,
-0.09139604866504669,
-0.044796109199523926,
0.056119654327631,
0.12276668846607208,
0.13486890494823456,
-0.06900151073932648,
-0.027823468670248985,
0.09413374960422516,
0.2483200579881668,
0.10618734359741211,
0.10326313227415085,
0.025228429585695267,
-0.030453728511929512,
-0.007687960751354694,
-0.08200464397668839,
-0.06824629008769989,
0.06877420097589493,
-0.08502400666475296,
0.04191507026553154,
-0.019157424569129944,
0.07914700359106064,
0.02656104601919651,
0.06204582750797272,
0.08482341468334198,
-0.24881820380687714,
-0.010549751110374928,
0.04117750748991966,
0.11900551617145538,
-0.02486600913107395,
0.0752619206905365,
0.0926140695810318,
-0.06634169071912766,
0.032070476561784744,
0.03976534679532051,
0.051013100892305374,
-0.0446789525449276,
0.006619176361709833,
-0.0319027379155159,
0.04839000478386879,
-0.00653535732999444,
0.13770724833011627,
-0.21492986381053925,
0.21907901763916016,
0.0018973114201799035,
-0.04505806416273117,
-0.0811794325709343,
-0.08346660435199738,
-0.021943973377346992,
0.04319146275520325,
0.19759081304073334,
0.05523398518562317,
-0.022462403401732445,
-0.11417112499475479,
-0.11769840866327286,
0.08969444781541824,
0.02629101276397705,
-0.02608315646648407,
0.023918893188238144,
0.06314001977443695,
0.01050363015383482,
-0.03136097639799118,
-0.05060332641005516,
-0.055368516594171524,
-0.05304588004946709,
-0.04004225134849548,
0.049856092780828476,
-0.09628047794103622,
-0.04573298618197441,
-0.053647223860025406,
-0.0705784410238266,
-0.057233039289712906,
-0.10914235562086105,
-0.07915032655000687,
-0.07416574656963348,
0.04246247932314873,
0.15016217529773712,
-0.15477709472179413,
-0.04266580939292908,
-0.01745486631989479,
-0.06814684718847275,
0.02376934327185154,
-0.1131729856133461,
0.11923490464687347,
-0.0984867736697197,
-0.13679665327072144,
-0.006115777418017387,
0.0956205278635025,
0.06145184859633446,
0.0787203311920166,
-0.060456935316324234,
0.00879709143191576,
-0.14054660499095917,
-0.11993928253650665,
0.03331809118390083,
0.10481862723827362,
0.023231718689203262,
0.13314807415008545,
-0.010063476860523224,
-0.08198440819978714,
0.01852833665907383,
0.002238529734313488,
0.17633481323719025,
0.19991004467010498,
-0.1216922253370285,
0.11058197170495987,
0.1436792016029358,
0.008275463245809078,
-0.28766298294067383,
-0.059647414833307266,
-0.010349075309932232,
0.04178084060549736,
-0.03157006576657295,
-0.15287625789642334,
0.1331603080034256,
0.06395809352397919,
-0.04636051505804062,
0.10858345776796341,
-0.32216885685920715,
-0.031188463792204857,
0.10257884860038757,
-0.00043251787428744137,
0.23023557662963867,
-0.006833684165030718,
-0.009190362878143787,
-0.0066284434869885445,
-0.18392902612686157,
0.11208843439817429,
-0.09551602602005005,
0.10285406559705734,
-0.10042901337146759,
0.1353607028722763,
0.0053552016615867615,
-0.060476988554000854,
0.12753529846668243,
0.05057361721992493,
0.05947563797235489,
-0.04049111530184746,
-0.07094059139490128,
0.006908590439707041,
-0.0044838665053248405,
0.12512902915477753,
-0.07732071727514267,
0.09044956415891647,
-0.17973332107067108,
0.029660290107131004,
-0.1411624252796173,
0.07625780254602432,
-0.024447474628686905,
-0.05109817534685135,
-0.07213421165943146,
0.02337278239428997,
-0.010405820794403553,
-0.016651060432195663,
0.11148898303508759,
0.04398992657661438,
-0.03186232969164848,
0.062330447137355804,
0.03203374892473221,
0.02383686974644661,
-0.017211677506566048,
-0.04717095568776131,
0.0027140090242028236,
0.1153946965932846,
-0.22174091637134552,
-0.023949570953845978,
0.14130927622318268,
0.03910532593727112,
0.001972038997337222,
0.05025646463036537,
-0.12363149970769882,
0.017237579450011253,
0.1055712103843689,
-0.20310285687446594,
-0.02445795200765133,
-0.007325370796024799,
0.09842897206544876,
0.014886671677231789,
0.05568094924092293,
0.11943237483501434,
-0.07620741426944733,
-0.033550623804330826,
-0.04319781810045242,
0.06306415051221848,
-0.06572221964597702,
0.14432445168495178,
0.15318728983402252,
0.04448942095041275,
-0.16801856458187103,
0.11752525717020035,
-0.030478550121188164,
0.07002675533294678,
0.0168721042573452,
0.0457121767103672,
-0.12530836462974548,
-0.08982407301664352,
-0.00005648188016493805,
0.09223346412181854,
-0.10389029234647751,
-0.06687301397323608,
-0.02995547465980053,
-0.07017727941274643,
0.022245332598686218,
0.009210994467139244,
0.0961868017911911,
0.005281046032905579,
0.006209570448845625,
-0.11331842094659805,
0.014886277727782726,
0.0759105235338211,
0.036596547812223434,
-0.015100382268428802,
-0.17972716689109802,
-0.10161524266004562,
-0.03194620832800865,
0.14487825334072113,
-0.046517036855220795,
0.03887437656521797,
-0.11852230131626129,
0.05131940171122551,
-0.08104798942804337,
0.03382115438580513,
0.007786818314343691,
0.0035278841387480497,
-0.07378444820642471,
-0.07973132282495499,
-0.05033239722251892,
0.03441416844725609,
-0.11954798549413681,
0.02961513213813305,
-0.02512168325483799,
0.05148134380578995,
-0.07048726081848145,
-0.020391004160046577,
0.06265044957399368,
-0.022394100204110146,
0.08303345739841461,
0.1502152532339096,
-0.02040158025920391,
0.06919237226247787,
-0.10739713162183762,
-0.003880695439875126,
0.06311747431755066,
0.06407816708087921,
0.10126099735498428,
0.019889626652002335,
0.043289922177791595,
-0.02885420247912407,
-0.004178911913186312,
0.005524289328604937,
0.05581332743167877,
-0.09656093269586563,
0.027785658836364746,
-0.08420638740062714,
-0.0021175583824515343,
-0.08217377215623856,
0.01885197125375271,
0.07230190932750702,
0.12155473977327347,
0.10389825701713562,
-0.001474734046496451,
0.06851819157600403,
-0.1102752760052681,
0.005366429686546326,
-0.02008955180644989,
-0.10225851833820343,
-0.09803429991006851,
0.01153318677097559,
0.06947200000286102,
-0.03267707675695419,
0.10691177099943161,
0.023193132132291794,
-0.15083535015583038,
-0.06771574169397354,
0.10211110860109329,
-0.018426865339279175,
-0.015306598506867886,
0.07437069714069366,
0.009396053850650787,
0.008035403676331043,
-0.02784757688641548,
-0.007104096002876759,
0.057664722204208374,
0.04956541955471039,
0.08720938861370087,
0.1279480904340744,
0.14051780104637146,
0.04138753563165665,
0.06353536248207092,
-0.05769640579819679,
-0.017495041713118553,
-0.0896393284201622,
-0.03823002055287361,
-0.01127584557980299,
-0.022260552272200584,
0.06365407258272171,
0.08585949242115021,
-0.02413187548518181,
-0.01799014024436474,
-0.037815943360328674,
-0.038068272173404694,
-0.12038217484951019,
-0.1683003455400467,
-0.0767306461930275,
-0.09353643655776978,
0.014837696217000484,
-0.08840473741292953,
-0.01640874519944191,
0.1404656171798706,
0.06905020028352737,
-0.023098375648260117,
0.08837203681468964,
0.04785323515534401,
-0.03320492431521416,
0.012693683616816998,
0.048268042504787445,
-0.03874942660331726,
0.04378749057650566,
0.037177518010139465,
0.08646156638860703,
-0.006416235119104385,
0.007995027117431164,
0.03581448271870613,
-0.039611514657735825,
-0.011077357456088066,
-0.040646057575941086,
-0.07207565009593964,
-0.08167238533496857,
0.029395250603556633,
0.01328775193542242,
0.12940727174282074,
0.05639248341321945,
0.018751367926597595,
-0.026019806042313576,
0.2201666235923767,
-0.047716643661260605,
-0.037700947374105453,
-0.038624491542577744,
0.14822472631931305,
-0.06050926819443703,
0.07636474817991257,
0.003561256919056177,
-0.11215139180421829,
-0.05346117913722992,
0.19350098073482513,
0.24853473901748657,
-0.08150506764650345,
0.03540477156639099,
0.003261161968111992,
0.012451549991965294,
-0.05375659093260765,
0.049072109162807465,
0.1295383870601654,
0.173532634973526,
-0.11845255643129349,
0.050688084214925766,
-0.12839540839195251,
0.00824576523154974,
0.013712442480027676,
-0.003784401575103402,
0.03142211586236954,
-0.08849413692951202,
-0.09992286562919617,
0.05248336121439934,
-0.22851751744747162,
-0.04698118939995766,
0.00565172778442502,
-0.1339595764875412,
-0.12265243381261826,
-0.04889588803052902,
0.10612212121486664,
0.11209498345851898,
0.0686550885438919,
-0.10715730488300323,
-0.026930760592222214,
0.01141499262303114,
0.021070649847388268,
-0.17454765737056732,
-0.059221651405096054,
0.05131412670016289,
-0.042578231543302536,
0.09892606735229492,
-0.01362665556371212,
0.14619576930999756,
0.08238568156957626,
-0.01059892401099205,
-0.12775744497776031,
0.06666862219572067,
0.009195017628371716,
0.013320605270564556,
-0.06622839719057083,
0.03634924069046974,
0.03793120011687279,
-0.032936785370111465,
0.08806861191987991,
-0.018256299197673798,
-0.024501077830791473,
0.11952138692140579,
0.061725813895463943,
-0.10052365809679031,
0.07301512360572815,
-0.023175768554210663,
0.09099467098712921,
0.08644063770771027,
-0.08317036926746368,
-0.021025625988841057,
-0.07375675439834595,
0.052715327590703964,
0.030410271137952805,
-0.08474679291248322,
0.053644515573978424,
-0.07809577882289886,
-0.014559127390384674,
0.026846496388316154,
0.0600350983440876,
-0.1098705306649208,
0.015194644220173359,
-0.09355238080024719,
-0.05106939375400543,
0.011364475823938847,
0.04201903194189072,
0.014332881197333336,
0.07502042502164841,
-0.007118704728782177,
0.02478877827525139,
0.005579323973506689,
0.02929113805294037,
-0.11922947317361832,
-0.11372426152229309
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-large-squad-model2
This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "roberta-large", "model-index": [{"name": "roberta-large-squad-model2", "results": []}]} | question-answering | varun-v-rao/roberta-large-squad-model2 | [
"transformers",
"tensorboard",
"safetensors",
"roberta",
"question-answering",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:roberta-large",
"license:mit",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:48:02+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us
|
# roberta-large-squad-model2
This model is a fine-tuned version of roberta-large on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 16
- seed: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# roberta-large-squad-model2\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 64\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n",
"# roberta-large-squad-model2\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 64\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
70,
34,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #roberta #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-roberta-large #license-mit #endpoints_compatible #region-us \n# roberta-large-squad-model2\n\nThis model is a fine-tuned version of roberta-large on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 64\n- eval_batch_size: 16\n- seed: 64\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.06344472616910934,
0.06228947266936302,
-0.0021089499350637197,
0.07660242915153503,
0.17077182233333588,
0.020322222262620926,
0.13673928380012512,
0.08526064455509186,
-0.12469545751810074,
0.055366531014442444,
0.07402071356773376,
0.07141181826591492,
0.02139977365732193,
0.11216602474451065,
-0.03865303844213486,
-0.23121628165245056,
0.003767475252971053,
-0.013273444958031178,
-0.10615558922290802,
0.1023726612329483,
0.09835109859704971,
-0.12501394748687744,
0.058810386806726456,
-0.009426718577742577,
-0.19642037153244019,
0.04434239864349365,
-0.004293587990105152,
-0.027642128989100456,
0.10908611118793488,
0.015791721642017365,
0.13443006575107574,
0.007454928010702133,
0.1397504210472107,
-0.2216459959745407,
0.01474433857947588,
0.08234050869941711,
0.027099788188934326,
0.06341218948364258,
0.03792911767959595,
-0.003499450394883752,
0.10914256423711777,
-0.14432962238788605,
0.0967516079545021,
0.024075739085674286,
-0.08356977254152298,
-0.17982885241508484,
-0.08268685638904572,
0.03841851279139519,
0.09223516285419464,
0.09652590751647949,
-0.012886996380984783,
0.16728763282299042,
-0.12788009643554688,
0.08122087270021439,
0.20232872664928436,
-0.27856865525245667,
-0.09296713769435883,
0.10057390481233597,
0.055453456938266754,
0.09488140046596527,
-0.1188417598605156,
-0.012099008075892925,
0.05039553716778755,
0.025736890733242035,
0.10061950981616974,
-0.028833521530032158,
-0.08524716645479202,
0.013242690823972225,
-0.15105657279491425,
0.01782876066863537,
0.14057865738868713,
0.05633316561579704,
-0.030638787895441055,
-0.055002957582473755,
-0.04785292223095894,
-0.03682611882686615,
-0.026596706360578537,
-0.05647910758852959,
0.05080963298678398,
-0.054500579833984375,
-0.11161831766366959,
-0.03998963162302971,
-0.08576375246047974,
-0.07384034246206284,
-0.014634950086474419,
0.11258997768163681,
0.04959249496459961,
0.011822294443845749,
-0.05344562605023384,
0.0872390940785408,
-0.02259858325123787,
-0.09593180567026138,
0.0011363317025825381,
0.004171022679656744,
-0.08703642338514328,
-0.07400012761354446,
-0.039681464433670044,
-0.04634000360965729,
0.04020627215504646,
0.1678137183189392,
-0.07203905284404755,
0.05199877917766571,
0.019341258332133293,
0.008538875728845596,
-0.03895144164562225,
0.12907013297080994,
-0.06076463684439659,
-0.06127849221229553,
0.005099068861454725,
0.07925146073102951,
0.009559870697557926,
0.015940984711050987,
-0.08017612993717194,
0.00223181932233274,
0.09061725437641144,
0.05390963703393936,
-0.06692296266555786,
0.03396935015916824,
-0.008721220307052135,
-0.006292812991887331,
-0.03272349759936333,
-0.10826822370290756,
0.03921214118599892,
-0.0014839995419606566,
-0.05482575669884682,
-0.026150327175855637,
0.004623659886419773,
0.015147166326642036,
0.0089099807664752,
0.1102072075009346,
-0.09218931943178177,
0.012073369696736336,
-0.10510839521884918,
-0.09631230682134628,
0.0019863767083734274,
-0.07277090102434158,
0.0019817089196294546,
-0.0798780620098114,
-0.16463159024715424,
-0.04732213914394379,
0.03434678539633751,
-0.04197361692786217,
-0.017130406573414803,
-0.05121482163667679,
-0.07678547501564026,
-0.015718162059783936,
0.0037527827080339193,
0.15183745324611664,
-0.04115424305200577,
0.07808509469032288,
0.020526958629488945,
0.04285135865211487,
0.003298016730695963,
0.020013902336359024,
-0.0913541316986084,
0.019460322335362434,
-0.1532919853925705,
0.03928549215197563,
-0.07783462107181549,
0.05935061722993851,
-0.1053600162267685,
-0.10062520205974579,
-0.016157139092683792,
0.005832592025399208,
0.05388571694493294,
0.08715670555830002,
-0.15936630964279175,
-0.0347476452589035,
0.19498994946479797,
-0.08296210318803787,
-0.0818096324801445,
0.10947137326002121,
-0.06554947048425674,
0.05301147326827049,
0.07125497609376907,
0.16727948188781738,
0.06376494467258453,
-0.1156897023320198,
-0.00029345948132686317,
-0.0255831778049469,
0.04144095629453659,
-0.005248343106359243,
0.05068909004330635,
0.00037712842458859086,
0.018187543377280235,
0.00045140538713894784,
-0.06290998309850693,
0.016574125736951828,
-0.10681357234716415,
-0.07297012209892273,
-0.04300767183303833,
-0.10332468152046204,
0.001358484267257154,
0.047821108251810074,
0.04523187130689621,
-0.09382298588752747,
-0.09266696125268936,
0.1338532269001007,
0.1140727698802948,
-0.05828630179166794,
0.0003833726223092526,
-0.07393644750118256,
0.03328276798129082,
-0.06039474159479141,
-0.027794763445854187,
-0.1809377819299698,
-0.11488444358110428,
0.0028988744597882032,
-0.01566067524254322,
0.04416901618242264,
0.05260440707206726,
0.07144425809383392,
0.06687889993190765,
-0.06107182800769806,
-0.005188618320971727,
-0.06468162685632706,
0.0051743704825639725,
-0.09351665526628494,
-0.21602098643779755,
-0.026886332780122757,
-0.027998924255371094,
0.1647166907787323,
-0.271613746881485,
0.02864588424563408,
-0.04900897294282913,
0.11845920234918594,
0.034939106553792953,
-0.03518163040280342,
-0.02926359511911869,
0.06013496592640877,
-0.00801540445536375,
-0.07457823306322098,
0.04195568710565567,
-0.014762229286134243,
-0.06424012780189514,
-0.08266469836235046,
-0.1424941122531891,
0.08166997879743576,
0.09002035856246948,
-0.005920404568314552,
-0.10426627099514008,
0.0072043524123728275,
-0.06917372345924377,
-0.03160351142287254,
-0.0876261442899704,
0.011171103455126286,
0.12273837625980377,
-0.01102255005389452,
0.12145427614450455,
-0.05882471799850464,
-0.054272279143333435,
-0.003066091798245907,
-0.027606230229139328,
0.030184712260961533,
0.09691274166107178,
0.11531953513622284,
-0.13200442492961884,
0.09897057712078094,
0.09569846838712692,
-0.09394823759794235,
0.1443803459405899,
-0.04677220806479454,
-0.07223958522081375,
-0.026755599305033684,
0.00946035422384739,
0.0018696235492825508,
0.13361871242523193,
-0.08327852934598923,
0.003135216888040304,
0.008766155689954758,
0.00985513161867857,
0.02845541387796402,
-0.18309763073921204,
-0.03193940967321396,
0.009533979929983616,
-0.029947010800242424,
0.009459288790822029,
-0.014436832629144192,
0.030535798519849777,
0.10145583748817444,
0.009191620163619518,
-0.026206418871879578,
0.0034084711223840714,
-0.006319792941212654,
-0.08073054254055023,
0.2049429714679718,
-0.07536155730485916,
-0.08340151607990265,
-0.09405628591775894,
0.025692837312817574,
-0.03367834538221359,
-0.025799313560128212,
0.02853519096970558,
-0.11366122961044312,
-0.04428255930542946,
-0.0927112028002739,
0.016137035563588142,
-0.012034346349537373,
-0.004490980412811041,
0.018614046275615692,
0.031174061819911003,
0.08078493922948837,
-0.12609001994132996,
0.011651860550045967,
-0.059484582394361496,
-0.1370740681886673,
0.018570035696029663,
0.05338432267308235,
0.12673839926719666,
0.11693942546844482,
-0.02142917737364769,
0.008328611962497234,
-0.032966017723083496,
0.21100255846977234,
-0.06567452102899551,
-0.003721555694937706,
0.10775283724069595,
0.013816576451063156,
0.0357683002948761,
0.13065963983535767,
0.04831495136022568,
-0.10617968440055847,
0.04423583671450615,
0.09676380455493927,
-0.030380047857761383,
-0.2365509271621704,
-0.03166224807500839,
-0.03850705176591873,
-0.07957890629768372,
0.06075245887041092,
0.0412481427192688,
0.007793231401592493,
0.05122930929064751,
0.007802349515259266,
0.037862230092287064,
-0.013183696195483208,
0.08150725066661835,
0.09082768857479095,
0.03703876584768295,
0.12314113229513168,
-0.048679597675800323,
-0.06561989337205887,
0.04694807156920433,
-0.02046516351401806,
0.3105227053165436,
-0.002677225274965167,
0.04636829346418381,
0.08299535512924194,
0.12241417914628983,
-0.01649874821305275,
0.028973275795578957,
0.008587315678596497,
-0.03777674585580826,
0.0015739663504064083,
-0.0525767058134079,
0.006228838115930557,
0.02003958448767662,
-0.04246329516172409,
0.060382548719644547,
-0.06900352239608765,
0.04710008203983307,
0.0515938326716423,
0.22520920634269714,
0.027612581849098206,
-0.27861055731773376,
-0.07436666637659073,
0.016168972477316856,
-0.035849496722221375,
-0.01845337264239788,
0.017028067260980606,
0.1406867951154709,
-0.10551134496927261,
0.028121205046772957,
-0.05893617123365402,
0.08374663442373276,
-0.009090092033147812,
0.011628348380327225,
0.03255018591880798,
0.15885747969150543,
-0.017370015382766724,
0.07752962410449982,
-0.22338052093982697,
0.2398461401462555,
0.010478179901838303,
0.12641990184783936,
-0.033607419580221176,
0.0015249913558363914,
0.017322571948170662,
0.07371990382671356,
0.08343960344791412,
-0.006463396362960339,
-0.0372348353266716,
-0.16786101460456848,
-0.031532492488622665,
0.06143888086080551,
0.12392163276672363,
-0.01832943595945835,
0.10633884370326996,
-0.049878284335136414,
0.008030041120946407,
0.06340433657169342,
-0.06807172298431396,
-0.1660253405570984,
-0.10226871818304062,
-0.03642817586660385,
-0.002026477362960577,
-0.07864846289157867,
-0.09056282043457031,
-0.0901772528886795,
-0.05475902184844017,
0.16439484059810638,
0.0058410195633769035,
-0.009308168664574623,
-0.12090665847063065,
0.12223928421735764,
0.08893662691116333,
-0.05814532935619354,
0.006613066419959068,
0.018737874925136566,
0.08790528029203415,
0.034407150000333786,
-0.05367375537753105,
0.06911338120698929,
-0.07664889097213745,
-0.15434488654136658,
-0.06263657659292221,
0.10299515724182129,
0.06698894500732422,
0.04437147080898285,
0.000586859998293221,
0.027600150555372238,
0.016337988898158073,
-0.09591133892536163,
0.003157916711643338,
0.06781347095966339,
0.06214097887277603,
0.06185517832636833,
-0.08923904597759247,
-0.013875281438231468,
-0.035931576043367386,
-0.02886967919766903,
0.12726062536239624,
0.22624826431274414,
-0.08433844894170761,
0.03241317719221115,
0.06168666481971741,
-0.07425183057785034,
-0.18314670026302338,
0.10344753414392471,
0.07580051571130753,
0.004131737630814314,
0.07278335839509964,
-0.14854785799980164,
0.16622158885002136,
0.1255762279033661,
-0.013485315255820751,
0.036059826612472534,
-0.3101952075958252,
-0.1288924664258957,
0.07025483250617981,
0.15505053102970123,
0.07420777529478073,
-0.16306957602500916,
-0.0174397099763155,
-0.022650200873613358,
-0.12367166578769684,
0.10174308717250824,
-0.16908486187458038,
0.1004737839102745,
0.0034165552351623774,
0.07618556171655655,
0.010406126268208027,
-0.05048104375600815,
0.1200452595949173,
0.01570403017103672,
0.1107797846198082,
-0.056829508394002914,
-0.010448573157191277,
0.11463193595409393,
-0.046994224190711975,
0.014394174329936504,
-0.046139851212501526,
0.058070871978998184,
-0.09124620258808136,
-0.03232662007212639,
-0.07559797167778015,
0.05361398309469223,
-0.055136580020189285,
-0.06633006036281586,
-0.06200988218188286,
0.06654389202594757,
0.044455043971538544,
-0.022264575585722923,
0.05907374247908592,
-0.007131391204893589,
0.1410309076309204,
0.03368755429983139,
0.10485969483852386,
0.0012817827519029379,
-0.050095804035663605,
0.004903522785753012,
-0.01698969304561615,
0.05247226729989052,
-0.12724067270755768,
0.02319439873099327,
0.11580227315425873,
0.04367317631840706,
0.14442865550518036,
0.054107800126075745,
-0.048572562634944916,
0.01673002727329731,
0.05667433887720108,
-0.10491533577442169,
-0.19004467129707336,
0.030114632099866867,
-0.06709802150726318,
-0.13297945261001587,
0.047182220965623856,
0.09568195790052414,
-0.05917109549045563,
-0.006956287194043398,
-0.021311817690730095,
0.018759537488222122,
-0.03997587785124779,
0.1787191778421402,
0.061296842992305756,
0.05693472921848297,
-0.07816202938556671,
0.07778317481279373,
0.06193007528781891,
-0.04247662425041199,
0.024482131004333496,
0.07950887084007263,
-0.08221330493688583,
-0.020650947466492653,
0.08714950084686279,
0.22183352708816528,
-0.04849400743842125,
-0.038234446197748184,
-0.1164676696062088,
-0.10500660538673401,
0.03319115936756134,
0.1519189029932022,
0.05952242761850357,
-0.03562641888856888,
-0.016403118148446083,
0.051693592220544815,
-0.16095121204853058,
0.0941125899553299,
0.03750927746295929,
0.07763577252626419,
-0.13151726126670837,
0.11210808157920837,
0.0097389817237854,
0.00586725166067481,
-0.016748540103435516,
0.03862006962299347,
-0.1308550089597702,
-0.013056885451078415,
-0.1335318684577942,
-0.025227311998605728,
-0.02141067013144493,
0.0044710650108754635,
0.0015269488794729114,
-0.05547350272536278,
-0.06793292611837387,
0.03931165114045143,
-0.07191583514213562,
-0.039561256766319275,
0.030004585161805153,
0.05579978600144386,
-0.1473122239112854,
-0.003177085891366005,
0.017836563289165497,
-0.07441620528697968,
0.060331929475069046,
0.03247760236263275,
0.03975437954068184,
0.0450110025703907,
-0.1799343079328537,
-0.010722202248871326,
0.03261498361825943,
0.01070721447467804,
0.07202073931694031,
-0.07429496198892593,
-0.009045866318047047,
-0.014753890223801136,
0.08879049122333527,
0.026273969560861588,
0.03284464403986931,
-0.10833677649497986,
0.03229636326432228,
-0.08434688299894333,
-0.06892414391040802,
-0.052689164876937866,
0.023302355781197548,
0.07593801617622375,
0.030285518616437912,
0.18154633045196533,
-0.10115576535463333,
0.035969410091638565,
-0.22172173857688904,
-0.03179103508591652,
-0.0007733523962087929,
-0.03489748388528824,
-0.08110993355512619,
-0.04074937850236893,
0.06560000032186508,
-0.06194436550140381,
0.12057231366634369,
0.011342487297952175,
0.07078732550144196,
0.04985331371426582,
-0.053046975284814835,
0.004373300354927778,
0.016742484644055367,
0.18585233390331268,
0.050796207040548325,
-0.0231753159314394,
0.051452599465847015,
0.0004412989073898643,
0.06676527112722397,
0.017953047528862953,
0.21224011480808258,
0.19679680466651917,
-0.06251112371683121,
0.04119717702269554,
0.07101651281118393,
-0.08096419274806976,
-0.10105419158935547,
0.13294149935245514,
-0.03269351273775101,
0.0633668527007103,
-0.04436763375997543,
0.17809562385082245,
0.15872542560100555,
-0.16675588488578796,
0.024639032781124115,
-0.07059107720851898,
-0.0965966060757637,
-0.11622244119644165,
-0.009591316804289818,
-0.1014774814248085,
-0.12236591428518295,
0.03881392627954483,
-0.1270437240600586,
0.010963737033307552,
0.10849057883024216,
0.02239695005118847,
0.012156459502875805,
0.17291635274887085,
0.0028217460494488478,
0.043838366866111755,
0.025268686935305595,
-0.0016311600338667631,
-0.016663549467921257,
-0.05102384462952614,
-0.0474492684006691,
0.03719155490398407,
-0.01966656744480133,
0.0614507794380188,
-0.05005890876054764,
-0.032097313553094864,
0.03288992866873741,
-0.026777416467666626,
-0.06665591895580292,
0.02819444052875042,
0.03537360951304436,
0.038020044565200806,
0.031248196959495544,
0.04612976685166359,
-0.0165055263787508,
-0.03413420170545578,
0.27307701110839844,
-0.0626501813530922,
-0.12281301617622375,
-0.12144487351179123,
0.2179994136095047,
0.02735479734838009,
0.0019300737185403705,
0.035165805369615555,
-0.11511211842298508,
0.022871581837534904,
0.18383796513080597,
0.18711180984973907,
-0.06670000404119492,
-0.006438910961151123,
-0.009590618312358856,
-0.021311847493052483,
-0.0646543949842453,
0.11462010443210602,
0.10604900866746902,
0.053059618920087814,
-0.0534856915473938,
-0.026180997490882874,
-0.01922851987183094,
-0.017603999003767967,
-0.06685598939657211,
0.03192206844687462,
0.056021548807621,
0.015358840115368366,
-0.03267879784107208,
0.08910319209098816,
-0.005275589879602194,
-0.1997523456811905,
0.07480952888727188,
-0.1528659462928772,
-0.16760151088237762,
-0.029231175780296326,
0.09650740027427673,
-0.03262856602668762,
0.05213601142168045,
-0.036427464336156845,
-0.02039390429854393,
0.07926687598228455,
-0.018830953165888786,
-0.05195162445306778,
-0.10228106379508972,
0.08613492548465729,
-0.07668287307024002,
0.21682099997997284,
-0.02701985463500023,
0.0631512850522995,
0.13134483993053436,
-0.0024844573345035315,
-0.09828098863363266,
0.07436630874872208,
0.05527131259441376,
-0.08600598573684692,
0.020906347781419754,
0.1365390121936798,
-0.051085904240608215,
0.100044384598732,
0.05631563067436218,
-0.14185547828674316,
0.013235678896307945,
-0.057974766939878464,
-0.04625288397073746,
-0.0882437452673912,
0.009873047471046448,
-0.08024885505437851,
0.1415095180273056,
0.20602738857269287,
-0.02472512610256672,
0.039406511932611465,
-0.07859916239976883,
0.0387626588344574,
0.06556160002946854,
0.10361623764038086,
-0.03705810382962227,
-0.22188390791416168,
0.03051612339913845,
0.005649283528327942,
0.0008915993385016918,
-0.29108381271362305,
-0.08237779885530472,
0.021655438467860222,
-0.03855894133448601,
-0.06904157251119614,
0.10301979631185532,
0.12421678006649017,
0.06125200539827347,
-0.045114677399396896,
-0.18058572709560394,
-0.05794084072113037,
0.16642990708351135,
-0.1422998607158661,
-0.06001778692007065
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-8.0bpw-h8-exl2 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T17:48:30+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
194,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06399306654930115,
-0.029307762160897255,
-0.004331388510763645,
0.03184039518237114,
0.018488328903913498,
0.03748767077922821,
0.10618676990270615,
0.07814439386129379,
0.07063937187194824,
0.07519446313381195,
-0.04871003329753876,
-0.02137892320752144,
0.07951480895280838,
0.10956592857837677,
0.007890605367720127,
-0.23595072329044342,
0.08250588178634644,
-0.03575211763381958,
-0.026412149891257286,
0.03893791511654854,
0.14763003587722778,
-0.05930296331644058,
0.05646873265504837,
-0.0034751244820654392,
-0.05727600306272507,
0.05131329223513603,
-0.026073675602674484,
-0.007598201744258404,
0.09508240222930908,
0.0628398060798645,
0.041950274258852005,
0.0221012644469738,
-0.018701795488595963,
-0.15748009085655212,
0.004922133404761553,
-0.017628159373998642,
-0.04106835648417473,
0.027759606018662453,
0.11616206914186478,
-0.045659661293029785,
0.11880303174257278,
-0.12680159509181976,
-0.018632367253303528,
0.08243369311094284,
-0.12379294633865356,
-0.10155918449163437,
-0.20841088891029358,
0.15503807365894318,
0.07979700714349747,
0.008151463232934475,
-0.04288718104362488,
0.07205401360988617,
-0.054492317140102386,
0.0433170422911644,
0.08753809332847595,
-0.2521143853664398,
-0.04418989270925522,
0.17082107067108154,
0.007189796306192875,
-0.04941372200846672,
-0.05787082388997078,
0.06996709108352661,
0.029280396178364754,
0.0067648859694600105,
-0.04790947958827019,
-0.04312778264284134,
0.2201330065727234,
-0.03288035839796066,
-0.1199050173163414,
-0.024701880291104317,
0.08322399854660034,
0.08842694014310837,
-0.04718442261219025,
-0.15207047760486603,
-0.011922257021069527,
-0.022255953401327133,
-0.04938272759318352,
-0.09886841475963593,
0.003574386006221175,
-0.02095438726246357,
0.07563887536525726,
-0.05726063251495361,
-0.04117260500788689,
0.0583573654294014,
-0.05984218046069145,
0.1629648655653,
-0.0014757480239495635,
-0.03302643075585365,
0.05737544223666191,
0.03662708401679993,
-0.2148275226354599,
-0.08567781001329422,
-0.04239440709352493,
-0.054734885692596436,
-0.1342768371105194,
-0.0431794635951519,
-0.04418531805276871,
-0.08217091858386993,
-0.007381834089756012,
0.23022499680519104,
-0.02492726407945156,
0.02804037742316723,
0.041754867881536484,
0.010314702056348324,
0.04419277235865593,
0.1185997873544693,
-0.05733427032828331,
-0.1965552270412445,
-0.02631615847349167,
-0.01892445981502533,
0.06395958364009857,
0.013402312994003296,
-0.02684073895215988,
-0.01519184559583664,
0.02383972890675068,
0.02890627458691597,
0.02422332763671875,
0.03168574720621109,
-0.07110860198736191,
-0.03037358447909355,
0.14692901074886322,
-0.10123810917139053,
0.04444355517625809,
-0.03552069887518883,
-0.03803017735481262,
0.07610528916120529,
-0.030228903517127037,
0.023446066305041313,
-0.01123465783894062,
0.14718079566955566,
-0.04342922940850258,
-0.02282293513417244,
-0.04051238298416138,
-0.05655231699347496,
0.056725140661001205,
-0.08175330609083176,
-0.06928885728120804,
-0.0921410471200943,
-0.07629581540822983,
-0.0781061053276062,
0.019695017486810684,
-0.04772234335541725,
-0.00727304769679904,
-0.00963415578007698,
-0.03471999987959862,
0.004572734236717224,
0.03700930252671242,
-0.03850525617599487,
-0.006236410234123468,
-0.05285993963479996,
-0.02465738169848919,
0.016344578936696053,
-0.07584261149168015,
0.020161736756563187,
-0.09869375079870224,
0.055850550532341,
-0.3218844532966614,
0.09342993050813675,
-0.11400728672742844,
0.009930809028446674,
-0.1314808875322342,
-0.0335058830678463,
0.016112782061100006,
0.005322059616446495,
0.013146123848855495,
0.11860170215368271,
-0.15911765396595,
-0.09210728108882904,
0.09710568189620972,
-0.13270564377307892,
-0.06123318895697594,
0.13335616886615753,
0.006097014062106609,
-0.007668835576623678,
0.0786266103386879,
0.1425439566373825,
0.14123733341693878,
-0.04838160052895546,
-0.13677071034908295,
-0.06624309718608856,
0.027206536382436752,
0.1544196456670761,
0.06648306548595428,
-0.0522027313709259,
-0.0182489063590765,
0.03598593920469284,
-0.031956933438777924,
-0.0026491705793887377,
-0.016834797337651253,
-0.05705594643950462,
-0.023978661745786667,
-0.002377238357439637,
0.096376933157444,
-0.015163181349635124,
0.002383065177127719,
-0.07838787883520126,
-0.09363235533237457,
0.039118777960538864,
0.0882512778043747,
0.041436195373535156,
0.01435884553939104,
-0.06161881238222122,
0.09994251281023026,
-0.03056234121322632,
0.022030595690011978,
-0.10238077491521835,
-0.15781597793102264,
0.03258318454027176,
-0.15215061604976654,
0.06616183370351791,
0.046265859156847,
0.06192070245742798,
0.027611956000328064,
-0.11285637319087982,
-0.06351099908351898,
-0.02237211912870407,
0.016440147534012794,
0.007992693223059177,
-0.12434909492731094,
-0.08924297988414764,
-0.04348140209913254,
0.22069425880908966,
-0.054984692484140396,
-0.007727441843599081,
0.06026673689484596,
0.2515117824077606,
0.018960142508149147,
-0.014410633593797684,
-0.05082327499985695,
0.07166947424411774,
0.00031130402931012213,
0.0197899267077446,
0.01994451880455017,
-0.04336968809366226,
-0.14637522399425507,
0.095527283847332,
-0.055198993533849716,
-0.02979397401213646,
0.06157594546675682,
0.054885584861040115,
-0.09651610255241394,
-0.011951263062655926,
-0.05258362367749214,
-0.029914764687418938,
0.006216008216142654,
-0.12069366872310638,
0.042220812290906906,
0.04080578312277794,
0.03252416476607323,
-0.0428742878139019,
-0.048637744039297104,
0.01096632331609726,
0.00038561708061024547,
-0.12152905762195587,
0.10360410064458847,
-0.08400224894285202,
-0.21976810693740845,
0.10374713689088821,
0.15290415287017822,
0.07434318959712982,
0.12372945249080658,
-0.013827105984091759,
-0.022371480241417885,
-0.1197407990694046,
0.0015785678988322616,
0.021130502223968506,
0.042604196816682816,
0.04024510458111763,
0.05408281460404396,
0.03447001054883003,
0.043456509709358215,
0.02863817848265171,
-0.02880154922604561,
0.04750415310263634,
-0.009588376618921757,
0.0040976181626319885,
0.12772449851036072,
0.10417608171701431,
0.02852170541882515,
0.08533679693937302,
0.03910043090581894,
-0.0404503159224987,
0.01288803294301033,
-0.02749737910926342,
-0.06698718667030334,
0.10686532407999039,
-0.10741190612316132,
-0.16609297692775726,
-0.17008981108665466,
-0.051623594015836716,
-0.18006007373332977,
-0.024816229939460754,
0.03015192784368992,
-0.03652867674827576,
-0.07785869389772415,
-0.07390360534191132,
0.13494938611984253,
0.03926362842321396,
-0.05240222439169884,
-0.05028913915157318,
-0.03344966843724251,
0.042253971099853516,
-0.10105009377002716,
-0.04346547648310661,
0.0018156386213377118,
-0.050230301916599274,
0.021803520619869232,
0.027263345196843147,
0.0939897820353508,
0.060186538845300674,
-0.00003084815762122162,
-0.02964363619685173,
-0.040875863283872604,
0.17850545048713684,
-0.1293039619922638,
0.11540871858596802,
0.04773532599210739,
-0.0027224498335272074,
0.06698734313249588,
0.16732339560985565,
0.05765581503510475,
-0.04658689722418785,
-0.011350571177899837,
0.06666777282953262,
-0.030819516628980637,
-0.19200317561626434,
-0.1769828051328659,
-0.06770068407058716,
0.007487337104976177,
-0.027668775990605354,
0.04515873268246651,
0.06839024275541306,
-0.0003817266842816025,
-0.09345950931310654,
-0.07025164365768433,
0.036232683807611465,
0.04945642128586769,
0.17080222070217133,
0.009348781779408455,
0.05902401730418205,
-0.06639035046100616,
-0.06039080023765564,
0.09840981662273407,
0.00460409838706255,
0.08121812343597412,
0.06772522628307343,
0.06868960708379745,
0.051509808748960495,
0.012179503217339516,
0.0003098213055636734,
0.04621569812297821,
0.015963001176714897,
-0.02683101035654545,
-0.011713831685483456,
-0.08623592555522919,
0.07211727648973465,
0.08716612309217453,
0.10278671234846115,
0.03722165524959564,
-0.051564089953899384,
-0.06990828365087509,
0.12818388640880585,
0.1740635186433792,
0.11096327751874924,
-0.12066809087991714,
-0.10149893164634705,
0.059367019683122635,
-0.012059872969985008,
0.020104901865124702,
-0.03295712172985077,
-0.051309432834386826,
-0.13940629363059998,
0.20302264392375946,
-0.018019717186689377,
0.06347235292196274,
0.0055109113454818726,
-0.0001949289726326242,
-0.02158956229686737,
0.06696530431509018,
0.017838791012763977,
0.050336673855781555,
-0.141648069024086,
0.14686071872711182,
0.010985979810357094,
0.021909072995185852,
-0.01216028444468975,
0.05184603109955788,
0.01702127605676651,
0.05610823258757591,
0.10946010798215866,
0.04561372473835945,
-0.09489621967077255,
-0.08675345033407211,
-0.04737729951739311,
-0.05841971933841705,
0.13532322645187378,
-0.0978664755821228,
0.08815576881170273,
-0.005218944977968931,
-0.04548734799027443,
-0.058141354471445084,
0.16146430373191833,
-0.15934163331985474,
-0.10354094952344894,
0.10525287687778473,
-0.04850301891565323,
0.028323976323008537,
-0.07308647781610489,
-0.04719609022140503,
-0.24660754203796387,
0.2158661186695099,
0.015384898521006107,
0.00384410098195076,
-0.09645523875951767,
-0.0031059239991009235,
0.16681231558322906,
-0.09507244825363159,
0.038081102073192596,
-0.0030803470872342587,
0.08820296823978424,
-0.09653045237064362,
-0.08165580779314041,
0.04465007409453392,
-0.07744460552930832,
-0.14185139536857605,
0.011764489114284515,
0.19781433045864105,
0.022616581991314888,
0.037971191108226776,
0.025954479351639748,
0.09160146117210388,
0.08607495576143265,
-0.10742948949337006,
0.08422011882066727,
0.0879930630326271,
-0.030707523226737976,
0.1495898962020874,
-0.033895764499902725,
-0.15872009098529816,
-0.05146879330277443,
-0.0411202572286129,
0.10311200469732285,
0.34257373213768005,
-0.06467219442129135,
0.08678370714187622,
0.08080018311738968,
-0.042148686945438385,
-0.19632753729820251,
0.007789987605065107,
0.01917961612343788,
0.02523370087146759,
0.00445806048810482,
-0.05062577873468399,
0.10844220966100693,
0.13072700798511505,
-0.020657971501350403,
-0.04788164049386978,
-0.20526863634586334,
-0.1290767788887024,
0.011098041199147701,
-0.06513512134552002,
-0.012189746834337711,
-0.06006240472197533,
-0.10408233106136322,
-0.07066939771175385,
-0.16428698599338531,
0.05405878648161888,
0.04084285721182823,
0.048617810010910034,
0.0022039911709725857,
-0.022370649501681328,
0.03627573698759079,
-0.018924014642834663,
0.18879440426826477,
0.06920293718576431,
0.02023874782025814,
-0.07239631563425064,
-0.04903245344758034,
0.11341241002082825,
-0.037153460085392,
0.09947021305561066,
0.04320525750517845,
-0.005356981884688139,
-0.11361989378929138,
0.012292295694351196,
-0.09132620692253113,
0.11476991325616837,
-0.09256471693515778,
-0.030837757512927055,
-0.08548001945018768,
0.140853151679039,
0.0238761268556118,
-0.015673771500587463,
0.12679731845855713,
-0.032148636877536774,
0.1091376468539238,
0.13080862164497375,
0.0701589584350586,
0.0031420262530446053,
-0.07472140341997147,
0.0012759605888277292,
-0.06170511618256569,
0.019098622724413872,
0.05546579137444496,
0.04380559176206589,
0.13516387343406677,
-0.01975567452609539,
0.14712172746658325,
-0.018929120153188705,
-0.13689684867858887,
-0.023719700053334236,
0.11753503233194351,
-0.11223208159208298,
-0.2821040451526642,
-0.04018678143620491,
-0.034599367529153824,
0.012139244005084038,
0.061747223138809204,
0.19457398355007172,
-0.03780088946223259,
-0.023348631337285042,
0.021921996027231216,
0.04527831822633743,
-0.09146711230278015,
0.08037439733743668,
-0.021092766895890236,
0.042854998260736465,
-0.06821004301309586,
0.10744444280862808,
0.07745382934808731,
-0.22681817412376404,
-0.025234894827008247,
0.09869363903999329,
-0.06257352977991104,
-0.07561001926660538,
-0.09376278519630432,
0.1380273848772049,
-0.1314297765493393,
-0.0158812515437603,
-0.020428206771612167,
-0.1451670229434967,
0.04510067030787468,
0.09574063867330551,
0.04472881183028221,
0.011867307126522064,
0.035759247839450836,
0.00018209819972980767,
0.01861182041466236,
0.03738055005669594,
0.09328993409872055,
0.10640279203653336,
-0.042188696563243866,
0.03816690295934677,
-0.08412777632474899,
0.015412991866469383,
-0.021022463217377663,
-0.01296116691082716,
-0.09465239942073822,
-0.07739552855491638,
-0.10236255824565887,
-0.06780892610549927,
-0.08637113869190216,
-0.02516632154583931,
-0.0233630258589983,
-0.023995107039809227,
0.004357992671430111,
-0.027227098122239113,
-0.07138573378324509,
-0.050982195883989334,
-0.0155563373118639,
0.08955743908882141,
-0.1038399413228035,
-0.0027917600236833096,
0.06826390326023102,
-0.051174331456422806,
0.0997040867805481,
0.024214884266257286,
0.0514301136136055,
0.03840871900320053,
-0.03028056025505066,
-0.03847609460353851,
0.01905941218137741,
0.006461322773247957,
0.05086956173181534,
-0.14872215688228607,
-0.012150841765105724,
-0.0032515712082386017,
-0.051708951592445374,
0.050017524510622025,
-0.0625491663813591,
-0.10236836224794388,
0.008505094796419144,
0.016627702862024307,
-0.09351813793182373,
-0.07604184001684189,
0.027614127844572067,
0.0729801207780838,
0.035474032163619995,
0.11152558028697968,
-0.058112598955631256,
0.05574757605791092,
-0.1357552707195282,
-0.01836516335606575,
-0.02554010972380638,
-0.1030651405453682,
0.016653461381793022,
-0.011252963915467262,
0.022870756685733795,
-0.0031690942123532295,
0.11289511620998383,
-0.026996877044439316,
-0.025828121230006218,
0.06756260991096497,
-0.028583889827132225,
-0.0780753642320633,
0.0507938489317894,
0.07868042588233948,
0.04812269285321236,
-0.0161933284252882,
-0.08120898902416229,
-0.0014988429611548781,
-0.05973583459854126,
-0.007319761905819178,
0.1114073172211647,
0.14946618676185608,
0.11061113327741623,
0.05770944431424141,
0.08207276463508606,
-0.04891999810934067,
-0.028829647228121758,
-0.009984364733099937,
-0.029658254235982895,
0.08592110872268677,
-0.02771167829632759,
0.046906858682632446,
0.07054735720157623,
-0.23260782659053802,
0.1542171984910965,
-0.04036210849881172,
0.003158458275720477,
-0.019460761919617653,
-0.11383481323719025,
-0.07883796095848083,
-0.042559150606393814,
0.010454006493091583,
-0.08244454860687256,
0.009923603385686874,
0.00909425225108862,
0.03983308747410774,
-0.018626000732183456,
0.14684614539146423,
-0.22303219139575958,
-0.020058225840330124,
0.1173686534166336,
0.01384452823549509,
-0.01986311748623848,
0.04652818292379379,
-0.01833965815603733,
-0.05564456060528755,
0.0526103749871254,
0.003485421882942319,
0.042003098875284195,
0.01768621988594532,
-0.009287383407354355,
-0.030709892511367798,
-0.0902811586856842,
0.015258834697306156,
-0.013983458280563354,
0.019041143357753754,
0.026359856128692627,
0.02583683282136917,
0.016558820381760597,
-0.0304630808532238,
0.1203962042927742,
-0.04186068847775459,
-0.08964694291353226,
-0.09655430167913437,
0.18649248778820038,
0.015295648016035557,
0.07767751067876816,
0.014468489214777946,
-0.11231113970279694,
0.003929391037672758,
0.011538079008460045,
0.2675095796585083,
0.06229402497410774,
0.0249533299356699,
0.022313563153147697,
0.0121700968593359,
0.03887152299284935,
0.019862765446305275,
-0.021693918853998184,
0.17263448238372803,
-0.035285405814647675,
0.2010856717824936,
-0.05225873738527298,
-0.046340879052877426,
0.0050239525735378265,
0.1740422248840332,
-0.01995527185499668,
-0.01296838466078043,
-0.015919938683509827,
0.11686369776725769,
-0.02941286377608776,
-0.15430203080177307,
0.030166465789079666,
-0.1680508852005005,
-0.1555458903312683,
-0.04883193224668503,
0.03437260910868645,
0.0537441223859787,
0.11792732775211334,
0.015498639084398746,
-0.03355927765369415,
0.11172260344028473,
0.013434622436761856,
-0.01853151060640812,
-0.12593184411525726,
0.06253454834222794,
-0.005848423577845097,
0.10684798657894135,
0.020717212930321693,
0.04494085907936096,
0.12986114621162415,
-0.024106910452246666,
-0.10693443566560745,
0.014985187910497189,
0.045302849262952805,
-0.08386611938476562,
0.0049399384297430515,
0.15897472202777863,
-0.014517411589622498,
0.11524591594934464,
0.05899004638195038,
-0.13243594765663147,
0.07087180018424988,
0.0894564762711525,
-0.041566379368305206,
-0.06131380423903465,
0.1714923232793808,
-0.08582644164562225,
0.13566255569458008,
0.23618710041046143,
-0.026338893920183182,
0.025297190994024277,
-0.03987700119614601,
0.044342536479234695,
0.07475463300943375,
0.12010806798934937,
-0.017106499522924423,
-0.16066502034664154,
0.06056363880634308,
0.023605575785040855,
0.04031899571418762,
-0.14200104773044586,
-0.11799366027116776,
-0.015118377283215523,
0.02434436045587063,
-0.030886957421898842,
0.12695416808128357,
0.08972851932048798,
-0.0224088616669178,
0.001209444715641439,
-0.1888996660709381,
-0.008432535454630852,
0.07094074040651321,
-0.045595064759254456,
-0.016666574403643608
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# working
This model is a fine-tuned version of [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.36.0
- Pytorch 2.0.0
- Datasets 2.16.0
- Tokenizers 0.15.0
| {"license": "mit", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "microsoft/phi-1_5", "model-index": [{"name": "working", "results": []}]} | text-generation | ManthanCisco/phi_Text2SQL_v1 | [
"transformers",
"tensorboard",
"safetensors",
"phi",
"text-generation",
"trl",
"sft",
"generated_from_trainer",
"custom_code",
"base_model:microsoft/phi-1_5",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:48:47+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #phi #text-generation #trl #sft #generated_from_trainer #custom_code #base_model-microsoft/phi-1_5 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# working
This model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 4
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.36.0
- Pytorch 2.0.0
- Datasets 2.16.0
- Tokenizers 0.15.0
| [
"# working\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 1\n- eval_batch_size: 1\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.05\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.36.0\n- Pytorch 2.0.0\n- Datasets 2.16.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#transformers #tensorboard #safetensors #phi #text-generation #trl #sft #generated_from_trainer #custom_code #base_model-microsoft/phi-1_5 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# working\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 1\n- eval_batch_size: 1\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.05\n- num_epochs: 1",
"### Training results",
"### Framework versions\n\n- Transformers 4.36.0\n- Pytorch 2.0.0\n- Datasets 2.16.0\n- Tokenizers 0.15.0"
] | [
75,
26,
6,
12,
8,
3,
129,
4,
34
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #phi #text-generation #trl #sft #generated_from_trainer #custom_code #base_model-microsoft/phi-1_5 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# working\n\nThis model is a fine-tuned version of microsoft/phi-1_5 on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 1\n- eval_batch_size: 1\n- seed: 42\n- gradient_accumulation_steps: 4\n- total_train_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.05\n- num_epochs: 1### Training results### Framework versions\n\n- Transformers 4.36.0\n- Pytorch 2.0.0\n- Datasets 2.16.0\n- Tokenizers 0.15.0"
] | [
-0.08880594372749329,
0.06419459730386734,
-0.002933547832071781,
0.07752366364002228,
0.14904847741127014,
0.030237630009651184,
0.11416822671890259,
0.10977454483509064,
-0.12066652625799179,
0.06878675520420074,
0.033893369138240814,
0.08786351978778839,
0.05131911113858223,
0.12991121411323547,
-0.053436797112226486,
-0.22713924944400787,
0.021908646449446678,
-0.021966585889458656,
-0.07094862312078476,
0.09560985118150711,
0.09313299506902695,
-0.10743088275194168,
0.06595451384782791,
0.023775305598974228,
-0.1477067768573761,
0.003894404973834753,
-0.0022317771799862385,
-0.03034619428217411,
0.11310546100139618,
0.012619683519005775,
0.12244009971618652,
0.018117869272828102,
0.14395302534103394,
-0.2006009817123413,
-0.0018209457630291581,
0.09792930632829666,
0.04643148183822632,
0.08323244005441666,
0.09611532837152481,
0.010727670975029469,
0.10755030810832977,
-0.1163518950343132,
0.10511757433414459,
0.02746039815247059,
-0.10649583488702774,
-0.20404136180877686,
-0.10754360258579254,
0.024014241993427277,
0.08960945904254913,
0.10182273387908936,
0.006801557261496782,
0.16255952417850494,
-0.06430304795503616,
0.08475344628095627,
0.21963845193386078,
-0.27402836084365845,
-0.06063095107674599,
0.08497900515794754,
0.06714529544115067,
0.04396451637148857,
-0.08794621378183365,
-0.013396996073424816,
0.02246496081352234,
0.027939623221755028,
0.09459460526704788,
-0.0026466096751391888,
-0.040664829313755035,
-0.030036164447665215,
-0.12581847608089447,
-0.01845001056790352,
0.04903237149119377,
0.03303046151995659,
-0.029752155765891075,
-0.07312159240245819,
-0.08573644608259201,
-0.11441878974437714,
-0.010936146602034569,
-0.06014901027083397,
0.0590360201895237,
-0.03263014554977417,
-0.046930260956287384,
-0.05472235009074211,
-0.07094351947307587,
-0.0617247074842453,
-0.006574017461389303,
0.12774237990379333,
0.0370861291885376,
0.022800832986831665,
-0.02969467081129551,
0.11707060784101486,
-0.013365584425628185,
-0.13026009500026703,
-0.004017503000795841,
0.010594679042696953,
-0.08997213840484619,
-0.0659978985786438,
-0.06019095331430435,
-0.0077252560295164585,
-0.006760456133633852,
0.14646963775157928,
-0.10640121251344681,
0.07260245084762573,
0.017358290031552315,
0.00032931132591329515,
-0.06467095762491226,
0.16340960562229156,
-0.027847349643707275,
-0.05008430778980255,
0.0099715581163764,
0.0782097727060318,
0.03153829649090767,
-0.0003502066829241812,
-0.0900019109249115,
-0.01773569919168949,
0.09025254845619202,
0.049745894968509674,
-0.04554436355829239,
0.0357309952378273,
-0.046293020248413086,
-0.019752291962504387,
0.03821440413594246,
-0.1314155012369156,
0.06152530759572983,
-0.005177818238735199,
-0.06964031606912613,
-0.020145094022154808,
0.03760595619678497,
0.01905764639377594,
-0.033512383699417114,
0.13143514096736908,
-0.05856757238507271,
0.01819688268005848,
-0.11162290722131729,
-0.07572870701551437,
0.00823444128036499,
-0.0866035595536232,
-0.008805928751826286,
-0.048115845769643784,
-0.2103816121816635,
-0.04974675551056862,
0.05775029584765434,
-0.07759114354848862,
-0.024619603529572487,
-0.014317275956273079,
-0.08806665986776352,
0.007523681968450546,
-0.023088669404387474,
0.15977391600608826,
-0.053963497281074524,
0.06436963379383087,
0.025585385039448738,
0.044409606605768204,
0.019846448674798012,
0.03328564018011093,
-0.08183525502681732,
0.016880961135029793,
-0.18792040646076202,
0.0593053475022316,
-0.059504009783267975,
0.013157793320715427,
-0.09440106898546219,
-0.0894279032945633,
-0.0007853953866288066,
-0.019107647240161896,
0.07364018261432648,
0.08559023588895798,
-0.17266754806041718,
-0.020148344337940216,
0.13454151153564453,
-0.11433225870132446,
-0.07334831357002258,
0.06572460383176804,
-0.026486536487936974,
0.023697400465607643,
0.05598558858036995,
0.12137968838214874,
0.022844726219773293,
-0.1627100557088852,
0.016657382249832153,
-0.009806322865188122,
0.04796074703335762,
0.0341508612036705,
0.045303624123334885,
-0.01763055846095085,
0.11961502581834793,
0.004011870827525854,
-0.057839758694171906,
-0.03181392699480057,
-0.07987295836210251,
-0.07046325504779816,
-0.06652649492025375,
-0.08040065318346024,
0.0038768076337873936,
0.043924376368522644,
0.03180660679936409,
-0.07080772519111633,
-0.1320727914571762,
0.1406158208847046,
0.11036361008882523,
-0.042814016342163086,
0.038529157638549805,
-0.08891408890485764,
0.02349558286368847,
-0.015601171180605888,
-0.04407818615436554,
-0.23168662190437317,
-0.11474451422691345,
0.01840358041226864,
-0.08299718052148819,
0.017516151070594788,
0.047138649970293045,
0.07296102494001389,
0.08368104696273804,
-0.06691057235002518,
-0.010757427662611008,
-0.10096705704927444,
0.017720447853207588,
-0.11360664665699005,
-0.21239173412322998,
-0.061850182712078094,
-0.019280115142464638,
0.1366066336631775,
-0.21019567549228668,
0.007054658140987158,
0.00024875267990864813,
0.17612387239933014,
0.0311602670699358,
-0.055782344192266464,
-0.041649553924798965,
0.047845546156167984,
-0.0031850324012339115,
-0.10143810510635376,
0.049980875104665756,
0.008799577131867409,
-0.054115284234285355,
-0.05621953681111336,
-0.17512328922748566,
0.04486517980694771,
0.101841039955616,
-0.006504757329821587,
-0.09036049991846085,
0.03830252215266228,
-0.05318176746368408,
-0.04729331284761429,
-0.07780422270298004,
0.007944618351757526,
0.17607970535755157,
0.019427239894866943,
0.12377884238958359,
-0.0681329071521759,
-0.07615356892347336,
-0.011270771734416485,
-0.011801978573203087,
0.021352244541049004,
0.0631740614771843,
0.05622894689440727,
-0.09586632251739502,
0.08714492619037628,
0.10444840788841248,
-0.0699923187494278,
0.09529804438352585,
-0.053798265755176544,
-0.07902222126722336,
-0.02168927900493145,
-0.024065565317869186,
0.0066559226252138615,
0.11239880323410034,
-0.07084889709949493,
0.030118897557258606,
0.02290274389088154,
0.04036340117454529,
0.03773555904626846,
-0.20648320019245148,
0.016477713361382484,
0.026794137433171272,
-0.04546037316322327,
0.026405131444334984,
-0.03535642847418785,
0.0418417863547802,
0.08393929153680801,
0.019720973446965218,
-0.020094087347388268,
0.01310139149427414,
-0.02428271807730198,
-0.08117145299911499,
0.16823284327983856,
-0.09186623990535736,
-0.11123653501272202,
-0.1033906638622284,
0.027465887367725372,
-0.03446033596992493,
-0.01973913609981537,
0.021038854494690895,
-0.0684085488319397,
-0.07470659166574478,
-0.08540695905685425,
-0.008921010419726372,
-0.0018895600223913789,
-0.018234243616461754,
0.06406175345182419,
0.020152676850557327,
0.10874152928590775,
-0.12605322897434235,
-0.004452509805560112,
-0.03640032187104225,
-0.10606491565704346,
-0.00533820316195488,
0.07007915526628494,
0.0612080916762352,
0.13169211149215698,
-0.029232855886220932,
0.01828322932124138,
-0.019405033439397812,
0.2283288836479187,
-0.08392726629972458,
0.007766821421682835,
0.1739204078912735,
-0.008693011477589607,
0.06007661297917366,
0.11537064611911774,
0.034356746822595596,
-0.10444895178079605,
0.02378401719033718,
0.08138123899698257,
-0.011660858057439327,
-0.2372351586818695,
-0.03972717374563217,
-0.014536323957145214,
-0.09544075280427933,
0.049965303391218185,
0.037150487303733826,
0.021764783188700676,
0.01953926496207714,
-0.005622606258839369,
0.006022661924362183,
0.0136337261646986,
0.07580466568470001,
0.1094793975353241,
0.06638474017381668,
0.11340809613466263,
-0.029187610372900963,
0.005708741024136543,
0.0573112927377224,
-0.012871832586824894,
0.21928536891937256,
-0.008868514560163021,
0.06620132923126221,
0.0395636223256588,
0.11046808958053589,
-0.009733445942401886,
0.04305513948202133,
0.021678974851965904,
-0.016371266916394234,
0.009122619405388832,
-0.07006718218326569,
-0.021323375403881073,
0.01593414507806301,
-0.0730368122458458,
0.04024193063378334,
-0.06827455013990402,
0.043222732841968536,
0.020837455987930298,
0.24882954359054565,
0.05668908730149269,
-0.2995974123477936,
-0.07658248394727707,
0.0026005252730101347,
-0.03755050525069237,
-0.045064665377140045,
-0.021023772656917572,
0.08142556250095367,
-0.13504622876644135,
0.08167582005262375,
-0.06898821145296097,
0.09686210751533508,
-0.03552086278796196,
0.020303040742874146,
0.0986151322722435,
0.12443666160106659,
0.015019089914858341,
0.050913456827402115,
-0.21311409771442413,
0.21779827773571014,
0.01165775302797556,
0.13910342752933502,
-0.041504569351673126,
0.048502251505851746,
0.015388409607112408,
0.08523070067167282,
0.06272251158952713,
-0.015630455687642097,
-0.06306054443120956,
-0.15429186820983887,
-0.07117578387260437,
0.019662054255604744,
0.11805374175310135,
-0.014453765004873276,
0.08910923451185226,
-0.059079863131046295,
0.005826916079968214,
0.056408971548080444,
-0.035089217126369476,
-0.1540147364139557,
-0.08360166847705841,
0.014097047969698906,
-0.03133581206202507,
-0.01818469911813736,
-0.07895680516958237,
-0.11128280311822891,
-0.07772672921419144,
0.19030915200710297,
0.015699084848165512,
-0.04714677482843399,
-0.14672169089317322,
0.0883263498544693,
0.11803574860095978,
-0.05083748698234558,
0.036307040601968765,
0.012024580501019955,
0.10745535790920258,
0.04121944680809975,
-0.08058420568704605,
0.07858767360448837,
-0.06617960333824158,
-0.22831058502197266,
-0.049062248319387436,
0.10874499380588531,
0.03562972694635391,
0.04355539381504059,
0.00015232732403092086,
0.05422409623861313,
0.011183291673660278,
-0.08133165538311005,
0.01816173829138279,
0.08007726818323135,
0.10211789608001709,
0.02513216994702816,
-0.05048603564500809,
0.01797429285943508,
-0.008395000360906124,
-0.029445825144648552,
0.06961032748222351,
0.23038828372955322,
-0.08680244535207748,
0.053543753921985626,
0.043408509343862534,
-0.08067720383405685,
-0.17714332044124603,
0.062244631350040436,
0.12314236909151077,
0.0075477976351976395,
0.0733964741230011,
-0.13669371604919434,
0.13693514466285706,
0.13287761807441711,
-0.045537132769823074,
0.009241935797035694,
-0.2898285686969757,
-0.1651064157485962,
0.05248009040951729,
0.12881286442279816,
0.008949531242251396,
-0.15192148089408875,
-0.03980661556124687,
-0.04344853758811951,
-0.16958820819854736,
0.0938652902841568,
-0.1154172420501709,
0.09986793994903564,
0.011402834206819534,
0.056330498307943344,
0.01299560721963644,
-0.039165377616882324,
0.14251694083213806,
0.012178861536085606,
0.09465504437685013,
-0.057818103581666946,
0.0476694256067276,
0.08607605844736099,
-0.06439585238695145,
0.010050981305539608,
0.002190026920288801,
0.05946444720029831,
-0.08792687952518463,
-0.019167866557836533,
-0.06807929277420044,
0.052401844412088394,
-0.06237436830997467,
-0.05697440356016159,
-0.0254803653806448,
0.049055520445108414,
0.05007688328623772,
-0.049398139119148254,
0.08207856863737106,
-0.008539468981325626,
0.12931619584560394,
0.1529073417186737,
0.09985090047121048,
-0.028814343735575676,
-0.08541436493396759,
0.020924080163240433,
-0.009359297342598438,
0.03318971022963524,
-0.08308615535497665,
0.03648323193192482,
0.12283630669116974,
0.021452195942401886,
0.11289618164300919,
0.042249515652656555,
-0.05332032963633537,
-0.006093738134950399,
0.04116097092628479,
-0.10467071086168289,
-0.13466601073741913,
0.024978209286928177,
-0.05628684163093567,
-0.09478267282247543,
0.047302309423685074,
0.11895869672298431,
-0.03845083713531494,
-0.007987573742866516,
-0.0031575628090649843,
0.022817937657237053,
-0.026527853682637215,
0.20196110010147095,
0.0229289922863245,
0.07024701684713364,
-0.0949559137225151,
0.14070461690425873,
0.06877216696739197,
-0.07633771002292633,
0.011403742246329784,
0.08598752319812775,
-0.10311751812696457,
0.00022825815540272743,
0.09990987926721573,
0.15251369774341583,
-0.045354943722486496,
-0.05400973930954933,
-0.1089688092470169,
-0.10713811963796616,
0.040126651525497437,
0.14607882499694824,
0.06498013436794281,
0.005200871732085943,
-0.023843441158533096,
0.04193813353776932,
-0.12570621073246002,
0.09097841382026672,
0.05559638887643814,
0.07995326817035675,
-0.15508835017681122,
0.1419127732515335,
0.02168288454413414,
0.007604887243360281,
-0.02175474725663662,
0.047678522765636444,
-0.09096498042345047,
-0.0261215977370739,
-0.1336870789527893,
-0.005766208749264479,
-0.020044274628162384,
-0.010916735976934433,
-0.009253040887415409,
-0.0685228705406189,
-0.03130847588181496,
0.05272774398326874,
-0.07533053308725357,
-0.060644667595624924,
-0.0016254555666819215,
0.016267357394099236,
-0.1316068023443222,
0.0018063714960590005,
0.014008790254592896,
-0.09126468747854233,
0.07208579778671265,
0.045845311135053635,
0.038972798734903336,
0.03898424655199051,
-0.1135735809803009,
-0.006684067193418741,
0.03492806851863861,
0.02159292809665203,
0.07312465459108353,
-0.09538247436285019,
-0.024903206154704094,
-0.010464675724506378,
0.05895961821079254,
0.013215664774179459,
0.07869233191013336,
-0.13235843181610107,
0.015604638494551182,
-0.05016082897782326,
-0.06120714545249939,
-0.037757810205221176,
0.021229784935712814,
0.11829005181789398,
0.020823992788791656,
0.18062062561511993,
-0.08753751218318939,
0.04159267619252205,
-0.22343793511390686,
-0.0224831011146307,
0.00422565545886755,
-0.026964373886585236,
-0.09418531507253647,
-0.035959869623184204,
0.09405223280191422,
-0.052122343331575394,
0.11134467273950577,
0.016093838959932327,
0.12979918718338013,
0.05543491616845131,
-0.05071938410401344,
-0.015194308944046497,
0.025453543290495872,
0.13460133969783783,
0.0550253801047802,
0.0009583483333699405,
0.11629410088062286,
-0.008915895596146584,
0.03957593813538551,
-0.005149479955434799,
0.24550341069698334,
0.17677417397499084,
-0.033887747675180435,
0.07938109338283539,
0.0742201879620552,
-0.10281331837177277,
-0.17014992237091064,
0.08517510443925858,
-0.06231195479631424,
0.09974317997694016,
-0.08183383196592331,
0.1492687314748764,
0.08569976687431335,
-0.1824924200773239,
0.049633853137493134,
-0.06619936227798462,
-0.08571381121873856,
-0.14289042353630066,
-0.02123592048883438,
-0.09176778793334961,
-0.13060027360916138,
0.011497320607304573,
-0.10994504392147064,
0.059757575392723083,
0.13824044167995453,
0.0034388303756713867,
0.0006756641669198871,
0.15007384121418,
-0.04099263623356819,
0.015979232266545296,
0.01975824125111103,
0.026108670979738235,
0.017019057646393776,
-0.061269089579582214,
-0.06872064620256424,
0.008150231093168259,
0.04307568818330765,
0.07019153237342834,
-0.03428330272436142,
-0.030008738860487938,
0.015332857146859169,
0.00761190801858902,
-0.07241935282945633,
0.04241810739040375,
0.029657596722245216,
0.03584444895386696,
0.03720377758145332,
0.03073212318122387,
0.02554287575185299,
-0.04576960206031799,
0.2879832684993744,
-0.09544061124324799,
-0.12892656028270721,
-0.10835815221071243,
0.2607702314853668,
0.02483939938247204,
-0.00001938887726282701,
0.0671139508485794,
-0.12078245729207993,
-0.014552599750459194,
0.14915338158607483,
0.13537350296974182,
-0.07503369450569153,
-0.016334237530827522,
0.0018116934224963188,
-0.019670603796839714,
-0.05153179168701172,
0.10772258043289185,
0.10626038908958435,
0.037941060960292816,
-0.05704471841454506,
-0.003564402461051941,
-0.001254344591870904,
-0.03426568955183029,
-0.07730457186698914,
0.05127362161874771,
0.0035804028157144785,
0.013842753134667873,
-0.02261502854526043,
0.07176219671964645,
0.02687055617570877,
-0.22059348225593567,
0.07969962060451508,
-0.1817552149295807,
-0.18252618610858917,
-0.00100877252407372,
0.10799199342727661,
-0.040406107902526855,
0.06540020555257797,
0.004994102753698826,
-0.015864480286836624,
0.10145223885774612,
-0.017678095027804375,
-0.029238084331154823,
-0.08505209535360336,
0.07958941161632538,
-0.1403852105140686,
0.21442069113254547,
-0.011329655535519123,
0.06882566958665848,
0.1032392680644989,
0.02681829407811165,
-0.12734639644622803,
0.06633614748716354,
0.0540597178041935,
-0.1323460340499878,
0.031036576256155968,
0.1334935426712036,
-0.03658191114664078,
0.05711118504405022,
0.04293837770819664,
-0.17901360988616943,
-0.0013069396372884512,
-0.012293179519474506,
-0.053170375525951385,
-0.04542883485555649,
-0.028475601226091385,
-0.07705914229154587,
0.1496565341949463,
0.21299318969249725,
-0.021294165402650833,
0.03792766109108925,
-0.06785063445568085,
0.02862527407705784,
0.031337276101112366,
0.09114523977041245,
-0.02825680561363697,
-0.251056045293808,
0.06077330559492111,
0.0360654853284359,
0.009301400743424892,
-0.16950002312660217,
-0.09515110403299332,
0.05752820521593094,
-0.05908602103590965,
-0.09225210547447205,
0.09825000166893005,
0.05140587314963341,
0.0320618562400341,
-0.031940631568431854,
-0.14546708762645721,
-0.03838850557804108,
0.1670692414045334,
-0.17381563782691956,
-0.058044951409101486
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-classification | selink/citation-classifier-roberta-base | [
"transformers",
"safetensors",
"roberta",
"text-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:51:23+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #roberta #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #roberta #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
47,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #roberta #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07097413390874863,
0.1667894870042801,
-0.003666079370304942,
0.021719951182603836,
0.1169515997171402,
0.008284984156489372,
0.07978618890047073,
0.10868363827466965,
-0.0287666916847229,
0.12472489476203918,
0.03958536684513092,
0.100002720952034,
0.11150532960891724,
0.19375544786453247,
0.0031385887414216995,
-0.2044016718864441,
0.0626462921500206,
-0.11493270099163055,
0.012976650148630142,
0.12250365316867828,
0.1423846185207367,
-0.10432620346546173,
0.07270357757806778,
-0.038304854184389114,
-0.02555735409259796,
-0.02992432750761509,
-0.06288715451955795,
-0.06023360788822174,
0.06700628250837326,
0.060681991279125214,
0.0691668838262558,
0.022982222959399223,
0.08022861182689667,
-0.2922787070274353,
0.0204449649900198,
0.07901842147111893,
0.0025627717841416597,
0.06292086839675903,
0.07761354744434357,
-0.07406430691480637,
0.11906187236309052,
-0.055618006736040115,
0.15586547553539276,
0.07508774101734161,
-0.09551873803138733,
-0.1806546300649643,
-0.08346612006425858,
0.09468797594308853,
0.16155311465263367,
0.05840203911066055,
-0.03498850762844086,
0.14096277952194214,
-0.08077409118413925,
0.016175461933016777,
0.06836850196123123,
-0.07082965970039368,
-0.05214150249958038,
0.05027921497821808,
0.07374554872512817,
0.09283687174320221,
-0.1307315081357956,
-0.010299519635736942,
0.04289179667830467,
0.018450511619448662,
0.10514726489782333,
0.023654451593756676,
0.11117657274007797,
0.028080156072974205,
-0.13854995369911194,
-0.0629403293132782,
0.12449787557125092,
0.032600224018096924,
-0.060118548572063446,
-0.23809053003787994,
-0.004739876836538315,
-0.03291091322898865,
-0.021559255197644234,
-0.04033812880516052,
0.04089386388659477,
-0.029875123873353004,
0.08085108548402786,
0.006228691898286343,
-0.07215629518032074,
-0.052459850907325745,
0.08968427777290344,
0.05656275153160095,
0.02484809421002865,
-0.02587738446891308,
0.02272983267903328,
0.1178721934556961,
0.09954223781824112,
-0.11430655419826508,
-0.06540504842996597,
-0.06454093754291534,
-0.08725714683532715,
-0.04798560217022896,
0.03668604791164398,
0.07030875980854034,
0.046011876314878464,
0.19716040790081024,
0.00269759027287364,
0.05310693010687828,
0.02948983944952488,
0.016264021396636963,
0.06394703686237335,
0.06814983487129211,
-0.05028722807765007,
-0.12879575788974762,
-0.03764699772000313,
0.11916788667440414,
0.006199078168720007,
-0.032489337027072906,
-0.0359247624874115,
0.05915093049407005,
0.05401566997170448,
0.11897829174995422,
0.06329395622015,
0.01798311620950699,
-0.06940232217311859,
-0.041600339114665985,
0.1802349090576172,
-0.15566512942314148,
0.020810957998037338,
0.015188573859632015,
-0.055285364389419556,
-0.042803313583135605,
0.023135531693696976,
0.009154262952506542,
-0.02835596352815628,
0.10190211236476898,
-0.06733403354883194,
-0.03976709768176079,
-0.10797561705112457,
-0.05405012145638466,
0.03397495672106743,
-0.023470325395464897,
-0.0295037180185318,
-0.04243861138820648,
-0.12201286107301712,
-0.07699050009250641,
0.06754051148891449,
-0.06153370812535286,
-0.06873670220375061,
-0.039611347019672394,
-0.0595494844019413,
0.012631349265575409,
0.002569467294961214,
0.12651456892490387,
-0.029271753504872322,
0.04821896180510521,
-0.05001894757151604,
0.06899536401033401,
0.13243289291858673,
0.03377855569124222,
-0.07108951359987259,
0.06568732112646103,
-0.21172066032886505,
0.10609432309865952,
-0.09376481175422668,
0.03139003366231918,
-0.16234688460826874,
-0.024603933095932007,
0.027305902913212776,
0.03785030171275139,
-0.01080681849271059,
0.14199338853359222,
-0.18069814145565033,
-0.0368097685277462,
0.18344907462596893,
-0.13205336034297943,
-0.09475254267454147,
0.06138899549841881,
-0.06121760979294777,
0.13521988689899445,
0.05350736901164055,
-0.023734936490654945,
0.05919842794537544,
-0.1313716471195221,
-0.025711048394441605,
-0.05855761095881462,
-0.008730711415410042,
0.14457696676254272,
0.06305380910634995,
-0.05395244061946869,
0.026081737130880356,
0.018570488318800926,
-0.019653121009469032,
-0.04788879305124283,
-0.03383433073759079,
-0.09769339114427567,
0.005564571358263493,
-0.08092508465051651,
0.01647305116057396,
-0.018450245261192322,
-0.08526520431041718,
-0.040688324719667435,
-0.1556665003299713,
0.002278752624988556,
0.09955150634050369,
0.00428511667996645,
-0.028858626261353493,
-0.09412826597690582,
0.0010344249894842505,
0.014487308450043201,
-0.011919399723410606,
-0.15230298042297363,
-0.04881799593567848,
0.026287198066711426,
-0.1660703718662262,
0.03276912495493889,
-0.05137382447719574,
0.03549332916736603,
0.044475916773080826,
-0.0461643822491169,
-0.024930957704782486,
0.014140721410512924,
0.017139434814453125,
-0.02374175749719143,
-0.24581307172775269,
-0.013825789093971252,
-0.04980151727795601,
0.18321004509925842,
-0.24926310777664185,
0.04708116501569748,
0.061669353395700455,
0.11851225793361664,
0.005235658492892981,
-0.047542329877614975,
0.04137692227959633,
-0.049956902861595154,
-0.04146651551127434,
-0.06634348630905151,
-0.003162689507007599,
-0.03229794278740883,
-0.04517984762787819,
0.041715674102306366,
-0.190790057182312,
-0.026901599019765854,
0.11227498203516006,
0.0757983848452568,
-0.172615647315979,
-0.07716702669858932,
-0.032237354665994644,
-0.06075936183333397,
-0.08794162422418594,
-0.04783210903406143,
0.1008552610874176,
0.040512196719646454,
0.05559884011745453,
-0.07266750186681747,
-0.05319931358098984,
0.01357761025428772,
-0.011024849489331245,
-0.03397562727332115,
0.09108693152666092,
0.08268309384584427,
-0.1214602142572403,
0.1049051284790039,
0.06639663875102997,
0.06267683953046799,
0.10698967427015305,
0.00657825730741024,
-0.09603405743837357,
-0.011494014412164688,
0.02610582485795021,
0.01399005576968193,
0.1426926553249359,
-0.07498639076948166,
0.030146000906825066,
0.042116980999708176,
-0.030949845910072327,
0.012327825650572777,
-0.1036502942442894,
0.0180202666670084,
0.03208072856068611,
-0.00865737535059452,
0.015741337090730667,
-0.056021902710199356,
0.013855318538844585,
0.10478801280260086,
0.03378775715827942,
0.028230587020516396,
0.01726776920258999,
-0.039543844759464264,
-0.12787796556949615,
0.1788790076971054,
-0.09654324501752853,
-0.25226548314094543,
-0.13508103787899017,
0.00398462638258934,
0.046175748109817505,
-0.011871901340782642,
0.01716979779303074,
-0.05890684202313423,
-0.10515041649341583,
-0.10667064040899277,
0.0201003048568964,
0.05383666977286339,
-0.0892900750041008,
-0.06228140741586685,
0.054920271039009094,
0.037761468440294266,
-0.12314439564943314,
0.023868469521403313,
0.045075058937072754,
-0.07041552662849426,
0.006107859779149294,
0.0533861480653286,
0.08385360985994339,
0.17951355874538422,
0.008954732678830624,
-0.01666683331131935,
0.009695366024971008,
0.2122192084789276,
-0.14731436967849731,
0.09077969938516617,
0.13881319761276245,
-0.05979595333337784,
0.08313421159982681,
0.2037704885005951,
0.028123976662755013,
-0.09402397274971008,
0.04063381254673004,
0.03390387445688248,
-0.04006393998861313,
-0.24267235398292542,
-0.07322345674037933,
0.003262192476540804,
-0.06732312589883804,
0.10439975559711456,
0.09025674313306808,
0.11683525890111923,
0.05397806316614151,
-0.10713037848472595,
-0.06759513914585114,
0.048913486301898956,
0.11977367848157883,
-0.031416673213243484,
0.004043765366077423,
0.0945456475019455,
-0.03565856069326401,
0.023265741765499115,
0.09114573150873184,
0.016503766179084778,
0.18390727043151855,
0.03800623491406441,
0.1316893994808197,
0.08565124869346619,
0.06437213718891144,
0.019635362550616264,
0.020543145015835762,
0.02446223795413971,
0.029573822394013405,
-0.020725078880786896,
-0.08386499434709549,
-0.01673775352537632,
0.14245417714118958,
0.024977976456284523,
0.0360303595662117,
0.0035881029907613993,
-0.03668458014726639,
0.06648463755846024,
0.16771891713142395,
0.012426458299160004,
-0.23092281818389893,
-0.06617839634418488,
0.0744035392999649,
-0.06997847557067871,
-0.11378128826618195,
-0.011021395213901997,
0.03138574957847595,
-0.1820969581604004,
0.04344824329018593,
-0.02262875810265541,
0.10113118588924408,
-0.11110582947731018,
-0.025930341333150864,
0.038102298974990845,
0.06534357368946075,
-0.03762030974030495,
0.07928402721881866,
-0.20931516587734222,
0.14360107481479645,
0.006824537646025419,
0.0662476122379303,
-0.1070002019405365,
0.08016478270292282,
0.02166800945997238,
0.003962947987020016,
0.1639549285173416,
-0.006579355336725712,
-0.08173416554927826,
-0.08530484884977341,
-0.07792343199253082,
-0.01293096225708723,
0.09758622944355011,
-0.1083054319024086,
0.08659143000841141,
-0.007826665416359901,
-0.031724635511636734,
-0.002251977799460292,
-0.11942502856254578,
-0.13174815475940704,
-0.1834975630044937,
0.04998116195201874,
-0.11704818159341812,
0.03951563686132431,
-0.11107321828603745,
-0.061925847083330154,
-0.03264854475855827,
0.19538627564907074,
-0.1968487948179245,
-0.08019812405109406,
-0.14597393572330475,
-0.0734013020992279,
0.11603168398141861,
-0.04134324938058853,
0.07953549921512604,
0.003630653489381075,
0.2027774453163147,
-0.0049596261233091354,
0.0008760678465478122,
0.0867764800786972,
-0.09453176707029343,
-0.2061263471841812,
-0.095054991543293,
0.13693954050540924,
0.12683691084384918,
0.045141350477933884,
-0.001808035303838551,
0.022449098527431488,
-0.003543051891028881,
-0.10818912833929062,
0.030619105324149132,
0.1492711752653122,
0.09657935798168182,
0.0439409464597702,
-0.025595664978027344,
-0.14485116302967072,
-0.10286740958690643,
-0.05655559152364731,
0.017665131017565727,
0.183364138007164,
-0.06921800225973129,
0.16631482541561127,
0.15526814758777618,
-0.06590348482131958,
-0.20979635417461395,
0.03614744171500206,
0.03232418745756149,
-0.008261965587735176,
0.03450188785791397,
-0.20602507889270782,
0.07341582328081131,
0.01584669202566147,
-0.057912733405828476,
0.13327911496162415,
-0.17143049836158752,
-0.1488124579191208,
0.08935882896184921,
0.0777517780661583,
-0.211406409740448,
-0.13250647485256195,
-0.09561139345169067,
-0.05425233766436577,
-0.1049741804599762,
0.08951318264007568,
-0.00034212550963275135,
0.00718674948439002,
0.035794805735349655,
0.02204621024429798,
0.017388639971613884,
-0.05294845998287201,
0.19216622412204742,
-0.0027261620853096247,
0.04696741700172424,
-0.08073770254850388,
-0.08108212053775787,
0.03557400405406952,
-0.06683739274740219,
0.0830080509185791,
-0.023084791377186775,
0.004679036792367697,
-0.11573878675699234,
-0.06346318125724792,
-0.04972510412335396,
0.03276039659976959,
-0.08775602281093597,
-0.09446880221366882,
-0.05614441633224487,
0.10359876602888107,
0.08980697393417358,
-0.03496086969971657,
-0.061577584594488144,
-0.09710686653852463,
0.07230715453624725,
0.22115971148014069,
0.18863879144191742,
0.07647473365068436,
-0.0675395280122757,
-0.0029117604717612267,
-0.023906316608190536,
0.054080698639154434,
-0.2153206318616867,
0.044365670531988144,
0.03903084248304367,
0.03029310703277588,
0.13431638479232788,
-0.023942356929183006,
-0.16170847415924072,
-0.044563427567481995,
0.06042470782995224,
-0.0675726905465126,
-0.15454024076461792,
0.003008861094713211,
0.08846548199653625,
-0.16023053228855133,
-0.053557612001895905,
0.02649034932255745,
-0.0350460447371006,
-0.027545088902115822,
0.0010348923970013857,
0.08224344998598099,
0.02413026988506317,
0.11039666831493378,
0.07115603238344193,
0.11174404621124268,
-0.10106326639652252,
0.08730421215295792,
0.0898137167096138,
-0.10840801149606705,
0.03910919278860092,
0.06986082345247269,
-0.06326653808355331,
-0.0329873263835907,
0.031108932569622993,
0.08609528094530106,
0.02910410612821579,
-0.07589155435562134,
-0.0003110404359176755,
-0.10957848280668259,
0.06651712954044342,
0.14283788204193115,
0.03319181129336357,
0.007173929363489151,
0.04601660370826721,
0.029003625735640526,
-0.10390203446149826,
0.11265801638364792,
0.041551440954208374,
0.0359308086335659,
-0.04995269328355789,
0.003519897349178791,
0.04222669079899788,
-0.011379980482161045,
-0.015903322026133537,
-0.039394035935401917,
-0.06654489040374756,
-0.009303842671215534,
-0.1557016372680664,
0.02798696979880333,
-0.0701727494597435,
0.008773589506745338,
0.016523631289601326,
-0.02984452061355114,
0.00421373825520277,
0.010279021225869656,
-0.07592474669218063,
-0.03921252489089966,
-0.00597892701625824,
0.10799336433410645,
-0.16164712607860565,
0.008396364748477936,
0.0864807590842247,
-0.12359115481376648,
0.0785754844546318,
-0.004872123710811138,
-0.008228191174566746,
0.016140099614858627,
-0.14468397200107574,
0.06475566327571869,
-0.009952590800821781,
0.002042346866801381,
0.02179797925055027,
-0.20463575422763824,
0.004303002264350653,
-0.04947707802057266,
-0.05808398127555847,
-0.0054858666844666,
-0.0371098592877388,
-0.11110534518957138,
0.10392162203788757,
0.016078852117061615,
-0.08070345222949982,
-0.01963779143989086,
0.054263342171907425,
0.11082571744918823,
-0.05670001730322838,
0.1372474879026413,
-0.0212235189974308,
0.05716216191649437,
-0.17756865918636322,
-0.01764446683228016,
-0.017806867137551308,
0.013603313826024532,
-0.03553657978773117,
-0.008313823491334915,
0.05301200598478317,
-0.016943611204624176,
0.2223619669675827,
-0.021118339151144028,
0.024224206805229187,
0.06494368612766266,
0.003768373280763626,
-0.014263768680393696,
0.09235488623380661,
0.04585632309317589,
0.016036443412303925,
0.01857670024037361,
0.017341768369078636,
-0.043770886957645416,
-0.012837734073400497,
-0.12971961498260498,
0.0822075679898262,
0.16555102169513702,
0.08159321546554565,
-0.0025153302121907473,
0.04910283908247948,
-0.11745497584342957,
-0.08390354365110397,
0.09864985197782516,
-0.03327968344092369,
-0.010939104482531548,
-0.05545666068792343,
0.1393168866634369,
0.15790824592113495,
-0.18001005053520203,
0.06630969792604446,
-0.06947531551122665,
-0.05813371390104294,
-0.10700751096010208,
-0.17336682975292206,
-0.06371892243623734,
-0.032686252146959305,
-0.007239010650664568,
-0.06101612001657486,
0.06587019562721252,
0.10927599668502808,
0.011966768652200699,
0.0038074126932770014,
0.08745987713336945,
-0.03549729287624359,
0.0032165138982236385,
0.04436643421649933,
0.053415216505527496,
0.018384460359811783,
-0.06577059626579285,
0.0062953270971775055,
0.005856741219758987,
0.03797626122832298,
0.05783457309007645,
0.030920613557100296,
-0.010121258907020092,
0.010502477176487446,
-0.017619112506508827,
-0.09953264892101288,
0.040120288729667664,
-0.02619595266878605,
-0.04532664641737938,
0.14967231452465057,
0.019792752340435982,
-0.0015246463008224964,
-0.020656753331422806,
0.23401693999767303,
-0.06644817441701889,
-0.07621476799249649,
-0.139224112033844,
0.14507612586021423,
-0.04275493323802948,
0.049066413193941116,
0.04825753718614578,
-0.10474846512079239,
0.03731349855661392,
0.1451096534729004,
0.14759132266044617,
-0.028853775933384895,
0.008417275734245777,
0.010314252227544785,
0.004880689084529877,
-0.023987766355276108,
0.05473437160253525,
0.04762427508831024,
0.1143009290099144,
-0.06655429303646088,
0.0966365858912468,
-0.004376910626888275,
-0.08667857199907303,
-0.01979442685842514,
0.1345251053571701,
0.00039378736983053386,
0.02452908828854561,
-0.07970208674669266,
0.11903424561023712,
-0.06584613770246506,
-0.254661500453949,
0.06372799724340439,
-0.06844745576381683,
-0.15212500095367432,
-0.020589474588632584,
0.0192123856395483,
-0.00045748791308142245,
0.021751869469881058,
0.06367123872041702,
-0.060426320880651474,
0.1489165723323822,
0.03729896992444992,
-0.07695354521274567,
-0.07863835990428925,
0.07556982338428497,
-0.08442215621471405,
0.3028976023197174,
0.005024306941777468,
0.04913705214858055,
0.09561952203512192,
-0.03656364977359772,
-0.13419130444526672,
0.03908977657556534,
0.0922386571764946,
-0.052813757210969925,
0.06462136656045914,
0.20481634140014648,
-0.011546511203050613,
0.11757413297891617,
0.07377773523330688,
-0.08157313615083694,
0.05285948887467384,
-0.08638615161180496,
-0.0914662778377533,
-0.09036537259817123,
0.09207529574632645,
-0.060405224561691284,
0.1555742472410202,
0.13106994330883026,
-0.046590786427259445,
0.0021229966077953577,
-0.02774938941001892,
0.05171699821949005,
-0.0023280514869838953,
0.1081797182559967,
0.027423536404967308,
-0.1910448670387268,
0.03156386315822601,
-0.01286277174949646,
0.09995201975107193,
-0.2582627832889557,
-0.0796453133225441,
0.039762452244758606,
-0.01255226694047451,
-0.05679939687252045,
0.12189354747533798,
0.05551491677761078,
0.05127199366688728,
-0.055598754435777664,
-0.0537382997572422,
-0.00422793859615922,
0.16415901482105255,
-0.1055506095290184,
-0.00439038872718811
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | token-classification | sahillihas/G2-finetuned-ner | [
"transformers",
"safetensors",
"distilbert",
"token-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T17:57:59+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
49,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07577246427536011,
0.160971537232399,
-0.0037754587829113007,
0.025772565975785255,
0.11726456880569458,
0.008670859038829803,
0.07505112886428833,
0.10672148317098618,
-0.015180999413132668,
0.12574146687984467,
0.037852395325899124,
0.1016918197274208,
0.10858695209026337,
0.19223041832447052,
-0.0040494236163794994,
-0.20760315656661987,
0.0604725256562233,
-0.11741560697555542,
0.010131745599210262,
0.12369410693645477,
0.13811981678009033,
-0.11246351152658463,
0.06881340593099594,
-0.04209144040942192,
-0.01980483904480934,
-0.035129256546497345,
-0.06221386045217514,
-0.052587710320949554,
0.06488312780857086,
0.05438476800918579,
0.06559453904628754,
0.021042533218860626,
0.0840018093585968,
-0.28357821702957153,
0.019257022067904472,
0.0766754001379013,
0.002646347275003791,
0.06359819322824478,
0.07652021199464798,
-0.07693555951118469,
0.09431595355272293,
-0.059984829276800156,
0.15352454781532288,
0.07422540336847305,
-0.09718401730060577,
-0.18205241858959198,
-0.08698616921901703,
0.10743562132120132,
0.1786193698644638,
0.05337109789252281,
-0.03495080769062042,
0.14324238896369934,
-0.07000283896923065,
0.017874419689178467,
0.06666957587003708,
-0.0738137811422348,
-0.05611572042107582,
0.06171523407101631,
0.07419267296791077,
0.09766413271427155,
-0.13268914818763733,
-0.009328355081379414,
0.039757777005434036,
0.01763407699763775,
0.11028990149497986,
0.017873499542474747,
0.12974926829338074,
0.03081592544913292,
-0.1462060660123825,
-0.059808772057294846,
0.10660259425640106,
0.03442228585481644,
-0.05801061540842056,
-0.24857386946678162,
-0.00732324318960309,
-0.03398886322975159,
-0.03147710859775543,
-0.045992717146873474,
0.04087558388710022,
-0.027321036905050278,
0.08885861188173294,
0.0030806211289018393,
-0.06717013567686081,
-0.05033329501748085,
0.09465020149946213,
0.06679657101631165,
0.026782603934407234,
-0.028168529272079468,
0.01867647096514702,
0.11993368715047836,
0.10629972815513611,
-0.11282733827829361,
-0.06030019000172615,
-0.06428343802690506,
-0.08693956583738327,
-0.04836411029100418,
0.03905215859413147,
0.0691521167755127,
0.05319429561495781,
0.20513659715652466,
-0.005697339307516813,
0.04798254370689392,
0.03319632261991501,
0.012943265959620476,
0.06999380886554718,
0.07388506829738617,
-0.057592663913965225,
-0.13561639189720154,
-0.032188303768634796,
0.1180354431271553,
0.0034096194431185722,
-0.029926057904958725,
-0.03362307325005531,
0.061736419796943665,
0.04901990294456482,
0.12567868828773499,
0.06872507929801941,
0.016199197620153427,
-0.07887519896030426,
-0.05266522243618965,
0.18039943277835846,
-0.15842315554618835,
0.026145054027438164,
0.01778857596218586,
-0.04808545112609863,
-0.03221158683300018,
0.01776694506406784,
0.010511110536754131,
-0.027442574501037598,
0.09513456374406815,
-0.064456507563591,
-0.0437372550368309,
-0.10800766944885254,
-0.05169657990336418,
0.03205830603837967,
-0.02027302049100399,
-0.02823185920715332,
-0.039158232510089874,
-0.12495775520801544,
-0.0780569463968277,
0.07112475484609604,
-0.06458895653486252,
-0.06230264902114868,
-0.035103172063827515,
-0.06434695422649384,
0.01328826043754816,
-0.0016070485580712557,
0.11751557141542435,
-0.02924034371972084,
0.05098489671945572,
-0.058480363339185715,
0.06735116988420486,
0.13876435160636902,
0.030480407178401947,
-0.06721892952919006,
0.06733744591474533,
-0.21017982065677643,
0.10650269687175751,
-0.0903177410364151,
0.030309993773698807,
-0.16417838633060455,
-0.020649932324886322,
0.03423565998673439,
0.03357144445180893,
-0.011650003492832184,
0.14477767050266266,
-0.17939403653144836,
-0.035702068358659744,
0.18933741748332977,
-0.12515437602996826,
-0.092041015625,
0.05768411606550217,
-0.06046192720532417,
0.13482485711574554,
0.05646788328886032,
-0.021687285974621773,
0.05586513131856918,
-0.13862864673137665,
-0.02437570132315159,
-0.06216869503259659,
-0.014887474477291107,
0.1528262197971344,
0.059848666191101074,
-0.046563770622015,
0.02752561867237091,
0.017916126176714897,
-0.026528963819146156,
-0.04967709258198738,
-0.03546646237373352,
-0.09774158149957657,
0.00822865217924118,
-0.07969193160533905,
0.02241206355392933,
-0.022354070097208023,
-0.09176263213157654,
-0.03786236792802811,
-0.15299317240715027,
0.01120391022413969,
0.09777288138866425,
-0.0018325879937037826,
-0.030514128506183624,
-0.09985145181417465,
-0.0009975290158763528,
0.014818643219769001,
-0.006299570668488741,
-0.15255902707576752,
-0.05487200245261192,
0.02453494630753994,
-0.16442841291427612,
0.027572009712457657,
-0.04693130776286125,
0.03860992193222046,
0.04300114884972572,
-0.04640088230371475,
-0.03466949611902237,
0.020573051646351814,
0.021380405873060226,
-0.02527492307126522,
-0.25819748640060425,
-0.015081126242876053,
-0.050718042999506,
0.1735834777355194,
-0.24959886074066162,
0.04765407368540764,
0.0635317862033844,
0.12164347618818283,
0.008648045361042023,
-0.043875813484191895,
0.03864365443587303,
-0.0540064312517643,
-0.03614342212677002,
-0.06847652792930603,
-0.006085617933422327,
-0.03506658226251602,
-0.04601104184985161,
0.036002401262521744,
-0.17769227921962738,
-0.026747822761535645,
0.11292517185211182,
0.07372257113456726,
-0.16646650433540344,
-0.07437358796596527,
-0.03605179488658905,
-0.05963003262877464,
-0.07967717945575714,
-0.05352560430765152,
0.08457031846046448,
0.047888897359371185,
0.050600916147232056,
-0.06766360253095627,
-0.05917840451002121,
0.012786190956830978,
-0.014223156496882439,
-0.030130434781312943,
0.08920878916978836,
0.06863047182559967,
-0.13044358789920807,
0.10779033601284027,
0.07565328478813171,
0.07318258285522461,
0.10535803437232971,
0.004109293222427368,
-0.09151213616132736,
-0.019049784168601036,
0.030527541413903236,
0.01623775251209736,
0.14909477531909943,
-0.058906275779008865,
0.037276990711688995,
0.042408693581819534,
-0.025859607383608818,
0.007144377566874027,
-0.09663669019937515,
0.020623620599508286,
0.027944007888436317,
-0.010041053406894207,
0.02522680163383484,
-0.05456255003809929,
0.016582027077674866,
0.10647457838058472,
0.0315246656537056,
0.029977189376950264,
0.01468842476606369,
-0.04244951531291008,
-0.1256953328847885,
0.17814458906650543,
-0.09753371775150299,
-0.24721357226371765,
-0.12567918002605438,
-0.009092925116419792,
0.03918062895536423,
-0.012046195566654205,
0.01987382024526596,
-0.05762042477726936,
-0.10939344018697739,
-0.1005193367600441,
0.03179878741502762,
0.06615057587623596,
-0.08552238345146179,
-0.0674377977848053,
0.05327526107430458,
0.04219662770628929,
-0.1277419775724411,
0.020916761830449104,
0.041558679193258286,
-0.07218944281339645,
0.008345745503902435,
0.056684307754039764,
0.07901664078235626,
0.1805250197649002,
0.011102218180894852,
-0.018453268334269524,
0.012500734999775887,
0.21901123225688934,
-0.1455126702785492,
0.09525687992572784,
0.13985903561115265,
-0.06373906135559082,
0.08310941606760025,
0.20266622304916382,
0.03211383894085884,
-0.10105792433023453,
0.03751940280199051,
0.03433764725923538,
-0.03450850397348404,
-0.24166366457939148,
-0.07809934765100479,
0.0032382020726799965,
-0.06378763914108276,
0.10437320917844772,
0.08564607799053192,
0.10547725111246109,
0.04683990776538849,
-0.11285781115293503,
-0.06474319845438004,
0.05201273784041405,
0.11905202269554138,
-0.028264760971069336,
-0.0015284299151971936,
0.09661765396595001,
-0.023179147392511368,
0.024211812764406204,
0.091147281229496,
0.024589717388153076,
0.18455083668231964,
0.04622681811451912,
0.13505716621875763,
0.0910676121711731,
0.06324100494384766,
0.014886823482811451,
0.021261626854538918,
0.01765025220811367,
0.02958594262599945,
-0.018347864970564842,
-0.08319743722677231,
-0.009293846786022186,
0.13596801459789276,
0.024112429469823837,
0.04030708968639374,
0.006161263212561607,
-0.048370931297540665,
0.07493776082992554,
0.17692242562770844,
0.014188401401042938,
-0.22553029656410217,
-0.06700198352336884,
0.07436924427747726,
-0.07410255074501038,
-0.11821255087852478,
-0.01875862292945385,
0.02883634716272354,
-0.18362410366535187,
0.03783649206161499,
-0.027923477813601494,
0.10164659470319748,
-0.11413569748401642,
-0.021443994715809822,
0.03836014121770859,
0.05620839074254036,
-0.032449860125780106,
0.07187992334365845,
-0.18838849663734436,
0.13979743421077728,
0.009049986489117146,
0.06722430139780045,
-0.1007412001490593,
0.08134089410305023,
0.017924414947628975,
0.005374805070459843,
0.16363629698753357,
-0.0033102943561971188,
-0.061859820038080215,
-0.10252897441387177,
-0.08680899441242218,
-0.015643585473299026,
0.09927050769329071,
-0.12645161151885986,
0.09391359239816666,
-0.00760313356295228,
-0.03509579226374626,
-0.0025038165040314198,
-0.13591554760932922,
-0.13381120562553406,
-0.17948365211486816,
0.04688849300146103,
-0.12670443952083588,
0.04449006915092468,
-0.1076912134885788,
-0.056078001856803894,
-0.041555143892765045,
0.19199959933757782,
-0.21791574358940125,
-0.08265700191259384,
-0.15180857479572296,
-0.06470925360918045,
0.11797353625297546,
-0.04333622008562088,
0.08231177926063538,
0.009503866545855999,
0.19813112914562225,
-0.0016607206780463457,
-0.004083174746483564,
0.0959770604968071,
-0.09965015947818756,
-0.2107795774936676,
-0.09851331263780594,
0.13872341811656952,
0.13654115796089172,
0.04202498495578766,
0.0035957242362201214,
0.026746563613414764,
-0.003834358649328351,
-0.1127103865146637,
0.03198874741792679,
0.15448687970638275,
0.11144908517599106,
0.038192201405763626,
-0.028366046026349068,
-0.13728831708431244,
-0.10052017867565155,
-0.04788326099514961,
0.009005418047308922,
0.19273622334003448,
-0.07073298841714859,
0.1647789627313614,
0.1582319587469101,
-0.06154803931713104,
-0.2109813690185547,
0.03008965775370598,
0.03392798826098442,
-0.003125392831861973,
0.04911292716860771,
-0.2020125389099121,
0.08147795498371124,
0.014267290011048317,
-0.05861039087176323,
0.12597185373306274,
-0.17893800139427185,
-0.1463114470243454,
0.0894082561135292,
0.07762204855680466,
-0.19969920814037323,
-0.1310804784297943,
-0.09587812423706055,
-0.04523595795035362,
-0.09975142776966095,
0.09141786396503448,
-0.006663811393082142,
0.00584616232663393,
0.03217587620019913,
0.01619444414973259,
0.016584929078817368,
-0.04955518618226051,
0.19393765926361084,
-0.0010097876656800508,
0.049837395548820496,
-0.07284197211265564,
-0.07579720765352249,
0.03499779477715492,
-0.07016132026910782,
0.08607465773820877,
-0.017582224681973457,
0.006899731699377298,
-0.11493121087551117,
-0.06381088495254517,
-0.04385467618703842,
0.0321214497089386,
-0.08662021160125732,
-0.09713311493396759,
-0.04749152809381485,
0.105962373316288,
0.0914415493607521,
-0.03657814860343933,
-0.06360591948032379,
-0.0900130569934845,
0.052277419716119766,
0.22165511548519135,
0.18128447234630585,
0.0717686340212822,
-0.07471393048763275,
-0.008312922902405262,
-0.021178819239139557,
0.05892517790198326,
-0.20861060917377472,
0.049327246844768524,
0.03792690858244896,
0.03388948366045952,
0.12989014387130737,
-0.025939077138900757,
-0.16148367524147034,
-0.052209001034498215,
0.05540921539068222,
-0.07436220347881317,
-0.16007542610168457,
0.006253175903111696,
0.08246254920959473,
-0.1577574759721756,
-0.040928490459918976,
0.0384182333946228,
-0.029118487611413002,
-0.031075680628418922,
0.0023355213925242424,
0.08367416262626648,
0.021480288356542587,
0.10684667527675629,
0.06465408951044083,
0.10775714367628098,
-0.10543544590473175,
0.06850111484527588,
0.08449427038431168,
-0.10566192865371704,
0.03645234927535057,
0.059928081929683685,
-0.06519834697246552,
-0.03475077450275421,
0.03517557680606842,
0.08707848936319351,
0.02939283289015293,
-0.06929907202720642,
0.006287590134888887,
-0.10996777564287186,
0.06547177582979202,
0.13484768569469452,
0.04042261838912964,
0.011329763568937778,
0.040748391300439835,
0.030845293775200844,
-0.09934883564710617,
0.12121784687042236,
0.05189082771539688,
0.03800792247056961,
-0.0530884675681591,
-0.017324361950159073,
0.03913010656833649,
-0.019996602088212967,
-0.017436852678656578,
-0.03976694867014885,
-0.07124674320220947,
-0.011837370693683624,
-0.16637970507144928,
0.020963486284017563,
-0.06557925045490265,
0.011502030305564404,
0.015106898732483387,
-0.03105640783905983,
0.00648743100464344,
0.012552679516375065,
-0.07388464361429214,
-0.045671455562114716,
-0.00442871730774641,
0.10947413742542267,
-0.16812850534915924,
0.008818828500807285,
0.08374110609292984,
-0.12818290293216705,
0.08439380675554276,
0.0020208086352795362,
-0.0032075955532491207,
0.02065693773329258,
-0.1317322701215744,
0.05960160121321678,
-0.006254117004573345,
0.006569796707481146,
0.03298337012529373,
-0.21374841034412384,
0.0022494937293231487,
-0.05006609857082367,
-0.06266371160745621,
-0.0007477401522919536,
-0.03451014682650566,
-0.11394216865301132,
0.10471237450838089,
0.014338361099362373,
-0.07764562219381332,
-0.021158380433917046,
0.048064179718494415,
0.10854631662368774,
-0.04941081255674362,
0.1473926305770874,
-0.01730068400502205,
0.059951625764369965,
-0.1836392730474472,
-0.02037329412996769,
-0.018616020679473877,
0.016425784677267075,
-0.037237949669361115,
-0.006127772852778435,
0.051674678921699524,
-0.01703222282230854,
0.2173253446817398,
-0.022195780649781227,
0.026851661503314972,
0.06280741095542908,
-0.002322237938642502,
-0.013443386182188988,
0.09835761785507202,
0.046587999910116196,
0.013757294043898582,
0.022333940491080284,
0.010798705741763115,
-0.04187694564461708,
-0.00982733629643917,
-0.13689175248146057,
0.07585512101650238,
0.16693128645420074,
0.08301392197608948,
-0.00876491330564022,
0.04851881042122841,
-0.10903714597225189,
-0.10301291197538376,
0.09611163288354874,
-0.03860481455922127,
-0.017003152519464493,
-0.050830189138650894,
0.13978728652000427,
0.15471777319908142,
-0.19259576499462128,
0.06503742933273315,
-0.0670342966914177,
-0.05641358345746994,
-0.10491608083248138,
-0.17902258038520813,
-0.06081055849790573,
-0.037293195724487305,
-0.01373780332505703,
-0.06118905171751976,
0.06020566076040268,
0.10252559930086136,
0.014830772764980793,
0.008825747296214104,
0.082810178399086,
-0.027097169309854507,
0.007841658778488636,
0.04027826339006424,
0.06163087487220764,
0.015406538732349873,
-0.06096850335597992,
0.008726970292627811,
0.0014340170891955495,
0.03696046769618988,
0.053261034190654755,
0.03251050412654877,
-0.012918577529489994,
0.006573197897523642,
-0.023761101067066193,
-0.10426638275384903,
0.04047070071101189,
-0.02524152584373951,
-0.052286889404058456,
0.15217694640159607,
0.022632325068116188,
-0.005827770568430424,
-0.022215187549591064,
0.2338925302028656,
-0.06722013652324677,
-0.0785137191414833,
-0.14054499566555023,
0.14897066354751587,
-0.03910091519355774,
0.05552518740296364,
0.046295974403619766,
-0.105179063975811,
0.03945622220635414,
0.13950063288211823,
0.14286132156848907,
-0.04025404527783394,
0.012145495042204857,
0.009853771887719631,
0.0038128900341689587,
-0.028584260493516922,
0.052456315606832504,
0.05106692388653755,
0.12483679503202438,
-0.06253138184547424,
0.09782347828149796,
-0.006969998124986887,
-0.09410524368286133,
-0.025307200849056244,
0.13257914781570435,
0.002008436480537057,
0.023570919409394264,
-0.0815717875957489,
0.12813320755958557,
-0.05384290590882301,
-0.2535746693611145,
0.07024679332971573,
-0.06422019749879837,
-0.15186867117881775,
-0.019949471578001976,
0.020098339766263962,
-0.004076872952282429,
0.023379439488053322,
0.06391935795545578,
-0.0649435967206955,
0.15813666582107544,
0.03603892773389816,
-0.06333170086145401,
-0.07421127706766129,
0.07901927083730698,
-0.07848652452230453,
0.3060082495212555,
0.007198201958090067,
0.05301123857498169,
0.09581229835748672,
-0.03854295611381531,
-0.13927392661571503,
0.028419770300388336,
0.09097424894571304,
-0.05137800797820091,
0.061749134212732315,
0.2061481475830078,
-0.010625248774886131,
0.11590193957090378,
0.07372608780860901,
-0.08649783581495285,
0.04505147039890289,
-0.09321032464504242,
-0.09480153769254684,
-0.08939364552497864,
0.09279339760541916,
-0.055773381143808365,
0.1508018523454666,
0.1253325492143631,
-0.048127513378858566,
0.008756249211728573,
-0.019953599199652672,
0.058552540838718414,
0.0011799887288361788,
0.11399417370557785,
0.028718046844005585,
-0.19564524292945862,
0.02838771604001522,
-0.0018802561098709702,
0.1010337695479393,
-0.23878240585327148,
-0.0895518958568573,
0.04254518821835518,
-0.0008608286152593791,
-0.060450565069913864,
0.12522582709789276,
0.05044018104672432,
0.043433722108602524,
-0.0508544035255909,
-0.04669932648539543,
-0.008167625404894352,
0.1615993231534958,
-0.10094329714775085,
-0.0017090580658987164
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-AWQ | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | 2024-02-08T18:01:25+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
197,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.07166692614555359,
-0.01856779120862484,
-0.004747839644551277,
0.041673798114061356,
0.027521833777427673,
0.04779491946101189,
0.1371924877166748,
0.07062480598688126,
0.08466701954603195,
0.08759720623493195,
-0.0507846362888813,
-0.02669893018901348,
0.06949843466281891,
0.09128747880458832,
-0.00007204919529613107,
-0.24146710336208344,
0.07936806231737137,
-0.04216822609305382,
-0.031431857496500015,
0.03770412504673004,
0.1437464952468872,
-0.050008922815322876,
0.05366573855280876,
-0.004730504006147385,
-0.0562172532081604,
0.036363597959280014,
-0.024722712114453316,
-0.01414013747125864,
0.10011827945709229,
0.0568009689450264,
0.050431590527296066,
0.018151763826608658,
-0.01583727076649666,
-0.1574489176273346,
0.0017065659631043673,
-0.021783573552966118,
-0.05198531597852707,
0.026153123006224632,
0.1145445853471756,
-0.04289707913994789,
0.10300558805465698,
-0.12474972754716873,
-0.015136501751840115,
0.07512025535106659,
-0.12584854662418365,
-0.12048150599002838,
-0.20401285588741302,
0.14822199940681458,
0.07863311469554901,
0.004033535253256559,
-0.03707587718963623,
0.0748719796538353,
-0.060603465884923935,
0.043737925589084625,
0.11193779855966568,
-0.2542112171649933,
-0.04448657110333443,
0.1809338629245758,
0.011578950099647045,
-0.05022599175572395,
-0.05205680802464485,
0.06751102209091187,
0.022408317774534225,
0.007748056203126907,
-0.04819576069712639,
-0.04811180755496025,
0.23282013833522797,
-0.04577168449759483,
-0.13417668640613556,
-0.025770103558897972,
0.09419859945774078,
0.09597891569137573,
-0.044829174876213074,
-0.16251683235168457,
-0.006201127078384161,
-0.01078060083091259,
-0.04928068816661835,
-0.0864228680729866,
-0.0017060514073818922,
-0.02288738638162613,
0.08867747336626053,
-0.028266215696930885,
-0.035385385155677795,
0.06508079916238785,
-0.043290480971336365,
0.16347505152225494,
-0.0023628456983715296,
-0.03455541655421257,
0.043034449219703674,
0.025033365935087204,
-0.20190224051475525,
-0.09658104926347733,
-0.02915882132947445,
-0.05415399372577667,
-0.12556849420070648,
-0.04047897085547447,
-0.04969988390803337,
-0.07777149230241776,
-0.003943731542676687,
0.2353157252073288,
-0.024604419246315956,
0.027479270473122597,
0.028405744582414627,
0.013507463969290257,
0.0377434566617012,
0.12795132398605347,
-0.048294804990291595,
-0.18439054489135742,
-0.029752491042017937,
-0.019375726580619812,
0.07812777161598206,
0.01373204030096531,
-0.032615479081869125,
-0.007996312342584133,
0.019229818135499954,
0.031586904078722,
0.018543295562267303,
0.03860402852296829,
-0.07945998758077621,
-0.02549120970070362,
0.13152454793453217,
-0.09707082808017731,
0.041251376271247864,
-0.02637331560254097,
-0.03593601658940315,
0.08585458993911743,
-0.026036109775304794,
0.021218789741396904,
-0.01904943957924843,
0.15953503549098969,
-0.0346175841987133,
-0.02231231890618801,
-0.02017483115196228,
-0.054505497217178345,
0.06332291662693024,
-0.07406073063611984,
-0.0570487417280674,
-0.10050495713949203,
-0.05141746997833252,
-0.08122805505990982,
0.031792040914297104,
-0.050462592393159866,
-0.0021242413204163313,
-0.0016031115083023906,
-0.04996504634618759,
0.02073086053133011,
0.038508981466293335,
-0.0597531832754612,
-0.0012625460512936115,
-0.053792502731084824,
-0.02438385784626007,
0.02440471760928631,
-0.0593913197517395,
0.023985227569937706,
-0.10469114780426025,
0.050482504069805145,
-0.3242761492729187,
0.08415523171424866,
-0.12036149203777313,
0.017790859565138817,
-0.13528020679950714,
-0.03617624565958977,
0.02661304362118244,
0.007084574084728956,
0.01803712360560894,
0.12879891693592072,
-0.15671409666538239,
-0.0918942466378212,
0.10375721752643585,
-0.13387979567050934,
-0.05863671377301216,
0.1258038729429245,
0.007876512594521046,
-0.011172925122082233,
0.07438617199659348,
0.14570368826389313,
0.13417822122573853,
-0.05062301456928253,
-0.11208087205886841,
-0.08211597055196762,
0.031422507017850876,
0.16398058831691742,
0.07695598900318146,
-0.057796064764261246,
0.010089975781738758,
0.02844400703907013,
-0.04282980412244797,
-0.017284182831645012,
-0.006598720792680979,
-0.05812206491827965,
-0.029946424067020416,
-0.009356892667710781,
0.09093672782182693,
-0.005748063791543245,
0.004233375657349825,
-0.07724923640489578,
-0.09976086020469666,
0.04016620293259621,
0.08392030745744705,
0.03062279336154461,
0.013633833266794682,
-0.0478052981197834,
0.09870649874210358,
-0.012274181470274925,
0.020570868626236916,
-0.10519386827945709,
-0.17221181094646454,
0.039511486887931824,
-0.17834673821926117,
0.05988134443759918,
0.04895787313580513,
0.05827634781599045,
0.02942204289138317,
-0.0929659754037857,
-0.06262516975402832,
-0.03823966160416603,
0.010320470668375492,
0.010905236005783081,
-0.12763884663581848,
-0.079424649477005,
-0.03823478892445564,
0.21644505858421326,
-0.05969979614019394,
-0.009742994792759418,
0.05346892029047012,
0.2567080557346344,
0.01481753122061491,
-0.010963077656924725,
-0.05421293154358864,
0.06589116156101227,
0.0013430870603770018,
0.013423752039670944,
0.026568399742245674,
-0.04893876612186432,
-0.13044172525405884,
0.0778152197599411,
-0.054626964032649994,
-0.019297393038868904,
0.06781565397977829,
0.05149891600012779,
-0.0892624482512474,
-0.021050142124295235,
-0.04410421848297119,
-0.026380112394690514,
0.006166961509734392,
-0.12192948907613754,
0.03333337604999542,
0.04112009331583977,
0.02553110383450985,
-0.03216144070029259,
-0.041846632957458496,
0.020595988258719444,
-0.0015515601262450218,
-0.13159988820552826,
0.10762070119380951,
-0.07231535017490387,
-0.1985868215560913,
0.10931753367185593,
0.15639609098434448,
0.07499983161687851,
0.1255371868610382,
-0.01865905337035656,
-0.019309168681502342,
-0.11882124096155167,
0.002003759378567338,
0.02907773293554783,
0.05062086507678032,
0.03181397169828415,
0.03428458422422409,
0.0335618332028389,
0.047900814563035965,
0.01767580956220627,
-0.027259889990091324,
0.031684510409832,
-0.006566612981259823,
0.007650402374565601,
0.11074890941381454,
0.10958029329776764,
0.016832789406180382,
0.0924636647105217,
0.040604762732982635,
-0.063636913895607,
0.019954746589064598,
-0.03644585236907005,
-0.07675156742334366,
0.09943929314613342,
-0.10312601923942566,
-0.16916556656360626,
-0.16307270526885986,
-0.05933355540037155,
-0.17813056707382202,
-0.029100004583597183,
0.036138057708740234,
-0.03930490463972092,
-0.07992920279502869,
-0.06851964443922043,
0.1346370428800583,
0.058466289192438126,
-0.054753366857767105,
-0.00986373983323574,
-0.04479872062802315,
0.04059246554970741,
-0.10420270264148712,
-0.04073433578014374,
-0.005826231557875872,
-0.04385700821876526,
0.008065430447459221,
0.02805226482450962,
0.0826081708073616,
0.06799381226301193,
0.010754055343568325,
-0.027217550203204155,
-0.030553758144378662,
0.19394133985042572,
-0.119123175740242,
0.11225245893001556,
0.06520520895719528,
-0.014087967574596405,
0.06482816487550735,
0.18299448490142822,
0.0718090608716011,
-0.051033902913331985,
-0.008411448448896408,
0.05225401371717453,
-0.03585704416036606,
-0.19407221674919128,
-0.18306823074817657,
-0.06369616836309433,
0.010265922173857689,
-0.015217662788927555,
0.04227472096681595,
0.07468270510435104,
0.0016454565338790417,
-0.08774523437023163,
-0.07763171941041946,
0.04026481509208679,
0.06102767586708069,
0.1757727414369583,
0.015612189657986164,
0.04541792720556259,
-0.06486240774393082,
-0.05733994022011757,
0.0940183624625206,
0.0063015129417181015,
0.06279605627059937,
0.07289133220911026,
0.06432249397039413,
0.03413931280374527,
-0.004464691039174795,
0.007793685421347618,
0.04280984401702881,
0.0046075014397501945,
-0.031769316643476486,
-0.009046059101819992,
-0.09013689309358597,
0.07866528630256653,
0.09145693480968475,
0.08355432748794556,
0.019883280619978905,
-0.07098124921321869,
-0.06529392302036285,
0.13076291978359222,
0.16756437718868256,
0.10196398943662643,
-0.12283069640398026,
-0.10278254747390747,
0.060479097068309784,
-0.014812919311225414,
0.031086120754480362,
-0.03404274210333824,
-0.04570675641298294,
-0.13509927690029144,
0.18957211077213287,
-0.020913276821374893,
0.05880563706159592,
0.007109012920409441,
0.00624396838247776,
0.004693005699664354,
0.06133235618472099,
0.01934615708887577,
0.03919937461614609,
-0.15104331076145172,
0.14347967505455017,
0.0034757587127387524,
0.018483491614460945,
-0.027245445176959038,
0.045628659427165985,
-0.0026385036762803793,
0.046254437416791916,
0.10560659319162369,
0.05201854184269905,
-0.11086054891347885,
-0.07126643508672714,
-0.026265688240528107,
-0.061014410108327866,
0.1471344232559204,
-0.09676079452037811,
0.07962802797555923,
-0.002087333705276251,
-0.045661117881536484,
-0.0564737431704998,
0.16169817745685577,
-0.14479972422122955,
-0.09653850644826889,
0.1114371195435524,
-0.044917576014995575,
0.01976080797612667,
-0.08063764125108719,
-0.05127454921603203,
-0.25548508763313293,
0.19047515094280243,
-0.0020668606739491224,
0.002891882322728634,
-0.10111324489116669,
-0.008823903277516365,
0.1456259936094284,
-0.09630976617336273,
0.05750128999352455,
-0.005270663183182478,
0.08375052362680435,
-0.09350375831127167,
-0.08390098810195923,
0.04129662364721298,
-0.08279844373464584,
-0.15130697190761566,
-0.00888698548078537,
0.2215437889099121,
0.024114584550261497,
0.033528972417116165,
0.025731200352311134,
0.08714527636766434,
0.10023680329322815,
-0.10970569401979446,
0.0801391750574112,
0.08529455959796906,
-0.024397820234298706,
0.13129834830760956,
-0.03969581797719002,
-0.14613623917102814,
-0.045773182064294815,
-0.04918927326798439,
0.10370392352342606,
0.32989591360092163,
-0.07563622295856476,
0.08599881827831268,
0.08223699033260345,
-0.026896363124251366,
-0.18759256601333618,
0.019910065457224846,
0.036614254117012024,
0.01737174391746521,
0.00885500106960535,
-0.061702512204647064,
0.11388777941465378,
0.1275063008069992,
-0.01975497044622898,
-0.04270234704017639,
-0.23315757513046265,
-0.12457060068845749,
0.004730218555778265,
-0.04686909541487694,
-0.025309395045042038,
-0.08713436126708984,
-0.10722968727350235,
-0.07478421926498413,
-0.18328820168972015,
0.07139819860458374,
0.04733116179704666,
0.03624711185693741,
0.011342871002852917,
-0.010099153034389019,
0.034775521606206894,
-0.021900873631238937,
0.1932504028081894,
0.09935834258794785,
0.016590164974331856,
-0.08487586677074432,
-0.056348420679569244,
0.10161729156970978,
-0.03657044842839241,
0.0970173254609108,
0.04091767221689224,
-0.006266639567911625,
-0.1131814494729042,
0.022327156737446785,
-0.10140369832515717,
0.11298160254955292,
-0.10201042145490646,
-0.02313254214823246,
-0.07965683937072754,
0.1435467004776001,
0.030004480853676796,
-0.016887271776795387,
0.12155310809612274,
-0.02601400949060917,
0.08929963409900665,
0.1376388669013977,
0.0676378682255745,
0.034696489572525024,
-0.09019912034273148,
-0.011523691937327385,
-0.05981307104229927,
0.01826051063835621,
0.040025800466537476,
0.043672844767570496,
0.13055400550365448,
-0.011025434359908104,
0.15035192668437958,
-0.01957176811993122,
-0.12762615084648132,
-0.029465116560459137,
0.11758231371641159,
-0.1211458146572113,
-0.2631201148033142,
-0.037575941532850266,
-0.04119125008583069,
0.016432154923677444,
0.07350212335586548,
0.20434686541557312,
-0.034057825803756714,
-0.02101954072713852,
0.01810077391564846,
0.052491091191768646,
-0.08820600807666779,
0.09010323882102966,
-0.017758112400770187,
0.04325520992279053,
-0.06467947363853455,
0.1095864549279213,
0.07124035805463791,
-0.20291148126125336,
-0.01677929237484932,
0.11272403597831726,
-0.06563327461481094,
-0.07275592535734177,
-0.1029103547334671,
0.12529356777668,
-0.148361936211586,
-0.008921816013753414,
-0.02332892268896103,
-0.13892322778701782,
0.04183046892285347,
0.11552514880895615,
0.039746250957250595,
0.002464479533955455,
0.03706410154700279,
-0.006167799234390259,
0.017322270199656487,
0.03622523322701454,
0.08276547491550446,
0.10242993384599686,
-0.049269236624240875,
0.024841003119945526,
-0.07733651995658875,
0.007323069032281637,
-0.017937319353222847,
-0.017314139753580093,
-0.12697076797485352,
-0.07650735974311829,
-0.09981229156255722,
-0.05792104825377464,
-0.07968639582395554,
-0.01957182213664055,
-0.03271479159593582,
-0.041927091777324677,
0.004599887877702713,
-0.023104215040802956,
-0.0676175206899643,
-0.039795275777578354,
-0.023995617404580116,
0.09295538067817688,
-0.11409192532300949,
0.002146559301763773,
0.058958664536476135,
-0.050302620977163315,
0.11067753285169601,
0.025820666924118996,
0.04491415247321129,
0.04857226833701134,
-0.04441788047552109,
-0.025481386110186577,
0.008861368522047997,
-0.008401160128414631,
0.05473468825221062,
-0.17071831226348877,
-0.013499503023922443,
-0.004324992187321186,
-0.05643303692340851,
0.04369330778717995,
-0.06834764778614044,
-0.10161814838647842,
0.006090106908231974,
0.0211369376629591,
-0.09219589829444885,
-0.07886380702257156,
0.02787606790661812,
0.08851932734251022,
0.027523431926965714,
0.11945012211799622,
-0.05355551466345787,
0.07242678105831146,
-0.1315307915210724,
-0.018528195098042488,
-0.016048824414610863,
-0.1018289253115654,
0.02828708291053772,
-0.002898699836805463,
0.020593682304024696,
-0.0057486859150230885,
0.11374329775571823,
-0.0168294794857502,
-0.02329329028725624,
0.0631520077586174,
-0.010374664328992367,
-0.08221432566642761,
0.058187808841466904,
0.07395084202289581,
0.04485054314136505,
-0.009556334465742111,
-0.08988865464925766,
-0.005175701342523098,
-0.055808551609516144,
-0.011543648317456245,
0.106855109333992,
0.1642477661371231,
0.10909119993448257,
0.04876912012696266,
0.1011897400021553,
-0.06141478568315506,
-0.03898105397820473,
-0.0001704300520941615,
-0.008491910062730312,
0.0921425074338913,
-0.04718974977731705,
0.04675251618027687,
0.06329746544361115,
-0.2306232899427414,
0.1440032571554184,
-0.052848752588033676,
0.00585526367649436,
-0.012586986646056175,
-0.13633717596530914,
-0.06504505127668381,
-0.05237798020243645,
0.009047366678714752,
-0.07830917090177536,
0.01411522924900055,
0.02100704424083233,
0.03929991275072098,
-0.024791916832327843,
0.13863351941108704,
-0.22034111618995667,
-0.019474877044558525,
0.12164667993783951,
0.013981896452605724,
-0.026152877137064934,
0.0496508963406086,
-0.019688468426465988,
-0.06726118922233582,
0.06631804257631302,
0.0052291033789515495,
0.04099326208233833,
0.024001702666282654,
-0.028310270980000496,
-0.04721837118268013,
-0.09266979247331619,
0.015445679426193237,
-0.015362463891506195,
-0.003832170506939292,
0.04202878847718239,
0.0322144478559494,
0.01984487473964691,
-0.026549549773335457,
0.11563486605882645,
-0.03820420801639557,
-0.08881393074989319,
-0.10545165836811066,
0.17544689774513245,
0.00585847906768322,
0.07692792266607285,
0.013110289350152016,
-0.11492262035608292,
0.008544947020709515,
0.0196751169860363,
0.26304495334625244,
0.05606941878795624,
0.035928186029195786,
0.022049689665436745,
0.01529813464730978,
0.039831340312957764,
0.023069337010383606,
-0.016041522845625877,
0.18279467523097992,
-0.021373625844717026,
0.20077632367610931,
-0.05168483778834343,
-0.04396864026784897,
0.009256161749362946,
0.1731610745191574,
-0.012568815611302853,
-0.014908827841281891,
-0.022075269371271133,
0.12189619243144989,
-0.03424420580267906,
-0.15636777877807617,
0.0209692120552063,
-0.17102083563804626,
-0.16333866119384766,
-0.04669443517923355,
0.03240116313099861,
0.06242522597312927,
0.12637504935264587,
0.02052353508770466,
-0.04278843104839325,
0.10590676963329315,
0.011864572763442993,
-0.003533759154379368,
-0.12654967606067657,
0.06130436807870865,
-0.027027951553463936,
0.11480697244405746,
0.01575396955013275,
0.0542067214846611,
0.13347263634204865,
-0.006136980839073658,
-0.10534203052520752,
-0.00002871052492992021,
0.059604085981845856,
-0.09689513593912125,
-0.014077813364565372,
0.16759124398231506,
-0.027202622964978218,
0.12471456080675125,
0.06352413445711136,
-0.1365993171930313,
0.0652024894952774,
0.10882023721933365,
-0.031912028789520264,
-0.06733845174312592,
0.16819308698177338,
-0.08782497048377991,
0.1461896300315857,
0.2453107386827469,
-0.02832711488008499,
0.011399230919778347,
-0.03763555735349655,
0.0453321672976017,
0.0719548761844635,
0.11854153126478195,
-0.003028048202395439,
-0.1681540161371231,
0.06275437772274017,
0.007728729862719774,
0.04761946201324463,
-0.12227306514978409,
-0.12306256592273712,
-0.01318181212991476,
0.03041958622634411,
-0.03339448571205139,
0.12228099256753922,
0.09585151076316833,
-0.02014251984655857,
-0.009588202461600304,
-0.17807522416114807,
-0.008098357357084751,
0.08778125792741776,
-0.045280177146196365,
-0.016020970419049263
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# zephyr-7b-beta-es-2000-es-agent
This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 2000
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | {"license": "mit", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "HuggingFaceH4/zephyr-7b-beta", "model-index": [{"name": "zephyr-7b-beta-es-2000-es-agent", "results": []}]} | null | Yaxin1992/zephyr-7b-beta-es-2000-es-agent | [
"peft",
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"license:mit",
"region:us"
] | 2024-02-08T18:03:54+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #generated_from_trainer #base_model-HuggingFaceH4/zephyr-7b-beta #license-mit #region-us
|
# zephyr-7b-beta-es-2000-es-agent
This model is a fine-tuned version of HuggingFaceH4/zephyr-7b-beta on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 2000
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.8.2
- Transformers 4.37.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1 | [
"# zephyr-7b-beta-es-2000-es-agent\n\nThis model is a fine-tuned version of HuggingFaceH4/zephyr-7b-beta on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 2000\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-HuggingFaceH4/zephyr-7b-beta #license-mit #region-us \n",
"# zephyr-7b-beta-es-2000-es-agent\n\nThis model is a fine-tuned version of HuggingFaceH4/zephyr-7b-beta on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 2000\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
47,
45,
6,
12,
8,
3,
102,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-HuggingFaceH4/zephyr-7b-beta #license-mit #region-us \n# zephyr-7b-beta-es-2000-es-agent\n\nThis model is a fine-tuned version of HuggingFaceH4/zephyr-7b-beta on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1e-05\n- train_batch_size: 1\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 2000\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.16.1\n- Tokenizers 0.15.1"
] | [
-0.09249429404735565,
-0.007297024130821228,
-0.0011073898058384657,
0.054167501628398895,
0.122227743268013,
0.01934105157852173,
0.14266502857208252,
0.1213046982884407,
-0.01213366724550724,
0.07049587368965149,
0.04875067621469498,
-0.03893095254898071,
0.07175980508327484,
0.1966795027256012,
-0.025661388412117958,
-0.2586599886417389,
0.020868755877017975,
0.0005290294648148119,
-0.015533357858657837,
0.0819845125079155,
0.11393728107213974,
-0.09006156027317047,
0.05288528650999069,
0.03616020455956459,
-0.15950177609920502,
0.04400038719177246,
-0.025596024468541145,
-0.03768638148903847,
0.11942371726036072,
0.033085282891988754,
0.12060873210430145,
-0.010630992241203785,
0.13055309653282166,
-0.2368122935295105,
0.016349103301763535,
0.07857123017311096,
0.03274605795741081,
0.06750959157943726,
0.07950115948915482,
-0.007112695369869471,
0.11975083500146866,
-0.12011636048555374,
0.11289431899785995,
0.04116901755332947,
-0.07332199066877365,
-0.2497168928384781,
-0.10669033229351044,
0.07202742993831635,
0.12011712789535522,
0.06861339509487152,
0.0033478797413408756,
0.12507107853889465,
-0.06845090538263321,
0.05150522291660309,
0.24524682760238647,
-0.23139336705207825,
-0.07522344589233398,
0.0062827467918396,
0.09841270744800568,
0.06469032168388367,
-0.129924014210701,
-0.033949095755815506,
0.06616135686635971,
0.039252862334251404,
0.09864585846662521,
-0.014729712158441544,
-0.04755070060491562,
-0.01622684858739376,
-0.13844242691993713,
0.022163642570376396,
0.10053576529026031,
0.0637865960597992,
-0.06406956911087036,
-0.12829042971134186,
-0.03147794306278229,
-0.02925008349120617,
-0.045874856412410736,
-0.01790623739361763,
0.028162360191345215,
-0.03701724484562874,
-0.0423399955034256,
-0.07157633453607559,
-0.08726461231708527,
-0.10049193352460861,
0.011092161759734154,
0.18299221992492676,
0.02960517629981041,
0.006318540778011084,
-0.02087578922510147,
0.09224779158830643,
-0.052785225212574005,
-0.07756895571947098,
-0.005289535503834486,
-0.009713987819850445,
-0.10154185444116592,
-0.05322955548763275,
-0.04515044391155243,
-0.052146587520837784,
0.01435835286974907,
0.12371215224266052,
-0.10011466592550278,
0.09658540785312653,
-0.012688199058175087,
0.02923102304339409,
-0.0316532738506794,
0.08406314253807068,
-0.004220715723931789,
-0.00742498179897666,
-0.013356568291783333,
0.11670677363872528,
0.004013375844806433,
-0.01007125061005354,
-0.07220975309610367,
-0.0444146953523159,
0.047153834253549576,
0.037809573113918304,
-0.06086358055472374,
0.017339618876576424,
-0.05844005569815636,
-0.008445262908935547,
-0.020082874223589897,
-0.11236301064491272,
0.05290639400482178,
0.009358324110507965,
-0.0609913095831871,
-0.10368050634860992,
0.04018205404281616,
0.022912047803401947,
0.01940734311938286,
0.10121935606002808,
-0.08460915833711624,
0.019156483933329582,
-0.08521551638841629,
-0.08643702417612076,
0.0072104125283658504,
-0.042159877717494965,
-0.012004815973341465,
-0.08711706846952438,
-0.16595834493637085,
-0.05696376413106918,
0.02741481363773346,
-0.044768452644348145,
-0.006862982641905546,
-0.030239347368478775,
-0.02490234375,
0.030380411073565483,
-0.020289277657866478,
0.14011245965957642,
-0.057920169085264206,
0.069316565990448,
-0.025556733831763268,
0.009330694563686848,
-0.026597891002893448,
0.024118104949593544,
-0.057534635066986084,
0.04635987803339958,
-0.09062257409095764,
0.042465001344680786,
-0.1332327276468277,
0.03353254869580269,
-0.11487800627946854,
-0.0684146136045456,
-0.023021699860692024,
0.008420840837061405,
0.10608629882335663,
0.0980515107512474,
-0.2021624594926834,
-0.05417906492948532,
0.1738983392715454,
-0.11846224218606949,
-0.09454525262117386,
0.09822966903448105,
-0.052727509289979935,
0.07719162106513977,
0.024259161204099655,
0.21073301136493683,
0.0849735215306282,
-0.18222448229789734,
0.04267632216215134,
-0.00959091167896986,
0.07370216399431229,
0.040220875293016434,
0.031710054725408554,
-0.057800356298685074,
-0.027189970016479492,
0.004308105446398258,
-0.059584833681583405,
0.058675773441791534,
-0.07539887726306915,
-0.046732284128665924,
-0.043974511325359344,
-0.08000528067350388,
0.052115559577941895,
0.01874208264052868,
0.01480122935026884,
-0.08926770836114883,
-0.0807698592543602,
0.13471361994743347,
0.13786324858665466,
-0.05152841657400131,
0.01961827091872692,
-0.052368249744176865,
0.01386407669633627,
0.008741567842662334,
-0.03217629715800285,
-0.1665755808353424,
-0.1168653815984726,
0.03291257470846176,
-0.052864931523799896,
0.025857293978333473,
-0.019473372027277946,
0.08319022506475449,
0.05999331548810005,
-0.055956196039915085,
-0.04555073007941246,
-0.11404872685670853,
0.012748311273753643,
-0.1110454872250557,
-0.19345536828041077,
-0.06620531529188156,
-0.04832093417644501,
0.18639373779296875,
-0.233634352684021,
0.026591140776872635,
-0.029838629066944122,
0.13206200301647186,
0.04974863678216934,
-0.04919455572962761,
-0.00962766818702221,
0.06679516285657883,
0.015499192290008068,
-0.08815543353557587,
0.053747907280921936,
0.025788120925426483,
-0.12945015728473663,
-0.003079095156863332,
-0.15015237033367157,
0.018704285845160484,
0.05811166390776634,
0.1154220923781395,
-0.12645602226257324,
-0.14463172852993011,
-0.07992807775735855,
-0.047288928180933,
-0.08979739248752594,
-0.006103323306888342,
0.17269189655780792,
0.028581513091921806,
0.12769171595573425,
-0.06922601908445358,
-0.05302009359002113,
-0.005736417602747679,
-0.033962566405534744,
-0.0033191146794706583,
0.07795540988445282,
0.027559665963053703,
-0.08659964054822922,
0.07675871253013611,
0.08785296976566315,
-0.06500576436519623,
0.1848425418138504,
-0.059237271547317505,
-0.09259134531021118,
-0.015655266121029854,
0.04446258023381233,
0.007920113392174244,
0.1541292667388916,
-0.027720510959625244,
0.007873067632317543,
0.023743396624922752,
0.021895678713917732,
0.03534319996833801,
-0.1884414106607437,
-0.015436967834830284,
0.003426480572670698,
-0.02769247442483902,
0.012010983191430569,
-0.00730258459225297,
0.027463247999548912,
0.09057813882827759,
0.024917632341384888,
0.0003732732147909701,
0.004392544273287058,
-0.013872504234313965,
-0.11388074606657028,
0.18657398223876953,
-0.13087455928325653,
-0.10237221419811249,
-0.1200048178434372,
0.01752316579222679,
-0.03268054872751236,
-0.03803110867738724,
0.01866794563829899,
-0.07975298911333084,
-0.03902222588658333,
-0.09929459542036057,
0.014982529915869236,
-0.03086228296160698,
-0.019352415576577187,
0.015109303407371044,
0.010674402117729187,
0.10625194758176804,
-0.1329350620508194,
0.007034048438072205,
-0.03409692645072937,
-0.05468433350324631,
0.011089017614722252,
0.03713952377438545,
0.07774744182825089,
0.11426819860935211,
-0.009486372582614422,
0.0036791146267205477,
-0.06960108876228333,
0.1823502779006958,
-0.0547725073993206,
-0.009270276874303818,
0.10785229504108429,
-0.009284396655857563,
0.06930624693632126,
0.10491336137056351,
0.039972007274627686,
-0.0916956290602684,
0.062253035604953766,
0.0753108561038971,
-0.014556633308529854,
-0.20855547487735748,
-0.050299230962991714,
-0.03812478855252266,
-0.04965212568640709,
0.09262841939926147,
0.05484289675951004,
-0.012646635994315147,
0.04086249694228172,
-0.017815833911299706,
0.0005743864458054304,
0.02794809639453888,
0.10587447136640549,
0.04256419092416763,
0.021843209862709045,
0.08519811183214188,
-0.023491326719522476,
-0.004486719146370888,
0.060518864542245865,
0.002321918262168765,
0.27416953444480896,
-0.014146885834634304,
0.06117276847362518,
0.06619581580162048,
0.13448798656463623,
-0.00825319904834032,
0.028719738125801086,
0.018323564901947975,
-0.021071402356028557,
0.007301210891455412,
-0.05242138355970383,
-0.019533682614564896,
0.06313358247280121,
-0.038011737167835236,
0.05522816255688667,
-0.1089770644903183,
-0.027399269863963127,
-0.0007004080107435584,
0.2827727198600769,
0.010026942938566208,
-0.2356993854045868,
-0.07270672917366028,
0.011697759851813316,
-0.04729670286178589,
-0.10493777692317963,
-0.01590583845973015,
0.13409917056560516,
-0.1394115388393402,
0.0715482160449028,
-0.07638853043317795,
0.09269192814826965,
0.006345656700432301,
-0.017950251698493958,
0.023118145763874054,
0.14592553675174713,
-0.029123542830348015,
0.055877890437841415,
-0.2157374769449234,
0.2384064793586731,
0.006799293216317892,
0.10734590888023376,
-0.02694806084036827,
0.03826972097158432,
0.03924721106886864,
0.03269912302494049,
0.0816759392619133,
-0.0016977485502138734,
-0.058389339596033096,
-0.16841909289360046,
-0.03482421115040779,
0.02677684649825096,
0.12321311980485916,
-0.013974569737911224,
0.07279249280691147,
-0.05391399562358856,
0.03214266896247864,
0.05057058483362198,
-0.011132941581308842,
-0.20481902360916138,
-0.123775415122509,
-0.011290453374385834,
0.019037440419197083,
-0.030683623626828194,
-0.13230246305465698,
-0.09417160600423813,
0.011349575594067574,
0.14273467659950256,
0.03345029428601265,
-0.03670121729373932,
-0.1544385850429535,
0.06134577840566635,
0.10584372282028198,
-0.04105116426944733,
0.030807407572865486,
0.0255038533359766,
0.15680336952209473,
-0.0313115157186985,
-0.04816428944468498,
0.06760125607252121,
-0.07842633128166199,
-0.15898266434669495,
-0.0825750082731247,
0.13855251669883728,
0.08636507391929626,
0.03945241495966911,
-0.00714783975854516,
0.01638357900083065,
0.016983358189463615,
-0.08723179250955582,
0.04219774529337883,
0.1464167982339859,
0.019024254754185677,
0.05103704333305359,
-0.107150137424469,
-0.0036007207818329334,
-0.024734698235988617,
-0.045655783265829086,
0.1199522539973259,
0.2809784412384033,
-0.07518845051527023,
0.09098877757787704,
0.0588226318359375,
-0.07640617340803146,
-0.16154006123542786,
0.08291007578372955,
0.1647905856370926,
0.03574329614639282,
0.07318949699401855,
-0.16986125707626343,
0.042336031794548035,
0.12093137949705124,
-0.02705426700413227,
0.034781839698553085,
-0.260668009519577,
-0.12967541813850403,
0.08579521626234055,
0.13138839602470398,
0.0074551887810230255,
-0.12921197712421417,
-0.047202400863170624,
-0.041564106941223145,
-0.1259559690952301,
0.08980602025985718,
-0.11146070808172226,
0.08937152475118637,
-0.021617228165268898,
0.08128606528043747,
0.037022992968559265,
-0.02119157277047634,
0.15634436905384064,
-0.04818816855549812,
0.10377392917871475,
-0.05445932224392891,
0.06615079939365387,
0.00619965186342597,
-0.07937189191579819,
0.03477802872657776,
-0.049426332116127014,
0.03403015434741974,
-0.13403910398483276,
-0.016782715916633606,
-0.07837733626365662,
0.10156035423278809,
-0.04125860705971718,
-0.05659889429807663,
-0.03742894530296326,
0.0683082640171051,
0.019064076244831085,
-0.03188807889819145,
0.05157310143113136,
-0.03396221622824669,
0.17953816056251526,
0.08520415425300598,
0.10619653016328812,
-0.008046615868806839,
-0.06541193276643753,
0.007399704307317734,
-0.036227498203516006,
0.05881580710411072,
-0.12304180860519409,
0.02304576151072979,
0.09987006336450577,
0.04358496144413948,
0.15719285607337952,
0.04412407800555229,
-0.10384494066238403,
0.016689758747816086,
0.056869421154260635,
-0.07515524327754974,
-0.13533256947994232,
-0.008807089179754257,
0.08084997534751892,
-0.16204003989696503,
-0.002477827947586775,
0.09901430457830429,
-0.04661370441317558,
-0.02381899021565914,
-0.012925935909152031,
0.005957827903330326,
-0.04197760671377182,
0.1509445160627365,
0.037623487412929535,
0.048308007419109344,
-0.0527697429060936,
0.11231519281864166,
0.04403227940201759,
-0.03206981346011162,
0.0533088780939579,
0.08670002967119217,
-0.06740908324718475,
-0.0081256702542305,
0.058552082628011703,
0.19837690889835358,
-0.008922580629587173,
-0.05375878885388374,
-0.06357384473085403,
-0.1302875578403473,
-0.009014468640089035,
0.13997939229011536,
0.014024556614458561,
-0.043331410735845566,
0.008964575827121735,
0.049650564789772034,
-0.11255665123462677,
0.07187984883785248,
0.039293669164180756,
0.08550035953521729,
-0.12249469757080078,
0.13623669743537903,
0.00906860176473856,
0.015483210794627666,
-0.005047297570854425,
0.023729316890239716,
-0.10742893069982529,
-0.019559074193239212,
-0.16264355182647705,
-0.004925414454191923,
-0.030942892655730247,
0.0050146738067269325,
-0.001359638525173068,
-0.0387277789413929,
-0.02151227742433548,
0.0383128821849823,
-0.07923068851232529,
-0.05395760014653206,
0.007807335816323757,
0.07207304984331131,
-0.10197535157203674,
-0.015768636018037796,
0.04240506887435913,
-0.0912468209862709,
0.06223118305206299,
0.05644528940320015,
0.053983211517333984,
0.03271359205245972,
-0.1357279121875763,
0.013126621022820473,
0.037846289575099945,
-0.013901866972446442,
0.046761129051446915,
-0.1007886677980423,
-0.04254499450325966,
-0.04438754543662071,
0.03104499913752079,
0.006663966458290815,
0.0656578466296196,
-0.11832942068576813,
-0.04797520488500595,
-0.03525371477007866,
-0.06927794218063354,
-0.0665106475353241,
0.030312396585941315,
0.07829178869724274,
0.09977550804615021,
0.11750947684049606,
-0.10713624209165573,
0.05040658637881279,
-0.1994282603263855,
-0.02829933911561966,
-0.018946316093206406,
-0.0028151401784271,
-0.008805635385215282,
-0.04440018907189369,
0.06700128316879272,
-0.037064675241708755,
0.09447936713695526,
-0.007563532795757055,
0.07934893667697906,
0.03221100941300392,
-0.09620869904756546,
0.032137203961610794,
0.005720071494579315,
0.2031380534172058,
0.014746595174074173,
-0.010957264341413975,
0.07972434163093567,
0.00014215466217137873,
0.033017825335264206,
0.06326155364513397,
0.1436665952205658,
0.1481330692768097,
-0.03295683488249779,
0.054466258734464645,
0.0827820748090744,
-0.1211409643292427,
-0.04610290750861168,
0.08523354679346085,
0.016436535865068436,
0.0629202276468277,
-0.06380946189165115,
0.2610306441783905,
0.1149744838476181,
-0.20290298759937286,
0.04316378012299538,
-0.0284134428948164,
-0.09025337547063828,
-0.10036268085241318,
-0.0068010264076292515,
-0.06484736502170563,
-0.16493451595306396,
0.024463361129164696,
-0.10158485174179077,
0.010547561571002007,
0.04800233989953995,
-0.009784912690520287,
0.04616282507777214,
0.1298765242099762,
-0.02222546562552452,
-0.014016405679285526,
0.07513708621263504,
0.013032889924943447,
-0.003211759030818939,
-0.10197751969099045,
-0.09743411839008331,
0.08054612576961517,
-0.0319107323884964,
0.04307396709918976,
-0.05433447286486626,
-0.012255923822522163,
0.03868462145328522,
0.005544126499444246,
-0.06118851155042648,
0.03354816883802414,
0.007940942421555519,
0.03246507793664932,
0.06380072981119156,
0.07002922147512436,
-0.0030082850717008114,
-0.04851735010743141,
0.2898244559764862,
-0.07825250178575516,
-0.06311295181512833,
-0.12881889939308167,
0.2060181200504303,
-0.014717633835971355,
-0.008307715877890587,
0.013860826380550861,
-0.1155429407954216,
0.013175207190215588,
0.12404072284698486,
0.08314398676156998,
-0.04742393642663956,
0.01695127785205841,
-0.02662503346800804,
-0.028849398717284203,
-0.08232523500919342,
0.12635551393032074,
0.08872712403535843,
0.00261265248991549,
-0.045685119926929474,
0.02474675141274929,
-0.020598648115992546,
-0.010886459611356258,
-0.07111424952745438,
0.08906013518571854,
0.005970703437924385,
-0.008104324340820312,
-0.049570851027965546,
0.07446187734603882,
0.032010383903980255,
-0.174949049949646,
0.029490014538168907,
-0.13369734585285187,
-0.1478555053472519,
-0.031721632927656174,
0.04776578024029732,
-0.01457983162254095,
0.056242797523736954,
-0.023839084431529045,
0.038231052458286285,
0.11609967797994614,
-0.029049102216959,
-0.008247032761573792,
-0.1486251950263977,
0.07892107218503952,
-0.11693979054689407,
0.22142091393470764,
-0.015244415029883385,
0.037125661969184875,
0.09177973866462708,
0.03649192675948143,
-0.11095505952835083,
0.06777233630418777,
0.044764526188373566,
-0.06997166574001312,
0.00916131678968668,
0.16579824686050415,
-0.05614030361175537,
0.09748616069555283,
0.039005521684885025,
-0.16064931452274323,
0.029533442109823227,
-0.04237869009375572,
-0.056711114943027496,
-0.08552132546901703,
0.039159294217824936,
-0.07440879195928574,
0.15009169280529022,
0.1711883544921875,
-0.04594409093260765,
-0.011169481091201305,
-0.06930355727672577,
0.030972875654697418,
0.06215367838740349,
0.12304142862558365,
-0.00579995708540082,
-0.20749862492084503,
0.022206660360097885,
0.10165257751941681,
0.0025852173566818237,
-0.2682204842567444,
-0.10879257321357727,
0.011866909451782703,
-0.057621680200099945,
-0.025115258991718292,
0.10696388781070709,
0.07823194563388824,
0.037838589400053024,
-0.034619297832250595,
-0.23599305748939514,
-0.0036734563764184713,
0.16114141047000885,
-0.08666296303272247,
-0.031179487705230713
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | token-classification | sahillihas/G3-finetuned-ner | [
"transformers",
"safetensors",
"distilbert",
"token-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T18:04:21+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
49,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #distilbert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.07577246427536011,
0.160971537232399,
-0.0037754587829113007,
0.025772565975785255,
0.11726456880569458,
0.008670859038829803,
0.07505112886428833,
0.10672148317098618,
-0.015180999413132668,
0.12574146687984467,
0.037852395325899124,
0.1016918197274208,
0.10858695209026337,
0.19223041832447052,
-0.0040494236163794994,
-0.20760315656661987,
0.0604725256562233,
-0.11741560697555542,
0.010131745599210262,
0.12369410693645477,
0.13811981678009033,
-0.11246351152658463,
0.06881340593099594,
-0.04209144040942192,
-0.01980483904480934,
-0.035129256546497345,
-0.06221386045217514,
-0.052587710320949554,
0.06488312780857086,
0.05438476800918579,
0.06559453904628754,
0.021042533218860626,
0.0840018093585968,
-0.28357821702957153,
0.019257022067904472,
0.0766754001379013,
0.002646347275003791,
0.06359819322824478,
0.07652021199464798,
-0.07693555951118469,
0.09431595355272293,
-0.059984829276800156,
0.15352454781532288,
0.07422540336847305,
-0.09718401730060577,
-0.18205241858959198,
-0.08698616921901703,
0.10743562132120132,
0.1786193698644638,
0.05337109789252281,
-0.03495080769062042,
0.14324238896369934,
-0.07000283896923065,
0.017874419689178467,
0.06666957587003708,
-0.0738137811422348,
-0.05611572042107582,
0.06171523407101631,
0.07419267296791077,
0.09766413271427155,
-0.13268914818763733,
-0.009328355081379414,
0.039757777005434036,
0.01763407699763775,
0.11028990149497986,
0.017873499542474747,
0.12974926829338074,
0.03081592544913292,
-0.1462060660123825,
-0.059808772057294846,
0.10660259425640106,
0.03442228585481644,
-0.05801061540842056,
-0.24857386946678162,
-0.00732324318960309,
-0.03398886322975159,
-0.03147710859775543,
-0.045992717146873474,
0.04087558388710022,
-0.027321036905050278,
0.08885861188173294,
0.0030806211289018393,
-0.06717013567686081,
-0.05033329501748085,
0.09465020149946213,
0.06679657101631165,
0.026782603934407234,
-0.028168529272079468,
0.01867647096514702,
0.11993368715047836,
0.10629972815513611,
-0.11282733827829361,
-0.06030019000172615,
-0.06428343802690506,
-0.08693956583738327,
-0.04836411029100418,
0.03905215859413147,
0.0691521167755127,
0.05319429561495781,
0.20513659715652466,
-0.005697339307516813,
0.04798254370689392,
0.03319632261991501,
0.012943265959620476,
0.06999380886554718,
0.07388506829738617,
-0.057592663913965225,
-0.13561639189720154,
-0.032188303768634796,
0.1180354431271553,
0.0034096194431185722,
-0.029926057904958725,
-0.03362307325005531,
0.061736419796943665,
0.04901990294456482,
0.12567868828773499,
0.06872507929801941,
0.016199197620153427,
-0.07887519896030426,
-0.05266522243618965,
0.18039943277835846,
-0.15842315554618835,
0.026145054027438164,
0.01778857596218586,
-0.04808545112609863,
-0.03221158683300018,
0.01776694506406784,
0.010511110536754131,
-0.027442574501037598,
0.09513456374406815,
-0.064456507563591,
-0.0437372550368309,
-0.10800766944885254,
-0.05169657990336418,
0.03205830603837967,
-0.02027302049100399,
-0.02823185920715332,
-0.039158232510089874,
-0.12495775520801544,
-0.0780569463968277,
0.07112475484609604,
-0.06458895653486252,
-0.06230264902114868,
-0.035103172063827515,
-0.06434695422649384,
0.01328826043754816,
-0.0016070485580712557,
0.11751557141542435,
-0.02924034371972084,
0.05098489671945572,
-0.058480363339185715,
0.06735116988420486,
0.13876435160636902,
0.030480407178401947,
-0.06721892952919006,
0.06733744591474533,
-0.21017982065677643,
0.10650269687175751,
-0.0903177410364151,
0.030309993773698807,
-0.16417838633060455,
-0.020649932324886322,
0.03423565998673439,
0.03357144445180893,
-0.011650003492832184,
0.14477767050266266,
-0.17939403653144836,
-0.035702068358659744,
0.18933741748332977,
-0.12515437602996826,
-0.092041015625,
0.05768411606550217,
-0.06046192720532417,
0.13482485711574554,
0.05646788328886032,
-0.021687285974621773,
0.05586513131856918,
-0.13862864673137665,
-0.02437570132315159,
-0.06216869503259659,
-0.014887474477291107,
0.1528262197971344,
0.059848666191101074,
-0.046563770622015,
0.02752561867237091,
0.017916126176714897,
-0.026528963819146156,
-0.04967709258198738,
-0.03546646237373352,
-0.09774158149957657,
0.00822865217924118,
-0.07969193160533905,
0.02241206355392933,
-0.022354070097208023,
-0.09176263213157654,
-0.03786236792802811,
-0.15299317240715027,
0.01120391022413969,
0.09777288138866425,
-0.0018325879937037826,
-0.030514128506183624,
-0.09985145181417465,
-0.0009975290158763528,
0.014818643219769001,
-0.006299570668488741,
-0.15255902707576752,
-0.05487200245261192,
0.02453494630753994,
-0.16442841291427612,
0.027572009712457657,
-0.04693130776286125,
0.03860992193222046,
0.04300114884972572,
-0.04640088230371475,
-0.03466949611902237,
0.020573051646351814,
0.021380405873060226,
-0.02527492307126522,
-0.25819748640060425,
-0.015081126242876053,
-0.050718042999506,
0.1735834777355194,
-0.24959886074066162,
0.04765407368540764,
0.0635317862033844,
0.12164347618818283,
0.008648045361042023,
-0.043875813484191895,
0.03864365443587303,
-0.0540064312517643,
-0.03614342212677002,
-0.06847652792930603,
-0.006085617933422327,
-0.03506658226251602,
-0.04601104184985161,
0.036002401262521744,
-0.17769227921962738,
-0.026747822761535645,
0.11292517185211182,
0.07372257113456726,
-0.16646650433540344,
-0.07437358796596527,
-0.03605179488658905,
-0.05963003262877464,
-0.07967717945575714,
-0.05352560430765152,
0.08457031846046448,
0.047888897359371185,
0.050600916147232056,
-0.06766360253095627,
-0.05917840451002121,
0.012786190956830978,
-0.014223156496882439,
-0.030130434781312943,
0.08920878916978836,
0.06863047182559967,
-0.13044358789920807,
0.10779033601284027,
0.07565328478813171,
0.07318258285522461,
0.10535803437232971,
0.004109293222427368,
-0.09151213616132736,
-0.019049784168601036,
0.030527541413903236,
0.01623775251209736,
0.14909477531909943,
-0.058906275779008865,
0.037276990711688995,
0.042408693581819534,
-0.025859607383608818,
0.007144377566874027,
-0.09663669019937515,
0.020623620599508286,
0.027944007888436317,
-0.010041053406894207,
0.02522680163383484,
-0.05456255003809929,
0.016582027077674866,
0.10647457838058472,
0.0315246656537056,
0.029977189376950264,
0.01468842476606369,
-0.04244951531291008,
-0.1256953328847885,
0.17814458906650543,
-0.09753371775150299,
-0.24721357226371765,
-0.12567918002605438,
-0.009092925116419792,
0.03918062895536423,
-0.012046195566654205,
0.01987382024526596,
-0.05762042477726936,
-0.10939344018697739,
-0.1005193367600441,
0.03179878741502762,
0.06615057587623596,
-0.08552238345146179,
-0.0674377977848053,
0.05327526107430458,
0.04219662770628929,
-0.1277419775724411,
0.020916761830449104,
0.041558679193258286,
-0.07218944281339645,
0.008345745503902435,
0.056684307754039764,
0.07901664078235626,
0.1805250197649002,
0.011102218180894852,
-0.018453268334269524,
0.012500734999775887,
0.21901123225688934,
-0.1455126702785492,
0.09525687992572784,
0.13985903561115265,
-0.06373906135559082,
0.08310941606760025,
0.20266622304916382,
0.03211383894085884,
-0.10105792433023453,
0.03751940280199051,
0.03433764725923538,
-0.03450850397348404,
-0.24166366457939148,
-0.07809934765100479,
0.0032382020726799965,
-0.06378763914108276,
0.10437320917844772,
0.08564607799053192,
0.10547725111246109,
0.04683990776538849,
-0.11285781115293503,
-0.06474319845438004,
0.05201273784041405,
0.11905202269554138,
-0.028264760971069336,
-0.0015284299151971936,
0.09661765396595001,
-0.023179147392511368,
0.024211812764406204,
0.091147281229496,
0.024589717388153076,
0.18455083668231964,
0.04622681811451912,
0.13505716621875763,
0.0910676121711731,
0.06324100494384766,
0.014886823482811451,
0.021261626854538918,
0.01765025220811367,
0.02958594262599945,
-0.018347864970564842,
-0.08319743722677231,
-0.009293846786022186,
0.13596801459789276,
0.024112429469823837,
0.04030708968639374,
0.006161263212561607,
-0.048370931297540665,
0.07493776082992554,
0.17692242562770844,
0.014188401401042938,
-0.22553029656410217,
-0.06700198352336884,
0.07436924427747726,
-0.07410255074501038,
-0.11821255087852478,
-0.01875862292945385,
0.02883634716272354,
-0.18362410366535187,
0.03783649206161499,
-0.027923477813601494,
0.10164659470319748,
-0.11413569748401642,
-0.021443994715809822,
0.03836014121770859,
0.05620839074254036,
-0.032449860125780106,
0.07187992334365845,
-0.18838849663734436,
0.13979743421077728,
0.009049986489117146,
0.06722430139780045,
-0.1007412001490593,
0.08134089410305023,
0.017924414947628975,
0.005374805070459843,
0.16363629698753357,
-0.0033102943561971188,
-0.061859820038080215,
-0.10252897441387177,
-0.08680899441242218,
-0.015643585473299026,
0.09927050769329071,
-0.12645161151885986,
0.09391359239816666,
-0.00760313356295228,
-0.03509579226374626,
-0.0025038165040314198,
-0.13591554760932922,
-0.13381120562553406,
-0.17948365211486816,
0.04688849300146103,
-0.12670443952083588,
0.04449006915092468,
-0.1076912134885788,
-0.056078001856803894,
-0.041555143892765045,
0.19199959933757782,
-0.21791574358940125,
-0.08265700191259384,
-0.15180857479572296,
-0.06470925360918045,
0.11797353625297546,
-0.04333622008562088,
0.08231177926063538,
0.009503866545855999,
0.19813112914562225,
-0.0016607206780463457,
-0.004083174746483564,
0.0959770604968071,
-0.09965015947818756,
-0.2107795774936676,
-0.09851331263780594,
0.13872341811656952,
0.13654115796089172,
0.04202498495578766,
0.0035957242362201214,
0.026746563613414764,
-0.003834358649328351,
-0.1127103865146637,
0.03198874741792679,
0.15448687970638275,
0.11144908517599106,
0.038192201405763626,
-0.028366046026349068,
-0.13728831708431244,
-0.10052017867565155,
-0.04788326099514961,
0.009005418047308922,
0.19273622334003448,
-0.07073298841714859,
0.1647789627313614,
0.1582319587469101,
-0.06154803931713104,
-0.2109813690185547,
0.03008965775370598,
0.03392798826098442,
-0.003125392831861973,
0.04911292716860771,
-0.2020125389099121,
0.08147795498371124,
0.014267290011048317,
-0.05861039087176323,
0.12597185373306274,
-0.17893800139427185,
-0.1463114470243454,
0.0894082561135292,
0.07762204855680466,
-0.19969920814037323,
-0.1310804784297943,
-0.09587812423706055,
-0.04523595795035362,
-0.09975142776966095,
0.09141786396503448,
-0.006663811393082142,
0.00584616232663393,
0.03217587620019913,
0.01619444414973259,
0.016584929078817368,
-0.04955518618226051,
0.19393765926361084,
-0.0010097876656800508,
0.049837395548820496,
-0.07284197211265564,
-0.07579720765352249,
0.03499779477715492,
-0.07016132026910782,
0.08607465773820877,
-0.017582224681973457,
0.006899731699377298,
-0.11493121087551117,
-0.06381088495254517,
-0.04385467618703842,
0.0321214497089386,
-0.08662021160125732,
-0.09713311493396759,
-0.04749152809381485,
0.105962373316288,
0.0914415493607521,
-0.03657814860343933,
-0.06360591948032379,
-0.0900130569934845,
0.052277419716119766,
0.22165511548519135,
0.18128447234630585,
0.0717686340212822,
-0.07471393048763275,
-0.008312922902405262,
-0.021178819239139557,
0.05892517790198326,
-0.20861060917377472,
0.049327246844768524,
0.03792690858244896,
0.03388948366045952,
0.12989014387130737,
-0.025939077138900757,
-0.16148367524147034,
-0.052209001034498215,
0.05540921539068222,
-0.07436220347881317,
-0.16007542610168457,
0.006253175903111696,
0.08246254920959473,
-0.1577574759721756,
-0.040928490459918976,
0.0384182333946228,
-0.029118487611413002,
-0.031075680628418922,
0.0023355213925242424,
0.08367416262626648,
0.021480288356542587,
0.10684667527675629,
0.06465408951044083,
0.10775714367628098,
-0.10543544590473175,
0.06850111484527588,
0.08449427038431168,
-0.10566192865371704,
0.03645234927535057,
0.059928081929683685,
-0.06519834697246552,
-0.03475077450275421,
0.03517557680606842,
0.08707848936319351,
0.02939283289015293,
-0.06929907202720642,
0.006287590134888887,
-0.10996777564287186,
0.06547177582979202,
0.13484768569469452,
0.04042261838912964,
0.011329763568937778,
0.040748391300439835,
0.030845293775200844,
-0.09934883564710617,
0.12121784687042236,
0.05189082771539688,
0.03800792247056961,
-0.0530884675681591,
-0.017324361950159073,
0.03913010656833649,
-0.019996602088212967,
-0.017436852678656578,
-0.03976694867014885,
-0.07124674320220947,
-0.011837370693683624,
-0.16637970507144928,
0.020963486284017563,
-0.06557925045490265,
0.011502030305564404,
0.015106898732483387,
-0.03105640783905983,
0.00648743100464344,
0.012552679516375065,
-0.07388464361429214,
-0.045671455562114716,
-0.00442871730774641,
0.10947413742542267,
-0.16812850534915924,
0.008818828500807285,
0.08374110609292984,
-0.12818290293216705,
0.08439380675554276,
0.0020208086352795362,
-0.0032075955532491207,
0.02065693773329258,
-0.1317322701215744,
0.05960160121321678,
-0.006254117004573345,
0.006569796707481146,
0.03298337012529373,
-0.21374841034412384,
0.0022494937293231487,
-0.05006609857082367,
-0.06266371160745621,
-0.0007477401522919536,
-0.03451014682650566,
-0.11394216865301132,
0.10471237450838089,
0.014338361099362373,
-0.07764562219381332,
-0.021158380433917046,
0.048064179718494415,
0.10854631662368774,
-0.04941081255674362,
0.1473926305770874,
-0.01730068400502205,
0.059951625764369965,
-0.1836392730474472,
-0.02037329412996769,
-0.018616020679473877,
0.016425784677267075,
-0.037237949669361115,
-0.006127772852778435,
0.051674678921699524,
-0.01703222282230854,
0.2173253446817398,
-0.022195780649781227,
0.026851661503314972,
0.06280741095542908,
-0.002322237938642502,
-0.013443386182188988,
0.09835761785507202,
0.046587999910116196,
0.013757294043898582,
0.022333940491080284,
0.010798705741763115,
-0.04187694564461708,
-0.00982733629643917,
-0.13689175248146057,
0.07585512101650238,
0.16693128645420074,
0.08301392197608948,
-0.00876491330564022,
0.04851881042122841,
-0.10903714597225189,
-0.10301291197538376,
0.09611163288354874,
-0.03860481455922127,
-0.017003152519464493,
-0.050830189138650894,
0.13978728652000427,
0.15471777319908142,
-0.19259576499462128,
0.06503742933273315,
-0.0670342966914177,
-0.05641358345746994,
-0.10491608083248138,
-0.17902258038520813,
-0.06081055849790573,
-0.037293195724487305,
-0.01373780332505703,
-0.06118905171751976,
0.06020566076040268,
0.10252559930086136,
0.014830772764980793,
0.008825747296214104,
0.082810178399086,
-0.027097169309854507,
0.007841658778488636,
0.04027826339006424,
0.06163087487220764,
0.015406538732349873,
-0.06096850335597992,
0.008726970292627811,
0.0014340170891955495,
0.03696046769618988,
0.053261034190654755,
0.03251050412654877,
-0.012918577529489994,
0.006573197897523642,
-0.023761101067066193,
-0.10426638275384903,
0.04047070071101189,
-0.02524152584373951,
-0.052286889404058456,
0.15217694640159607,
0.022632325068116188,
-0.005827770568430424,
-0.022215187549591064,
0.2338925302028656,
-0.06722013652324677,
-0.0785137191414833,
-0.14054499566555023,
0.14897066354751587,
-0.03910091519355774,
0.05552518740296364,
0.046295974403619766,
-0.105179063975811,
0.03945622220635414,
0.13950063288211823,
0.14286132156848907,
-0.04025404527783394,
0.012145495042204857,
0.009853771887719631,
0.0038128900341689587,
-0.028584260493516922,
0.052456315606832504,
0.05106692388653755,
0.12483679503202438,
-0.06253138184547424,
0.09782347828149796,
-0.006969998124986887,
-0.09410524368286133,
-0.025307200849056244,
0.13257914781570435,
0.002008436480537057,
0.023570919409394264,
-0.0815717875957489,
0.12813320755958557,
-0.05384290590882301,
-0.2535746693611145,
0.07024679332971573,
-0.06422019749879837,
-0.15186867117881775,
-0.019949471578001976,
0.020098339766263962,
-0.004076872952282429,
0.023379439488053322,
0.06391935795545578,
-0.0649435967206955,
0.15813666582107544,
0.03603892773389816,
-0.06333170086145401,
-0.07421127706766129,
0.07901927083730698,
-0.07848652452230453,
0.3060082495212555,
0.007198201958090067,
0.05301123857498169,
0.09581229835748672,
-0.03854295611381531,
-0.13927392661571503,
0.028419770300388336,
0.09097424894571304,
-0.05137800797820091,
0.061749134212732315,
0.2061481475830078,
-0.010625248774886131,
0.11590193957090378,
0.07372608780860901,
-0.08649783581495285,
0.04505147039890289,
-0.09321032464504242,
-0.09480153769254684,
-0.08939364552497864,
0.09279339760541916,
-0.055773381143808365,
0.1508018523454666,
0.1253325492143631,
-0.048127513378858566,
0.008756249211728573,
-0.019953599199652672,
0.058552540838718414,
0.0011799887288361788,
0.11399417370557785,
0.028718046844005585,
-0.19564524292945862,
0.02838771604001522,
-0.0018802561098709702,
0.1010337695479393,
-0.23878240585327148,
-0.0895518958568573,
0.04254518821835518,
-0.0008608286152593791,
-0.060450565069913864,
0.12522582709789276,
0.05044018104672432,
0.043433722108602524,
-0.0508544035255909,
-0.04669932648539543,
-0.008167625404894352,
0.1615993231534958,
-0.10094329714775085,
-0.0017090580658987164
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
#Mahna_Mahna, Fine Tuning Mistral 7b in Google Colab with QLoRA(tutorial_reproduction).ipynb
#Tutorial by Ali Mobarekati, 2024
#https://medium.com/@codersama/fine-tuning-mistral-7b-in-google-colab-with-qlora-complete-guide-60e12d437cca
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | null | KJTDHISAW/Enlighten_Instruct | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | 2024-02-08T18:07:31+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for Model ID
#Mahna_Mahna, Fine Tuning Mistral 7b in Google Colab with QLoRA(tutorial_reproduction).ipynb
#Tutorial by Ali Mobarekati, 2024
#URL
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
31,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.06646376848220825,
0.2168014943599701,
-0.00225935154594481,
0.023818302899599075,
0.1271018385887146,
-0.001635765191167593,
0.04218708351254463,
0.13324736058712006,
-0.020175931975245476,
0.11144465953111649,
0.046588581055402756,
0.09377603232860565,
0.09928803145885468,
0.18404334783554077,
0.04859916493296623,
-0.2059975117444992,
0.007056170143187046,
-0.09090408682823181,
0.014076028019189835,
0.1116579994559288,
0.13719257712364197,
-0.10291384905576706,
0.08272874355316162,
-0.04045208916068077,
-0.02019004337489605,
0.00012576708104461432,
-0.09259183704853058,
-0.07032395154237747,
0.06885425746440887,
0.06264153122901917,
0.051234472543001175,
0.001456156256608665,
0.09140396863222122,
-0.2864592671394348,
0.017265573143959045,
0.08406311273574829,
0.0027674848679453135,
0.06290827691555023,
0.07236549258232117,
-0.07389893382787704,
0.11328595131635666,
-0.08021481335163116,
0.13019037246704102,
0.08625296503305435,
-0.062064990401268005,
-0.23071379959583282,
-0.07525765895843506,
0.0963398814201355,
0.12251301854848862,
0.06215599179267883,
-0.022921854630112648,
0.15455181896686554,
-0.06248689442873001,
0.012971068732440472,
0.1294165402650833,
-0.11526761949062347,
-0.05572471022605896,
0.061741601675748825,
0.11775490641593933,
0.10740239918231964,
-0.14110268652439117,
-0.0017287094378843904,
0.04900608956813812,
0.029121357947587967,
0.08589313924312592,
0.022661056369543076,
0.12003941088914871,
0.04652795568108559,
-0.13695219159126282,
-0.04037507623434067,
0.12011898308992386,
0.038862764835357666,
-0.06446044892072678,
-0.2168138176202774,
-0.006778308190405369,
-0.0601806715130806,
-0.014732478186488152,
-0.07019448280334473,
0.039128515869379044,
-0.02470310963690281,
0.07317749410867691,
-0.04465159401297569,
-0.1063927412033081,
-0.0421026237308979,
0.0892222449183464,
0.07748593389987946,
0.011527054943144321,
-0.02519804798066616,
0.04627908393740654,
0.13455867767333984,
0.05402068421244621,
-0.10399353504180908,
-0.07017925381660461,
-0.06942764669656754,
-0.09420394152402878,
-0.04035796597599983,
0.056760527193546295,
0.031942449510097504,
0.02665667235851288,
0.22703726589679718,
0.016653569415211678,
0.04155244305729866,
0.0224777739495039,
0.01032855175435543,
0.043662428855895996,
0.0955500528216362,
-0.05303520709276199,
-0.15660029649734497,
-0.04072032496333122,
0.09077946096658707,
-0.0027527001220732927,
-0.036689214408397675,
-0.03966725245118141,
0.03849169611930847,
0.06843466311693192,
0.13122352957725525,
0.07552056759595871,
-0.017929591238498688,
-0.04813180863857269,
-0.030096933245658875,
0.23523783683776855,
-0.1493375599384308,
0.04426715523004532,
-0.02271856553852558,
-0.01804111897945404,
-0.03908449783921242,
0.03597262129187584,
0.022118929773569107,
-0.000004518366949923802,
0.09706240892410278,
-0.058981191366910934,
-0.05378659814596176,
-0.10168042778968811,
-0.03272576630115509,
0.04088849574327469,
-0.013975566253066063,
-0.010589460842311382,
-0.09025166928768158,
-0.09490354359149933,
-0.04766594246029854,
0.05537205561995506,
-0.05123869329690933,
-0.03770573064684868,
0.009465423412621021,
-0.08151785284280777,
-0.005444355774670839,
-0.005417742300778627,
0.10699385404586792,
-0.03222226724028587,
0.04445803165435791,
-0.027600755915045738,
0.05225523188710213,
0.09919606149196625,
0.031576547771692276,
-0.0773419588804245,
0.0561848059296608,
-0.22559374570846558,
0.07503069192171097,
-0.11481974273920059,
0.04335082694888115,
-0.1704932004213333,
-0.042439818382263184,
0.005444696638733149,
0.0139949731528759,
0.013206101022660732,
0.12720820307731628,
-0.19255615770816803,
-0.01654396951198578,
0.13260798156261444,
-0.09212633967399597,
-0.118110790848732,
0.07884611934423447,
-0.029701577499508858,
0.1624738723039627,
0.04682036489248276,
-0.027025915682315826,
0.09224298596382141,
-0.16434773802757263,
-0.07092688232660294,
-0.00949116237461567,
-0.01727987825870514,
0.12109188735485077,
0.07512219995260239,
-0.05991523340344429,
0.046571120619773865,
0.02832140028476715,
-0.038078423589468,
-0.04424772411584854,
-0.050857074558734894,
-0.10884185880422592,
-0.01070026308298111,
-0.08987759798765182,
0.04065500199794769,
-0.01250192429870367,
-0.07916021347045898,
-0.029885273426771164,
-0.18612512946128845,
-0.0030564051121473312,
0.10038342326879501,
0.0035033065360039473,
-0.005652366206049919,
-0.08666291832923889,
0.026358824223279953,
-0.03112892620265484,
-0.008404186926782131,
-0.16764774918556213,
-0.04399421438574791,
0.046902090311050415,
-0.16094985604286194,
0.020117372274398804,
-0.06413903087377548,
0.06334125250577927,
0.03641495108604431,
-0.05590536445379257,
-0.0248766727745533,
-0.01730942726135254,
0.011945613659918308,
-0.05083848536014557,
-0.18994836509227753,
-0.056277405470609665,
-0.037882111966609955,
0.149809330701828,
-0.25956398248672485,
0.032966937869787216,
0.051140617579221725,
0.14649195969104767,
0.00406361510977149,
-0.05115427449345589,
0.01429014839231968,
-0.05360214412212372,
-0.054652128368616104,
-0.06746816635131836,
-0.006135428790003061,
-0.027576493099331856,
-0.05147203803062439,
0.019243421033024788,
-0.1755700707435608,
-0.021410830318927765,
0.09424154460430145,
0.12876708805561066,
-0.1486445665359497,
-0.018640631809830666,
-0.048725154250860214,
-0.06339836865663528,
-0.0715010017156601,
-0.07038594037294388,
0.10712739825248718,
0.0513901449739933,
0.04796046018600464,
-0.07435787469148636,
-0.07092321664094925,
0.02726263552904129,
0.006906150374561548,
-0.03382374346256256,
0.08727246522903442,
0.05199531093239784,
-0.09209315478801727,
0.0756213590502739,
0.1092359870672226,
0.07177663594484329,
0.09363535046577454,
0.01574566215276718,
-0.11756632477045059,
-0.028492970392107964,
0.036266472190618515,
0.02740776725113392,
0.1465986967086792,
-0.05952361226081848,
0.04016614332795143,
0.04494241625070572,
-0.04170418903231621,
0.022319864481687546,
-0.08787637203931808,
0.024075502529740334,
0.025203049182891846,
-0.0034381982404738665,
0.06284574419260025,
-0.02525499276816845,
-0.0050758360885083675,
0.07016654312610626,
0.047779910266399384,
0.04621000960469246,
0.009655474685132504,
-0.01720241829752922,
-0.1047825813293457,
0.16950392723083496,
-0.0951867327094078,
-0.269941508769989,
-0.17632324993610382,
0.026197833940386772,
0.04035249724984169,
-0.022378476336598396,
0.031619444489479065,
-0.07056326419115067,
-0.10630585998296738,
-0.1060405746102333,
-0.002429972169920802,
0.01714223250746727,
-0.06364088505506516,
-0.0741225928068161,
0.07348573952913284,
0.04382912442088127,
-0.14902326464653015,
0.038552410900592804,
0.055694397538900375,
-0.057955220341682434,
-0.0233661737293005,
0.09118817001581192,
0.12397737801074982,
0.14583967626094818,
-0.021366750821471214,
-0.028626007959246635,
0.029004426673054695,
0.19620531797409058,
-0.13469526171684265,
0.10371150821447372,
0.13814030587673187,
-0.04545360431075096,
0.08360563963651657,
0.1560150384902954,
0.029186224564909935,
-0.08317049592733383,
0.05044832453131676,
0.04082648828625679,
-0.043159641325473785,
-0.2666129767894745,
-0.0534592866897583,
0.012832709588110447,
-0.06255637854337692,
0.09786593168973923,
0.10183793306350708,
0.11542957276105881,
0.034910861402750015,
-0.07166364789009094,
-0.043925940990448,
-0.0058974819257855415,
0.11737963557243347,
-0.05490213260054588,
-0.012639665976166725,
0.07686592638492584,
-0.05086168646812439,
0.005355054512619972,
0.10266812145709991,
0.02973790094256401,
0.17442677915096283,
0.020399179309606552,
0.11231429129838943,
0.06195578724145889,
0.08633565157651901,
0.0007386076031252742,
0.02951662428677082,
0.05147615820169449,
0.017203815281391144,
-0.002300140680745244,
-0.10421168059110641,
-0.006156572140753269,
0.1449710875749588,
0.028103826567530632,
0.029669636860489845,
-0.0018948549404740334,
-0.005003341939300299,
0.05121048167347908,
0.1746254414319992,
-0.011592294089496136,
-0.22072425484657288,
-0.0845772922039032,
0.06936841458082199,
-0.06218599155545235,
-0.12968985736370087,
-0.026130788028240204,
0.045467354357242584,
-0.17519839107990265,
0.026703642681241035,
-0.027433741837739944,
0.0919293761253357,
-0.09345759451389313,
-0.02221956104040146,
0.03687324374914169,
0.084866963326931,
-0.014529162086546421,
0.08703910559415817,
-0.14498743414878845,
0.11886418610811234,
0.02978132851421833,
0.09024628251791,
-0.11081171780824661,
0.07909037172794342,
-0.007550720125436783,
0.009180475026369095,
0.19379350543022156,
-0.011335089802742004,
-0.03514958545565605,
-0.08774717897176743,
-0.11210042238235474,
-0.013537433929741383,
0.12687496840953827,
-0.1243172138929367,
0.08773399889469147,
-0.015198243781924248,
-0.044079482555389404,
0.00937260314822197,
-0.12100647389888763,
-0.17273177206516266,
-0.19628387689590454,
0.05585884302854538,
-0.09575839340686798,
0.025643249973654747,
-0.11914430558681488,
-0.07089093327522278,
-0.02952558360993862,
0.241120383143425,
-0.1745356321334839,
-0.06510113179683685,
-0.1468164622783661,
-0.046294767409563065,
0.1662203073501587,
-0.04437198117375374,
0.0718095526099205,
-0.0208172257989645,
0.20345525443553925,
0.005988610442727804,
-0.004939318168908358,
0.06724198162555695,
-0.08892562240362167,
-0.16873881220817566,
-0.06771010160446167,
0.1510489284992218,
0.11680185794830322,
0.04907919466495514,
-0.002248800592496991,
0.0011772146681323647,
-0.016943959519267082,
-0.1137804463505745,
-0.0033210667315870523,
0.16037839651107788,
0.03878779336810112,
0.025986969470977783,
-0.05243593826889992,
-0.08797456324100494,
-0.06899320334196091,
-0.06853509694337845,
0.06221301481127739,
0.19590823352336884,
-0.10376439243555069,
0.1700313836336136,
0.147536963224411,
-0.07305635511875153,
-0.23175598680973053,
0.035342130810022354,
0.04983805492520332,
0.0014306638622656465,
0.04886869341135025,
-0.18252557516098022,
0.10521943867206573,
0.019543392583727837,
-0.05505957826972008,
0.13485197722911835,
-0.1557481735944748,
-0.1552847921848297,
0.0722852572798729,
0.03904085233807564,
-0.22423844039440155,
-0.1354004591703415,
-0.09622503817081451,
-0.05825018882751465,
-0.14065024256706238,
0.06054598465561867,
-0.002136280992999673,
0.015948504209518433,
0.03500790148973465,
-0.0015643214574083686,
0.027123261243104935,
-0.058935679495334625,
0.18609118461608887,
-0.004065449349582195,
0.020676052197813988,
-0.060264769941568375,
-0.0478842556476593,
0.09839435666799545,
-0.06130504235625267,
0.12208222597837448,
0.004057085141539574,
0.01594383642077446,
-0.10362856835126877,
-0.048314861953258514,
-0.04328322783112526,
0.05154227837920189,
-0.07548051327466965,
-0.10070807486772537,
-0.043625857681035995,
0.08841723203659058,
0.07005169242620468,
-0.03383097052574158,
0.00549331633374095,
-0.07189501076936722,
0.10019614547491074,
0.17795267701148987,
0.17573626339435577,
0.009926567785441875,
-0.07241068035364151,
0.01677953451871872,
-0.04142116755247116,
0.044231921434402466,
-0.2513144314289093,
0.03756171092391014,
0.06098250672221184,
0.029438555240631104,
0.09217222779989243,
-0.020435843616724014,
-0.1820858269929886,
-0.04050002992153168,
0.08094815909862518,
-0.05452597141265869,
-0.22617179155349731,
-0.019085140898823738,
0.0954197570681572,
-0.2020406424999237,
-0.007372708059847355,
0.03995226323604584,
-0.048725228756666183,
-0.023169852793216705,
0.00010950004070764408,
0.06317184865474701,
0.002471912419423461,
0.09773622453212738,
0.0735151618719101,
0.09715340286493301,
-0.08337292820215225,
0.10562895983457565,
0.10150538384914398,
-0.09572599828243256,
0.03605884686112404,
0.06754924356937408,
-0.05300498008728027,
-0.043293699622154236,
0.03665391728281975,
0.033023297786712646,
0.005234600510448217,
-0.060321882367134094,
0.013913018628954887,
-0.036497246474027634,
0.044923391193151474,
0.08326134830713272,
0.03754979372024536,
-0.013354414142668247,
0.06462216377258301,
0.03401726484298706,
-0.10898099094629288,
0.10366570204496384,
0.01731540448963642,
0.04105307161808014,
-0.08384523540735245,
-0.019968897104263306,
0.035425446927547455,
0.030576206743717194,
-0.01765924133360386,
-0.02306121215224266,
-0.02860277332365513,
-0.01614218018949032,
-0.14299540221691132,
-0.023106401786208153,
-0.07243485748767853,
0.006181265693157911,
0.014656842686235905,
-0.031884219497442245,
-0.011233693920075893,
0.02475680410861969,
-0.06979699432849884,
-0.07426341623067856,
-0.006949664559215307,
0.09833318740129471,
-0.15115703642368317,
0.008848577737808228,
0.06907843053340912,
-0.11088496446609497,
0.08190931379795074,
-0.008411259390413761,
0.016245156526565552,
0.022527478635311127,
-0.15448406338691711,
0.05601610988378525,
0.0008648968650959432,
0.01916889287531376,
0.025886621326208115,
-0.16471809148788452,
0.004104440100491047,
-0.04661374166607857,
-0.02149827405810356,
-0.00004464812809601426,
-0.02647159807384014,
-0.12325995415449142,
0.06858719140291214,
-0.015622655861079693,
-0.035931166261434555,
-0.02701525390148163,
0.0539589487016201,
0.07888586074113846,
-0.027474910020828247,
0.10445091128349304,
-0.008690856397151947,
0.04941811040043831,
-0.16801609098911285,
-0.02470702864229679,
-0.04982255399227142,
0.019377702847123146,
0.009884213097393513,
-0.007693959400057793,
0.04183054715394974,
-0.00976533442735672,
0.21883612871170044,
-0.05075952783226967,
0.1607085019350052,
0.05847611650824547,
-0.017352959141135216,
-0.0007513365126214921,
0.06180921941995621,
0.05997028574347496,
0.04658793285489082,
0.009480604901909828,
0.023740366101264954,
-0.022450892254710197,
-0.006695089396089315,
-0.15932634472846985,
0.01890849508345127,
0.14999441802501678,
0.06301083415746689,
0.024745315313339233,
0.05866100639104843,
-0.12775006890296936,
-0.12135478109121323,
0.09311001747846603,
-0.026755332946777344,
0.00928465835750103,
-0.08245618641376495,
0.1358020007610321,
0.14980104565620422,
-0.14000412821769714,
0.05256148427724838,
-0.06134212389588356,
-0.05217423290014267,
-0.10388828068971634,
-0.12032219022512436,
-0.05887215584516525,
-0.053666237741708755,
0.002330566756427288,
-0.03760887682437897,
0.054546963423490524,
0.03344334661960602,
-0.009351172484457493,
-0.00022941511997487396,
0.13597318530082703,
-0.019751882180571556,
-0.0028988157864660025,
0.048313532024621964,
0.03693558648228645,
0.02373051457107067,
-0.05275435373187065,
0.02940409444272518,
0.02539868652820587,
0.032232340425252914,
0.06546790152788162,
0.033412106335163116,
-0.047448933124542236,
0.03804153576493263,
-0.0025254099164158106,
-0.11207924783229828,
0.019641218706965446,
-0.00460948096588254,
-0.0742158442735672,
0.1268945336341858,
0.0407399944961071,
0.010224059224128723,
-0.03741471841931343,
0.24361543357372284,
-0.06653323769569397,
-0.06378097087144852,
-0.13251738250255585,
0.10491154342889786,
-0.0027236645109951496,
0.06476365029811859,
0.023412218317389488,
-0.1284150779247284,
0.005243356805294752,
0.13858191668987274,
0.12181595712900162,
0.0045748427510261536,
0.009228081442415714,
0.0518609918653965,
0.0025186820421367884,
-0.06998204439878464,
0.054019294679164886,
0.06992026418447495,
0.12919506430625916,
-0.07847554981708527,
0.07680778950452805,
0.0006860480643808842,
-0.08370215445756912,
-0.02947772853076458,
0.11312682181596756,
-0.0409729965031147,
0.03491825982928276,
-0.047444481402635574,
0.10916327685117722,
-0.05787910893559456,
-0.29412412643432617,
0.02350960113108158,
-0.09588567912578583,
-0.15202060341835022,
-0.018367812037467957,
0.05944539234042168,
-0.02624768204987049,
0.018029648810625076,
0.06971040368080139,
-0.06011629104614258,
0.20098382234573364,
0.0335683599114418,
-0.07864278554916382,
-0.0664360448718071,
0.04837050288915634,
-0.06564252078533173,
0.2949807047843933,
0.008418165147304535,
0.02863333560526371,
0.10770907253026962,
-0.03253700211644173,
-0.18271861970424652,
0.010723991319537163,
0.1133992001414299,
-0.08056149631738663,
0.08200647681951523,
0.19000613689422607,
-0.012578671798110008,
0.1209007054567337,
0.05294662341475487,
-0.047376248985528946,
0.04217283055186272,
-0.03389401361346245,
-0.051268599927425385,
-0.10752558708190918,
0.058453381061553955,
-0.05909625440835953,
0.15447644889354706,
0.10152646154165268,
-0.05671518296003342,
-0.004550917539745569,
-0.05555408447980881,
0.04875178262591362,
0.01804669201374054,
0.12263146042823792,
0.02951994352042675,
-0.1865430772304535,
0.032826557755470276,
-0.01144319772720337,
0.10186848044395447,
-0.25588861107826233,
-0.08421015739440918,
0.08833149075508118,
-0.011924264021217823,
-0.05105875805020332,
0.10560628771781921,
0.057650718837976456,
0.04243382066488266,
-0.043439045548439026,
-0.10480839014053345,
-0.02186836116015911,
0.14663739502429962,
-0.1469624787569046,
-0.025013303384184837
] |
null | null | transformers |
<img src="aya-fig1.png" alt="Aya model summary image" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
# Model Card for Aya 101
## Model Summary
> The Aya model is a massively multilingual generative language model that follows instructions in 101 languages.
> Aya outperforms [mT0](https://huggingface.co/bigscience/mt0-xxl) and [BLOOMZ](https://huggingface.co/bigscience/bloomz) a wide variety of automatic and human evaluations despite covering double the number of languages.
> The Aya model is trained using [xP3x](https://huggingface.co/datasets/CohereForAI/xP3x), [Aya Dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset), [Aya Collection](https://huggingface.co/datasets/CohereForAI/aya_collection), a subset of [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses) and ShareGPT-Command.
> We release the checkpoints under a Apache-2.0 license to further our mission of multilingual technologies empowering a
> multilingual world.
- **Developed by:** [Cohere For AI]((https://cohere.for.ai))
- **Model type:** a Transformer style autoregressive massively multilingual language model.
- **Paper**: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](https://arxiv.org/abs/2402.07827)
- **Point of Contact**: Cohere For AI: [cohere.for.ai](https://cohere.for.ai)
- **Languages**: Refer to the list of languages in the `language` section of this model card.
- **License**: Apache-2.0
- **Model**: [Aya-101](https://huggingface.co/CohereForAI/aya-101)
- **Model Size**: 13 billion parameters
- **Datasets**: [xP3x](https://huggingface.co/datasets/CohereForAI/xP3x), [Aya Dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset), [Aya Collection](https://huggingface.co/datasets/CohereForAI/aya_collection), [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses), ShareGPT-Command.
## Use
```bash
# pip install -q transformers
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
checkpoint = "CohereForAI/aya-101"
tokenizer = AutoTokenizer.from_pretrained(checkpoint)
aya_model = AutoModelForSeq2SeqLM.from_pretrained(checkpoint)
# Turkish to English translation
tur_inputs = tokenizer.encode("Translate to English: Aya cok dilli bir dil modelidir.", return_tensors="pt")
tur_outputs = aya_model.generate(tur_inputs, max_new_tokens=128)
print(tokenizer.decode(tur_outputs[0]))
# Aya is a multi-lingual language model
# Q: Why are there so many languages in India?
hin_inputs = tokenizer.encode("भारत में इतनी सारी भाषाएँ क्यों हैं?", return_tensors="pt")
hin_outputs = aya_model.generate(hin_inputs, max_new_tokens=128)
print(tokenizer.decode(hin_outputs[0]))
# Expected output: भारत में कई भाषाएँ हैं और विभिन्न भाषाओं के बोली जाने वाले लोग हैं। यह विभिन्नता भाषाई विविधता और सांस्कृतिक विविधता का परिणाम है। Translates to "India has many languages and people speaking different languages. This diversity is the result of linguistic diversity and cultural diversity."
```
## Model Details
### Finetuning
- Architecture: Same as [mt5-xxl](https://huggingface.co/google/mt5-xxl)
- Number of Samples seen during Finetuning: 25M
- Batch size: 256
- Hardware: TPUv4-128
- Software: T5X, Jax
### Data Sources
The Aya model is trained on the following datasets:
- [xP3x](https://huggingface.co/datasets/CohereForAI/xP3x)
- [Aya Dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset)
- [Aya Collection](https://huggingface.co/datasets/CohereForAI/aya_collection)
- [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses)
- ShareGPT-Command
All datasets are subset to the 101 languages supported by [mT5](https://huggingface.co/google/mt5-xxl). See the [paper](https://arxiv.org/abs/2402.07827) for details about filtering and pruning.
## Evaluation
We refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance.
## Bias, Risks, and Limitations
For a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer to Sections 6 and 7 of our paper: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](https://arxiv.org/abs/2402.07827).
We hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.
## Citation
**BibTeX:**
```
@article{üstün2024aya,
title={Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model},
author={Ahmet Üstün and Viraat Aryabumi and Zheng-Xin Yong and Wei-Yin Ko and Daniel D'souza and Gbemileke Onilude and Neel Bhandari and Shivalika Singh and Hui-Lee Ooi and Amr Kayid and Freddie Vargus and Phil Blunsom and Shayne Longpre and Niklas Muennighoff and Marzieh Fadaee and Julia Kreutzer and Sara Hooker},
journal={arXiv preprint arXiv:2402.07827},
year={2024}
}
```
## Languages Covered
Below is the list of languages used in finetuning the Aya Model. We group languages into higher-, mid-, and lower-resourcedness based on a language classification by [Joshi et. al, 2020](https://microsoft.github.io/linguisticdiversity/). For further details, we refer to our [paper](https://arxiv.org/abs/2402.07827)
| ISO Code | Language Name | Script | Family | Subgrouping | Resourcedness |
| :------- | :-------------- | :----------: | :-------------: | :---------------: | :-----------: |
| afr | Afrikaans | Latin | Indo-European | Germanic | Mid |
| amh | Amharic | Ge'ez | Afro-Asiatic | Semitic | Low |
| ara | Arabic | Arabic | Afro-Asiatic | Semitic | High |
| aze | Azerbaijani | Arabic/Latin | Turkic | Common Turkic | Low |
| bel | Belarusian | Cyrillic | Indo-European | Balto-Slavic | Mid |
| ben | Bengali | Bengali | Indo-European | Indo-Aryan | Mid |
| bul | Bulgarian | Cyrillic | Indo-European | Balto-Slavic | Mid |
| cat | Catalan | Latin | Indo-European | Italic | High |
| ceb | Cebuano | Latin | Austronesian | Malayo-Polynesian | Mid |
| ces | Czech | Latin | Indo-European | Balto-Slavic | High |
| cym | Welsh | Latin | Indo-European | Celtic | Low |
| dan | Danish | Latin | Indo-European | Germanic | Mid |
| deu | German | Latin | Indo-European | Germanic | High |
| ell | Greek | Greek | Indo-European | Graeco-Phrygian | Mid |
| eng | English | Latin | Indo-European | Germanic | High |
| epo | Esperanto | Latin | Constructed | Esperantic | Low |
| est | Estonian | Latin | Uralic | Finnic | Mid |
| eus | Basque | Latin | Basque | - | High |
| fin | Finnish | Latin | Uralic | Finnic | High |
| fil | Tagalog | Latin | Austronesian | Malayo-Polynesian | Mid |
| fra | French | Latin | Indo-European | Italic | High |
| fry | Western Frisian | Latin | Indo-European | Germanic | Low |
| gla | Scottish Gaelic | Latin | Indo-European | Celtic | Low |
| gle | Irish | Latin | Indo-European | Celtic | Low |
| glg | Galician | Latin | Indo-European | Italic | Mid |
| guj | Gujarati | Gujarati | Indo-European | Indo-Aryan | Low |
| hat | Haitian Creole | Latin | Indo-European | Italic | Low |
| hau | Hausa | Latin | Afro-Asiatic | Chadic | Low |
| heb | Hebrew | Hebrew | Afro-Asiatic | Semitic | Mid |
| hin | Hindi | Devanagari | Indo-European | Indo-Aryan | High |
| hun | Hungarian | Latin | Uralic | - | High |
| hye | Armenian | Armenian | Indo-European | Armenic | Low |
| ibo | Igbo | Latin | Atlantic-Congo | Benue-Congo | Low |
| ind | Indonesian | Latin | Austronesian | Malayo-Polynesian | Mid |
| isl | Icelandic | Latin | Indo-European | Germanic | Low |
| ita | Italian | Latin | Indo-European | Italic | High |
| jav | Javanese | Latin | Austronesian | Malayo-Polynesian | Low |
| jpn | Japanese | Japanese | Japonic | Japanesic | High |
| kan | Kannada | Kannada | Dravidian | South Dravidian | Low |
| kat | Georgian | Georgian | Kartvelian | Georgian-Zan | Mid |
| kaz | Kazakh | Cyrillic | Turkic | Common Turkic | Mid |
| khm | Khmer | Khmer | Austroasiatic | Khmeric | Low |
| kir | Kyrgyz | Cyrillic | Turkic | Common Turkic | Low |
| kor | Korean | Hangul | Koreanic | Korean | High |
| kur | Kurdish | Latin | Indo-European | Iranian | Low |
| lao | Lao | Lao | Tai-Kadai | Kam-Tai | Low |
| lav | Latvian | Latin | Indo-European | Balto-Slavic | Mid |
| lat | Latin | Latin | Indo-European | Italic | Mid |
| lit | Lithuanian | Latin | Indo-European | Balto-Slavic | Mid |
| ltz | Luxembourgish | Latin | Indo-European | Germanic | Low |
| mal | Malayalam | Malayalam | Dravidian | South Dravidian | Low |
| mar | Marathi | Devanagari | Indo-European | Indo-Aryan | Low |
| mkd | Macedonian | Cyrillic | Indo-European | Balto-Slavic | Low |
| mlg | Malagasy | Latin | Austronesian | Malayo-Polynesian | Low |
| mlt | Maltese | Latin | Afro-Asiatic | Semitic | Low |
| mon | Mongolian | Cyrillic | Mongolic-Khitan | Mongolic | Low |
| mri | Maori | Latin | Austronesian | Malayo-Polynesian | Low |
| msa | Malay | Latin | Austronesian | Malayo-Polynesian | Mid |
| mya | Burmese | Myanmar | Sino-Tibetan | Burmo-Qiangic | Low |
| nep | Nepali | Devanagari | Indo-European | Indo-Aryan | Low |
| nld | Dutch | Latin | Indo-European | Germanic | High |
| nor | Norwegian | Latin | Indo-European | Germanic | Low |
| nso | Northern Sotho | Latin | Atlantic-Congo | Benue-Congo | Low |
| nya | Chichewa | Latin | Atlantic-Congo | Benue-Congo | Low |
| ory | Oriya | Oriya | Indo-European | Indo-Aryan | Low |
| pan | Punjabi | Gurmukhi | Indo-European | Indo-Aryan | Low |
| pes | Persian | Arabic | Indo-European | Iranian | High |
| pol | Polish | Latin | Indo-European | Balto-Slavic | High |
| por | Portuguese | Latin | Indo-European | Italic | High |
| pus | Pashto | Arabic | Indo-European | Iranian | Low |
| ron | Romanian | Latin | Indo-European | Italic | Mid |
| rus | Russian | Cyrillic | Indo-European | Balto-Slavic | High |
| sin | Sinhala | Sinhala | Indo-European | Indo-Aryan | Low |
| slk | Slovak | Latin | Indo-European | Balto-Slavic | Mid |
| slv | Slovenian | Latin | Indo-European | Balto-Slavic | Mid |
| smo | Samoan | Latin | Austronesian | Malayo-Polynesian | Low |
| sna | Shona | Latin | Indo-European | Indo-Aryan | Low |
| snd | Sindhi | Arabic | Indo-European | Indo-Aryan | Low |
| som | Somali | Latin | Afro-Asiatic | Cushitic | Low |
| sot | Southern Sotho | Latin | Atlantic-Congo | Benue-Congo | Low |
| spa | Spanish | Latin | Indo-European | Italic | High |
| sqi | Albanian | Latin | Indo-European | Albanian | Low |
| srp | Serbian | Cyrillic | Indo-European | Balto-Slavic | High |
| sun | Sundanese | Latin | Austronesian | Malayo-Polynesian | Low |
| swa | Swahili | Latin | Atlantic-Congo | Benue-Congo | Low |
| swe | Swedish | Latin | Indo-European | Germanic | High |
| tam | Tamil | Tamil | Dravidian | South Dravidian | Mid |
| tel | Telugu | Telugu | Dravidian | South Dravidian | Low |
| tgk | Tajik | Cyrillic | Indo-European | Iranian | Low |
| tha | Thai | Thai | Tai-Kadai | Kam-Tai | Mid |
| tur | Turkish | Latin | Turkic | Common Turkic | High |
| twi | Twi | Latin | Atlantic-Congo | Niger-Congo | Low |
| ukr | Ukrainian | Cyrillic | Indo-European | Balto-Slavic | Mid |
| urd | Urdu | Arabic | Indo-European | Indo-Aryan | Mid |
| uzb | Uzbek | Latin | Turkic | Common Turkic | Mid |
| vie | Vietnamese | Latin | Austroasiatic | Vietic | High |
| xho | Xhosa | Latin | Atlantic-Congo | Benue-Congo | Low |
| yid | Yiddish | Hebrew | Indo-European | Germanic | Low |
| yor | Yoruba | Latin | Atlantic-Congo | Benue-Congo | Low |
| zho | Chinese | Han | Sino-Tibetan | Sinitic | High |
| zul | Zulu | Latin | Atlantic-Congo | Benue-Congo | Low |
## Model Card Contact
For errors in this model card, contact Ahmet or Viraat, `{ahmet, viraat} at cohere dot com`.
| {"language": ["afr", "amh", "ara", "aze", "bel", "ben", "bul", "cat", "ceb", "ces", "cym", "dan", "deu", "ell", "eng", "epo", "est", "eus", "fin", "fil", "fra", "fry", "gla", "gle", "glg", "guj", "hat", "hau", "heb", "hin", "hun", "hye", "ibo", "ind", "isl", "ita", "jav", "jpn", "kan", "kat", "kaz", "khm", "kir", "kor", "kur", "lao", "lav", "lat", "lit", "ltz", "mal", "mar", "mkd", "mlg", "mlt", "mon", "mri", "msa", "mya", "nep", "nld", "nor", "nso", "nya", "ory", "pan", "pes", "pol", "por", "pus", "ron", "rus", "sin", "slk", "slv", "smo", "sna", "snd", "som", "sot", "spa", "sqi", "srp", "sun", "swa", "swe", "tam", "tel", "tgk", "tha", "tur", "twi", "ukr", "urd", "uzb", "vie", "xho", "yid", "yor", "zho", "zul"], "license": "apache-2.0", "datasets": ["CohereForAI/xP3x", "CohereForAI/aya_dataset", "CohereForAI/aya_collection", "DataProvenanceInitiative/Commercially-Verified-Licenses", "CohereForAI/aya_evaluation_suite"], "metrics": ["accuracy", "bleu"]} | text2text-generation | CohereForAI/aya-101 | [
"transformers",
"safetensors",
"t5",
"text2text-generation",
"afr",
"amh",
"ara",
"aze",
"bel",
"ben",
"bul",
"cat",
"ceb",
"ces",
"cym",
"dan",
"deu",
"ell",
"eng",
"epo",
"est",
"eus",
"fin",
"fil",
"fra",
"fry",
"gla",
"gle",
"glg",
"guj",
"hat",
"hau",
"heb",
"hin",
"hun",
"hye",
"ibo",
"ind",
"isl",
"ita",
"jav",
"jpn",
"kan",
"kat",
"kaz",
"khm",
"kir",
"kor",
"kur",
"lao",
"lav",
"lat",
"lit",
"ltz",
"mal",
"mar",
"mkd",
"mlg",
"mlt",
"mon",
"mri",
"msa",
"mya",
"nep",
"nld",
"nor",
"nso",
"nya",
"ory",
"pan",
"pes",
"pol",
"por",
"pus",
"ron",
"rus",
"sin",
"slk",
"slv",
"smo",
"sna",
"snd",
"som",
"sot",
"spa",
"sqi",
"srp",
"sun",
"swa",
"swe",
"tam",
"tel",
"tgk",
"tha",
"tur",
"twi",
"ukr",
"urd",
"uzb",
"vie",
"xho",
"yid",
"yor",
"zho",
"zul",
"dataset:CohereForAI/xP3x",
"dataset:CohereForAI/aya_dataset",
"dataset:CohereForAI/aya_collection",
"dataset:DataProvenanceInitiative/Commercially-Verified-Licenses",
"dataset:CohereForAI/aya_evaluation_suite",
"arxiv:2402.07827",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] | 2024-02-08T18:08:25+00:00 | [
"2402.07827"
] | [
"afr",
"amh",
"ara",
"aze",
"bel",
"ben",
"bul",
"cat",
"ceb",
"ces",
"cym",
"dan",
"deu",
"ell",
"eng",
"epo",
"est",
"eus",
"fin",
"fil",
"fra",
"fry",
"gla",
"gle",
"glg",
"guj",
"hat",
"hau",
"heb",
"hin",
"hun",
"hye",
"ibo",
"ind",
"isl",
"ita",
"jav",
"jpn",
"kan",
"kat",
"kaz",
"khm",
"kir",
"kor",
"kur",
"lao",
"lav",
"lat",
"lit",
"ltz",
"mal",
"mar",
"mkd",
"mlg",
"mlt",
"mon",
"mri",
"msa",
"mya",
"nep",
"nld",
"nor",
"nso",
"nya",
"ory",
"pan",
"pes",
"pol",
"por",
"pus",
"ron",
"rus",
"sin",
"slk",
"slv",
"smo",
"sna",
"snd",
"som",
"sot",
"spa",
"sqi",
"srp",
"sun",
"swa",
"swe",
"tam",
"tel",
"tgk",
"tha",
"tur",
"twi",
"ukr",
"urd",
"uzb",
"vie",
"xho",
"yid",
"yor",
"zho",
"zul"
] | TAGS
#transformers #safetensors #t5 #text2text-generation #afr #amh #ara #aze #bel #ben #bul #cat #ceb #ces #cym #dan #deu #ell #eng #epo #est #eus #fin #fil #fra #fry #gla #gle #glg #guj #hat #hau #heb #hin #hun #hye #ibo #ind #isl #ita #jav #jpn #kan #kat #kaz #khm #kir #kor #kur #lao #lav #lat #lit #ltz #mal #mar #mkd #mlg #mlt #mon #mri #msa #mya #nep #nld #nor #nso #nya #ory #pan #pes #pol #por #pus #ron #rus #sin #slk #slv #smo #sna #snd #som #sot #spa #sqi #srp #sun #swa #swe #tam #tel #tgk #tha #tur #twi #ukr #urd #uzb #vie #xho #yid #yor #zho #zul #dataset-CohereForAI/xP3x #dataset-CohereForAI/aya_dataset #dataset-CohereForAI/aya_collection #dataset-DataProvenanceInitiative/Commercially-Verified-Licenses #dataset-CohereForAI/aya_evaluation_suite #arxiv-2402.07827 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
| 
Model Card for Aya 101
======================
Model Summary
-------------
>
> The Aya model is a massively multilingual generative language model that follows instructions in 101 languages.
> Aya outperforms mT0 and BLOOMZ a wide variety of automatic and human evaluations despite covering double the number of languages.
> The Aya model is trained using xP3x, Aya Dataset, Aya Collection, a subset of DataProvenance collection and ShareGPT-Command.
> We release the checkpoints under a Apache-2.0 license to further our mission of multilingual technologies empowering a
> multilingual world.
>
>
>
* Developed by: Cohere For AI)
* Model type: a Transformer style autoregressive massively multilingual language model.
* Paper: Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model
* Point of Contact: Cohere For AI: URL
* Languages: Refer to the list of languages in the 'language' section of this model card.
* License: Apache-2.0
* Model: Aya-101
* Model Size: 13 billion parameters
* Datasets: xP3x, Aya Dataset, Aya Collection, DataProvenance collection, ShareGPT-Command.
Use
---
Model Details
-------------
### Finetuning
* Architecture: Same as mt5-xxl
* Number of Samples seen during Finetuning: 25M
* Batch size: 256
* Hardware: TPUv4-128
* Software: T5X, Jax
### Data Sources
The Aya model is trained on the following datasets:
* xP3x
* Aya Dataset
* Aya Collection
* DataProvenance collection
* ShareGPT-Command
All datasets are subset to the 101 languages supported by mT5. See the paper for details about filtering and pruning.
Evaluation
----------
We refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance.
Bias, Risks, and Limitations
----------------------------
For a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer to Sections 6 and 7 of our paper: Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model.
We hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.
BibTeX:
Languages Covered
-----------------
Below is the list of languages used in finetuning the Aya Model. We group languages into higher-, mid-, and lower-resourcedness based on a language classification by Joshi et. al, 2020. For further details, we refer to our paper
Model Card Contact
------------------
For errors in this model card, contact Ahmet or Viraat, '{ahmet, viraat} at cohere dot com'.
| [
"### Finetuning\n\n\n* Architecture: Same as mt5-xxl\n* Number of Samples seen during Finetuning: 25M\n* Batch size: 256\n* Hardware: TPUv4-128\n* Software: T5X, Jax",
"### Data Sources\n\n\nThe Aya model is trained on the following datasets:\n\n\n* xP3x\n* Aya Dataset\n* Aya Collection\n* DataProvenance collection\n* ShareGPT-Command\n\n\nAll datasets are subset to the 101 languages supported by mT5. See the paper for details about filtering and pruning.\n\n\nEvaluation\n----------\n\n\nWe refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance.\n\n\nBias, Risks, and Limitations\n----------------------------\n\n\nFor a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer to Sections 6 and 7 of our paper: Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model.\n\n\nWe hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.\n\n\nBibTeX:\n\n\nLanguages Covered\n-----------------\n\n\nBelow is the list of languages used in finetuning the Aya Model. We group languages into higher-, mid-, and lower-resourcedness based on a language classification by Joshi et. al, 2020. For further details, we refer to our paper\n\n\n\nModel Card Contact\n------------------\n\n\nFor errors in this model card, contact Ahmet or Viraat, '{ahmet, viraat} at cohere dot com'."
] | [
"TAGS\n#transformers #safetensors #t5 #text2text-generation #afr #amh #ara #aze #bel #ben #bul #cat #ceb #ces #cym #dan #deu #ell #eng #epo #est #eus #fin #fil #fra #fry #gla #gle #glg #guj #hat #hau #heb #hin #hun #hye #ibo #ind #isl #ita #jav #jpn #kan #kat #kaz #khm #kir #kor #kur #lao #lav #lat #lit #ltz #mal #mar #mkd #mlg #mlt #mon #mri #msa #mya #nep #nld #nor #nso #nya #ory #pan #pes #pol #por #pus #ron #rus #sin #slk #slv #smo #sna #snd #som #sot #spa #sqi #srp #sun #swa #swe #tam #tel #tgk #tha #tur #twi #ukr #urd #uzb #vie #xho #yid #yor #zho #zul #dataset-CohereForAI/xP3x #dataset-CohereForAI/aya_dataset #dataset-CohereForAI/aya_collection #dataset-DataProvenanceInitiative/Commercially-Verified-Licenses #dataset-CohereForAI/aya_evaluation_suite #arxiv-2402.07827 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n",
"### Finetuning\n\n\n* Architecture: Same as mt5-xxl\n* Number of Samples seen during Finetuning: 25M\n* Batch size: 256\n* Hardware: TPUv4-128\n* Software: T5X, Jax",
"### Data Sources\n\n\nThe Aya model is trained on the following datasets:\n\n\n* xP3x\n* Aya Dataset\n* Aya Collection\n* DataProvenance collection\n* ShareGPT-Command\n\n\nAll datasets are subset to the 101 languages supported by mT5. See the paper for details about filtering and pruning.\n\n\nEvaluation\n----------\n\n\nWe refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance.\n\n\nBias, Risks, and Limitations\n----------------------------\n\n\nFor a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer to Sections 6 and 7 of our paper: Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model.\n\n\nWe hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.\n\n\nBibTeX:\n\n\nLanguages Covered\n-----------------\n\n\nBelow is the list of languages used in finetuning the Aya Model. We group languages into higher-, mid-, and lower-resourcedness based on a language classification by Joshi et. al, 2020. For further details, we refer to our paper\n\n\n\nModel Card Contact\n------------------\n\n\nFor errors in this model card, contact Ahmet or Viraat, '{ahmet, viraat} at cohere dot com'."
] | [
388,
52,
346
] | [
"passage: TAGS\n#transformers #safetensors #t5 #text2text-generation #afr #amh #ara #aze #bel #ben #bul #cat #ceb #ces #cym #dan #deu #ell #eng #epo #est #eus #fin #fil #fra #fry #gla #gle #glg #guj #hat #hau #heb #hin #hun #hye #ibo #ind #isl #ita #jav #jpn #kan #kat #kaz #khm #kir #kor #kur #lao #lav #lat #lit #ltz #mal #mar #mkd #mlg #mlt #mon #mri #msa #mya #nep #nld #nor #nso #nya #ory #pan #pes #pol #por #pus #ron #rus #sin #slk #slv #smo #sna #snd #som #sot #spa #sqi #srp #sun #swa #swe #tam #tel #tgk #tha #tur #twi #ukr #urd #uzb #vie #xho #yid #yor #zho #zul #dataset-CohereForAI/xP3x #dataset-CohereForAI/aya_dataset #dataset-CohereForAI/aya_collection #dataset-DataProvenanceInitiative/Commercially-Verified-Licenses #dataset-CohereForAI/aya_evaluation_suite #arxiv-2402.07827 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n### Finetuning\n\n\n* Architecture: Same as mt5-xxl\n* Number of Samples seen during Finetuning: 25M\n* Batch size: 256\n* Hardware: TPUv4-128\n* Software: T5X, Jax"
] | [
-0.11557380855083466,
0.1760784387588501,
-0.010248380713164806,
0.051504530012607574,
0.049763016402721405,
0.02448815107345581,
0.09948809444904327,
0.11077605187892914,
0.004310523625463247,
0.15540149807929993,
0.11106549203395844,
0.11620792746543884,
0.1013154536485672,
0.08953925967216492,
0.03869448974728584,
-0.16301922500133514,
0.01749277673661709,
-0.03977310657501221,
-0.06422831863164902,
0.11265718936920166,
0.08252716064453125,
-0.10400548577308655,
0.12707170844078064,
-0.054890166968107224,
0.03035922907292843,
-0.024692382663488388,
-0.08572176098823547,
-0.0260690376162529,
0.04031437262892723,
0.02007756382226944,
-0.015974048525094986,
0.055728912353515625,
0.030178667977452278,
-0.17458011209964752,
0.02819371223449707,
0.049326226115226746,
-0.03874342143535614,
0.03451041877269745,
0.06474315375089645,
-0.06484765559434891,
0.1252240389585495,
-0.10118938237428665,
-0.05452049523591995,
0.017128638923168182,
-0.1648964136838913,
-0.18909691274166107,
-0.08388946950435638,
0.10412715375423431,
0.027773035690188408,
0.047967422753572464,
-0.048762254416942596,
0.11775633692741394,
-0.06154630705714226,
0.061933767050504684,
0.18426939845085144,
-0.12240834534168243,
-0.07610411942005157,
0.04287688061594963,
0.06031125783920288,
0.06996715068817139,
-0.07638728618621826,
0.030416272580623627,
0.03293713182210922,
-0.010868877172470093,
0.008057937026023865,
-0.001779621234163642,
0.05691716820001602,
0.03334067389369011,
-0.08480743318796158,
-0.052581410855054855,
0.18125486373901367,
0.06047246232628822,
0.06642141938209534,
-0.07990899682044983,
-0.024374360218644142,
-0.12550118565559387,
-0.0362517423927784,
-0.01827990449965,
0.02939276583492756,
-0.009076740592718124,
-0.04360201209783554,
0.1482936441898346,
-0.07979948073625565,
0.03088896907866001,
0.049399714916944504,
-0.08045952767133713,
0.004381205420941114,
-0.0010481084464117885,
-0.059867095202207565,
0.04851061850786209,
0.1002454161643982,
-0.16908414661884308,
-0.0005425821291282773,
0.010443612933158875,
0.003538386197760701,
0.023506954312324524,
0.0876193717122078,
0.11950109153985977,
0.13361643254756927,
0.1239963099360466,
-0.04087654873728752,
0.07420460134744644,
0.09899763017892838,
0.03960123285651207,
0.03555445373058319,
0.024311453104019165,
-0.09731069207191467,
-0.184335857629776,
-0.12189072370529175,
0.07696456462144852,
0.010274242609739304,
-0.00249165459536016,
0.035140734165906906,
0.049128029495477676,
0.047984544187784195,
0.05162009596824646,
0.09678217023611069,
0.07715144753456116,
-0.07678677886724472,
-0.013561952859163284,
0.1551397144794464,
-0.06454892456531525,
0.012753551825881004,
0.07539087533950806,
-0.004101856146007776,
-0.035557687282562256,
-0.053920257836580276,
0.015043750405311584,
0.03587930649518967,
0.0689845159649849,
-0.07481475919485092,
-0.04692893102765083,
-0.0008305644150823355,
-0.03445058315992355,
0.07463288307189941,
-0.0907876268029213,
0.040227390825748444,
-0.10776800662279129,
-0.10502631217241287,
-0.08964679390192032,
0.01735358126461506,
-0.05340782552957535,
-0.012186394073069096,
-0.04185105115175247,
-0.09984321147203445,
0.01037056278437376,
-0.01219497062265873,
0.09274331480264664,
-0.09387112408876419,
0.0877537801861763,
0.007332442793995142,
0.08049692213535309,
0.07897581160068512,
-0.004357486497610807,
-0.10039058327674866,
0.06546530872583389,
-0.011823626235127449,
0.03272779658436775,
-0.08311732113361359,
0.0966913029551506,
-0.12260287255048752,
-0.06335876882076263,
-0.08618064224720001,
0.0064762351103127,
0.04380817711353302,
0.20741447806358337,
-0.2161683291196823,
-0.08051639050245285,
0.15339243412017822,
-0.07797650247812271,
-0.16341926157474518,
0.08448711037635803,
0.06294746696949005,
0.0638907253742218,
0.07276328653097153,
0.10930480808019638,
0.039371341466903687,
-0.10585255175828934,
-0.1660461276769638,
-0.07115960121154785,
0.06934205442667007,
0.09188508242368698,
0.11196570098400116,
0.007403952535241842,
0.12699776887893677,
0.0362883061170578,
-0.07535085082054138,
0.008827490732073784,
-0.09023071080446243,
-0.08254584670066833,
0.0676405057311058,
-0.09619559347629547,
-0.002110956935212016,
0.10580246895551682,
-0.04001227393746376,
-0.07575048506259918,
-0.055046092718839645,
-0.12599359452724457,
0.09237722307443619,
-0.004657548386603594,
-0.023623844608664513,
-0.1161501407623291,
0.07844040542840958,
0.05669749155640602,
0.00845594797283411,
-0.09288603067398071,
0.0586906336247921,
-0.0029789581894874573,
0.011204713024199009,
-0.03194912523031235,
0.05731387436389923,
0.10868345946073532,
-0.010553867556154728,
-0.03835468366742134,
-0.06725950539112091,
0.09189970791339874,
0.014148879796266556,
-0.04916905239224434,
-0.2221391350030899,
0.028979627415537834,
-0.03578423708677292,
0.1405653953552246,
-0.15178897976875305,
0.037108585238456726,
0.12886498868465424,
0.2004091441631317,
0.028508225455880165,
-0.07407211512327194,
-0.02575324848294258,
0.0045242574997246265,
0.000003511304157655104,
-0.045321978628635406,
0.041523829102516174,
-0.014829895459115505,
-0.013851460069417953,
0.07738763093948364,
-0.15954571962356567,
0.09138724207878113,
0.09254366159439087,
-0.004159221891313791,
-0.03385605290532112,
0.07759491354227066,
-0.02215798757970333,
-0.03857727721333504,
0.04070395231246948,
-0.049460429698228836,
-0.0011842030799016356,
0.0010764605831354856,
0.03350328281521797,
-0.0638844221830368,
-0.06589145213365555,
0.048044320195913315,
-0.0013105840189382434,
-0.0829232856631279,
0.1865660846233368,
0.034155309200286865,
-0.1945570558309555,
0.18254457414150238,
0.18219609558582306,
0.10644467920064926,
0.17917127907276154,
-0.009514933452010155,
-0.03657772019505501,
-0.07363291829824448,
-0.0005383977550081909,
0.017487727105617523,
0.13374628126621246,
-0.17073822021484375,
-0.014685089699923992,
0.010904152877628803,
-0.016698135063052177,
0.014513135887682438,
-0.03859966993331909,
-0.02498636581003666,
-0.028510870411992073,
-0.018914170563220978,
0.06284855306148529,
0.05354482680559158,
-0.06648609787225723,
0.082514688372612,
-0.03893371671438217,
-0.008037553168833256,
-0.061674512922763824,
-0.03312767297029495,
-0.08707641065120697,
0.15174955129623413,
-0.08652309328317642,
-0.2107497751712799,
0.04425312578678131,
-0.020410675555467606,
0.0691007524728775,
-0.03060419112443924,
0.03022710606455803,
-0.12034281343221664,
-0.025911377742886543,
-0.041160859167575836,
0.03577163442969322,
-0.013128247112035751,
-0.022770918905735016,
0.043735288083553314,
-0.015354045666754246,
0.024391641840338707,
-0.08144064247608185,
-0.004650040064007044,
0.019679522141814232,
-0.008535572327673435,
0.07449055463075638,
-0.03151295334100723,
0.11291280388832092,
0.0366494245827198,
0.04240690544247627,
0.007227644789963961,
0.009726060554385185,
0.2616457939147949,
-0.14444464445114136,
0.031944409012794495,
0.030825773254036903,
0.04246388003230095,
0.034598927944898605,
0.13859449326992035,
0.03978179767727852,
-0.07866154611110687,
-0.023783164098858833,
0.03945194557309151,
-0.004712789785116911,
-0.11321910470724106,
-0.04491814598441124,
-0.04966099560260773,
0.11090056598186493,
0.0688972920179367,
0.07975729554891586,
-0.00983903557062149,
0.03221006318926811,
-0.08376380056142807,
-0.10403380542993546,
0.011140340939164162,
-0.00001661573514866177,
0.0038793631829321384,
-0.03915612772107124,
0.04921736940741539,
-0.017544075846672058,
-0.01933840848505497,
0.07533810287714005,
0.012250645086169243,
-0.012676378712058067,
0.008420666679739952,
0.10502288490533829,
0.07869354635477066,
0.13429464399814606,
-0.00838833674788475,
-0.03535185754299164,
0.05141843482851982,
-0.0038151175249367952,
-0.015690691769123077,
-0.06844527274370193,
-0.04369809478521347,
0.020207718014717102,
0.06578399986028671,
-0.031863417476415634,
0.006327685434371233,
0.006279533728957176,
0.10109125077724457,
0.08717318624258041,
0.10659942775964737,
-0.22507232427597046,
-0.039822664111852646,
0.06766089797019958,
0.029095688834786415,
-0.028841666877269745,
-0.010426623746752739,
0.07796680927276611,
-0.06559894233942032,
0.10942982137203217,
0.009399465285241604,
0.05657012015581131,
-0.08670124411582947,
0.04152379930019379,
0.03627227619290352,
-0.062409717589616776,
0.006968497298657894,
0.026813454926013947,
-0.17665626108646393,
0.18687930703163147,
0.02518792450428009,
0.04154393449425697,
-0.04871107637882233,
-0.034083541482686996,
0.01776829920709133,
0.019567226991057396,
0.14098863303661346,
0.039583392441272736,
-0.13231734931468964,
-0.10402653366327286,
-0.08349079638719559,
0.022508367896080017,
0.06999076902866364,
-0.10258173197507858,
0.04779138043522835,
0.027579154819250107,
-0.04444405063986778,
-0.0424853190779686,
0.02262750267982483,
-0.10380110889673233,
-0.057365141808986664,
0.08374129980802536,
-0.028921866789460182,
0.04979075491428375,
-0.04337441548705101,
-0.04505586996674538,
-0.207528755068779,
0.048154570162296295,
-0.13188669085502625,
-0.003192705102264881,
-0.03592309355735779,
0.037825483828783035,
0.13407793641090393,
-0.09646402299404144,
-0.03168284893035889,
0.005543985404074192,
-0.03037678450345993,
0.0065988944843411446,
0.020796209573745728,
0.04863889887928963,
-0.09841825813055038,
-0.21310824155807495,
-0.07186198979616165,
0.0530322790145874,
0.09038219600915909,
0.04020095616579056,
-0.03659180551767349,
0.07581330090761185,
-0.04674268141388893,
-0.0586104616522789,
0.12479978054761887,
0.025566818192601204,
0.03823089599609375,
-0.009912731125950813,
0.042601216584444046,
-0.07699508219957352,
-0.07631905376911163,
-0.07925470173358917,
-0.002633903408423066,
0.3210941255092621,
-0.043837323784828186,
0.06883944571018219,
0.13966478407382965,
-0.04689103364944458,
-0.21863652765750885,
-0.10844402760267258,
-0.045168276876211166,
0.030421871691942215,
0.08118502050638199,
-0.10712504386901855,
-0.004770738538354635,
0.047414712607860565,
-0.013720007613301277,
0.022032517939805984,
-0.26195982098579407,
-0.06302722543478012,
0.07797742635011673,
0.00991906225681305,
-0.05397462844848633,
-0.1951911896467209,
-0.06571236252784729,
0.03146875277161598,
-0.10964848101139069,
-0.017286641523241997,
0.022441327571868896,
0.009818769060075283,
-0.018490690737962723,
-0.05429961904883385,
0.024052752181887627,
-0.05146632343530655,
0.1768437922000885,
-0.0019366515334695578,
0.0037100992631167173,
-0.1083701029419899,
-0.07020561397075653,
0.07236846536397934,
-0.030380334705114365,
-0.01919875666499138,
-0.12217661738395691,
0.03243761509656906,
-0.06752602010965347,
0.038457974791526794,
-0.12854957580566406,
0.026208946481347084,
-0.06144467741250992,
0.003963472321629524,
-0.08447995781898499,
0.09321719408035278,
0.06693266332149506,
0.02659471333026886,
0.13851211965084076,
-0.04847787320613861,
0.10677509754896164,
0.04771553352475166,
0.10392221063375473,
0.049785297363996506,
-0.00903953518718481,
-0.06020933389663696,
-0.005421290639787912,
0.04201379418373108,
-0.11261183768510818,
0.013376045040786266,
0.1011015921831131,
0.030070021748542786,
0.06092222034931183,
-0.017943425104022026,
-0.12496506422758102,
-0.01989610306918621,
0.13675525784492493,
-0.0944782942533493,
-0.14856627583503723,
-0.002852234523743391,
-0.01668948121368885,
-0.06678759306669235,
0.004706891719251871,
0.09389976412057877,
-0.016800325363874435,
-0.008601375855505466,
-0.019755523651838303,
0.11350644379854202,
0.0024231846909970045,
0.1968022584915161,
-0.0062306663021445274,
0.0397367924451828,
-0.10900123417377472,
0.06214006990194321,
0.05768199265003204,
-0.08591628074645996,
-0.011427626945078373,
0.0804552510380745,
-0.04996456205844879,
-0.07006261497735977,
-0.014264844357967377,
0.07163101434707642,
0.0827639102935791,
-0.04502367228269577,
-0.07237397134304047,
-0.04666948691010475,
0.06593941152095795,
0.07638846337795258,
0.018174365162849426,
0.06369436532258987,
0.09364087879657745,
0.02082381211221218,
0.021164605394005775,
0.11622250080108643,
0.10771729797124863,
0.007101008668541908,
-0.09515875577926636,
-0.009950942359864712,
-0.03096320480108261,
0.02998981438577175,
0.023361390456557274,
0.031124193221330643,
-0.1253114938735962,
-0.005882119759917259,
-0.018271906301379204,
-0.025098392739892006,
-0.13459308445453644,
-0.025219548493623734,
-0.010816352441906929,
-0.07557137310504913,
-0.0748574510216713,
-0.06271864473819733,
-0.04152262210845947,
-0.019622735679149628,
0.016988618299365044,
0.16159787774085999,
-0.10636626929044724,
-0.020265202969312668,
0.09032430499792099,
-0.07931113243103027,
0.08594053983688354,
0.019974887371063232,
0.01780148595571518,
0.02220105566084385,
-0.1364714503288269,
-0.026817820966243744,
0.028185749426484108,
0.02339702472090721,
-0.011150856502354145,
-0.07977985590696335,
-0.059163060039281845,
-0.03826405853033066,
-0.011414380744099617,
0.031191043555736542,
-0.008747325278818607,
-0.06280746310949326,
0.13198035955429077,
0.013074115850031376,
-0.019151262938976288,
-0.042151451110839844,
0.058778177946805954,
0.04008988291025162,
-0.02826942689716816,
0.0773138627409935,
-0.05197959020733833,
0.009827579371631145,
-0.18287330865859985,
0.03269590437412262,
0.00912005640566349,
-0.09957515448331833,
-0.0251714326441288,
0.017667129635810852,
0.09693499654531479,
-0.035449422895908356,
0.06318730860948563,
-0.032396603375673294,
-0.15666693449020386,
0.034102726727724075,
0.038177747279405594,
-0.07447729259729385,
0.02246151678264141,
0.08796494454145432,
0.002411137567833066,
-0.009609091095626354,
-0.06061043590307236,
0.02489471063017845,
-0.030575331300497055,
0.04800602048635483,
0.1037169098854065,
0.20325706899166107,
0.07213131338357925,
0.024823587387800217,
0.005085045006126165,
-0.13622473180294037,
0.009087583981454372,
0.07455875724554062,
-0.16072429716587067,
0.040256183594465256,
-0.02470412477850914,
0.06485145539045334,
0.17895038425922394,
-0.12747150659561157,
0.015284973196685314,
-0.14185841381549835,
-0.028486119583249092,
-0.02795916236937046,
-0.1617085039615631,
-0.06698152422904968,
-0.0065453494898974895,
0.0205547958612442,
-0.09239260107278824,
0.05361964926123619,
0.0692489966750145,
0.022325485944747925,
-0.027532758191227913,
0.03486282005906105,
0.0751882791519165,
-0.024979256093502045,
0.03279807046055794,
0.029771333560347557,
-0.025470562279224396,
-0.11003369092941284,
0.03523788973689079,
0.002640813123434782,
0.07089226692914963,
0.003854949027299881,
-0.007381095550954342,
-0.07938248664140701,
0.03382023423910141,
0.004059026017785072,
-0.12062938511371613,
-0.014064029790461063,
-0.012509257532656193,
0.003344174474477768,
0.06749642640352249,
0.02569766528904438,
0.05667947605252266,
0.03442753478884697,
0.09962465614080429,
-0.035774532705545425,
-0.04749029874801636,
-0.06236746534705162,
-0.011140123009681702,
-0.0994950532913208,
-0.010124845430254936,
0.051769040524959564,
-0.0842069536447525,
0.03946763277053833,
0.1805759221315384,
0.21517297625541687,
-0.038747962564229965,
0.048906225711107254,
-0.0029815086163580418,
0.017785098403692245,
-0.03289071470499039,
-0.029753655195236206,
0.034240588545799255,
0.17804449796676636,
-0.08663768321275711,
0.08319317549467087,
-0.042526789009571075,
-0.010650348849594593,
-0.024417303502559662,
-0.001787908491678536,
0.009881790727376938,
-0.010333696380257607,
-0.04542297124862671,
0.12745419144630432,
-0.10558253526687622,
-0.13811394572257996,
0.03693266212940216,
-0.1930367350578308,
-0.10008255392313004,
-0.033952247351408005,
0.07647771388292313,
0.10538828372955322,
0.059729550033807755,
-0.0069621168076992035,
-0.016825439408421516,
0.01380064245313406,
0.023015685379505157,
-0.09047005325555801,
-0.03009195625782013,
0.02194460667669773,
-0.044541068375110626,
0.18404175341129303,
-0.06446412950754166,
0.021883996203541756,
0.10925301909446716,
-0.04079241305589676,
-0.08854493498802185,
-0.0014800641220062971,
0.0942835882306099,
-0.05149814859032631,
-0.03162413835525513,
0.02921474725008011,
0.025548847392201424,
0.07828696072101593,
0.13398516178131104,
-0.009163273498415947,
0.02884436585009098,
0.13952729105949402,
0.02618158794939518,
-0.06807572394609451,
0.11937306076288223,
0.00025816133711487055,
0.13202384114265442,
0.07147857546806335,
-0.043127626180648804,
-0.014388005249202251,
-0.012991797178983688,
0.041601985692977905,
0.0024073245003819466,
0.021094156429171562,
-0.03057899698615074,
-0.1808033436536789,
-0.01871316134929657,
-0.06555218249559402,
0.06400173157453537,
-0.021189410239458084,
-0.04290403798222542,
0.0468500554561615,
-0.014037254266440868,
-0.03360281139612198,
0.10191372781991959,
0.06374610215425491,
-0.008259485475718975,
-0.026774486526846886,
-0.05280951038002968,
-0.0059226881712675095,
0.14997684955596924,
-0.0862422063946724,
-0.06394054740667343
] |
null | null | transformers |
# Wiedervereinigung-7b-dpo

This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
```json
{
"first_turn": 7.3,
"second_turn": 6.925,
"categories": {
"writing": 8.425,
"roleplay": 8.6,
"reasoning": 5.4,
"math": 4.35,
"coding": 4.3,
"extraction": 7.975,
"stem": 8.5,
"humanities": 9.35
},
"average": 7.1125
}
```
Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
* [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
* [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
* [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
* [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
All the actual heavylifting has been done by the creators of these models.
## 🧩 Configuration
```yaml
models:
- model: LeoLM/leo-mistral-hessianai-7b
# No parameters necessary for base model
- model: DiscoResearch/DiscoLM_German_7b_v1
parameters:
density: 0.6
weight: 0.25
- model: DRXD1000/Phoenix
parameters:
density: 0.6
weight: 0.25
- model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
parameters:
density: 0.6
weight: 0.25
- model: malteos/hermeo-7b
parameters:
density: 0.6
weight: 0.25
merge_method: dare_ties
base_model: LeoLM/leo-mistral-hessianai-7b
parameters:
int8_mask: true
dtype: bfloat16
```
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["de", "en"], "license": "apache-2.0", "tags": ["merge", "mergekit", "lazymergekit", "DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"], "base_model": ["DiscoResearch/DiscoLM_German_7b_v1", "DRXD1000/Phoenix", "VAGOsolutions/SauerkrautLM-7b-v1-mistral", "malteos/hermeo-7b"]} | text-generation | LoneStriker/Wiedervereinigung-7b-dpo-GPTQ | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"DiscoResearch/DiscoLM_German_7b_v1",
"DRXD1000/Phoenix",
"VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"malteos/hermeo-7b",
"de",
"en",
"base_model:DiscoResearch/DiscoLM_German_7b_v1",
"base_model:DRXD1000/Phoenix",
"base_model:VAGOsolutions/SauerkrautLM-7b-v1-mistral",
"base_model:malteos/hermeo-7b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T18:15:39+00:00 | [] | [
"de",
"en"
] | TAGS
#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Wiedervereinigung-7b-dpo
!image/png
This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
using our german fork of LLaMA-Factory.
## mt-bench-de
Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
Wiedervereinigung-7b itself is a LazyMergekit merge of:
* DiscoResearch/DiscoLM_German_7b_v1
* DRXD1000/Phoenix
* VAGOsolutions/SauerkrautLM-7b-v1-mistral
* malteos/hermeo-7b
All the actual heavylifting has been done by the creators of these models.
## Configuration
## Usage
| [
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
"TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.",
"## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.",
"## Configuration",
"## Usage"
] | [
194,
148,
120,
4,
3
] | [
"passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #DiscoResearch/DiscoLM_German_7b_v1 #DRXD1000/Phoenix #VAGOsolutions/SauerkrautLM-7b-v1-mistral #malteos/hermeo-7b #de #en #base_model-DiscoResearch/DiscoLM_German_7b_v1 #base_model-DRXD1000/Phoenix #base_model-VAGOsolutions/SauerkrautLM-7b-v1-mistral #base_model-malteos/hermeo-7b #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Wiedervereinigung-7b-dpo\n\n!image/png\n\nThis is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.\nIt is a merge of the best german 7B models with 7b parameters as a dare_ties merge. \nSince the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model. \nTherefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo \nusing our german fork of LLaMA-Factory.## mt-bench-de\n\nIs the merged model good? Well, of course. But it is even better with the help of some dpo tuning.\n\n\n\n\nWiedervereinigung-7b itself is a LazyMergekit merge of:\n* DiscoResearch/DiscoLM_German_7b_v1\n* DRXD1000/Phoenix\n* VAGOsolutions/SauerkrautLM-7b-v1-mistral\n* malteos/hermeo-7b\n\nAll the actual heavylifting has been done by the creators of these models.## Configuration## Usage"
] | [
-0.06399306654930115,
-0.029307762160897255,
-0.004331388510763645,
0.03184039518237114,
0.018488328903913498,
0.03748767077922821,
0.10618676990270615,
0.07814439386129379,
0.07063937187194824,
0.07519446313381195,
-0.04871003329753876,
-0.02137892320752144,
0.07951480895280838,
0.10956592857837677,
0.007890605367720127,
-0.23595072329044342,
0.08250588178634644,
-0.03575211763381958,
-0.026412149891257286,
0.03893791511654854,
0.14763003587722778,
-0.05930296331644058,
0.05646873265504837,
-0.0034751244820654392,
-0.05727600306272507,
0.05131329223513603,
-0.026073675602674484,
-0.007598201744258404,
0.09508240222930908,
0.0628398060798645,
0.041950274258852005,
0.0221012644469738,
-0.018701795488595963,
-0.15748009085655212,
0.004922133404761553,
-0.017628159373998642,
-0.04106835648417473,
0.027759606018662453,
0.11616206914186478,
-0.045659661293029785,
0.11880303174257278,
-0.12680159509181976,
-0.018632367253303528,
0.08243369311094284,
-0.12379294633865356,
-0.10155918449163437,
-0.20841088891029358,
0.15503807365894318,
0.07979700714349747,
0.008151463232934475,
-0.04288718104362488,
0.07205401360988617,
-0.054492317140102386,
0.0433170422911644,
0.08753809332847595,
-0.2521143853664398,
-0.04418989270925522,
0.17082107067108154,
0.007189796306192875,
-0.04941372200846672,
-0.05787082388997078,
0.06996709108352661,
0.029280396178364754,
0.0067648859694600105,
-0.04790947958827019,
-0.04312778264284134,
0.2201330065727234,
-0.03288035839796066,
-0.1199050173163414,
-0.024701880291104317,
0.08322399854660034,
0.08842694014310837,
-0.04718442261219025,
-0.15207047760486603,
-0.011922257021069527,
-0.022255953401327133,
-0.04938272759318352,
-0.09886841475963593,
0.003574386006221175,
-0.02095438726246357,
0.07563887536525726,
-0.05726063251495361,
-0.04117260500788689,
0.0583573654294014,
-0.05984218046069145,
0.1629648655653,
-0.0014757480239495635,
-0.03302643075585365,
0.05737544223666191,
0.03662708401679993,
-0.2148275226354599,
-0.08567781001329422,
-0.04239440709352493,
-0.054734885692596436,
-0.1342768371105194,
-0.0431794635951519,
-0.04418531805276871,
-0.08217091858386993,
-0.007381834089756012,
0.23022499680519104,
-0.02492726407945156,
0.02804037742316723,
0.041754867881536484,
0.010314702056348324,
0.04419277235865593,
0.1185997873544693,
-0.05733427032828331,
-0.1965552270412445,
-0.02631615847349167,
-0.01892445981502533,
0.06395958364009857,
0.013402312994003296,
-0.02684073895215988,
-0.01519184559583664,
0.02383972890675068,
0.02890627458691597,
0.02422332763671875,
0.03168574720621109,
-0.07110860198736191,
-0.03037358447909355,
0.14692901074886322,
-0.10123810917139053,
0.04444355517625809,
-0.03552069887518883,
-0.03803017735481262,
0.07610528916120529,
-0.030228903517127037,
0.023446066305041313,
-0.01123465783894062,
0.14718079566955566,
-0.04342922940850258,
-0.02282293513417244,
-0.04051238298416138,
-0.05655231699347496,
0.056725140661001205,
-0.08175330609083176,
-0.06928885728120804,
-0.0921410471200943,
-0.07629581540822983,
-0.0781061053276062,
0.019695017486810684,
-0.04772234335541725,
-0.00727304769679904,
-0.00963415578007698,
-0.03471999987959862,
0.004572734236717224,
0.03700930252671242,
-0.03850525617599487,
-0.006236410234123468,
-0.05285993963479996,
-0.02465738169848919,
0.016344578936696053,
-0.07584261149168015,
0.020161736756563187,
-0.09869375079870224,
0.055850550532341,
-0.3218844532966614,
0.09342993050813675,
-0.11400728672742844,
0.009930809028446674,
-0.1314808875322342,
-0.0335058830678463,
0.016112782061100006,
0.005322059616446495,
0.013146123848855495,
0.11860170215368271,
-0.15911765396595,
-0.09210728108882904,
0.09710568189620972,
-0.13270564377307892,
-0.06123318895697594,
0.13335616886615753,
0.006097014062106609,
-0.007668835576623678,
0.0786266103386879,
0.1425439566373825,
0.14123733341693878,
-0.04838160052895546,
-0.13677071034908295,
-0.06624309718608856,
0.027206536382436752,
0.1544196456670761,
0.06648306548595428,
-0.0522027313709259,
-0.0182489063590765,
0.03598593920469284,
-0.031956933438777924,
-0.0026491705793887377,
-0.016834797337651253,
-0.05705594643950462,
-0.023978661745786667,
-0.002377238357439637,
0.096376933157444,
-0.015163181349635124,
0.002383065177127719,
-0.07838787883520126,
-0.09363235533237457,
0.039118777960538864,
0.0882512778043747,
0.041436195373535156,
0.01435884553939104,
-0.06161881238222122,
0.09994251281023026,
-0.03056234121322632,
0.022030595690011978,
-0.10238077491521835,
-0.15781597793102264,
0.03258318454027176,
-0.15215061604976654,
0.06616183370351791,
0.046265859156847,
0.06192070245742798,
0.027611956000328064,
-0.11285637319087982,
-0.06351099908351898,
-0.02237211912870407,
0.016440147534012794,
0.007992693223059177,
-0.12434909492731094,
-0.08924297988414764,
-0.04348140209913254,
0.22069425880908966,
-0.054984692484140396,
-0.007727441843599081,
0.06026673689484596,
0.2515117824077606,
0.018960142508149147,
-0.014410633593797684,
-0.05082327499985695,
0.07166947424411774,
0.00031130402931012213,
0.0197899267077446,
0.01994451880455017,
-0.04336968809366226,
-0.14637522399425507,
0.095527283847332,
-0.055198993533849716,
-0.02979397401213646,
0.06157594546675682,
0.054885584861040115,
-0.09651610255241394,
-0.011951263062655926,
-0.05258362367749214,
-0.029914764687418938,
0.006216008216142654,
-0.12069366872310638,
0.042220812290906906,
0.04080578312277794,
0.03252416476607323,
-0.0428742878139019,
-0.048637744039297104,
0.01096632331609726,
0.00038561708061024547,
-0.12152905762195587,
0.10360410064458847,
-0.08400224894285202,
-0.21976810693740845,
0.10374713689088821,
0.15290415287017822,
0.07434318959712982,
0.12372945249080658,
-0.013827105984091759,
-0.022371480241417885,
-0.1197407990694046,
0.0015785678988322616,
0.021130502223968506,
0.042604196816682816,
0.04024510458111763,
0.05408281460404396,
0.03447001054883003,
0.043456509709358215,
0.02863817848265171,
-0.02880154922604561,
0.04750415310263634,
-0.009588376618921757,
0.0040976181626319885,
0.12772449851036072,
0.10417608171701431,
0.02852170541882515,
0.08533679693937302,
0.03910043090581894,
-0.0404503159224987,
0.01288803294301033,
-0.02749737910926342,
-0.06698718667030334,
0.10686532407999039,
-0.10741190612316132,
-0.16609297692775726,
-0.17008981108665466,
-0.051623594015836716,
-0.18006007373332977,
-0.024816229939460754,
0.03015192784368992,
-0.03652867674827576,
-0.07785869389772415,
-0.07390360534191132,
0.13494938611984253,
0.03926362842321396,
-0.05240222439169884,
-0.05028913915157318,
-0.03344966843724251,
0.042253971099853516,
-0.10105009377002716,
-0.04346547648310661,
0.0018156386213377118,
-0.050230301916599274,
0.021803520619869232,
0.027263345196843147,
0.0939897820353508,
0.060186538845300674,
-0.00003084815762122162,
-0.02964363619685173,
-0.040875863283872604,
0.17850545048713684,
-0.1293039619922638,
0.11540871858596802,
0.04773532599210739,
-0.0027224498335272074,
0.06698734313249588,
0.16732339560985565,
0.05765581503510475,
-0.04658689722418785,
-0.011350571177899837,
0.06666777282953262,
-0.030819516628980637,
-0.19200317561626434,
-0.1769828051328659,
-0.06770068407058716,
0.007487337104976177,
-0.027668775990605354,
0.04515873268246651,
0.06839024275541306,
-0.0003817266842816025,
-0.09345950931310654,
-0.07025164365768433,
0.036232683807611465,
0.04945642128586769,
0.17080222070217133,
0.009348781779408455,
0.05902401730418205,
-0.06639035046100616,
-0.06039080023765564,
0.09840981662273407,
0.00460409838706255,
0.08121812343597412,
0.06772522628307343,
0.06868960708379745,
0.051509808748960495,
0.012179503217339516,
0.0003098213055636734,
0.04621569812297821,
0.015963001176714897,
-0.02683101035654545,
-0.011713831685483456,
-0.08623592555522919,
0.07211727648973465,
0.08716612309217453,
0.10278671234846115,
0.03722165524959564,
-0.051564089953899384,
-0.06990828365087509,
0.12818388640880585,
0.1740635186433792,
0.11096327751874924,
-0.12066809087991714,
-0.10149893164634705,
0.059367019683122635,
-0.012059872969985008,
0.020104901865124702,
-0.03295712172985077,
-0.051309432834386826,
-0.13940629363059998,
0.20302264392375946,
-0.018019717186689377,
0.06347235292196274,
0.0055109113454818726,
-0.0001949289726326242,
-0.02158956229686737,
0.06696530431509018,
0.017838791012763977,
0.050336673855781555,
-0.141648069024086,
0.14686071872711182,
0.010985979810357094,
0.021909072995185852,
-0.01216028444468975,
0.05184603109955788,
0.01702127605676651,
0.05610823258757591,
0.10946010798215866,
0.04561372473835945,
-0.09489621967077255,
-0.08675345033407211,
-0.04737729951739311,
-0.05841971933841705,
0.13532322645187378,
-0.0978664755821228,
0.08815576881170273,
-0.005218944977968931,
-0.04548734799027443,
-0.058141354471445084,
0.16146430373191833,
-0.15934163331985474,
-0.10354094952344894,
0.10525287687778473,
-0.04850301891565323,
0.028323976323008537,
-0.07308647781610489,
-0.04719609022140503,
-0.24660754203796387,
0.2158661186695099,
0.015384898521006107,
0.00384410098195076,
-0.09645523875951767,
-0.0031059239991009235,
0.16681231558322906,
-0.09507244825363159,
0.038081102073192596,
-0.0030803470872342587,
0.08820296823978424,
-0.09653045237064362,
-0.08165580779314041,
0.04465007409453392,
-0.07744460552930832,
-0.14185139536857605,
0.011764489114284515,
0.19781433045864105,
0.022616581991314888,
0.037971191108226776,
0.025954479351639748,
0.09160146117210388,
0.08607495576143265,
-0.10742948949337006,
0.08422011882066727,
0.0879930630326271,
-0.030707523226737976,
0.1495898962020874,
-0.033895764499902725,
-0.15872009098529816,
-0.05146879330277443,
-0.0411202572286129,
0.10311200469732285,
0.34257373213768005,
-0.06467219442129135,
0.08678370714187622,
0.08080018311738968,
-0.042148686945438385,
-0.19632753729820251,
0.007789987605065107,
0.01917961612343788,
0.02523370087146759,
0.00445806048810482,
-0.05062577873468399,
0.10844220966100693,
0.13072700798511505,
-0.020657971501350403,
-0.04788164049386978,
-0.20526863634586334,
-0.1290767788887024,
0.011098041199147701,
-0.06513512134552002,
-0.012189746834337711,
-0.06006240472197533,
-0.10408233106136322,
-0.07066939771175385,
-0.16428698599338531,
0.05405878648161888,
0.04084285721182823,
0.048617810010910034,
0.0022039911709725857,
-0.022370649501681328,
0.03627573698759079,
-0.018924014642834663,
0.18879440426826477,
0.06920293718576431,
0.02023874782025814,
-0.07239631563425064,
-0.04903245344758034,
0.11341241002082825,
-0.037153460085392,
0.09947021305561066,
0.04320525750517845,
-0.005356981884688139,
-0.11361989378929138,
0.012292295694351196,
-0.09132620692253113,
0.11476991325616837,
-0.09256471693515778,
-0.030837757512927055,
-0.08548001945018768,
0.140853151679039,
0.0238761268556118,
-0.015673771500587463,
0.12679731845855713,
-0.032148636877536774,
0.1091376468539238,
0.13080862164497375,
0.0701589584350586,
0.0031420262530446053,
-0.07472140341997147,
0.0012759605888277292,
-0.06170511618256569,
0.019098622724413872,
0.05546579137444496,
0.04380559176206589,
0.13516387343406677,
-0.01975567452609539,
0.14712172746658325,
-0.018929120153188705,
-0.13689684867858887,
-0.023719700053334236,
0.11753503233194351,
-0.11223208159208298,
-0.2821040451526642,
-0.04018678143620491,
-0.034599367529153824,
0.012139244005084038,
0.061747223138809204,
0.19457398355007172,
-0.03780088946223259,
-0.023348631337285042,
0.021921996027231216,
0.04527831822633743,
-0.09146711230278015,
0.08037439733743668,
-0.021092766895890236,
0.042854998260736465,
-0.06821004301309586,
0.10744444280862808,
0.07745382934808731,
-0.22681817412376404,
-0.025234894827008247,
0.09869363903999329,
-0.06257352977991104,
-0.07561001926660538,
-0.09376278519630432,
0.1380273848772049,
-0.1314297765493393,
-0.0158812515437603,
-0.020428206771612167,
-0.1451670229434967,
0.04510067030787468,
0.09574063867330551,
0.04472881183028221,
0.011867307126522064,
0.035759247839450836,
0.00018209819972980767,
0.01861182041466236,
0.03738055005669594,
0.09328993409872055,
0.10640279203653336,
-0.042188696563243866,
0.03816690295934677,
-0.08412777632474899,
0.015412991866469383,
-0.021022463217377663,
-0.01296116691082716,
-0.09465239942073822,
-0.07739552855491638,
-0.10236255824565887,
-0.06780892610549927,
-0.08637113869190216,
-0.02516632154583931,
-0.0233630258589983,
-0.023995107039809227,
0.004357992671430111,
-0.027227098122239113,
-0.07138573378324509,
-0.050982195883989334,
-0.0155563373118639,
0.08955743908882141,
-0.1038399413228035,
-0.0027917600236833096,
0.06826390326023102,
-0.051174331456422806,
0.0997040867805481,
0.024214884266257286,
0.0514301136136055,
0.03840871900320053,
-0.03028056025505066,
-0.03847609460353851,
0.01905941218137741,
0.006461322773247957,
0.05086956173181534,
-0.14872215688228607,
-0.012150841765105724,
-0.0032515712082386017,
-0.051708951592445374,
0.050017524510622025,
-0.0625491663813591,
-0.10236836224794388,
0.008505094796419144,
0.016627702862024307,
-0.09351813793182373,
-0.07604184001684189,
0.027614127844572067,
0.0729801207780838,
0.035474032163619995,
0.11152558028697968,
-0.058112598955631256,
0.05574757605791092,
-0.1357552707195282,
-0.01836516335606575,
-0.02554010972380638,
-0.1030651405453682,
0.016653461381793022,
-0.011252963915467262,
0.022870756685733795,
-0.0031690942123532295,
0.11289511620998383,
-0.026996877044439316,
-0.025828121230006218,
0.06756260991096497,
-0.028583889827132225,
-0.0780753642320633,
0.0507938489317894,
0.07868042588233948,
0.04812269285321236,
-0.0161933284252882,
-0.08120898902416229,
-0.0014988429611548781,
-0.05973583459854126,
-0.007319761905819178,
0.1114073172211647,
0.14946618676185608,
0.11061113327741623,
0.05770944431424141,
0.08207276463508606,
-0.04891999810934067,
-0.028829647228121758,
-0.009984364733099937,
-0.029658254235982895,
0.08592110872268677,
-0.02771167829632759,
0.046906858682632446,
0.07054735720157623,
-0.23260782659053802,
0.1542171984910965,
-0.04036210849881172,
0.003158458275720477,
-0.019460761919617653,
-0.11383481323719025,
-0.07883796095848083,
-0.042559150606393814,
0.010454006493091583,
-0.08244454860687256,
0.009923603385686874,
0.00909425225108862,
0.03983308747410774,
-0.018626000732183456,
0.14684614539146423,
-0.22303219139575958,
-0.020058225840330124,
0.1173686534166336,
0.01384452823549509,
-0.01986311748623848,
0.04652818292379379,
-0.01833965815603733,
-0.05564456060528755,
0.0526103749871254,
0.003485421882942319,
0.042003098875284195,
0.01768621988594532,
-0.009287383407354355,
-0.030709892511367798,
-0.0902811586856842,
0.015258834697306156,
-0.013983458280563354,
0.019041143357753754,
0.026359856128692627,
0.02583683282136917,
0.016558820381760597,
-0.0304630808532238,
0.1203962042927742,
-0.04186068847775459,
-0.08964694291353226,
-0.09655430167913437,
0.18649248778820038,
0.015295648016035557,
0.07767751067876816,
0.014468489214777946,
-0.11231113970279694,
0.003929391037672758,
0.011538079008460045,
0.2675095796585083,
0.06229402497410774,
0.0249533299356699,
0.022313563153147697,
0.0121700968593359,
0.03887152299284935,
0.019862765446305275,
-0.021693918853998184,
0.17263448238372803,
-0.035285405814647675,
0.2010856717824936,
-0.05225873738527298,
-0.046340879052877426,
0.0050239525735378265,
0.1740422248840332,
-0.01995527185499668,
-0.01296838466078043,
-0.015919938683509827,
0.11686369776725769,
-0.02941286377608776,
-0.15430203080177307,
0.030166465789079666,
-0.1680508852005005,
-0.1555458903312683,
-0.04883193224668503,
0.03437260910868645,
0.0537441223859787,
0.11792732775211334,
0.015498639084398746,
-0.03355927765369415,
0.11172260344028473,
0.013434622436761856,
-0.01853151060640812,
-0.12593184411525726,
0.06253454834222794,
-0.005848423577845097,
0.10684798657894135,
0.020717212930321693,
0.04494085907936096,
0.12986114621162415,
-0.024106910452246666,
-0.10693443566560745,
0.014985187910497189,
0.045302849262952805,
-0.08386611938476562,
0.0049399384297430515,
0.15897472202777863,
-0.014517411589622498,
0.11524591594934464,
0.05899004638195038,
-0.13243594765663147,
0.07087180018424988,
0.0894564762711525,
-0.041566379368305206,
-0.06131380423903465,
0.1714923232793808,
-0.08582644164562225,
0.13566255569458008,
0.23618710041046143,
-0.026338893920183182,
0.025297190994024277,
-0.03987700119614601,
0.044342536479234695,
0.07475463300943375,
0.12010806798934937,
-0.017106499522924423,
-0.16066502034664154,
0.06056363880634308,
0.023605575785040855,
0.04031899571418762,
-0.14200104773044586,
-0.11799366027116776,
-0.015118377283215523,
0.02434436045587063,
-0.030886957421898842,
0.12695416808128357,
0.08972851932048798,
-0.0224088616669178,
0.001209444715641439,
-0.1888996660709381,
-0.008432535454630852,
0.07094074040651321,
-0.045595064759254456,
-0.016666574403643608
] |
null | null | null |
# **Llama 2**
| {"license": "llama2", "extra_gated_prompt": "### LLAMA 2 COMMUNITY LICENSE AGREEMENT", "extra_gated_fields": {"First Name": "text", "Last Name": "text", "Date of birth": "date_picker", "Country": "country", "Affiliation": "text", "I accept the terms and conditions": "checkbox", "Test request?": "checkbox", "geo": "ip_location"}, "extra_gated_description": "Llama 2 License and Acceptable Use Policy", "extra_gated_button_content": "I Accept Llama 2 License and AUP"} | null | huggingface-test1/test-model-1 | [
"license:llama2",
"region:us"
] | 2024-02-08T18:17:47+00:00 | [] | [] | TAGS
#license-llama2 #region-us
|
# Llama 2
| [
"# Llama 2"
] | [
"TAGS\n#license-llama2 #region-us \n",
"# Llama 2"
] | [
13,
4
] | [
"passage: TAGS\n#license-llama2 #region-us \n# Llama 2"
] | [
0.0347660668194294,
0.04838087409734726,
-0.010789492167532444,
0.04750456288456917,
0.021611440926790237,
0.047081537544727325,
0.14112889766693115,
0.0841500386595726,
0.19810737669467926,
0.009923181496560574,
0.09098634123802185,
0.0960492268204689,
0.026568012312054634,
-0.08888708055019379,
-0.012679772451519966,
-0.20264777541160583,
0.00850878469645977,
-0.053865768015384674,
0.032567091286182404,
0.0379580482840538,
-0.024908823892474174,
-0.03202476352453232,
0.03808555752038956,
-0.01972559280693531,
-0.05426160618662834,
0.055888477712869644,
0.06258847564458847,
-0.03833494335412979,
0.09537599235773087,
0.03844328969717026,
0.09947733581066132,
0.03695826977491379,
0.04421147331595421,
-0.24460293352603912,
0.01672130450606346,
-0.11111027747392654,
-0.09674878418445587,
0.011295712552964687,
0.00577987264841795,
-0.05073132365942001,
0.05429922789335251,
0.09842664003372192,
-0.005737581290304661,
0.021217823028564453,
-0.1956784874200821,
-0.12929032742977142,
-0.06598198413848877,
0.03278135508298874,
0.11000551283359528,
0.006857221946120262,
0.09500223398208618,
0.06294935196638107,
-0.18380172550678253,
-0.04327453300356865,
0.18925301730632782,
-0.34422487020492554,
0.04086639732122421,
0.1353931725025177,
0.05085889622569084,
0.04319314286112785,
-0.008956966921687126,
0.12373761087656021,
0.05199270695447922,
-0.07385245710611343,
-0.11929232627153397,
-0.07931195199489594,
0.03391243889927864,
0.10906146466732025,
-0.013055671937763691,
-0.039488427340984344,
0.16787229478359222,
0.009611403569579124,
0.008890829980373383,
0.09170883893966675,
-0.025598857551813126,
-0.030639057978987694,
0.016659336164593697,
0.025091489776968956,
0.055963654071092606,
0.13301627337932587,
0.1808372437953949,
-0.051078539341688156,
-0.1350889652967453,
-0.023477083072066307,
-0.2382728010416031,
0.15063637495040894,
0.011859850026667118,
0.11616939306259155,
-0.15455034375190735,
-0.01078640017658472,
-0.1400097906589508,
-0.027523145079612732,
-0.01605357602238655,
0.019531169906258583,
0.10167371481657028,
0.04937145113945007,
-0.030535610392689705,
0.06754561513662338,
0.11652959883213043,
0.09632724523544312,
0.052635811269283295,
0.006346814800053835,
-0.06462301313877106,
0.17051240801811218,
-0.028415609151124954,
0.018158582970499992,
0.15730337798595428,
0.10540129989385605,
-0.0035198864061385393,
-0.1379386931657791,
0.1269301027059555,
-0.044415269047021866,
-0.1702149212360382,
-0.031939659267663956,
-0.1457187682390213,
0.19088709354400635,
-0.03299960866570473,
-0.1114891842007637,
-0.08314575254917145,
0.0396723672747612,
-0.007468887139111757,
-0.07603570818901062,
0.01829424500465393,
0.0070495521649718285,
0.016885025426745415,
-0.04614808410406113,
-0.06659029424190521,
0.05490141734480858,
0.12297700345516205,
0.06374147534370422,
-0.10295013338327408,
0.028613626956939697,
-0.00853279884904623,
0.10157620161771774,
0.12592574954032898,
-0.05057166889309883,
0.04039866104722023,
-0.1292734146118164,
-0.12098000943660736,
0.02340571954846382,
0.027224931865930557,
-0.05865160748362541,
0.04791145771741867,
0.05907626450061798,
0.08194612711668015,
-0.016690688207745552,
-0.07477706670761108,
-0.09628208726644516,
-0.05928971618413925,
0.12127960473299026,
-0.00996905006468296,
0.045711416751146317,
-0.22120997309684753,
-0.017313355579972267,
-0.038375526666641235,
0.10849146544933319,
-0.0013604062842205167,
-0.09153701364994049,
-0.028920244425535202,
0.11393612623214722,
-0.02304394543170929,
0.07178240269422531,
-0.08774664252996445,
0.025213588029146194,
-0.062416452914476395,
0.10909591615200043,
-0.150508850812912,
-0.06458377093076706,
0.13756905496120453,
-0.13913938403129578,
-0.1075153574347496,
-0.016399463638663292,
0.01597736030817032,
0.11321871727705002,
0.0527290515601635,
0.4004785418510437,
-0.03853253647685051,
-0.21749837696552277,
0.12086068093776703,
0.08793581277132034,
-0.09621847420930862,
-0.2649899423122406,
0.14820782840251923,
-0.11059888452291489,
-0.08493208140134811,
0.019397828727960587,
0.003652423620223999,
0.05149358883500099,
-0.018132006749510765,
-0.10534969717264175,
0.031554341316223145,
-0.014676437713205814,
-0.05208525061607361,
-0.0014696614816784859,
0.055743858218193054,
-0.062321119010448456,
0.055878203362226486,
0.08515732735395432,
0.040620166808366776,
0.13515645265579224,
0.030542032793164253,
-0.10082407295703888,
0.1025952473282814,
0.022222043946385384,
-0.03221813589334488,
0.010482686571776867,
-0.09881187975406647,
-0.02649134397506714,
-0.06464815884828568,
0.08058744668960571,
0.04715766757726669,
0.033632416278123856,
-0.04720514267683029,
-0.016870060935616493,
0.058498699218034744,
0.05539222061634064,
0.045797545462846756,
0.016950011253356934,
-0.06687598675489426,
0.06637568771839142,
-0.01124501135200262,
-0.026250135153532028,
-0.047788865864276886,
-0.025717729702591896,
0.11600136756896973,
-0.02511805295944214,
0.015696080401539803,
0.01614917814731598,
-0.036861319094896317,
-0.00960781704634428,
0.037594884634017944,
0.03718855232000351,
0.14667217433452606,
-0.005564257502555847,
-0.07458807528018951,
0.20107850432395935,
-0.04857517406344414,
0.2043275535106659,
0.1291641891002655,
-0.03853590786457062,
0.024759724736213684,
-0.11677282303571701,
0.0006328438175842166,
0.006361842155456543,
0.014931418001651764,
-0.017543569207191467,
-0.02981579862535,
-0.018043912947177887,
0.06202513724565506,
-0.03136761859059334,
0.008903677575290203,
-0.03826633095741272,
-0.0947004184126854,
-0.103599414229393,
0.050029460340738297,
0.2642568349838257,
-0.12810233235359192,
0.0969076156616211,
0.4119666814804077,
0.10037491470575333,
0.13840191066265106,
-0.08243215829133987,
-0.0038870859425514936,
-0.10891545563936234,
0.04380275681614876,
0.011615293100476265,
0.08770273625850677,
0.01112134475260973,
0.009756835177540779,
0.006227277684956789,
0.021182622760534286,
0.03133406117558479,
-0.20233549177646637,
-0.16878637671470642,
-0.03224610164761543,
-0.0308656208217144,
-0.10543162375688553,
0.0895453467965126,
-0.06726306676864624,
0.05920355021953583,
0.05197470635175705,
-0.1110144555568695,
0.09876104444265366,
-0.005036753602325916,
-0.030917195603251457,
0.04277316853404045,
-0.10677957534790039,
-0.14851787686347961,
-0.18934266269207,
-0.11183420568704605,
0.041005104780197144,
0.0204943735152483,
0.05123239383101463,
-0.04458994045853615,
0.010650134645402431,
0.0625462681055069,
0.005151220131665468,
-0.13220132887363434,
-0.009400560520589352,
0.0059451390989124775,
0.1273510754108429,
-0.06569285690784454,
-0.061597954481840134,
-0.08762473613023758,
-0.06566426903009415,
-0.061418790370225906,
0.06894323974847794,
-0.0959208607673645,
0.058324117213487625,
0.16823038458824158,
0.07720493525266647,
0.07993117719888687,
-0.0036693583242595196,
0.06227065250277519,
-0.038381703197956085,
-0.068010114133358,
0.1427822858095169,
-0.031595755368471146,
0.06600423902273178,
0.22964246571063995,
0.09214428812265396,
-0.1479392945766449,
-0.041790544986724854,
-0.01096482016146183,
-0.15563331544399261,
-0.2142682820558548,
-0.02737196907401085,
-0.026030205190181732,
0.061768513172864914,
0.003956901375204325,
0.09119097888469696,
0.07122839242219925,
0.030796300619840622,
0.14484314620494843,
-0.11420052498579025,
-0.02379765175282955,
0.009270554408431053,
0.2721927762031555,
-0.05849899724125862,
0.0027044794987887144,
-0.10904198884963989,
-0.050747185945510864,
0.12387290596961975,
0.16346515715122223,
0.09224671125411987,
0.25741562247276306,
0.024741049855947495,
0.14487449824810028,
0.08852533251047134,
0.08871302008628845,
-0.03355390578508377,
0.03525519371032715,
-0.014848367311060429,
-0.061803244054317474,
-0.03144676983356476,
0.02954960986971855,
0.08545360714197159,
0.05570496991276741,
-0.1424909234046936,
0.0032946886494755745,
-0.19836823642253876,
-0.027278359979391098,
0.008110582828521729,
0.11295586824417114,
0.02913258783519268,
0.13430553674697876,
0.1472642719745636,
-0.02021341398358345,
0.01996196247637272,
0.1318151354789734,
-0.027590468525886536,
-0.07697348296642303,
0.0632394328713417,
0.0744972825050354,
0.05206475406885147,
0.014423321932554245,
0.04762712121009827,
-0.11260958015918732,
-0.2235565483570099,
0.06924952566623688,
0.10024965554475784,
-0.14015185832977295,
0.28565096855163574,
0.0459117591381073,
-0.10937339067459106,
0.0026531126350164413,
-0.0763266459107399,
-0.012223688885569572,
0.10228999704122543,
0.1441807746887207,
0.06377233564853668,
-0.10887664556503296,
-0.11984692513942719,
-0.032738812267780304,
-0.0007602040423080325,
0.10437455773353577,
-0.0445086769759655,
-0.08117133378982544,
-0.06279924511909485,
0.040057141333818436,
-0.026552041992545128,
0.14001615345478058,
-0.005877719726413488,
-0.053602900356054306,
0.015603932552039623,
0.07866168767213821,
-0.018947357311844826,
-0.056041598320007324,
-0.00853903777897358,
-0.017487220466136932,
0.08710737526416779,
-0.015286179259419441,
0.011828328482806683,
-0.059365469962358475,
-0.21904230117797852,
0.02404455840587616,
-0.08434325456619263,
-0.012016567401587963,
-0.03518782928586006,
-0.10382452607154846,
-0.10700473189353943,
-0.11500027030706406,
0.10548441857099533,
0.007071739993989468,
0.04217927157878876,
-0.048399847000837326,
0.21472245454788208,
-0.08202607184648514,
0.051650695502758026,
0.032532211393117905,
0.05343322083353996,
0.025452136993408203,
-0.08729015290737152,
0.09825737029314041,
-0.030984992161393166,
-0.026271970942616463,
-0.056975267827510834,
-0.11163993179798126,
0.08342760801315308,
0.03292808309197426,
-0.13085328042507172,
0.2213132530450821,
0.3636651337146759,
-0.025027403607964516,
0.2627878785133362,
0.16866722702980042,
-0.1356252282857895,
-0.1886873096227646,
-0.18641333281993866,
-0.2036379873752594,
-0.06813476979732513,
0.16617432236671448,
-0.1281425654888153,
0.04385045915842056,
0.18277491629123688,
-0.12644927203655243,
0.2399892807006836,
-0.3762728273868561,
-0.0754655972123146,
0.05574267730116844,
-0.02337026223540306,
0.5756284594535828,
-0.15068285167217255,
-0.149977445602417,
-0.04315086081624031,
-0.17273952066898346,
0.1314510852098465,
0.02851686254143715,
0.08424325287342072,
-0.031365472823381424,
0.05808698385953903,
0.03708883747458458,
0.011304776184260845,
0.2296164631843567,
-0.009415117092430592,
0.03888731449842453,
-0.09055229276418686,
-0.22404788434505463,
0.10355953127145767,
-0.04618779942393303,
-0.03400407359004021,
-0.05270657688379288,
-0.08399561792612076,
-0.15352553129196167,
-0.00830678641796112,
-0.014223359525203705,
0.05369805172085762,
0.01667339727282524,
-0.062260061502456665,
-0.07688743621110916,
0.0012738292571157217,
-0.06141148880124092,
-0.026170535013079643,
0.22924882173538208,
-0.050514061003923416,
0.07150311768054962,
0.07537572830915451,
-0.10751524567604065,
-0.04395391792058945,
-0.0076787956058979034,
-0.11590205132961273,
-0.06321016699075699,
0.05364733561873436,
-0.1854311227798462,
0.030573861673474312,
0.09741061180830002,
-0.005973407998681068,
0.049210429191589355,
0.04527230188250542,
-0.04874371737241745,
0.0540580190718174,
0.1912793219089508,
-0.030674990266561508,
-0.10945437848567963,
0.03742282837629318,
0.059051264077425,
0.1890682876110077,
-0.041619908064603806,
0.08114851266145706,
0.07500869780778885,
0.0025231661275029182,
0.013047312386333942,
0.019047314301133156,
-0.11722184717655182,
-0.06732334941625595,
0.0534028597176075,
-0.022318318486213684,
-0.0676521509885788,
0.1177552118897438,
0.06940539926290512,
0.05240297317504883,
-0.027469459921121597,
0.14159134030342102,
0.021654969081282616,
-0.04173802584409714,
-0.1576356142759323,
0.014415298588573933,
-0.24035021662712097,
-0.08606821298599243,
-0.0010965282563120127,
-0.08883257210254669,
-0.02529635839164257,
0.18195490539073944,
0.031336016952991486,
0.07452503591775894,
0.0640823021531105,
-0.026969091966748238,
0.13733820617198944,
-0.058992840349674225,
-0.13626496493816376,
-0.013881680555641651,
-0.12449570745229721,
-0.05836782976984978,
0.03706863522529602,
0.11728833615779877,
-0.06865891814231873,
-0.09141325950622559,
-0.19844895601272583,
0.10548968613147736,
-0.13717179000377655,
0.0005871859611943364,
-0.06537799537181854,
-0.019253846257925034,
0.04799128323793411,
-0.059980880469083786,
-0.08695776760578156,
-0.02997998148202896,
-0.15190191566944122,
0.04766744747757912,
-0.01016752328723669,
0.07653671503067017,
-0.06690020859241486,
-0.09947741776704788,
0.13385465741157532,
0.033654578030109406,
0.06216869875788689,
0.028527740389108658,
0.06194472312927246,
0.1540437638759613,
-0.11880259215831757,
-0.017368214204907417,
0.14606526494026184,
-0.001367705874145031,
-0.009718122892081738,
0.010730482637882233,
-0.04114363342523575,
0.04920439049601555,
-0.015867702662944794,
0.06297178566455841,
-0.05549822375178337,
-0.14292630553245544,
-0.06744266301393509,
-0.00015614714357070625,
-0.2730069160461426,
0.04972328990697861,
-0.15232554078102112,
0.11181734502315521,
0.0034903238993138075,
0.07401207834482193,
0.06154124066233635,
0.030988015234470367,
0.031464748084545135,
0.0059429011307656765,
-0.014324445277452469,
-0.03954319283366203,
-0.03451140224933624,
-0.06836847960948944,
-0.11007707566022873,
-0.0219002403318882,
0.28129735589027405,
0.0255593229085207,
-0.16194160282611847,
0.032983001321554184,
0.12288340926170349,
0.008349745534360409,
-0.024415016174316406,
0.2682700753211975,
0.10123465955257416,
0.018424328416585922,
-0.1165573000907898,
0.01294022798538208,
-0.04983069375157356,
-0.1400976926088333,
0.07098051905632019,
0.08875910192728043,
0.0025721192359924316,
0.03243798762559891,
0.163546621799469,
-0.02617342211306095,
-0.023919565603137016,
-0.07188280671834946,
0.014872142113745213,
-0.06627510488033295,
0.016798987984657288,
0.0988616794347763,
0.21267175674438477,
-0.03369157016277313,
0.009272076189517975,
-0.05153371021151543,
0.03178452327847481,
-0.21987326443195343,
-0.0889134630560875,
-0.02868920937180519,
-0.03265279531478882,
0.06504280120134354,
0.06966549158096313,
0.07839027792215347,
0.2148970365524292,
-0.004470566753298044,
0.011103985831141472,
-0.05212774872779846,
-0.11159531027078629,
-0.07149849832057953,
-0.0816286951303482,
-0.009023023769259453,
0.04198561981320381,
-0.11933521926403046,
-0.032879509031772614,
-0.0735882967710495,
-0.1115991473197937,
-0.09624231606721878,
0.03593141585588455,
0.03524588420987129,
-0.07736480236053467,
-0.18253304064273834,
-0.03087722137570381,
-0.06392757594585419,
0.05498943477869034,
0.03464142605662346,
0.07317165285348892,
0.020017817616462708,
0.007196938153356314,
0.023914190009236336,
0.11163678765296936,
0.00002721626151469536,
-0.04769616574048996,
0.0935446098446846,
0.09643857926130295,
-0.034600768238306046,
0.10340926796197891,
-0.10200592130422592,
0.0028726302552968264,
-0.016709739342331886,
0.18143481016159058,
0.20550133287906647,
-0.07023075968027115,
0.029354844242334366,
-0.008762063458561897,
0.04775972664356232,
0.1035841554403305,
0.1335255354642868,
-0.018515540286898613,
0.2831133306026459,
-0.07667792588472366,
-0.012080462649464607,
-0.04016867280006409,
0.03047800622880459,
-0.15009243786334991,
0.10818193107843399,
-0.0035667873453348875,
-0.040944453328847885,
-0.07771094888448715,
0.10244017839431763,
-0.1070045530796051,
0.1377362459897995,
0.1843005120754242,
-0.19839975237846375,
-0.037922926247119904,
-0.020078545436263084,
0.10660550743341446,
-0.011231385171413422,
0.09334148466587067,
-0.11534233391284943,
-0.13631875813007355,
-0.12193375825881958,
0.08369871228933334,
-0.39127644896507263,
-0.1706816554069519,
0.10401938855648041,
0.19713029265403748,
0.14085189998149872,
-0.012087507173418999,
0.0861675962805748,
-0.004553290084004402,
0.03567054867744446,
-0.014675142243504524,
0.14649656414985657,
0.08444751799106598,
-0.029144296422600746,
-0.11864250898361206,
-0.16745726764202118,
-0.02771373651921749,
-0.03886020928621292,
0.08947645872831345,
0.10785847902297974,
-0.010155273601412773,
0.09119290113449097,
-0.07591553032398224,
-0.01316596195101738,
0.005222475156188011,
-0.15252435207366943,
0.013267514295876026,
-0.040891118347644806,
-0.00101861625444144,
-0.052532296627759933,
-0.05160721018910408,
-0.02056743949651718,
0.12295172363519669,
-0.21949222683906555,
-0.07846366614103317,
0.11216063797473907,
-0.015156978741288185,
0.11791340261697769,
0.006055242381989956,
-0.14559409022331238,
-0.051031846553087234,
-0.09938272088766098,
0.0737423300743103,
-0.13916584849357605,
0.05263558030128479,
0.16503863036632538,
0.005176732316613197,
0.029684606939554214,
-0.36513465642929077,
0.08347870409488678,
-0.030126985162496567,
-0.05454377457499504,
-0.05046136677265167
] |
null | null | transformers |
# **Llama 2 7b Spanish Version**
Orale amigos hispanohablantes, esta es una de las tantas pruebas que voy a seguir haciendo para tener un modelo en español que no alucine y empiece
a hablar en idioma "gringo" de la nada. Aportes, ayuda, condolencias...son bienvenidas!!! Gracias Zuck por querer tirar el monopolio de OpenAi (envidia ?)
regalando tu modelo al mundo!
Dataset usado para el entrenamiento: bertin-project/alpaca-spanish
Tiempo de entrenamiento: 4hs
Cantidad de registros procesados del dataset: 10k
# **Llama 2**
Llama 2 es una colección de modelos de texto generativo preentrenados y afinados que varían en escala desde 7 mil millones hasta 70 mil millones de parámetros. Este es el repositorio del modelo afinado de 7B, optimizado para casos de uso de diálogo y convertido al formato de Hugging Face Transformers. Los enlaces a otros modelos se pueden encontrar en el índice al final.
## Detalles del Modelo
*Nota: El uso de este modelo está regido por la licencia Meta. Para descargar los pesos del modelo y el tokenizador, por favor visita el [sitio web](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) y acepta nuestra Licencia antes de solicitar acceso aquí.*
Meta desarrolló y lanzó públicamente la familia de modelos de lenguaje grande Llama 2 (LLMs), una colección de modelos de texto generativo preentrenados y afinados que varían en escala desde 7 mil millones hasta 70 mil millones de parámetros. Nuestros LLMs afinados, llamados Llama-2-Chat, están optimizados para casos de uso de diálogo. Los modelos Llama-2-Chat superan a los modelos de chat de código abierto en la mayoría de los benchmarks que probamos, y en nuestras evaluaciones humanas de utilidad y seguridad, están a la par con algunos modelos cerrados populares como ChatGPT y PaLM.
**Desarrolladores del Modelo:** Meta
**Variaciones:** Llama 2 viene en una variedad de tamaños de parámetros, incluyendo 7B, 13B y 70B, así como variaciones preentrenadas y afinadas.
**Entrada:** Los modelos solo reciben texto de entrada.
**Salida:** Los modelos generan texto únicamente.
**Arquitectura del Modelo:** Llama 2 es un modelo de lenguaje auto-regresivo que utiliza una arquitectura de transformer optimizada. Las versiones ajustadas utilizan ajuste fino supervisado (SFT) y aprendizaje por refuerzo con retroalimentación humana (RLHF) para alinearse con las preferencias humanas en cuanto a utilidad y seguridad.
||Datos de Entrenamiento|Parámetros|Longitud del Contenido|GQA|Tokens|LR|
|---|---|---|---|---|---|---|
|Llama 2|*Una nueva combinación de datos disponibles públicamente en línea*|7B|4k|✗|2.0T|3.0 x 10<sup>-4</sup>|
|Llama 2|*Una nueva combinación de datos disponibles públicamente en línea*|13B|4k|✗|2.0T|3.0 x 10<sup>-4</sup>|
|Llama 2|*Una nueva combinación de datos disponibles públicamente en línea*|70B|4k|✔|2.0T|1.5 x 10<sup>-4</sup>|
*Familia de modelos Llama 2.* Los recuentos de tokens se refieren solo a los datos de preentrenamiento. Todos los modelos se entrenan con un tamaño de lote global de 4 millones de tokens. Los modelos más grandes - 70B - utilizan Atención de Consulta Agrupada (GQA) para mejorar la escalabilidad de la inferencia.
**Fechas del Modelo:** Llama 2 fue entrenado entre enero de 2023 y julio de 2023.
**Estado:** Este es un modelo estático entrenado en un conjunto de datos sin conexión. Se lanzarán futuras versiones de los modelos ajustados a medida que mejoremos la seguridad del modelo con la retroalimentación de la comunidad.
**Licencia:** Una licencia comercial personalizada está disponible en: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/)
**Artículo de Investigación:** ["Llama-2: Fundación Abierta y Modelos de Chat Afinados"](arxiv.org/abs/2307.09288)
## Uso Previsto
**Casos de Uso Previstos:** Llama 2 está destinado para uso comercial e investigación en inglés. Los modelos ajustados están destinados para chat similar a un asistente, mientras que los modelos preentrenados pueden adaptarse para una variedad de tareas de generación de lenguaje natural.
Para obtener las características y el rendimiento esperados para las versiones de chat, se debe seguir un formato específico, que incluye las etiquetas `INST` y `<<SYS>>`, los tokens `BOS` y `EOS`, y los espacios en blanco y saltos de línea entre ellos (recomendamos llamar a `strip()` en las entradas para evitar espacios dobles). Consulta nuestro código de referencia en github para más detalles: [`chat_completion`](https://github.com/facebookresearch/llama/blob/main/llama/generation.py#L212).
**Usos Fuera de Alcance:** Uso de cualquier manera que viole leyes o regulaciones aplicables (incluyendo leyes de cumplimiento comercial). Uso en idiomas que no sean inglés. Uso de cualquier otra manera que esté prohibida por la Política de Uso Aceptable y el Acuerdo de Licencia para Llama 2.
## Hardware y Software
**Factores de Entrenamiento:** Utilizamos bibliotecas de entrenamiento personalizadas, el Super Cluster de Investigación de Meta y clústeres de producción para el preentrenamiento. El ajuste fino, la anotación y la evaluación también se realizaron en cómputo en la nube de terceros.
**Huella de Carbono:** El preentrenamiento utilizó un total acumulado de 3.3 millones de horas de GPU de cómputo en hardware de tipo A100-80GB (TDP de 350-400W). Las emisiones totales estimadas fueron de 539 tCO2eq, el 100% de las cuales fueron compensadas por el programa de sostenibilidad de Meta.
||Tiempo (horas de GPU)|Consumo de Energía (W)|Carbono Emitido (tCO<sub>2</sub>eq)|
|---|---|---|---|
|Llama 2 7B|184320|400|31.22|
|Llama 2 13B|368640|400|62.44|
|Llama 2 70B|1720320|400|291.42|
|Total|3311616||539.00|
**Emisiones de CO<sub>2</sub> durante el preentrenamiento.** Tiempo: tiempo total de GPU necesario para entrenar cada modelo. Consumo de Energía: capacidad de energía pico por dispositivo GPU para las GPU utilizadas ajustadas para la eficiencia del uso de energía. El 100% de las emisiones se compensan directamente mediante el programa de sostenibilidad de Meta, y dado que estamos lanzando estos modelos abiertamente, los costos de preentrenamiento no deben ser asumidos por otros.
## Datos de Entrenamiento
**Descripción General:** Llama 2 fue preentrenado en 2 billones de tokens de datos de fuentes disponibles públicamente. Los datos de ajuste fino incluyen conjuntos de datos de instrucciones disponibles públicamente, así como más de un millón de nuevos ejemplos anotados por humanos. Ni los datos de preentrenamiento ni los datos de ajuste fino incluyen datos de usuarios de Meta.
**Actualidad de los Datos:** Los datos de preentrenamiento tienen un corte hasta septiembre de 2022, pero algunos datos de ajuste son más recientes, hasta julio de 2023.
## Resultados de Evaluación
En esta sección, reportamos los resultados de los modelos Llama 1 y Llama 2 en benchmarks académicos estándar. Para todas las evaluaciones, utilizamos nuestra biblioteca de evaluaciones internas.
|Modelo|Tamaño|Código|Razonamiento con Sentido Común|Conocimiento del Mundo|Comprensión de Lectura|Matemáticas|MMLU|BBH|Evaluación AGI|
|---|---|---|---|---|---|---|---|---|---|
|Llama 1|7B|14.1|60.8|46.2|58.5|6.95|35.1|30.3|23.9|
|Llama 1|13B|18.9|66.1|52.6|62.3|10.9|46.9|37.0|33.9|
|Llama 1|33B|26.0|70.0|58.4|67.6|21.4|57.8|39.8|41.7|
|Llama 1|65B|30.7|70.7|60.5|68.6|30.8|63.4|43.5|47.6|
|Llama 2|7B|16.8|63.9|48.9|61.3|14.6|45.3|32.6|29.3|
|Llama 2|13B|24.5|66.9|55.4|65.8|28.7|54.8|39.4|39.1|
|Llama 2|70B|**37.5**|**71.9**|**63.6**|**69.4**|**35.2**|**68.9**|**51.2**|**54.2**|
**Rendimiento general en benchmarks académicos agrupados.** *Código:* Reportamos el promedio de pase@1 de nuestros modelos en HumanEval y MBPP. *Razonamiento con Sentido Común:* Reportamos el promedio de PIQA, SIQA, HellaSwag, WinoGrande, ARC fácil y desafío, OpenBookQA y CommonsenseQA. Reportamos resultados de 7 disparos para CommonSenseQA y resultados de 0 disparos para todos los demás benchmarks. *Conocimiento del Mundo:* Evaluamos el rendimiento de 5 disparos en NaturalQuestions y TriviaQA y reportamos el promedio. *Comprensión de Lectura:* Para comprensión de lectura, reportamos el promedio de 0 disparos en SQuAD, QuAC y BoolQ. *Matemáticas:* Reportamos el promedio de los benchmarks GSM8K (8 disparos) y MATH (4 disparos) en el top 1.
|||TruthfulQA|Toxigen|
|---|---|---|---|
|Llama 1|7B|27.42|23.00|
|Llama 1|13B|41.74|23.08|
|Llama 1|33B|44.19|22.57|
|Llama 1|65B|48.71|21.77|
|Llama 2|7B|33.29|**21.25**|
|Llama 2|13B|41.86|26.10|
|Llama 2|70B|**50.18**|24.60|
**Evaluación de LLMs preentrenados en benchmarks automáticos de seguridad.** Para TruthfulQA, presentamos el porcentaje de generaciones que son tanto verídicas como informativas (cuanto mayor, mejor). Para ToxiGen, presentamos el porcentaje de generaciones tóxicas (cuanto menor, mejor).
|||TruthfulQA|Toxigen|
|---|---|---|---|
|Llama-2-Chat|7B|57.04|**0.00**|
|Llama-2-Chat|13B|62.18|**0.00**|
|Llama-2-Chat|70B|**64.14**|0.01|
**Evaluación de LLMs afinados en diferentes conjuntos de datos de seguridad.** Mismas definiciones de métricas que arriba.
## Consideraciones Éticas y Limitaciones
Llama 2 es una nueva tecnología que conlleva riesgos con su uso. Las pruebas realizadas hasta la fecha han sido en inglés, y no han cubierto, ni podrían cubrir, todos los escenarios. Por estas razones, como con todos los LLMs, las salidas potenciales de Llama 2 no pueden predecirse de antemano, y el modelo puede en algunos casos producir respuestas inexactas, sesgadas u otras respuestas objetables a las solicitudes de los usuarios. Por lo tanto, antes de implementar cualquier aplicación de Llama 2, los desarrolladores deben realizar pruebas de seguridad y ajuste adaptadas a sus aplicaciones específicas del modelo.
Consulta la Guía de Uso Responsable disponible en [https://ai.meta.com/llama/responsible-use-guide/](https://ai.meta.com/llama/responsible-use-guide)
## Reportar Problemas
Por favor, reporta cualquier "error" de software u otros problemas con los modelos a través de alguno de los siguientes medios:
- Reportar problemas con el modelo: [github.com/facebookresearch/llama](http://github.com/facebookresearch/llama)
- Reportar contenido problemático generado por el modelo: [developers.facebook.com/llama_output_feedback](http://developers.facebook.com/llama_output_feedback)
- Reportar errores y problemas de seguridad: [facebook.com/whitehat/info](http://facebook.com/whitehat/info)
## Índice de Modelos Llama
|Modelo|Llama2|Llama2-hf|Llama2-chat|Llama2-chat-hf|
|---|---|---|---|---|
|7B| [Enlace](https://huggingface.co/llamaste/Llama-2-7b) | [Enlace](https://huggingface.co/llamaste/Llama-2-7b-hf) | [Enlace](https://huggingface.co/llamaste/Llama-2-7b-chat) | [Enlace](https://huggingface.co/llamaste/Llama-2-7b-chat-hf)|
|13B| [Enlace](https://huggingface.co/llamaste/Llama-2-13b) | [Enlace](https://huggingface.co/llamaste/Llama-2-13b-hf) | [Enlace](https://huggingface.co/llamaste/Llama-2-13b-chat) | [Enlace](https://huggingface.co/llamaste/Llama-2-13b-hf)|
|70B| [Enlace](https://huggingface.co/llamaste/Llama-2-70b) | [Enlace](https://huggingface.co/llamaste/Llama-2-70b-hf) | [Enlace](https://huggingface.co/llamaste/Llama-2-70b-chat) | [Enlace](https://huggingface.co/llamaste/Llama-2-70b-hf)| | {"language": ["es"], "license": "llama2"} | text-generation | Kukedlc/Llama-7b-spanish | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"es",
"arxiv:2307.09288",
"license:llama2",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T18:18:32+00:00 | [
"2307.09288"
] | [
"es"
] | TAGS
#transformers #pytorch #safetensors #llama #text-generation #es #arxiv-2307.09288 #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
| Llama 2 7b Spanish Version
==========================
Orale amigos hispanohablantes, esta es una de las tantas pruebas que voy a seguir haciendo para tener un modelo en español que no alucine y empiece
a hablar en idioma "gringo" de la nada. Aportes, ayuda, condolencias...son bienvenidas!!! Gracias Zuck por querer tirar el monopolio de OpenAi (envidia ?)
regalando tu modelo al mundo!
Dataset usado para el entrenamiento: bertin-project/alpaca-spanish
Tiempo de entrenamiento: 4hs
Cantidad de registros procesados del dataset: 10k
Llama 2
=======
Llama 2 es una colección de modelos de texto generativo preentrenados y afinados que varían en escala desde 7 mil millones hasta 70 mil millones de parámetros. Este es el repositorio del modelo afinado de 7B, optimizado para casos de uso de diálogo y convertido al formato de Hugging Face Transformers. Los enlaces a otros modelos se pueden encontrar en el índice al final.
Detalles del Modelo
-------------------
*Nota: El uso de este modelo está regido por la licencia Meta. Para descargar los pesos del modelo y el tokenizador, por favor visita el sitio web y acepta nuestra Licencia antes de solicitar acceso aquí.*
Meta desarrolló y lanzó públicamente la familia de modelos de lenguaje grande Llama 2 (LLMs), una colección de modelos de texto generativo preentrenados y afinados que varían en escala desde 7 mil millones hasta 70 mil millones de parámetros. Nuestros LLMs afinados, llamados Llama-2-Chat, están optimizados para casos de uso de diálogo. Los modelos Llama-2-Chat superan a los modelos de chat de código abierto en la mayoría de los benchmarks que probamos, y en nuestras evaluaciones humanas de utilidad y seguridad, están a la par con algunos modelos cerrados populares como ChatGPT y PaLM.
Desarrolladores del Modelo: Meta
Variaciones: Llama 2 viene en una variedad de tamaños de parámetros, incluyendo 7B, 13B y 70B, así como variaciones preentrenadas y afinadas.
Entrada: Los modelos solo reciben texto de entrada.
Salida: Los modelos generan texto únicamente.
Arquitectura del Modelo: Llama 2 es un modelo de lenguaje auto-regresivo que utiliza una arquitectura de transformer optimizada. Las versiones ajustadas utilizan ajuste fino supervisado (SFT) y aprendizaje por refuerzo con retroalimentación humana (RLHF) para alinearse con las preferencias humanas en cuanto a utilidad y seguridad.
*Familia de modelos Llama 2.* Los recuentos de tokens se refieren solo a los datos de preentrenamiento. Todos los modelos se entrenan con un tamaño de lote global de 4 millones de tokens. Los modelos más grandes - 70B - utilizan Atención de Consulta Agrupada (GQA) para mejorar la escalabilidad de la inferencia.
Fechas del Modelo: Llama 2 fue entrenado entre enero de 2023 y julio de 2023.
Estado: Este es un modelo estático entrenado en un conjunto de datos sin conexión. Se lanzarán futuras versiones de los modelos ajustados a medida que mejoremos la seguridad del modelo con la retroalimentación de la comunidad.
Licencia: Una licencia comercial personalizada está disponible en: URL
Artículo de Investigación: "Llama-2: Fundación Abierta y Modelos de Chat Afinados"
Uso Previsto
------------
Casos de Uso Previstos: Llama 2 está destinado para uso comercial e investigación en inglés. Los modelos ajustados están destinados para chat similar a un asistente, mientras que los modelos preentrenados pueden adaptarse para una variedad de tareas de generación de lenguaje natural.
Para obtener las características y el rendimiento esperados para las versiones de chat, se debe seguir un formato específico, que incluye las etiquetas 'INST' y '<>', los tokens 'BOS' y 'EOS', y los espacios en blanco y saltos de línea entre ellos (recomendamos llamar a 'strip()' en las entradas para evitar espacios dobles). Consulta nuestro código de referencia en github para más detalles: 'chat\_completion'.
Usos Fuera de Alcance: Uso de cualquier manera que viole leyes o regulaciones aplicables (incluyendo leyes de cumplimiento comercial). Uso en idiomas que no sean inglés. Uso de cualquier otra manera que esté prohibida por la Política de Uso Aceptable y el Acuerdo de Licencia para Llama 2.
Hardware y Software
-------------------
Factores de Entrenamiento: Utilizamos bibliotecas de entrenamiento personalizadas, el Super Cluster de Investigación de Meta y clústeres de producción para el preentrenamiento. El ajuste fino, la anotación y la evaluación también se realizaron en cómputo en la nube de terceros.
Huella de Carbono: El preentrenamiento utilizó un total acumulado de 3.3 millones de horas de GPU de cómputo en hardware de tipo A100-80GB (TDP de 350-400W). Las emisiones totales estimadas fueron de 539 tCO2eq, el 100% de las cuales fueron compensadas por el programa de sostenibilidad de Meta.
Emisiones de CO2 durante el preentrenamiento. Tiempo: tiempo total de GPU necesario para entrenar cada modelo. Consumo de Energía: capacidad de energía pico por dispositivo GPU para las GPU utilizadas ajustadas para la eficiencia del uso de energía. El 100% de las emisiones se compensan directamente mediante el programa de sostenibilidad de Meta, y dado que estamos lanzando estos modelos abiertamente, los costos de preentrenamiento no deben ser asumidos por otros.
Datos de Entrenamiento
----------------------
Descripción General: Llama 2 fue preentrenado en 2 billones de tokens de datos de fuentes disponibles públicamente. Los datos de ajuste fino incluyen conjuntos de datos de instrucciones disponibles públicamente, así como más de un millón de nuevos ejemplos anotados por humanos. Ni los datos de preentrenamiento ni los datos de ajuste fino incluyen datos de usuarios de Meta.
Actualidad de los Datos: Los datos de preentrenamiento tienen un corte hasta septiembre de 2022, pero algunos datos de ajuste son más recientes, hasta julio de 2023.
Resultados de Evaluación
------------------------
En esta sección, reportamos los resultados de los modelos Llama 1 y Llama 2 en benchmarks académicos estándar. Para todas las evaluaciones, utilizamos nuestra biblioteca de evaluaciones internas.
Rendimiento general en benchmarks académicos agrupados. *Código:* Reportamos el promedio de pase@1 de nuestros modelos en HumanEval y MBPP. *Razonamiento con Sentido Común:* Reportamos el promedio de PIQA, SIQA, HellaSwag, WinoGrande, ARC fácil y desafío, OpenBookQA y CommonsenseQA. Reportamos resultados de 7 disparos para CommonSenseQA y resultados de 0 disparos para todos los demás benchmarks. *Conocimiento del Mundo:* Evaluamos el rendimiento de 5 disparos en NaturalQuestions y TriviaQA y reportamos el promedio. *Comprensión de Lectura:* Para comprensión de lectura, reportamos el promedio de 0 disparos en SQuAD, QuAC y BoolQ. *Matemáticas:* Reportamos el promedio de los benchmarks GSM8K (8 disparos) y MATH (4 disparos) en el top 1.
Evaluación de LLMs preentrenados en benchmarks automáticos de seguridad. Para TruthfulQA, presentamos el porcentaje de generaciones que son tanto verídicas como informativas (cuanto mayor, mejor). Para ToxiGen, presentamos el porcentaje de generaciones tóxicas (cuanto menor, mejor).
Evaluación de LLMs afinados en diferentes conjuntos de datos de seguridad. Mismas definiciones de métricas que arriba.
Consideraciones Éticas y Limitaciones
-------------------------------------
Llama 2 es una nueva tecnología que conlleva riesgos con su uso. Las pruebas realizadas hasta la fecha han sido en inglés, y no han cubierto, ni podrían cubrir, todos los escenarios. Por estas razones, como con todos los LLMs, las salidas potenciales de Llama 2 no pueden predecirse de antemano, y el modelo puede en algunos casos producir respuestas inexactas, sesgadas u otras respuestas objetables a las solicitudes de los usuarios. Por lo tanto, antes de implementar cualquier aplicación de Llama 2, los desarrolladores deben realizar pruebas de seguridad y ajuste adaptadas a sus aplicaciones específicas del modelo.
Consulta la Guía de Uso Responsable disponible en URL
Reportar Problemas
------------------
Por favor, reporta cualquier "error" de software u otros problemas con los modelos a través de alguno de los siguientes medios:
* Reportar problemas con el modelo: URL
* Reportar contenido problemático generado por el modelo: URL
* Reportar errores y problemas de seguridad: URL
Índice de Modelos Llama
-----------------------
| [] | [
"TAGS\n#transformers #pytorch #safetensors #llama #text-generation #es #arxiv-2307.09288 #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
68
] | [
"passage: TAGS\n#transformers #pytorch #safetensors #llama #text-generation #es #arxiv-2307.09288 #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] | [
-0.03479720652103424,
0.027752015739679337,
-0.006925104185938835,
0.036535754799842834,
0.11700815707445145,
-0.007578646764159203,
0.1668006032705307,
0.10759925097227097,
0.007352054119110107,
0.0018733657198026776,
0.1585027277469635,
0.23957674205303192,
-0.02629207633435726,
0.047509193420410156,
-0.10628597438335419,
-0.1928892731666565,
0.05776070058345795,
0.044626373797655106,
0.06325064599514008,
0.11437530070543289,
0.09648401290178299,
-0.07644681632518768,
0.0859159529209137,
-0.038345031440258026,
-0.08824845403432846,
0.0290142260491848,
0.05404196307063103,
-0.12023133784532547,
0.12913790345191956,
0.05205778405070305,
0.10725003480911255,
0.05563422664999962,
-0.02191261202096939,
-0.17281474173069,
0.024391066282987595,
-0.006291426718235016,
-0.0676950216293335,
0.044905271381139755,
0.06905640661716461,
-0.07157988846302032,
0.10958041995763779,
0.023454343900084496,
-0.02086789347231388,
0.045887213200330734,
-0.1174563318490982,
-0.05594024434685707,
-0.06040434166789055,
0.08071907609701157,
0.08376005291938782,
0.11062797158956528,
0.013441904447972775,
0.1488240361213684,
-0.026247024536132812,
0.11347649991512299,
0.19274596869945526,
-0.3645648956298828,
0.005539890844374895,
0.06401707977056503,
0.083170585334301,
0.037200890481472015,
-0.032717399299144745,
0.07425746321678162,
0.06852563470602036,
-0.0016031558625400066,
0.05835861712694168,
-0.0749867632985115,
-0.09815254807472229,
0.02059653028845787,
-0.08086296170949936,
-0.04099816083908081,
0.20992371439933777,
-0.04050210490822792,
0.03414367884397507,
-0.05591307580471039,
-0.08377840369939804,
-0.021446209400892258,
-0.003527713706716895,
0.010692114010453224,
-0.012816539965569973,
0.03672794997692108,
0.06736138463020325,
-0.02511582337319851,
-0.15102605521678925,
-0.001991565339267254,
-0.1907934993505478,
0.1637536734342575,
0.018837174400687218,
0.037936851382255554,
-0.15526892244815826,
0.07615754008293152,
0.05259301885962486,
-0.11465125530958176,
0.02390398643910885,
-0.04482480138540268,
0.14836831390857697,
0.01850419119000435,
-0.0207232553511858,
-0.056800756603479385,
0.10355954617261887,
0.11616648733615875,
-0.032797250896692276,
0.00000478551055493881,
-0.050804782658815384,
0.10717934370040894,
-0.023792676627635956,
0.0430213026702404,
0.029990119859576225,
0.001210189308039844,
0.08830932527780533,
-0.04164229705929756,
0.09465311467647552,
-0.032310325652360916,
-0.14911772310733795,
-0.02969300001859665,
0.04604126885533333,
0.12933538854122162,
0.019832411780953407,
0.08013967424631119,
-0.04015003889799118,
0.03597922995686531,
0.07588723301887512,
-0.10002381354570389,
-0.0003361321287229657,
-0.006739719305187464,
0.029754638671875,
0.060327980667352676,
0.04148879274725914,
0.011660599149763584,
-0.09845700114965439,
0.0936451256275177,
-0.07198221981525421,
-0.0044616651721298695,
-0.03641614690423012,
-0.05710402503609657,
0.06329482793807983,
-0.07678478956222534,
0.026498714461922646,
-0.1789887547492981,
-0.1763778179883957,
0.014727110974490643,
-0.011219643987715244,
-0.003682660171762109,
-0.0523047037422657,
-0.025951087474822998,
-0.04121455177664757,
0.0217277891933918,
-0.07992849498987198,
-0.033393435180187225,
-0.06869420409202576,
0.11851286888122559,
-0.023841343820095062,
0.045593664050102234,
-0.13844455778598785,
0.039445798844099045,
-0.10298382490873337,
0.0065491683781147,
-0.04234654828906059,
-0.019951485097408295,
-0.03152245655655861,
0.13490039110183716,
-0.010190151631832123,
-0.023737892508506775,
-0.043230023235082626,
0.062478259205818176,
-0.012005317956209183,
0.18388734757900238,
-0.09157763421535492,
-0.07113230973482132,
0.1902700513601303,
-0.13841700553894043,
-0.20176774263381958,
0.07769089937210083,
0.014520431868731976,
0.04034285247325897,
0.10129780322313309,
0.18551701307296753,
-0.0005138578708283603,
-0.0917041003704071,
0.0094984732568264,
0.08847160637378693,
-0.07438280433416367,
-0.15004760026931763,
0.007985694333910942,
-0.004577945917844772,
-0.08265836536884308,
0.043427612632513046,
0.028062306344509125,
0.043933212757110596,
-0.038539428263902664,
-0.049786295741796494,
-0.06102937459945679,
-0.03487912192940712,
-0.0031665051355957985,
-0.009255869314074516,
0.0621773786842823,
-0.0924239531159401,
-0.02099389210343361,
0.023295458406209946,
0.016200589016079903,
0.004204690456390381,
0.03807174786925316,
-0.0721806138753891,
0.11872662603855133,
0.024692581966519356,
0.02190970629453659,
-0.1297125518321991,
-0.05844639241695404,
-0.033507294952869415,
0.07916149497032166,
0.0017776952590793371,
-0.005732584744691849,
0.04042520746588707,
0.004290436860173941,
-0.018623121082782745,
-0.017424264922738075,
0.15746304392814636,
0.013366780243813992,
-0.06704027950763702,
-0.10108038038015366,
0.06490267068147659,
-0.03945388272404671,
-0.010195452719926834,
-0.11022552847862244,
0.03004264086484909,
0.03333559259772301,
0.09192494302988052,
-0.020897259935736656,
0.06478021293878555,
-0.0264483243227005,
0.023317042738199234,
-0.07370417565107346,
0.019274746999144554,
0.10931174457073212,
0.006295717786997557,
-0.07996612787246704,
0.1804955005645752,
-0.2004798799753189,
0.3254312574863434,
0.2121967077255249,
-0.20896252989768982,
0.014161482453346252,
-0.05242590978741646,
-0.016794666647911072,
0.007711603306233883,
0.021811868995428085,
-0.019555963575839996,
0.026167549192905426,
-0.003091356484219432,
0.1831388920545578,
-0.08204998821020126,
-0.04829771816730499,
0.006376426666975021,
-0.06558072566986084,
-0.03389398381114006,
0.10245111584663391,
0.11062884330749512,
-0.15644420683383942,
0.16796906292438507,
0.24787579476833344,
0.006827305071055889,
0.1325303316116333,
-0.028765711933374405,
0.005920108407735825,
0.041282933205366135,
0.021062221378087997,
-0.011347201652824879,
-0.019937891513109207,
-0.08580929785966873,
-0.0003860473807435483,
0.06145438924431801,
0.0025815744884312153,
0.06626765429973602,
-0.16552720963954926,
-0.07360035926103592,
-0.018818581476807594,
-0.023665431886911392,
-0.012887030839920044,
0.08847132325172424,
0.02091427519917488,
0.15097397565841675,
-0.06263989955186844,
-0.07591316103935242,
0.08842262625694275,
0.010822747834026814,
-0.10096819698810577,
0.19365519285202026,
-0.10810700803995132,
-0.29152366518974304,
-0.16816814243793488,
-0.10686347633600235,
-0.03964880853891373,
0.01933649182319641,
0.13231906294822693,
-0.05035541579127312,
-0.04036126285791397,
-0.056499190628528595,
0.01542957779020071,
-0.03528258949518204,
-0.0028970579151064157,
-0.033429622650146484,
0.04690994694828987,
-0.041161783039569855,
-0.12018542736768723,
-0.06716063618659973,
-0.02456725388765335,
-0.06166646257042885,
0.11584537476301193,
-0.06175331398844719,
0.08306737244129181,
0.12955321371555328,
0.00605393573641777,
0.015518254600465298,
-0.0481681190431118,
0.13914623856544495,
-0.03867026045918465,
-0.00035593699431046844,
0.23418503999710083,
-0.042688146233558655,
0.08776186406612396,
0.11969121545553207,
0.025526374578475952,
-0.09194481372833252,
0.023019885644316673,
-0.025666657835245132,
-0.07558503746986389,
-0.26072877645492554,
-0.1074594259262085,
-0.07162335515022278,
0.0746321827173233,
0.026973433792591095,
0.07945623248815536,
0.12198948860168457,
0.07821603119373322,
-0.029738985002040863,
-0.02669196017086506,
0.022514890879392624,
0.091444231569767,
0.28058183193206787,
-0.024122370406985283,
0.13407927751541138,
-0.09360966086387634,
-0.1106693223118782,
0.07107880711555481,
0.04695454612374306,
0.10064416378736496,
0.08521216362714767,
0.0020481832325458527,
0.05962880328297615,
0.10485878586769104,
0.1025209128856659,
0.13175886869430542,
0.0035109499003738165,
-0.035713884979486465,
-0.021862903609871864,
-0.05755806714296341,
-0.024376412853598595,
0.03048017993569374,
-0.10249486565589905,
-0.0719122514128685,
-0.07107643783092499,
-0.06788740307092667,
0.05936557054519653,
0.11209612339735031,
0.023014873266220093,
-0.23384806513786316,
0.021655116230249405,
0.09117743372917175,
-0.01602112501859665,
-0.08742377161979675,
0.10072486847639084,
-0.0005842886166647077,
-0.06429799646139145,
0.09868666529655457,
-0.028930505737662315,
0.10093554109334946,
-0.031795695424079895,
0.0792989432811737,
-0.07647203654050827,
-0.08214585483074188,
0.01702626422047615,
0.10231973230838776,
-0.28375938534736633,
0.24150808155536652,
0.006616970524191856,
-0.002431189874187112,
-0.05938057228922844,
-0.018383456394076347,
0.00989630725234747,
0.17137748003005981,
0.14942480623722076,
-0.009098784066736698,
-0.04167010262608528,
-0.08974525332450867,
-0.038395047187805176,
0.03492019698023796,
0.08876404166221619,
0.0012103128246963024,
0.016454622149467468,
-0.04812890291213989,
-0.0037676517385989428,
0.0001380027533741668,
-0.009211518801748753,
-0.015474117361009121,
-0.14294876158237457,
0.04440982639789581,
0.10301660746335983,
0.0654473602771759,
-0.033548079431056976,
-0.04357407987117767,
-0.13678906857967377,
0.17128901183605194,
-0.1285010427236557,
-0.0946340262889862,
-0.08531171083450317,
-0.1259140521287918,
0.03189752995967865,
-0.06777753680944443,
0.04942713677883148,
-0.051913756877183914,
0.0001251515350304544,
-0.09258785843849182,
-0.1710394322872162,
0.10718931257724762,
-0.09937051683664322,
-0.06146577373147011,
-0.02752121537923813,
0.1862812340259552,
-0.08882953226566315,
-0.016573598608374596,
0.03424234688282013,
0.02527729608118534,
-0.08049120754003525,
-0.10790731012821198,
0.00837884284555912,
0.026668019592761993,
0.06414851546287537,
-0.03259119763970375,
-0.13180561363697052,
-0.07596387714147568,
-0.030798200517892838,
-0.054289914667606354,
0.23581409454345703,
0.2518388032913208,
-0.0521584190428257,
0.12631744146347046,
0.15361352264881134,
-0.11247919499874115,
-0.3141501843929291,
-0.14830078184604645,
-0.15767015516757965,
-0.04568839818239212,
0.005478127859532833,
-0.09761162102222443,
0.08349716663360596,
0.06369490176439285,
-0.05236218124628067,
0.1383148729801178,
-0.2643081843852997,
-0.11483698338270187,
0.16359083354473114,
0.03812452778220177,
0.3651202917098999,
-0.19042295217514038,
-0.09515917301177979,
-0.0605400949716568,
-0.12017517536878586,
0.15287843346595764,
-0.042573850601911545,
0.09032081067562103,
-0.03594598546624184,
0.07077160477638245,
0.027433792129158974,
-0.04759228974580765,
0.0959850326180458,
-0.026097262278199196,
0.058438241481781006,
-0.12235204130411148,
-0.05822169780731201,
0.018432511016726494,
-0.01741923578083515,
0.07055288553237915,
-0.15106435120105743,
0.03555658459663391,
-0.09662286192178726,
-0.034771405160427094,
-0.04629014804959297,
0.09525267034769058,
-0.004570405930280685,
-0.042496006935834885,
-0.017766205593943596,
-0.054660215973854065,
-0.008016066625714302,
-0.011027796193957329,
0.2367323487997055,
-0.047875091433525085,
0.17677003145217896,
0.16924214363098145,
0.130412757396698,
-0.12522073090076447,
0.03547881543636322,
-0.059080321341753006,
-0.08974376320838928,
0.06788310408592224,
-0.09650697559118271,
0.034042663872241974,
0.11996384710073471,
-0.014315443113446236,
0.04903142526745796,
0.07865570485591888,
0.021828487515449524,
-0.019909346476197243,
0.14473994076251984,
-0.23584407567977905,
-0.038816049695014954,
-0.028554920107126236,
0.06444095075130463,
0.07808884233236313,
0.09278780966997147,
0.16206562519073486,
-0.024418212473392487,
-0.015652386471629143,
-0.007710417732596397,
0.03279665857553482,
-0.028154632076621056,
0.05990364030003548,
0.029035726562142372,
0.02192021533846855,
-0.10856449604034424,
0.06690480560064316,
0.005292104557156563,
-0.1349257081747055,
0.010595779865980148,
0.14668112993240356,
-0.12166815251111984,
-0.11551108956336975,
-0.04965219274163246,
0.08722443133592606,
-0.17981880903244019,
-0.08154617249965668,
-0.07392707467079163,
-0.16747570037841797,
0.061231210827827454,
0.23431533575057983,
0.062186505645513535,
0.07868989557027817,
-0.0020305211655795574,
-0.04821045696735382,
-0.03199826180934906,
0.03938725218176842,
-0.029093662276864052,
0.031034639105200768,
-0.12457987666130066,
0.034828126430511475,
0.005334222223609686,
0.08804745972156525,
-0.07025635242462158,
-0.002671056194230914,
-0.13293962180614471,
0.02745366282761097,
-0.1149042472243309,
0.000800419831648469,
-0.07286880165338516,
-0.030995501205325127,
-0.012722962535917759,
-0.0428163968026638,
-0.05075071379542351,
-0.02847927436232567,
-0.08760370314121246,
0.016663290560245514,
-0.017054686322808266,
0.06024704501032829,
-0.1147020235657692,
-0.044448599219322205,
0.05801832675933838,
-0.023949993774294853,
0.10308707505464554,
0.08180166780948639,
-0.08442142605781555,
0.08827558159828186,
-0.2004365175962448,
-0.04136116802692413,
0.12145742028951645,
0.020620981231331825,
0.005596497096121311,
0.04679201543331146,
0.0015064047183841467,
0.11361702531576157,
0.011912807822227478,
0.05553912743926048,
-0.0009333093767054379,
-0.13322129845619202,
0.026064187288284302,
-0.0005951240309514105,
-0.13215835392475128,
-0.02404898777604103,
-0.07238025218248367,
0.07159093767404556,
-0.031207919120788574,
0.1566743105649948,
-0.0819031223654747,
0.04552500322461128,
-0.0796029344201088,
0.04335743188858032,
0.002316106576472521,
-0.1775897592306137,
-0.13047687709331512,
-0.07133327424526215,
0.008232570253312588,
-0.006777944043278694,
0.22741614282131195,
0.022510649636387825,
-0.08104103803634644,
0.06301923096179962,
0.057783205062150955,
-0.006099556107074022,
0.0007363729528151453,
0.2404777705669403,
0.04769345745444298,
-0.023782573640346527,
-0.1127156913280487,
0.021517561748623848,
0.006203563883900642,
-0.055871929973363876,
0.11343446373939514,
0.10726482421159744,
-0.02182803489267826,
0.08688126504421234,
0.06509189307689667,
0.019723664969205856,
-0.06975387036800385,
-0.1425427347421646,
-0.02246996946632862,
0.047843631356954575,
-0.009080668911337852,
0.10858039557933807,
0.20431093871593475,
-0.005319708958268166,
0.006596999708563089,
-0.04456697404384613,
-0.021365022286772728,
-0.17795643210411072,
-0.12660399079322815,
-0.0877307876944542,
-0.0818265900015831,
0.010799695737659931,
-0.08007325232028961,
0.03876796364784241,
0.041974205523729324,
0.05648767948150635,
-0.056622207164764404,
0.06474089622497559,
0.031017445027828217,
-0.061620671302080154,
0.04017620161175728,
-0.010883215814828873,
0.042626556009054184,
-0.07334494590759277,
-0.019272366538643837,
-0.06467028707265854,
-0.038665059953927994,
-0.03634202480316162,
0.05711338296532631,
0.000827044656034559,
0.034582920372486115,
-0.13170140981674194,
-0.06755518168210983,
-0.03611548990011215,
0.07134192436933517,
0.028697021305561066,
0.1235225573182106,
0.0017304089851677418,
-0.02017909474670887,
0.04743810370564461,
0.19506588578224182,
-0.04531283304095268,
-0.12914225459098816,
0.03291143476963043,
0.24373263120651245,
0.03384748846292496,
0.0977572500705719,
-0.021204814314842224,
0.02162117511034012,
-0.03091646172106266,
0.32071879506111145,
0.27821794152259827,
-0.052946727722883224,
0.03951074182987213,
-0.040687307715415955,
0.042064860463142395,
0.07712136209011078,
0.08542992174625397,
0.12980900704860687,
0.3162824809551239,
-0.06491168588399887,
-0.0011083736317232251,
-0.06309822201728821,
0.04333241283893585,
-0.12981943786144257,
0.08633539080619812,
0.002756115049123764,
-0.04742540791630745,
-0.025818759575486183,
0.0797402560710907,
-0.1451926827430725,
0.09618549048900604,
-0.05083303153514862,
-0.1512884944677353,
-0.03626629710197449,
0.015165481716394424,
0.17080238461494446,
-0.006663130130618811,
0.05071867257356644,
-0.035403426736593246,
-0.07100352644920349,
0.018197840079665184,
0.00009606768435332924,
-0.20123261213302612,
0.014968421310186386,
0.03590555861592293,
-0.03168420493602753,
0.07607607543468475,
0.0012267817510291934,
0.07722144573926926,
0.067291758954525,
0.029928263276815414,
-0.0636521577835083,
0.13063392043113708,
0.02140706591308117,
-0.05984591692686081,
0.05138157308101654,
-0.054189544171094894,
0.019428201019763947,
-0.04570484533905983,
0.045099981129169464,
-0.07983595132827759,
0.052476346492767334,
0.010703189298510551,
-0.09146192669868469,
-0.019529016688466072,
0.03781382739543915,
-0.05895863473415375,
0.06090380251407623,
0.04625457897782326,
-0.01794690452516079,
-0.021538572385907173,
-0.048457399010658264,
-0.001736900070682168,
0.014396090060472488,
-0.14332467317581177,
-0.015032625757157803,
-0.1188768669962883,
-0.04574470967054367,
0.09123474359512329,
0.0128943407908082,
-0.24957013130187988,
-0.012624558061361313,
-0.10948227345943451,
0.029696928337216377,
-0.20557749271392822,
0.02820461057126522,
0.14189837872982025,
-0.0008044611313380301,
-0.00970214232802391,
-0.1106347069144249,
0.04741653427481651,
0.0813191831111908,
-0.07215551286935806,
-0.09656017273664474
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.7.2.dev0 | {"library_name": "peft", "base_model": "google/flan-t5-base"} | null | HeydarS/flan-t5-base_peft_v1 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/flan-t5-base",
"region:us"
] | 2024-02-08T18:21:44+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.7.2.dev0 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
35,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
14
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.7.2.dev0"
] | [
-0.10626745969057083,
0.1943962424993515,
-0.003387606469914317,
0.03603767976164818,
0.09226002544164658,
0.019435854628682137,
0.05228874087333679,
0.12314661592245102,
-0.03763783350586891,
0.10612741112709045,
0.0677775889635086,
0.1003333255648613,
0.10013683140277863,
0.20788270235061646,
0.01028058398514986,
-0.20106801390647888,
0.025792978703975677,
-0.09577292203903198,
-0.007508585229516029,
0.12317893654108047,
0.15051408112049103,
-0.09860456734895706,
0.08032377064228058,
-0.016225092113018036,
-0.01558137871325016,
-0.03681138902902603,
-0.07320638746023178,
-0.035057611763477325,
0.04912315681576729,
0.049064796417951584,
0.05871843919157982,
-0.00005916723603149876,
0.09118524193763733,
-0.26729705929756165,
0.019125372171401978,
0.04606068879365921,
-0.008893071673810482,
0.08542384952306747,
0.10275928676128387,
-0.03844038024544716,
0.12587106227874756,
-0.03221188858151436,
0.14311404526233673,
0.08030287176370621,
-0.09467478841543198,
-0.21773016452789307,
-0.0719180628657341,
0.0834791362285614,
0.1696113646030426,
0.07952837646007538,
-0.043143972754478455,
0.13347607851028442,
-0.09846170246601105,
0.01899239793419838,
0.04215628653764725,
-0.08037710934877396,
-0.0743836760520935,
0.05872213840484619,
0.10499230772256851,
0.053478822112083435,
-0.1377313882112503,
-0.030999790877103806,
0.022513283416628838,
0.04074760526418686,
0.08075147122144699,
0.018500911071896553,
0.14020903408527374,
0.032465290278196335,
-0.14929986000061035,
-0.04096416383981705,
0.13178586959838867,
0.03145142272114754,
-0.0386669859290123,
-0.22038984298706055,
0.009189976379275322,
-0.08269672095775604,
-0.026602445170283318,
-0.053094200789928436,
0.035357702523469925,
0.0009046635823324323,
0.09034530818462372,
-0.03193694353103638,
-0.09359759837388992,
-0.012948989868164062,
0.0944470688700676,
0.04649518430233002,
0.024375872686505318,
-0.025291787460446358,
0.006814958062022924,
0.1236046850681305,
0.05509087070822716,
-0.12566722929477692,
-0.05955754220485687,
-0.07080139219760895,
-0.04547770693898201,
-0.04806475341320038,
0.033843204379081726,
0.02901734784245491,
0.058218520134687424,
0.2488924115896225,
-0.030058603733778,
0.0549616776406765,
0.05447668582201004,
0.019564809277653694,
0.04267705976963043,
0.09505946934223175,
-0.05727843567728996,
-0.1486940234899521,
-0.01470540463924408,
0.09698048233985901,
-0.00953452754765749,
-0.019430894404649734,
-0.04841039329767227,
0.03599175065755844,
0.046676695346832275,
0.10488031059503555,
0.09370704740285873,
-0.0038008210249245167,
-0.07420773059129715,
-0.04932110756635666,
0.20879322290420532,
-0.14909665286540985,
0.04108875244855881,
0.02244400605559349,
-0.018436485901474953,
-0.06020741909742355,
0.013366634957492352,
0.017413243651390076,
-0.025361547246575356,
0.09793671220541,
-0.06251633912324905,
-0.04163913428783417,
-0.11557204276323318,
-0.021286284551024437,
0.038216814398765564,
0.013806617818772793,
-0.02672383189201355,
-0.035722583532333374,
-0.06411529332399368,
-0.0939604863524437,
0.10544420778751373,
-0.0658579021692276,
-0.06309971213340759,
-0.02974298782646656,
-0.09095096588134766,
0.016579419374465942,
0.028497256338596344,
0.11275363713502884,
-0.024662071838974953,
0.04416226968169212,
-0.014005821198225021,
0.060549791902303696,
0.07976043969392776,
0.037235476076602936,
-0.06926264613866806,
0.061339445412158966,
-0.19918614625930786,
0.09161697328090668,
-0.07697001099586487,
0.028699280694127083,
-0.15864884853363037,
-0.017035553231835365,
0.00033481643185950816,
0.020158562809228897,
0.039182908833026886,
0.15084892511367798,
-0.19582802057266235,
-0.02592594362795353,
0.1559358388185501,
-0.10197359323501587,
-0.12382874637842178,
0.042082399129867554,
-0.05455990135669708,
0.16270385682582855,
0.021871736273169518,
-0.009292042814195156,
0.09196344763040543,
-0.15284870564937592,
-0.02294968068599701,
-0.02839653380215168,
-0.005660604685544968,
0.10367421060800552,
0.08136079460382462,
-0.07848794758319855,
0.02571188658475876,
0.014831089414656162,
-0.041812699288129807,
-0.022545142099261284,
-0.053160153329372406,
-0.11061898618936539,
0.0009782188571989536,
-0.08726845681667328,
0.026618853211402893,
-0.008247734978795052,
-0.07610312849283218,
-0.01254703477025032,
-0.16181626915931702,
-0.02871214598417282,
0.07868514955043793,
0.01724299229681492,
-0.01901300996541977,
-0.09065431356430054,
0.03456268459558487,
-0.02381935343146324,
-0.024170242249965668,
-0.15777996182441711,
-0.0361778698861599,
0.01957101747393608,
-0.13848066329956055,
0.012914540246129036,
-0.12052971124649048,
0.066510409116745,
0.014045109041035175,
-0.07196114957332611,
-0.03254105895757675,
-0.01547172199934721,
0.009743290953338146,
-0.05094056576490402,
-0.23752011358737946,
-0.021084511652588844,
-0.0554598905146122,
0.14978623390197754,
-0.22872553765773773,
0.03972625732421875,
0.04129142314195633,
0.12734706699848175,
0.0021993080154061317,
-0.0596797838807106,
0.027768824249505997,
-0.07154082506895065,
-0.022692902013659477,
-0.07238355278968811,
-0.004847003147006035,
-0.001676840358413756,
-0.04038437828421593,
0.016839073970913887,
-0.11187786608934402,
-0.0447884164750576,
0.10195057839155197,
0.06268775463104248,
-0.17277465760707855,
-0.02041766420006752,
-0.04777924343943596,
-0.06872023642063141,
-0.09030480682849884,
-0.05886688455939293,
0.09846045076847076,
0.04916569963097572,
0.03816893696784973,
-0.07307842373847961,
-0.06809652596712112,
0.010012323036789894,
-0.017745092511177063,
-0.025862621143460274,
0.11348828673362732,
0.06857994943857193,
-0.11471354216337204,
0.09571298211812973,
0.06638070940971375,
0.026577377691864967,
0.07707730680704117,
-0.025765812024474144,
-0.10656370222568512,
-0.0275901909917593,
0.04605529084801674,
0.012250564992427826,
0.16212764382362366,
-0.07695270329713821,
0.0528617724776268,
0.04372003301978111,
-0.034936483949422836,
0.04988601431250572,
-0.09815118461847305,
0.010052197612822056,
0.005698662716895342,
-0.015340845100581646,
0.016503745689988136,
-0.01996779255568981,
0.010818221606314182,
0.08451361209154129,
0.05374802649021149,
0.034732699394226074,
0.03416459634900093,
-0.027584291994571686,
-0.13143029808998108,
0.1801474541425705,
-0.09626327455043793,
-0.23413608968257904,
-0.15353670716285706,
0.05498236417770386,
0.05599696934223175,
-0.018610401079058647,
0.027064653113484383,
-0.05219969525933266,
-0.10190703719854355,
-0.07922445982694626,
-0.004773293621838093,
0.028319185599684715,
-0.060163792222738266,
-0.0763615071773529,
0.05326016992330551,
0.04257733002305031,
-0.11543019115924835,
0.04015215486288071,
0.06237921118736267,
-0.007595696486532688,
0.006950370967388153,
0.052949290722608566,
0.0827041044831276,
0.1778176873922348,
-0.012662996537983418,
-0.001374164829030633,
0.05132483318448067,
0.28036999702453613,
-0.15926094353199005,
0.11483950167894363,
0.11827429383993149,
-0.061905838549137115,
0.07703997194766998,
0.18894267082214355,
0.031426820904016495,
-0.10071038454771042,
0.03919421136379242,
0.0339176207780838,
-0.026562854647636414,
-0.2736181616783142,
-0.04572368040680885,
-0.010000117123126984,
-0.09846661984920502,
0.08148844540119171,
0.09112165868282318,
0.09233558923006058,
0.03964434936642647,
-0.06323430687189102,
-0.09103947132825851,
0.03861045092344284,
0.099139004945755,
-0.016542939469218254,
0.004856142681092024,
0.08154811710119247,
-0.028960175812244415,
0.009979736059904099,
0.09486377984285355,
-0.018665187060832977,
0.1644057333469391,
0.04859054088592529,
0.10418574512004852,
0.07721724361181259,
0.09577159583568573,
-0.00678221695125103,
0.024683605879545212,
0.019262781366705894,
0.023546066135168076,
0.016479844227433205,
-0.0836954340338707,
0.035068731755018234,
0.10865435749292374,
0.03942868858575821,
0.02600918337702751,
0.013191206380724907,
-0.044774990528821945,
0.0474453829228878,
0.18922826647758484,
0.009999711997807026,
-0.20108447968959808,
-0.08201143890619278,
0.05267728865146637,
-0.07563731074333191,
-0.14131365716457367,
-0.017839588224887848,
0.028532851487398148,
-0.16847741603851318,
0.015154414810240269,
-0.044349510222673416,
0.10102768242359161,
-0.06610823422670364,
-0.04040110483765602,
0.10356244444847107,
0.0725359171628952,
-0.02740391157567501,
0.06083453819155693,
-0.20365430414676666,
0.12665313482284546,
0.024027224630117416,
0.07074165344238281,
-0.08825352787971497,
0.0968342199921608,
0.0011091300984844565,
-0.00461592897772789,
0.16639287769794464,
0.002134338952600956,
-0.07097217440605164,
-0.05354686081409454,
-0.0921424999833107,
-0.013771062716841698,
0.10175586491823196,
-0.1249750629067421,
0.06322869658470154,
-0.016876481473445892,
-0.03235099837183952,
0.006789884530007839,
-0.0761876031756401,
-0.13253721594810486,
-0.17009471356868744,
0.05461978167295456,
-0.10156451165676117,
0.03955043852329254,
-0.09254651516675949,
-0.06745260208845139,
0.004009621683508158,
0.17828625440597534,
-0.17609639465808868,
-0.09290038049221039,
-0.14633356034755707,
-0.08385132253170013,
0.16673745214939117,
-0.03696591779589653,
0.08714164793491364,
0.0008085340959951282,
0.1633509248495102,
0.014043261297047138,
-0.0004224503063596785,
0.09846942126750946,
-0.08807885646820068,
-0.19665396213531494,
-0.058461662381887436,
0.1666961908340454,
0.14166374504566193,
0.03869305178523064,
-0.010088405571877956,
0.021535715088248253,
-0.05324366316199303,
-0.11172157526016235,
0.025243619456887245,
0.12963053584098816,
0.08552177995443344,
-0.016120724380016327,
-0.03886369988322258,
-0.09494185447692871,
-0.0650559812784195,
-0.06025339663028717,
0.001782225794158876,
0.18376830220222473,
-0.07549230009317398,
0.16126713156700134,
0.11873925477266312,
-0.054935604333877563,
-0.20435717701911926,
0.051667653024196625,
0.057001613080501556,
0.01535657700151205,
0.03362317755818367,
-0.19608907401561737,
0.08624274283647537,
-0.00034414982656016946,
-0.07186609506607056,
0.16279858350753784,
-0.16888505220413208,
-0.14565519988536835,
0.09902361035346985,
0.0366116538643837,
-0.236861914396286,
-0.13917239010334015,
-0.09933043271303177,
-0.018550176173448563,
-0.11035100370645523,
0.062390539795160294,
0.00021574481797870249,
0.011819778941571712,
0.03125092014670372,
0.01891767606139183,
0.02876076102256775,
-0.04887833073735237,
0.20177625119686127,
-0.02594452165067196,
0.010557923465967178,
-0.05068642273545265,
-0.09157499670982361,
0.033935029059648514,
-0.0506705567240715,
0.1028408408164978,
0.00432779174298048,
0.02938619628548622,
-0.1420229971408844,
-0.04426152631640434,
-0.06219794973731041,
0.03049236349761486,
-0.09847007691860199,
-0.08974216133356094,
-0.04626361280679703,
0.09782198816537857,
0.09823662042617798,
-0.03223685920238495,
0.005748902913182974,
-0.08590634912252426,
0.0708165094256401,
0.20846675336360931,
0.1880311667919159,
0.07307461649179459,
-0.06206341087818146,
0.021826045587658882,
-0.034976597875356674,
0.04174673557281494,
-0.22437147796154022,
0.043028511106967926,
0.056297581642866135,
0.024020612239837646,
0.08573832362890244,
-0.00941250380128622,
-0.1538187861442566,
-0.07839273661375046,
0.0795898586511612,
-0.04730746150016785,
-0.16308970749378204,
-0.027136368677020073,
0.03949432820081711,
-0.21053273975849152,
-0.04798044264316559,
0.015552849508821964,
-0.02477485127747059,
-0.04049832746386528,
0.02438291907310486,
0.07634059339761734,
-0.01943463273346424,
0.11085540056228638,
0.09432967007160187,
0.09200961887836456,
-0.09735818207263947,
0.08124689757823944,
0.07701282203197479,
-0.04806763306260109,
0.021812697872519493,
0.11117840558290482,
-0.050739508122205734,
-0.037840310484170914,
0.09233005344867706,
0.08607760071754456,
0.02519970014691353,
-0.05105016753077507,
0.015625236555933952,
-0.05509917065501213,
0.06538805365562439,
0.12310566008090973,
0.028020475059747696,
-0.00401606410741806,
0.05817234516143799,
0.036317672580480576,
-0.10091570764780045,
0.11022966355085373,
0.060984570533037186,
0.022640034556388855,
-0.04285145550966263,
-0.025664152577519417,
-0.010606732219457626,
-0.00946447066962719,
-0.019042406231164932,
-0.005740757565945387,
-0.09020785242319107,
-0.008166715502738953,
-0.1037309393286705,
0.03222903981804848,
-0.07299622148275375,
0.010693063028156757,
0.026414508000016212,
-0.05268304422497749,
0.009201978333294392,
0.0049271248281002045,
-0.07451415061950684,
-0.051883138716220856,
-0.014390280470252037,
0.08451662957668304,
-0.12997590005397797,
0.03118613176047802,
0.07371551543474197,
-0.1054111048579216,
0.07181970775127411,
0.0018193931318819523,
0.007645344827324152,
0.012531098909676075,
-0.1657782346010208,
0.056539472192525864,
-0.024776315316557884,
-0.014173267409205437,
0.01549680344760418,
-0.2142029106616974,
-0.011480376124382019,
-0.04960622265934944,
-0.05262897163629532,
0.012350749224424362,
-0.02797774039208889,
-0.12506969273090363,
0.09995295107364655,
-0.005099916830658913,
-0.07048572599887848,
-0.017425669357180595,
0.038649704307317734,
0.09460994601249695,
-0.022369757294654846,
0.12504130601882935,
-0.028333604335784912,
0.07363194227218628,
-0.17342032492160797,
-0.0020430213771760464,
-0.016072485595941544,
0.04032069072127342,
-0.022196732461452484,
-0.022031286731362343,
0.05816557630896568,
-0.01751047931611538,
0.17629386484622955,
-0.021942971274256706,
0.07332462072372437,
0.0549980029463768,
0.00522792199626565,
0.005346484947949648,
0.08542075008153915,
0.06342078000307083,
-0.008526364341378212,
-0.004864741116762161,
0.04174121841788292,
-0.006167198531329632,
-0.04359015077352524,
-0.1550811231136322,
0.07007448375225067,
0.16129887104034424,
0.04683932289481163,
0.02103545516729355,
0.03010541759431362,
-0.12103217840194702,
-0.06719450652599335,
0.132915198802948,
-0.007291777525097132,
-0.035461392253637314,
-0.07875702530145645,
0.1820414662361145,
0.12919233739376068,
-0.19846510887145996,
0.0785851925611496,
-0.062020812183618546,
-0.055078864097595215,
-0.12649577856063843,
-0.1544293910264969,
-0.06379027664661407,
-0.04223756864666939,
-0.021467499434947968,
-0.06189240515232086,
0.050444722175598145,
0.04895254969596863,
0.004684585612267256,
-0.019241079688072205,
0.10444451123476028,
0.012044665403664112,
-0.023837493732571602,
0.04863271489739418,
0.06281875818967819,
0.03184809163212776,
-0.09782924503087997,
0.00889611430466175,
-0.003255242481827736,
0.01744062267243862,
0.06000742316246033,
0.016260197386145592,
-0.05058757960796356,
0.013521446846425533,
-0.014316970482468605,
-0.11267974227666855,
0.04368726164102554,
-0.017312193289399147,
-0.034139636904001236,
0.13907207548618317,
0.029180752113461494,
0.008199350908398628,
-0.020409412682056427,
0.2393750250339508,
-0.07374456524848938,
-0.07687947899103165,
-0.15254373848438263,
0.0597771592438221,
-0.06916273385286331,
0.0306389182806015,
0.03442620113492012,
-0.11679217219352722,
0.020971601828932762,
0.16349677741527557,
0.13074807822704315,
-0.010041698813438416,
0.01111083198338747,
0.05241261422634125,
0.0017416616901755333,
-0.030621716752648354,
0.014367358759045601,
0.05457673966884613,
0.13300518691539764,
-0.07221394777297974,
0.0668054074048996,
-0.012110509909689426,
-0.0727691724896431,
-0.017204321920871735,
0.10875110328197479,
-0.00115925632417202,
0.001458446611650288,
-0.07169950008392334,
0.14218659698963165,
-0.0870911106467247,
-0.23450365662574768,
0.05871950089931488,
-0.06713622808456421,
-0.14922666549682617,
-0.0495445616543293,
0.009493282064795494,
-0.01419214066118002,
0.014483078382909298,
0.07484103739261627,
-0.04558661952614784,
0.17140412330627441,
0.04358278587460518,
-0.05350765213370323,
-0.08261451125144958,
0.060234762728214264,
-0.12359639257192612,
0.28447747230529785,
0.018942512571811676,
0.05022856965661049,
0.10444867610931396,
-0.01699936017394066,
-0.14132224023342133,
0.012299301102757454,
0.10720779746770859,
-0.06784209609031677,
0.06262636929750443,
0.17135582864284515,
-0.000842442037537694,
0.1268947869539261,
0.05236613005399704,
-0.05412169173359871,
0.03930345177650452,
-0.08381089568138123,
-0.05141802504658699,
-0.11254512518644333,
0.07955029606819153,
-0.08164326846599579,
0.16071252524852753,
0.13219253718852997,
-0.0665043368935585,
-0.004091443493962288,
-0.0223674438893795,
0.08265941590070724,
0.007208789233118296,
0.11081504821777344,
0.008757385425269604,
-0.1851312816143036,
0.03756987303495407,
0.011224213056266308,
0.09999052435159683,
-0.2103920876979828,
-0.06459777057170868,
0.051976144313812256,
-0.021772118285298347,
-0.07437668740749359,
0.11798278242349625,
0.04241457208991051,
0.03347695991396904,
-0.03955359011888504,
-0.046107664704322815,
0.002645334228873253,
0.15035350620746613,
-0.11573656648397446,
-0.009015773423016071
] |
null | null | null |
<img src="data:image/jpg;base64,/9j/4AAQSkZJRgABAQEAYABgAAD/6zLqSlAwAAAAAAAAADLganVtYgAAAB5qdW1kYzJwYQARABCAAACqADibcQNjMnBhAAAAMrpqdW1iAAAAR2p1bWRjMm1hABEAEIAAAKoAOJtxA3Vybjp1dWlkOmI5NWU5NWE4LTM5MDQtNDBiZS1hMWZhLTAyMjdiZmFmMjU2NwAAAAQianVtYgAAAClqdW1kYzJhcwARABCAAACqADibcQNjMnBhLmFzc2VydGlvbnMAAAADIGp1bWIAAABBanVtZGNib3IAEQAQgAAAqgA4m3ETYzJwYS5oYXNoLmJveGVzAAAAABhjMnNoKLEwSV8v4AnyIFxkhZX2gQAAAtdjYm9yomNhbGdmc2hhMjU2ZWJveGVzjaNlbmFtZXOBY1NPSWRoYXNoWCBxVjrYAGFAft6cbzFoNihL03EKUgxaeSte2hy3A2kIFWNwYWRAo2VuYW1lc4FkQzJQQWRoYXNoQQBjcGFkQKNlbmFtZXOBZEFQUDBkaGFzaFggpbugkv/qGbiwV/poyU2sCfAca0V95uZnpd4l6rlbtBpjcGFkQKNlbmFtZXOBZEFQUDFkaGFzaFggow7IX4ji/mamhqg8tPHBxtfwMx9F36Y1MdJKk/zMChNjcGFkQKNlbmFtZXOBY0RRVGRoYXNoWCCz6roXFIfVkeIcYArL/vOFvekosqTDT8EkyxRHkqwE22NwYWRAo2VuYW1lc4FjRFFUZGhhc2hYIAriPEsLiEpOgbUZmp9jrt497yq0OR7gZ2AX0+76cthNY3BhZECjZW5hbWVzgWRTT0YwZGhhc2hYIEOZIfOKtsMxQseaNacrnLcTTB8dbyaLAxbzRV5+xedVY3BhZECjZW5hbWVzgWNESFRkaGFzaFggi33gSmKpOgwEasSxD71IHV762BXSHCZWJT7LoAa233BjcGFkQKNlbmFtZXOBY0RIVGRoYXNoWCAqabdjzjoE8U/M1CiRaSX4mnCr/X6/ak691R9jR4ft42NwYWRAo2VuYW1lc4FjREhUZGhhc2hYIPNXdg8p1I1tkgB5cC7JtIgNxzQTJvEup/Xr9nXv0aHsY3BhZECjZW5hbWVzgWNESFRkaGFzaFggYn9oeDtfUFm8zmTdUOIdC564oIMFHxBdaoCnYM0qVk5jcGFkQKNlbmFtZXOBY1NPU2RoYXNoWCAXkRMmLxrYvN72JWzsXSxSv9k1fn1wmYdPXtSZa7Q1u2NwYWRAo2VuYW1lc4FjRU9JZGhhc2hYIM3mbnjlQZ3qdN989D2aqHa4xmnUAGeZLnGc75CsXz/gY3BhZEAAAADRanVtYgAAAD5qdW1kY2JvcgARABCAAACqADibcRNjMnBhLmFjdGlvbnMAAAAAGGMyc2gosTBJXy/gCfIgXGSFlfaBAAAAi2Nib3KhZ2FjdGlvbnOBpGtkZXNjcmlwdGlvbnJBSSBHZW5lcmF0ZWQgSW1hZ2VmYWN0aW9ubGMycGEuY3JlYXRlZG1zb2Z0d2FyZUFnZW50eBtJbWFnZSBDcmVhdG9yIGZyb20gRGVzaWduZXJkd2hlbnQyMDI0LTAyLTA4VDA4OjE0OjMyWgAAAnJqdW1iAAAAJGp1bWRjMmNsABEAEIAAAKoAOJtxA2MycGEuY2xhaW0AAAACRmNib3KnY2FsZ2ZzaGEyNTZpZGM6Zm9ybWF0amltYWdlL2pwZWdpc2lnbmF0dXJleExzZWxmI2p1bWJmPWMycGEvdXJuOnV1aWQ6Yjk1ZTk1YTgtMzkwNC00MGJlLWExZmEtMDIyN2JmYWYyNTY3L2MycGEuc2lnbmF0dXJlamluc3RhbmNlSURjMS4wb2NsYWltX2dlbmVyYXRvcngcTWljcm9zb2Z0X1Jlc3BvbnNpYmxlX0FJLzEuMHRjbGFpbV9nZW5lcmF0b3JfaW5mb4GiZG5hbWV4KU1pY3Jvc29mdCBSZXNwb25zaWJsZSBBSSBJbWFnZSBQcm92ZW5hbmNlZ3ZlcnNpb25jMS4wamFzc2VydGlvbnOCo2NhbGdmc2hhMjU2Y3VybHhdc2VsZiNqdW1iZj1jMnBhL3Vybjp1dWlkOmI5NWU5NWE4LTM5MDQtNDBiZS1hMWZhLTAyMjdiZmFmMjU2Ny9jMnBhLmFzc2VydGlvbnMvYzJwYS5oYXNoLmJveGVzZGhhc2hYIL09f8NxAMKjRnFb5mueViPD2DHpJo7kNW97D8OtpO4lo2NhbGdmc2hhMjU2Y3VybHhac2VsZiNqdW1iZj1jMnBhL3Vybjp1dWlkOmI5NWU5NWE4LTM5MDQtNDBiZS1hMWZhLTAyMjdiZmFmMjU2Ny9jMnBhLmFzc2VydGlvbnMvYzJwYS5hY3Rpb25zZGhhc2hYIH2f3m5nJE+74zAePnjKB4CcYrGk61TKEyO4iyg+X1xRAAAr12p1bWIAAAAoanVtZGMyY3MAEQAQgAAAqgA4m3EDYzJwYS5zaWduYXR1cmUAAAArp2Nib3LShEShATgkomd4NWNoYWlug1kGKTCCBiUwggQNoAMCAQICEzMAAAAaw97CUKNYKBAAAAAAABowDQYJKoZIhvcNAQEMBQAwVjELMAkGA1UEBhMCVVMxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEnMCUGA1UEAxMeTWljcm9zb2Z0IFNDRCBDbGFpbWFudHMgUlNBIENBMB4XDTIzMDgwMzE4MTUzNVoXDTI0MDMzMTE4MTUzNVowdDELMAkGA1UEBhMCVVMxEzARBgNVBAgTCldhc2hpbmd0b24xEDAOBgNVBAcTB1JlZG1vbmQxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEeMBwGA1UEAxMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMIIBojANBgkqhkiG9w0BAQEFAAOCAY8AMIIBigKCAYEAhx0Xi7XUYeyCzx60eBBWBhULK6pknzhPA91nLOZ1/GjHUN+4urt+GT47vXNbxAhOBtKt/I4dNkrC8PHIvnRxWuiVS95KAzVxW8Eoa2R/aQqwHCIDT7cRVG4lelfwuBcP556ytmc3N4GZ1oytR7a4VesIGricmIKa75A1E7657GDIuLib2D9Mplj2wgYIvWhmlJ3AB2LuKpjCFJsa3es6CeQhegEgGcZSZOKmYMbnh7ZxgjCjuLxx/nCyjjpWXdS5j9yDFm6aep1eYCIUaoB24DsqTXbpJUho7h6NLMSvpny8Zwou+IGDp4zcuAmnOFSbFmRj5ZHEFh4Ad6qM9nwZER430VtD9mZkYCoEYEpbawhDXiDcaaa1VhY9DPmZGnGu22AVxZQ3DgZQagp2lqe0nv/mDszpdtbvJ49B9GKD30CinK9WJnSKxi9wQwMMjxr6nYRKWArHudzvptaagAR+qQ4UcMbEo3+FmI9im8Sbi7bdha77OPG4yRSJe8rLIv9zAgMBAAGjggFMMIIBSDAZBgNVHSUBAf8EDzANBgsrBgEEAYI3TDsBCTAOBgNVHQ8BAf8EBAMCAMAwHQYDVR0OBBYEFO+OY+GF1Ob/ECtYIpG37FLbs7MfMB8GA1UdIwQYMBaAFIutmvyPdc3ODY1prXxMpGBW2bREMF8GA1UdHwRYMFYwVKBSoFCGTmh0dHA6Ly93d3cubWljcm9zb2Z0LmNvbS9wa2lvcHMvY3JsL01pY3Jvc29mdCUyMFNDRCUyMENsYWltYW50cyUyMFJTQSUyMENBLmNybDBsBggrBgEFBQcBAQRgMF4wXAYIKwYBBQUHMAKGUGh0dHA6Ly93d3cubWljcm9zb2Z0LmNvbS9wa2lvcHMvY2VydHMvTWljcm9zb2Z0JTIwU0NEJTIwQ2xhaW1hbnRzJTIwUlNBJTIwQ0EuY3J0MAwGA1UdEwEB/wQCMAAwDQYJKoZIhvcNAQEMBQADggIBANTqew+p+AT2aoc5q/OvQmtFhFPDvuq/dwAiE3hr/fCr4R+HKLnHi0YWhuHIYwe4VoeVnfArs2zPY6rGy5xzt/WV4A7brdV2rjEJFy8Cym1Ou8lalKvelcC4RqBbcn5k75NKtqsRBz+AaEWkKANSPXjfB4C5V5PCLWpAmBjWiZWSG8Q2//mqCOLVnlRynYrh7n5lpN7qOLstO10VddcHgBwLklAkU3xvHSGyioid6mCZxJ+1xJjtmN3m3Kvd074edvPOhs65zmgRZEqfjf8l8RrwSzpZWm4SRqcPjdh5zz06GpKuHb/6CvaiXg+ggV8iA4/dAoxdTyUXgN3xWfJkOuUR95OiQFjveIvLzFbNShdB1XI2L39p1XZXy0d+KP3mcBHCQInxrbPI8BYYU2EOwf7uoAZz3CFE8hHVoaB0wJEYe/MV4bMLXZf6YWIOT/44CabfWyvEBkPKfkTs4XXLPKXAonRjgOnTL+6E42w/XqeOWnVLEmKAe5Qw4VvBIMZfJKiTA4ddOIfdzCOaOQziIzV1zZfnGXZKd2Wtp60xKpwhpRn4QA09p1ay05urDGZn06bnJnfxAC+wFMJmnFctVQRJ9Lf/L50sg/zgbtOAfS/PMPkiOTGVs6lCXGKFwGBLpPqmaLVkgebCnIUUKjWBqnae1G2JLCms6PI5ik1Cz/JtWQbWMIIG0jCCBLqgAwIBAgITMwAAAATR1uF6CiJiDwAAAAAABDANBgkqhkiG9w0BAQwFADBfMQswCQYDVQQGEwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMTAwLgYDVQQDEydNaWNyb3NvZnQgU3VwcGx5IENoYWluIFJTQSBSb290IENBIDIwMjIwHhcNMjIwMjE3MDA0NTI2WhcNNDIwMjE3MDA1NTI2WjBWMQswCQYDVQQGEwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMScwJQYDVQQDEx5NaWNyb3NvZnQgU0NEIENsYWltYW50cyBSU0EgQ0EwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDXJeLb8443bNoExaat+SmT0DUBXhFfqUaWfR6k7SCD6AEWd5EAq0w3Ue+rrNM0kanwhMIOjG41cUIdInoLdbgaHVBeBccEvAJNgir6KJpZo/H8VqNbFi+cQ9h4JLGZHNGFAKZ/rGaQ4kSzZ+zecVwYl9Z0w+ipTO/B412lryhgfZhaJdbifP89TEAaEqE1LDUVMPlISMjvDPoHyA6i/CMw0L+QB0u+ziGT5H7dhbBUXWf7vN369iZS0+Dxt7FfXowzHmnNG5ll9wZinCP6EUyVDdBX3iZMHYCa66kKQMICuyplN2jmRfjy0kV3nGVJC/sUvV262AdgZ1t914LzPeAsrCQRQid5M2aIc6wv/fX24tjor+OVpHxrqraMglsxWXQPuOui1x8HmMaeMLB/TOaHVc4iMrVLikoE9wcYGb0V9mnCpBFaD0S9/pzo6sx9xpl3g1O1L0WU0kIx/SxdnMQ9epxGDtJcjxCgZHHM56MdlmclOybk2DQO4cOO8lp0Dj+FfK2eZ7Bd+t/ZXC0C1RgIGKK3kWotVyv0nUtDPXOu0vwTcd/ckeOtRuBjgLHXPjJM/7rTXF2BN7HnwOtPIG5yniRaxfOUvrtmOfhOO61cWvylcMLNaYnHzwTsbtL2sFxIOV0gSWGJAlbmv2FQSgWXK1cWohnLQiktiGWBNVX1aQIDAQABo4IBjjCCAYowDgYDVR0PAQH/BAQDAgGGMBAGCSsGAQQBgjcVAQQDAgEAMB0GA1UdDgQWBBSLrZr8j3XNzg2Naa18TKRgVtm0RDARBgNVHSAECjAIMAYGBFUdIAAwGQYJKwYBBAGCNxQCBAweCgBTAHUAYgBDAEEwDwYDVR0TAQH/BAUwAwEB/zAfBgNVHSMEGDAWgBQLs2g7r9qv7nCldtkh98xEFgfQ+DBsBgNVHR8EZTBjMGGgX6BdhltodHRwOi8vd3d3Lm1pY3Jvc29mdC5jb20vcGtpb3BzL2NybC9NaWNyb3NvZnQlMjBTdXBwbHklMjBDaGFpbiUyMFJTQSUyMFJvb3QlMjBDQSUyMDIwMjIuY3JsMHkGCCsGAQUFBwEBBG0wazBpBggrBgEFBQcwAoZdaHR0cDovL3d3dy5taWNyb3NvZnQuY29tL3BraW9wcy9jZXJ0cy9NaWNyb3NvZnQlMjBTdXBwbHklMjBDaGFpbiUyMFJTQSUyMFJvb3QlMjBDQSUyMDIwMjIuY3J0MA0GCSqGSIb3DQEBDAUAA4ICAQBpxEcsFAQ9oJ+nX90jxdm0zaRrpzJVb7sSue3McYCfUjf3mnVTtU/AsWpQAWg6C/z65YH0AEeBV8qlDb+QoGDWjAc5Q9JKHzoIdJxJA07Q0GwJHAA8mPCkG9OJDMATcdXHe5Rr3gyWFWADZfcBEPdExeaQrAlXycwDlRtYpI86tjoqzb/9OKs4hl8FfK8kjm+9XeWCdiwlmPBMl6GdH/otPRkHzxtWuFv2ZPfVsIzsA04/QwhmUmY8OrCeKMTD4rY4aOrmgjR7MghQRfDoDNAueUDs5yYVdfkb5z6u3kpXPP5H/AsGrY5U3hvjmTQjGKvqc9vaSTsb9tHXk5g+6EQRK+OC8UE5K/lE+bhBiExQTCfwJeWgahGxTeXy807rrE3KZUg4j80lnaMt0DNbMSxhlPF9zLzMK8edCPctFwKvfWMOoE9mTf9giYJ3V2g45mQKOZfk93VcUkcNazLd+iiUzFlYTB8NLmu4Sc1Lqgr507Wtip3UEANCuVZCt/KyK3xupM40vubUyWHk3QxwPwaXy5/3kGxDtKzy7hVTCc5ILiHBnNQyvjiNYU8zOt2Fs/JkIMpPy7sqAvurpoGsSxv/0od0ns4p3Zg2ZIskuaI66ccB/6qLeAp1AAwTbV7lOgtiqhl9vKS9CyFLH4Sd44M+ZiGKGlunx26nOprkYnZSxVkFszCCBa8wggOXoAMCAQICEGgo1Ux+XNq9QzmuDMFaKjUwDQYJKoZIhvcNAQEMBQAwXzELMAkGA1UEBhMCVVMxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEwMC4GA1UEAxMnTWljcm9zb2Z0IFN1cHBseSBDaGFpbiBSU0EgUm9vdCBDQSAyMDIyMB4XDTIyMDIxNzAwMTIzNloXDTQ3MDIxNzAwMjEwOVowXzELMAkGA1UEBhMCVVMxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEwMC4GA1UEAxMnTWljcm9zb2Z0IFN1cHBseSBDaGFpbiBSU0EgUm9vdCBDQSAyMDIyMIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAniUBZhkfZDTBnTkjYh1xi1bqJdKbH+8nAYK/d9iUM4MYSJtQnnuWZMLQw19F/zKc6BcXvXjtdZhfOgYIKxw3m0ZKkAqwr0aSPjOJKvq45zJj8yPHbtIU+yZY7v4GhFT6wR83qtvU7FYqv0m9zOsC7cZO/KwZtRI1aRWJF02jaOpsHimaCfPOeiHGCdEZ6o8wRmk7aAQrfIot1mNd6m3WOZ69Bj5b7i8RWyhrp1KkaF5MpOquziO/TDZx2oFFUI7Khs7/U8O4Q7Mk7gd6orT6xwode8ZSNTHsCB+EgJJb+LHaOdbJ5+WJBH5Rf/TmamRHSer47Kb2oENT/trDIyTYJdoTLCq3P5TedxxMeBxq+ZqP62oVd3etSYTOEEDHmUgP1ZYegJxzoTihA2/TTSDQtUPk9y54D073vL9l2m2QC1u/3uonJ5lk+Dl8cz3WIdLu1vNTES5Vw9zq8SlX3lGheHOQCy/1yXU2643SbY55XboaOP/fGQGo0sjR1vLrivUu0cyTE5uckHhlY3kExPGen4w682QM/pgdk+KPVqVjUyO4bnMWRRq293sPzaQy/1r+lo3hh3jbcIOoJIVpIMJtEg3lefYqWc/Wq+eB5qCxiC0IjAuxz9dsNq+e+QNn2UFzqatFuHFgWBjUFixlutEF3pLFUBARkM5HzPuvvyPAnwUCAwEAAaNnMGUwDgYDVR0PAQH/BAQDAgGGMA8GA1UdEwEB/wQFMAMBAf8wHQYDVR0OBBYEFAuzaDuv2q/ucKV22SH3zEQWB9D4MBAGCSsGAQQBgjcVAQQDAgEAMBEGA1UdIAQKMAgwBgYEVR0gADANBgkqhkiG9w0BAQwFAAOCAgEASMc3///BaFfXi0NmRjomay/o+t5ooY9H8T00lXraVTH0ldI4Xyy6j6WNUTFqiVVobCtCZGqFJKBRA8fd0XJY7WwejNiRxedJEZ0ZejdYHndE+8IImELETeObig7PQEVPG4BwWYyTgegP1cgmlan3H3cGuMrvnPvoZtrlOeDS0mLDp9S2GJonmyZQSnKX1bNbKqT9Xy9+5mKjJ2YM+tkZzBEdMagBUgjmVAyZYyvq2ITUtAgW775KW4hY9AYWoOt6XeHHRNa7L1VWJfCeDOQPEtvAf69WXcaJDnGpVhLkuZyoZB61R5WSrtBwyJN9fFpY8QXxSrhschiprh9XmSZ0ZvUdD99d8Oc3W1+68LTv5GMHfh8yGGmpcFqS+XmcWNR+v3JdU0YrbqOZYNaFjGZ3Fnav4sUYW+JdCDbWZjcXZfAuz6HlvOaNDWW0VlNdn8ivTm5Rz4i+kuow+yzndT9CYMRx55efc8efytG4bCPqUCgdDkPM9akbQOummOXlD8WSL6WWx9f6PBjuHRthA/2G5yRBM73Y87ZgfPMcggPVYK/f9CCk5IEGIlrMhTN9ZPjkuL+AF9T7IT9jruePtxdE7HIuNckL0IEd6XIDCUHZ3wlI5s23shxgJRlS8z0SSe2dlCKOcSj4wQdUc904CLSFjxRsqgCvQKu1h862OVxz+ZBmc2lnVHN0oWl0c3RUb2tlbnOBoWN2YWxZFzQwghcwMAMCAQAwghcnBgkqhkiG9w0BBwKgghcYMIIXFAIBAzEPMA0GCWCGSAFlAwQCAQUAMHcGCyqGSIb3DQEJEAEEoGgEZjBkAgEBBglghkgBhv1sBwEwMTANBglghkgBZQMEAgEFAAQgtbgZ8oIWmS5XENkuY1VBNvFYGj1GMTUpNuIyT0vMtCACECZn80XjVP6AQSxxVKEpiNkYDzIwMjQwMjA4MDgxNDMxWqCCEwkwggbCMIIEqqADAgECAhAFRK/zlJ0IOaa/2z9f5WEWMA0GCSqGSIb3DQEBCwUAMGMxCzAJBgNVBAYTAlVTMRcwFQYDVQQKEw5EaWdpQ2VydCwgSW5jLjE7MDkGA1UEAxMyRGlnaUNlcnQgVHJ1c3RlZCBHNCBSU0E0MDk2IFNIQTI1NiBUaW1lU3RhbXBpbmcgQ0EwHhcNMjMwNzE0MDAwMDAwWhcNMzQxMDEzMjM1OTU5WjBIMQswCQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQsIEluYy4xIDAeBgNVBAMTF0RpZ2lDZXJ0IFRpbWVzdGFtcCAyMDIzMIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAo1NFhx2DjlusPlSzI+DPn9fl0uddoQ4J3C9Io5d6OyqcZ9xiFVjBqZMRp82qsmrdECmKHmJjadNYnDVxvzqX65RQjxwg6seaOy+WZuNp52n+W8PWKyAcwZeUtKVQgfLPywemMGjKg0La/H8JJJSkghraarrYO8pd3hkYhftF6g1hbJ3+cV7EBpo88MUueQ8bZlLjyNY+X9pD04T10Mf2SC1eRXWWdf7dEKEbg8G45lKVtUfXeCk5a+B4WZfjRCtK1ZXO7wgX6oJkTf8j48qG7rSkIWRw69XloNpjsy7pBe6q9iT1HbybHLK3X9/w7nZ9MZllR1WdSiQvrCuXvp/k/XtzPjLuUjT71Lvr1KAsNJvj3m5kGQc3AZEPHLVRzapMZoOIaGK7vEEbeBlt5NkP4FhB+9ixLOFRr7StFQYU6mIIE9NpHnxkTZ0P387RXoyqq1AVybPKvNfEO2hEo6U7Qv1zfe7dCv95NBB+plwKWEwAPoVpdceDZNZ1zY8SdlalJPrXxGshuugfNJgvOuprAbD3+yqG7HtSOKmYCaFxsmxxrz64b5bV4RAT/mFHCoz+8LbH1cfebCTwv0KCyqBxPZySkwS0aXAnDU+3tTbRyV8IpHCj7ArxES5k4MsiK8rxKBMhSVF+BmbTO77665E42FEHypS34lCh8zrTioPLQHsCAwEAAaOCAYswggGHMA4GA1UdDwEB/wQEAwIHgDAMBgNVHRMBAf8EAjAAMBYGA1UdJQEB/wQMMAoGCCsGAQUFBwMIMCAGA1UdIAQZMBcwCAYGZ4EMAQQCMAsGCWCGSAGG/WwHATAfBgNVHSMEGDAWgBS6FtltTYUvcyl2mi91jGogj57IbzAdBgNVHQ4EFgQUpbbvE+fvzdBkodVWqWUxo97V40kwWgYDVR0fBFMwUTBPoE2gS4ZJaHR0cDovL2NybDMuZGlnaWNlcnQuY29tL0RpZ2lDZXJ0VHJ1c3RlZEc0UlNBNDA5NlNIQTI1NlRpbWVTdGFtcGluZ0NBLmNybDCBkAYIKwYBBQUHAQEEgYMwgYAwJAYIKwYBBQUHMAGGGGh0dHA6Ly9vY3NwLmRpZ2ljZXJ0LmNvbTBYBggrBgEFBQcwAoZMaHR0cDovL2NhY2VydHMuZGlnaWNlcnQuY29tL0RpZ2lDZXJ0VHJ1c3RlZEc0UlNBNDA5NlNIQTI1NlRpbWVTdGFtcGluZ0NBLmNydDANBgkqhkiG9w0BAQsFAAOCAgEAgRrW3qCptZgXvHCNT4o8aJzYJf/LLOTN6l0ikuyMIgKpuM+AqNnn48XtJoKKcS8Y3U623mzX4WCcK+3tPUiOuGu6fF29wmE3aEl3o+uQqhLXJ4Xzjh6S2sJAOJ9dyKAuJXglnSoFeoQpmLZXeY/bJlYrsPOnvTcM2Jh2T1a5UsK2nTipgedtQVyMadG5K8TGe8+c+njikxp2oml101DkRBK+IA2eqUTQ+OVJdwhaIcW0z5iVGlS6ubzBaRm6zxbygzc0brBBJt3eWpdPM43UjXd9dUWhpVgmagNF3tlQtVCMr1a9TMXhRsUo063nQwBw3syYnhmJA+rUkTfvTVLzyWAhxFZH7doRS4wyw4jmWOK22z75X7BC1o/jF5HRqsBV44a/rCcsQdCaM0qoNtS5cpZ+l3k4SF/Kwtw9Mt911jZnWon49qfH5U81PAC9vpwqbHkB3NpE5jreODsHXjlY9HxzMVWggBHLFAx+rrz+pOt5Zapo1iLKO+uagjVXKBbLafIymrLS2Dq4sUaGa7oX/cR3bBVsrquvczroSUa31X/MtjjA2Owc9bahuEMs305MfR5ocMB3CtQC4Fxguyj/OOVSWtasFyIjTvTs0xf7UGv/B3cfcZdEQcm4RtNsMnxYL2dHZeUbc7aZ+WssBkbvQR7w8F/g29mtkIBEr4AQQYowggauMIIElqADAgECAhAHNje3JFR82Ees/ShmKl5bMA0GCSqGSIb3DQEBCwUAMGIxCzAJBgNVBAYTAlVTMRUwEwYDVQQKEwxEaWdpQ2VydCBJbmMxGTAXBgNVBAsTEHd3dy5kaWdpY2VydC5jb20xITAfBgNVBAMTGERpZ2lDZXJ0IFRydXN0ZWQgUm9vdCBHNDAeFw0yMjAzMjMwMDAwMDBaFw0zNzAzMjIyMzU5NTlaMGMxCzAJBgNVBAYTAlVTMRcwFQYDVQQKEw5EaWdpQ2VydCwgSW5jLjE7MDkGA1UEAxMyRGlnaUNlcnQgVHJ1c3RlZCBHNCBSU0E0MDk2IFNIQTI1NiBUaW1lU3RhbXBpbmcgQ0EwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDGhjUGSbPBPXJJUVXHJQPE8pE3qZdRodbSg9GeTKJtoLDMg/la9hGhRBVCX6SI82j6ffOciQt/nR+eDzMfUBMLJnOWbfhXqAJ9/UO0hNoR8XOxs+4rgISKIhjf69o9xBd/qxkrPkLcZ47qUT3w1lbU5ygt69OxtXXnHwZljZQp09nsad/ZkIdGAHvbREGJ3HxqV3rwN3mfXazL6IRktFLydkf3YYMZ3V+0VAshaG43IbtArF+y3kp9zvU5EmfvDqVjbOSmxR3NNg1c1eYbqMFkdECnwHLFuk4fsbVYTXn+149zk6wsOeKlSNbwsDETqVcplicu9Yemj052FVUmcJgmf6AaRyBD40NjgHt1biclkJg6OBGz9vae5jtb7IHeIhTZgirHkr+g3uM+onP65x9abJTyUpURK1h0QCirc0PO30qhHGs4xSnzyqqWc0Jon7ZGs506o9UD4L/wojzKQtwYSH8UNM/STKvvmz3+DrhkKvp1KCRB7UK/BZxmSVJQ9FHzNklNiyDSLFc1eSuo80VgvCONWPfcYd6T/jnA+bIwpUzX6ZhKWD7TA4j+s4/TXkt2ElGTyYwMO1uKIqjBJgj5FBASA31fI7tk42PgpuE+9sJ0sj8eCXbsq11GdeJgo1gJASgADoRU7s7pXcheMBK9Rp6103a50g5rmQzSM7TNsQIDAQABo4IBXTCCAVkwEgYDVR0TAQH/BAgwBgEB/wIBADAdBgNVHQ4EFgQUuhbZbU2FL3MpdpovdYxqII+eyG8wHwYDVR0jBBgwFoAU7NfjgtJxXWRM3y5nP+e6mK4cD08wDgYDVR0PAQH/BAQDAgGGMBMGA1UdJQQMMAoGCCsGAQUFBwMIMHcGCCsGAQUFBwEBBGswaTAkBggrBgEFBQcwAYYYaHR0cDovL29jc3AuZGlnaWNlcnQuY29tMEEGCCsGAQUFBzAChjVodHRwOi8vY2FjZXJ0cy5kaWdpY2VydC5jb20vRGlnaUNlcnRUcnVzdGVkUm9vdEc0LmNydDBDBgNVHR8EPDA6MDigNqA0hjJodHRwOi8vY3JsMy5kaWdpY2VydC5jb20vRGlnaUNlcnRUcnVzdGVkUm9vdEc0LmNybDAgBgNVHSAEGTAXMAgGBmeBDAEEAjALBglghkgBhv1sBwEwDQYJKoZIhvcNAQELBQADggIBAH1ZjsCTtm+YqUQiAX5m1tghQuGwGC4QTRPPMFPOvxj7x1Bd4ksp+3CKDaopafxpwc8dB+k+YMjYC+VcW9dth/qEICU0MWfNthKWb8RQTGIdDAiCqBa9qVbPFXONASIlzpVpP0d3+3J0FNf/q0+KLHqrhc1DX+1gtqpPkWaeLJ7giqzl/Yy8ZCaHbJK9nXzQcAp876i8dU+6WvepELJd6f8oVInw1YpxdmXazPByoyP6wCeCRK6ZJxurJB4mwbfeKuv2nrF5mYGjVoarCkXJ38SNoOeY+/umnXKvxMfBwWpx2cYTgAnEtp/Nh4cku0+jSbl3ZpHxcpzpSwJSpzd+k1OsOx0ISQ+UzTl63f8lY5knLD0/a6fxZsNBzU+2QJshIUDQtxMkzdwdeDrknq3lNHGS1yZr5Dhzq6YBT70/O3itTK37xJV77QpfMzmHQXh6OOmc4d0j/R0o08f56PGYX/sr2H7yRp11LB4nLCbbbxV7HhmLNriT1ObyF5lZynDwN7+YAN8gFk8n+2BnFqFmut1VwDophrCYoCvtlUG3OtUVmDG0YgkPCr2B2RP+v6TR81fZvAT6gt4y3wSJ8ADNXcL50CN/AAvkdgIm2fBldkKmKYcJRyvmfxqkhQ/8mJb2VVQrH4D6wPIOK+XW+6kvRBVK5xMOHds3OBqhK/bt1nz8MIIFjTCCBHWgAwIBAgIQDpsYjvnQLefv21DiCEAYWjANBgkqhkiG9w0BAQwFADBlMQswCQYDVQQGEwJVUzEVMBMGA1UEChMMRGlnaUNlcnQgSW5jMRkwFwYDVQQLExB3d3cuZGlnaWNlcnQuY29tMSQwIgYDVQQDExtEaWdpQ2VydCBBc3N1cmVkIElEIFJvb3QgQ0EwHhcNMjIwODAxMDAwMDAwWhcNMzExMTA5MjM1OTU5WjBiMQswCQYDVQQGEwJVUzEVMBMGA1UEChMMRGlnaUNlcnQgSW5jMRkwFwYDVQQLExB3d3cuZGlnaWNlcnQuY29tMSEwHwYDVQQDExhEaWdpQ2VydCBUcnVzdGVkIFJvb3QgRzQwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQC/5pBzaN675F1KPDAiMGkz7MKnJS7JIT3yithZwuEppz1Yq3aaza57G4QNxDAf8xukOBbrVsaXbR2rsnnyyhHS5F/WBTxSD1Ifxp4VpX6+n6lXFllVcq9ok3DCsrp1mWpzMpTREEQQLt+C8weE5nQ7bXHiLQwb7iDVySAdYyktzuxeTsiT+CFhmzTrBcZe7FsavOvJz82sNEBfsXpm7nfISKhmV1efVFiODCu3T6cw2Vbuyntd463JT17lNecxy9qTXtyOj4DatpGYQJB5w3jHtrHEtWoYOAMQjdjUN6QuBX2I9YI+EJFwq1WCQTLX2wRzKm6RAXwhTNS8rhsDdV14Ztk6MUSaM0C/CNdaSaTC5qmgZ92kJ7yhTzm1EVgX9yRcRo9k98FpiHaYdj1ZXUJ2h4mXaXpI8OCiEhtmmnTK3kse5w5jrubU75KSOp493ADkRSWJtppEGSt+wJS00mFt6zPZxd9LBADMfRyVw4/3IbKyEbe7f/LVjHAsQWCqsWMYRJUadmJ+9oCw++hkpjPRiQfhvbfmQ6QYuKZ3AeEPlAwhHbJUKSWJbOUOUlFHdL4mrLZBdd56rF+NP8m800ERElvlEFDrMcXKchYiCd98THU/Y+whX8QgUWtvsauGi0/C1kVfnSD8oR7FwI+isX4KJpn15GkvmB0t9dmpsh3lGwIDAQABo4IBOjCCATYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQU7NfjgtJxXWRM3y5nP+e6mK4cD08wHwYDVR0jBBgwFoAUReuir/SSy4IxLVGLp6chnfNtyA8wDgYDVR0PAQH/BAQDAgGGMHkGCCsGAQUFBwEBBG0wazAkBggrBgEFBQcwAYYYaHR0cDovL29jc3AuZGlnaWNlcnQuY29tMEMGCCsGAQUFBzAChjdodHRwOi8vY2FjZXJ0cy5kaWdpY2VydC5jb20vRGlnaUNlcnRBc3N1cmVkSURSb290Q0EuY3J0MEUGA1UdHwQ+MDwwOqA4oDaGNGh0dHA6Ly9jcmwzLmRpZ2ljZXJ0LmNvbS9EaWdpQ2VydEFzc3VyZWRJRFJvb3RDQS5jcmwwEQYDVR0gBAowCDAGBgRVHSAAMA0GCSqGSIb3DQEBDAUAA4IBAQBwoL9DXFXnOF+go3QbPbYW1/e/Vwe9mqyhhyzshV6pGrsi+IcaaVQi7aSId229GhT0E0p6Ly23OO/0/4C5+KH38nLeJLxSA8hO0Cre+i1Wz/n096wwepqLsl7Uz9FDRJtDIeuWcqFItJnLnU+nBgMTdydE1Od/6Fmo8L8vC6bp8jQ87PcDx4eo0kxAGTVGamlUsLihVo7spNU96LHc/RzY9HdaXFSMb++hUD38dglohJ9vytsgjTVgHAIDyyCwrFigDkBjxZgiwbJZ9VVrzyerbHbObyMt9H5xaiNrIv8SuFQtJ37YOtnwtoeW/VvRXKwYw02fc7cBqZ9Xql4o4rmUMYIDdjCCA3ICAQEwdzBjMQswCQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQsIEluYy4xOzA5BgNVBAMTMkRpZ2lDZXJ0IFRydXN0ZWQgRzQgUlNBNDA5NiBTSEEyNTYgVGltZVN0YW1waW5nIENBAhAFRK/zlJ0IOaa/2z9f5WEWMA0GCWCGSAFlAwQCAQUAoIHRMBoGCSqGSIb3DQEJAzENBgsqhkiG9w0BCRABBDAcBgkqhkiG9w0BCQUxDxcNMjQwMjA4MDgxNDMxWjArBgsqhkiG9w0BCRACDDEcMBowGDAWBBRm8CsywsLJD4JdzqqKycZPGZzPQDAvBgkqhkiG9w0BCQQxIgQgtjbcUw8JQ+4BJRYa3i6d02c3ponVYOJR1O+/tvEoOcIwNwYLKoZIhvcNAQkQAi8xKDAmMCQwIgQg0vbkbe10IszR1EBXaEE2b4KK2lWarjMWr00amtQMeCgwDQYJKoZIhvcNAQEBBQAEggIASj+0yT/ruWTV+BtWc7aN6dgh5DBlH8jj8pAijzf8u23eMA8nWRDXHCHEa+VMg9yq+mBSeO8dGY9y7T4CJc3C3BLG9FPrMv0a9h8QAoFBqtGak5HrkZ85VTUUbmDpjWdVBtuiY5Y6Rj979Gm0jJGr7HNtFL/kXBy1C0kmP1iKMsymeF/+0oZw1Cgj3wY+JBqdTiDj8sCyDe2bPLCUl9LOMoVOPzRWjVK8jiZGKty5RbnirqLzXtfu9dk82EkFZyvTVOXg3lJ8kyqBfpgmB7m8TM/RemZ4cHwpaiKP6pB3FjQvzC1DPsiD1N+ngD+Zw2PCRWPTaIJ9VLZkFQA/CepE6WV9qiz7lhzfXK+bG/HC737MZrwDvK2wSF6m0Cb+dRDfBVfleX0FCRKcEa8s/qqstFDCG/gm7Mzrsdns4Mw/0SKVqN1BvX9JJlOqhlzlNahdg9heLB8U4uxnvbasHNcVauiLqAU3GG9l/NkZyWXd14qaYpvNLMpg/6KIs6MxSTttNg1EMgA7ir9cv1zsdA2gyJAlf0SIMjESFEA3/di4RJ512rzQ4sCeEiu0mDW1zUgU7WSjki4QvkWpNnnYASQ6SShPr2n6azwe3rtWE12pTFyY84t+6/0aeIjjTc+0xDos491Kx8R8WqZG/Emc74+QN7s+hRmEXD9txmXz2GQbqDr2WQGAUq8+ncyKpuNTAm/GKfZBOpMH8QVsBUuEstuKIB+vUps3glqhtZ99pu3/Pxq78wHQ6/zZBxEV26sgrZ1QQ4zMAwL79Ch/1eAcIMsN6dJjz3iRprWj+yOkT6wruMU00BJvQTYRsefPJsz0p62rbYBwMWmNWwdvRj2S1+uT8E6VeZRy6NwPABXIga5yuYOK7xephG5Y+LmXTzB7JZVXVuTRmsf2kUHuOmSn6SB0US5YMsEkUxO70HHh+aWbLEALx7ZbuDdPLaflymzkB5EFEQR5BpPLPNABhJRPXalbscliue9zkwH+GkrdwZUbLSj6784QHizhcbjrN83RyBBKnD2cXcXFHtwcpU5TZH0om9JBoUG29oimvDklsR7pfxazB1dePj0lhgrPGkMOscXZMloQtxXYUXrujm2+TDSEayNN2ML/FuduuyA5xNXuzk1/hXk26bB3pjZI45eRJZsGKAHmp6l+m025WywniuyzIC6L3iEMBpLXoF8PKv9uU4OAHKKJ/+EARkV4aWYAAE1NACoAAAAIAAQBEgADAAAAAQABAABREAABAAAAAQEAAABREQAEAAAAAQAAAABREgAEAAAAAQAAAAAAAAAA/9sAQwACAQECAQECAgICAgICAgMFAwMDAwMGBAQDBQcGBwcHBgcHCAkLCQgICggHBwoNCgoLDAwMDAcJDg8NDA4LDAwM/9sAQwECAgIDAwMGAwMGDAgHCAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwM/8AAEQgBgAGAAwEiAAIRAQMRAf/EAB4AAAEEAwEBAQAAAAAAAAAAAAcEBQYIAAMJAgEK/8QAShAAAQMDAgQEBAMFBgUDAwIHAQIDEQQFIQYSAAcxQQgTIlEJMmFxFEKBFSORofAKUrHB0eEWJDNi8UNygjRjkheyGCUmNXOTov/EAB0BAAICAwEBAQAAAAAAAAAAAAUGBAcAAgMBCAn/xAA8EQABAwMCBAMFBgYBBAMAAAABAAIDBAUREiEGMUFREyJhMnGBkaEHFCOxwfAVM0JS0eFyFiRi8TRDgv/aAAwDAQACEQMRAD8A4H09NtZlaQWyRJ/KMT26cb0tq2oWN6QpJInGAJx3mD/LjzCwyNoIjoZwSIMR0nJMHhUwz+JlCEqUFTuSJ2wOhM4EGf5A8YvcLUyFNvKQoNpbLZThJ2mMds5xMxnrjjammV+8CUpCQFBRSsbSQDiZzGMA/wCPChmhIdb2lbRcbITkjaFHdux+UDMf6yLbVfwrrjbfhXNeKI670a7bam+GzJ08p8i4pIeDI+U7VObklflFMlG1fSeMG5ws2HNVIWyEMbEuEhtAQJVIbThSs4GZJkZxPSSVFutNTd7ghNI0p2qqCUttNq3uKJIThIzmQBJA6kA9tNW0k1bxQgK80TOTtE5MgSBmegwIjJ4lvh+54Xnw7859Ia5sKbeu/aPudPdKJq4MJqKdT7KgQhxCyApJMbhjqciOPCt1FqqiNnfdZqEqp3UOLQpK0H0LkDaQQNpAKusdDgTx5bbKU72kuJQopSFNq2CMgAnp1AOc578TjxDeIC++J/nxqnmFqk0SdQ6zrnblcG6GnTTsB50ZKGx8qIAx1J6knJhbCvN8tbgcPlkDdv6AiDJESIImft9OPVim3LPlzo2/aR1pVag1kzpi6WO0Jq7BQrtjtWdRV34hCBSJWjDCvLWpe907BsIycCGsraQ5tB3ISZCUyhKgkzujoTG7qMAHjxT0riqcpHnOBI3uKbTvAgnJIOIKhnHQdONrLKW9ydrZbQN4I3FOU9szmInBET9eMWYPNamWi+ny0oWFYwFbiI6CTO5JI3YmJEYMcOFDaFqrEoBaSdqgSJM/3s98T8pO3qCOPlv/ABDlS5TtpO5agNskFw4BTAMHvJOYERmOLH+KjxBVHjk1fo5Vj5caR5dI0zpqk085T6cpvLTdXGpH4l0CNzy5jaJOACVEDjXBcdLdypcMIx4smzRzQKoBQ2J5MNF6sc9IQhCZKjgJ2wQkTJEeomJAnglctOVGtOeV/boqChqairq2lQxRslyrqPLSVqXgblbUBSiTGASRHSVWfwqXDl3oBOpbtY7oxa6x1dC3cH6Yhj8UgJK2ypQCdwCkyO24GAFZQp17qOgq1P2S5XDTLCWnGjU0rymXglSdriSUkH1IJlMDCyOnzT/uIhAdU8+y2huclQDHQnDRzPT5pAzqjT/LTSt9tLdgtN8vt2NOmmu9Qp41VpdbUpbpp9iocU7OwlY9KUgpAUTxD3bDWXVhutvtQaOmQmWmHCAsY2wEgbUKJR7SYJMYlydqqHSr6gyyHKlTeHHEhbpJ6EJMmTP0iJ7glirkVV/qvNW6tJWYC1EzI6wR1HUykfrkDiO+WR5wB8OgUiOlhaNUjs/mT6LRdtUt20KZtqPKUpBJf+eqWDiCuJiPykRj6cR9GmlXSqU68A6+4ZWpaip1RJIlUmIx1JOD2McTOl0s1SHaUbAoztgAoIiMAwf/AJGehM9ndm2/hmYQhunQUEBWwbtogyTAHY5wcdiJ4FVdcyLYnJ9E12fheesIeW6GevNRG16FYoktufg2kK3BKlurjHulMdMHIj29jw70FgpwMArDacwkJCgATugTkZkn+MmOJDT6eLiSt8EiRKSQQn2UojGBmJ7/AK8bBTBK0bFNLJO6CQkHtMGJJgZAHf2wDnrHyeisi3cOUtJggZPc/omlmjW69KtjCQQcqIASSenWIGfaB1OOFNNRIUryxuU6QQVJEg5AGSMA5k/6cL6WhFepCZdrJXtQE7giCQBhIk5MCTOffh6p9Opo2UpuNQWkLI2UrABcBIyITIMjvk4MweIfiNbuSj7IZHZAGE1ULL7lak06lB0kLbLa9uwSBuDnpxkRtBycA8PFJpR2qQaisecWVwrc4grWfTgJ6AAAGCT1AxBI4kto0VWvUwXT0LdspSQhSn8uL6dJOQr2GBnAzDsnQrpfQlbqnnFnJdcJ3SSYjtuxI2g9sHjk6qc7ZuyltpYmfzDkqNt21q3FMBE/KVbt7y1A4BBMDM9CMZgjr7fcVUNpT+9d2gJMqJQM/NMACD1AjH/u4m9HyyLeahUpWUpIJLSGup9SiBJjMyQN2ZiQ40OlKFhgLdq2UuFtUBJLmw9d3sSZJyokkGJnjeKn1nLyskr44hpjCHtBYqmt2bKbaFdCdoySJHQ+oAGDtMEAz1hxouV9W8w2XXvMDZKClAhsH+6CSZ79ATjBzAnzVEyHHW7XQV9UtRUQ6GcmB1hUA9Ae4wJ6QfSNFXevXL7KGGkEgh55C9mdoAAmFmQJBAwczEmaeKhj3eclBZ66skOIxgKFv8urdblDzqlKXEAqaS49ucyTO4JQYA7hRn5faApGmbYwhfk2x+rLajsWphxTgUEypJKyUJAETMkbjIBMif0HL0opAk1Tz3REsJ8tKTtMoMCZkZj2PcyF9Dy1pnVpKmGycICFQSgAQlJgEgmRCQCPlPUmStPU0PJoQ6aluJ8xchqiGnNtNZaJtDw3JHnh1J2gGAYyn1HoY6SZzx8coa95tUsU7AdSFjctYWiQJ+VO8iCTKpAiBPY3WXkzdK8rNsoapllaipZaSQtwjAlUQR6R0Ix0ESOHK2+EK8XVlLtS1VO9AFLISASmPnMEGAQRHuAevBinbSybtG6GTC4M2yFW1+g8+RuqXGnipxJp2ylTm2SFDIUqRtkmEiB1kDhHU2pCH0rKa1pPVC1BIKyDG6JiSSOpPVX1ItNcPC01p1pa6tt/yvNBkVKVlZ6JnIlUScdYGO/EduPh+t1zcWilcDq/Una2tCVpEDCQkgAR1Ix1gSTJVlsiePZQievrIz5noAptX4Yuf8yd4CoKXd8wCrfKMKCYEFUDG6IIl0p6N1LDxAZVsQNgKpS3khZhRB9QM9hMkCeJ5qvk87YVLSrzXElKCsIdCWUlO0JJ+Ug5UIOcmY3cQOvVU6ceQXFuqidrqfQRClQoYkHGY9OJkxxxko3xHHRdYKqCbyzDdPNPVqZRsU2PKb8xfypSQkSreoLMBICspIO7ESrI9uimXTz5LjwS1sb8xJQreoEpSVAKkzuEJSnaVAJJJALSzqNl5lrz0hQQYS4y2VrRHdRmCcGCVKJkfYfBdKF6rU1TlttZAStRJC1n3BBMH6ghOBAIIPHPxnMGFL/hdPIOhC83+gFtJVS4QnclaPPQrcADsBTgTuVJUTuAkbk44Yfx7bdQlBU++wg7UbylCRG5KFAidhUmO4kySolW3h6dubbwSA40HGR09OFgq7iAB0ESSTmdoHEa1JRLQ2tTZSph1K1gqTvcCgDtVHclJ7jpIGJ3eGoaeaXrnYXwjXFuFovGkUB1yoYD29LimlNEq3IWVmSd6CrJIMmSIUQUgAG1/wAMf4hWhfB9S64puYnLyk1kvUFoVQUTrjLSnGnJWT8yYCF7huUnMgnuriptnuiC35Lqwwy635jZHq2gCAdzmFAzORuhEBU44W3SxLvCC+2pzzwkjCyoMkkqGVZBEA5MDeDAMgcWz6HZSbW2tk7CCN1IblcGrzcal5tX4Yu1DikNNna3TFREIiYCgCB0O336kSzlhW6KRZ9S0OrtNXG4uXO1OMWh2jqvJbt1VvTDziNqi42Etx5ZOBJyBgTW6/uUbbjdQ46H2lek7ilVQkH5ljChAIVGDKwQDuJ4mFoLFzQW14cQnaosqS2VkoSYiRPpkykCdpAJzw50NS2pb5TuqxudI6ifh/JMWvdPNsXRuo/fho5b2LEgESQUwNqe/VOUjpBmO1LJ8ko/+oS6oICSna4tQgjESidwBKTKj6RuiOCbW0jyGXaZ5FQp1n0hbidqXNgJS2qBiDIwABJMKk8RW425lFMC5tQiSFLgIM7SgkkpPp2tgZE5BABniNX0RwcIlZ7i1zgCU58tuVeqdW6Hvt8syGa+i0q2h+rXtQ4Utrc8uQlRB2iVDbIBUYUDuIDaKug1bZlvNFtxLHoqKZx1LikM99ywUlaIWkqXBJUCkAAyN7ddX6fttfSU1RV0tFXiKqnTvK9iin0FI6hJk7VAdZyDJg1nuT2ndR1IBYUyVB1TTjW5CkFRGO8jPqSesAH1AcLUjSOafGVDRgKrLISrZucSoKwB0BJAEfQfX6Hrji9fwIdLeHHVfi6LHiVuFHS6V/ZrrtuFc+ult7tcgphFQ42QUNbCrbkJKh168Und05WW610VbU09S3RVwWlh9Tag06pBG8BREKiUzGRjHGuhqPKRBUVoJCUysJ3IJ6KAk5n+BPvBFjnugQ3GFYf4n1n5QaY8cWtKXkRV/j+WdPXE2uoQ4txtYLY8wNqVClNpcKwg4JSBnM8AWludU20GEVLrlO1K1Nq9KFEp2pUUkgExPqIBgAdJ49WOxV+qq+noaSlqK2teO1qlYZUt1z/tCEg7iCCekE5zGElTRqp3nGVtIbfaWUKCkjr0IBxI/iAOnWePXHJyF61uBgr5gKVkDYZ2kj5ZJk9zETJA6jOQOPrjCFLSt5JdOyTuUPlMGDjKiD1x3zGePqArytwTMynCgZBmf1JAInHv24+fgy75yk+ajcduVEAKSEqVJIJwIzgDB6Y41Wy3lSENrRLYKN0ygnqACD+UnAgED5hMjj40ve6y+SqW1eYpRj0ztEmczg+o/SPpJeZ/J7UvI/WYsGq7FW6evKqZm4fgK5sNuhipbD7S9g6JU0tKk9ilU4HRBoi101/1HaaGquNLaWq+qZpl1j6VLZoEKWlCqhSW5UpttOTtBnarBPHhKwKZjxCUo8LCOWStF6TU9/xN/wASJ1R+HJvoQGC3+BDswaaVBZTAO8AnrxB6ChqLg80huUIQPLS4tMGN24dP0jp0mRB4k+uuX9FovmRe7RZ79R6os1srnKaivdOwunYvLKFEIqW23QHUoWmFBK8gKEgzwusNkSEpW4lxxpEQFhUKOIgD3PTsCYg540LnOOGopTUTD55dgvGjdCN1q22UIacClpbUBg5k7U4gCAVdD8pGCRJq0RdRyuVSVlpeeoq2mM079M2WngQqQUg53JiAobYGIn1JH9Jf02WnCaZpt2rCJ3YDKMgqkDA+sYjp9fNbqI0b3n1S3HnypSkJdXjptATIKoPsE9EmIyQaoZWU27Bl/foFBuERqvLL5YxyHUokXnmbqLmJam9Ou3O6V1pVU/imbOxufRU1akBHm+WSf3isJmNygAPYAX6t1LV01SaYU7tArcoN0oSUPMhJAIUoiQN0CEgeoK7weFegucGqeXvMOx6q03cqqxXnTtY3W2usYBbXQPNxtcQgzuKTHzDb3z0OrVGprnrHVFxvl7rai73661DlZWVLh/e1Dq1lS1rKY+ZaiSBiT9OINwuMbDrkOXH9/BE7HYKms8lOzTGOvRM9tsq3ypSm0sJUtSy4lUBJkEK9UdAJgndIntw60lKzSOhbaVuLlJUsY9P3MRJjH2/TVS0j1SpMoKvLMBAAhIk4wDEQegxP1PDwlhq3HzKoh12EkoC5UjBypR6CPoTjHQErNVcppfw2HS399Vatn4XoqH8WUeI/5/IL1QW0oZlKPw7LYMuBO51Ak9AqMn2xMdT2d9M2Fq73qjomBRsvXKoTRoeuDvkU7bqykSt5e0IRMkqIAGJMcRS56pLxLTQDhSTASkhIVkyAclWCZVnqRHb5QWisury5CnlLELgSEzJgnpBkjJzkfcW8MbvzTfTySvO2B2A/VPN8ffp7pUUhLVU7SPLaUttaVs7kqIJSoDaoYJkdeuZMprTb03l5CFioqVgE+RTNlWCcndnp75gfTPEn0nydqbwFOPtF1TKd/lIEhInBMfljvAHTMxM6tnLgtN+Q2hnytqVLRO0EFQOUIKirJ7qHePbjmY3lusjA9f8AakGpiiOjVv8AM/RRK0aYcUPLWGaZLYJUinT5iwDG4FZhLc4mMn744mWmtJmjUTR0rKEPN7vxMALB2/KXFSpSvqmB6fbiVW3QKqOm3PpYbT5nloCUh56okY2IHpBIHaevYZMoouVVyvAUC63Z6dZSUPVKw/UuQSYS2J2CFkDcZABH14FTV9DFvI7V7uXzXV332UYiGkeu5+X+VC3GQwz5wWCXEpQp0J8pG/cYlSyVHviY7YBgKba1VVq1U9tQupWiGgtgOBtBkR6sGOgxgkiAdsCdL03ovlwr8RdK43KtUYUKypK1Ff5gltA9MkCE4PywcCd45mvVVOpuwWGrca/6ZWG00rY6BJCjJIj6ie2M8QH3USkCBmG/L811ZRPYCZ35P77KKUHLGrrktqqm6pUFCYa/doaMCSColWYEQB8w6Azw9UHLmnpdw20TCgU5SDUFZBg+o4nBzJgwAeNdwvl+q0OKcTbaZskifMWsogTG4qCR1j5sSMRPDLc7hdKyWxXklxZCvJSC0nEk7SnpEH0jogGQJI7Nkc/YkD4rGtib6/BTb9mW2gC0KrKYkD0hcgge23APQASZ9RgYke6essbK2wblRgJSoN7ErMGJGf1UcbYgduIExpOuqXUOVT5JCSvy1O7IHRRUEdc4+/c9OFNPpOqbP7tVKzuJLBNSRIjqfQVdzB9jg54lxQU5Pnl+S3Mrx7MaMmjbHpjUSwmtvVtZS56pdKhtSMEQJET7RI6Zjgu6WsugtO04dYq7ZXKWAUCmgboJmOpkGepmR3nioCdI11SwEpuYfbwAhNQAnfOT2AjHZQz3yeG6psPkqUlC/Ic7KZqFbiR12wQNoxjAEjrJHBWntkDjmOYrhI6V3tN2VyNR+Ke3aaBZprfbggApG50AnbAMgdZj+U/Xgb618VLmoQUtOmkB9IAn0me5yYx3iO/1rk9cK+jQECreUW0eYlIeUpSQkbkxGRIJjdPQZgYQp1umld8ioShCEkqKgsuBMpEqEplRGOoUYB7DDLSB1OPKcqPJTwP2e1Ei+81K91TynEIUtzcFnapQB7Y6EHPbKsRwz1HNV3cUllaUOKUpQ3KlRAAAg5IgTtAz9ZJ4Y6S6t1ZG9R8tSUqUpLmAowAOhTgj1dQADkHjWKFis/duH0q9UISV7UeknMyJ9cYPf6AFReKiPBPJQJrBSTDDTupzauZzt42pqQthIKWQFrK0pIwQoiVR6QduSQU5PCa8aH/byEI8pDqXk7liVKEqKkgewJAgRAGDiSBmgKLTNu5e6ht1ys9bW6kqXKZdlujNeG6eiQFq/EB1qD5hWlMJ/ulJxmeJXoKjS5XooaxtUrIUyogTugBQCSYjt0kYmCQOD9HcmVAAPVJV0sM9Ll7OQQjvnLe6WBa1hlTjZMJKkgpa6CYSSAO8xkgfKQOInqKwvU4adbbWh5lYQ2EJI3yeqlHcQcntMfX1G3l8at7VCmlcp2WnmEiS7hPsIBBGCDmO5mTB4F+t9BU9wbcLZ8hodF7e4VEjMgTOZTOZmABLmtrXeZiXHXeWLyHbCAVf+/p/MZVuUVQptDcEE9JgFJMx1EY6AyOPDVM5d2ko8k1AUoMnH4iVbySZGSokkx1zOCY4m995UuwssL8t9e8nIASvaREqO04URMgg5mThjunL26aTWzV1FDVU9JVJK2lOMnynEoISqJAlMhQgHE5JieBE9A8cgitLxHFIzw5Xc1G7houronAlXmMu0z8pcUlaA0CdkZSRAASSZlISoAnHBc5DcmblzOU1T0jDrymQViEuK2FO3aACNqkztMEfQ54nPIvw/wBZzopEsW6ldqapgraUlpkykgg+YpM4MkAZIj+JtD8P6tpPAl4i6N3VlA4advzCWHkbTvM5g+kEQrb2MduPI7VO9viuadPdArlUxxsdJCQSqSc+PCNdeXVNUViqR2kS0sJqGlpU25S7p2QDlUBM+3pyJIkMaavNVaakhZLJZPljakgoVIj2JAEZTAgAeoxx0t+MD4tdLeIzmm1dLFRJom22m6Z1p5fluVK5j1pR1klIwT0H145tavs4RfatbRSX0tytyQsqbAKR0EgGRP0I9G7HHtLL9zqcA9shK91oTX0Ilc3BR2qeUl2oeSdj1y+3TfsDU9ZUW2hcarkuLQ8wE7g40DKAJQQcAwM/lIvvdtVSvuJBCgo+WlGFKxgIBCSBBUev/biCOPHKu/Oqo6hl5biUEHzSUHf0T7YAyDug4Ik5jh11NStqXUtKTKVkqKULB2naRMgfbMdiYERxYxayenDlSkTpaWqLHnkVH7NVpYuAQHUtE7mlOOGPLWVdYB65T6hmFE9AE8NvOPkVqCxcpaTmGikpxpM3T9g+Yl9LjqH1t+aELbJCtxCHDJEEglIECNrla7R3ZtbUuKCkk7AFgplRJHypQjcQCJgkZHzHj1qQp1HpN6gR5raS835jaTu8xwgHfs/KrJQkbQR6j2PCNXQhhLVa9uqTLEDnkq4aq8UWtOY/InRnLK7XyorNG6DdrKnT9uU20G6FypWFVCypI3LUqCJcJhOAIPEERUqWhtSVhuTKilUDBAnMdSBj69gOE9PSKShtwqb2qMkHqARicdJA6HEn6nhbb2UpqFlbZWoLQlQCsznI+mBgnv8AXhaJyu4GOSJng18W2qfA34jNPc0NFKtKtT6YLxo0XOlFXSFTjSmylTaiCYCgPSQpJSCkgieIjzE5h1vOHmFftVXkJevF9uD11r3GW0stuVFQ8t10oQnCQVLwlPRI9hAj7TZqKZHlILq1J2JgKyI9RB94nr2H3ixHKn4ZXNvnV4OtY88tN6a/aXLrQFQ7SXer85tp4oQG1urbbnc4GkKQVkTtmZPTjwDKwnCryw0HSinQpKglsAq3AhU5UB1G337yPoeNiKZTziCtLzyOu1YzmQMdZJmBPf3k8bHEkupR5qvMYWImTtUUjOCTOP1gdJx6dZ8wy/8AuS6lKj5gKVEESDke+T7gmJ7YsSysra3V1Y6/Vvv17qkwuoq3S44oISEJlSiTtSlITGdohIwI4c6CjZpaqAp1+rXCAEJlZO0QkADH5oz9Zzw1jz31oaTuBUCodA2kREe394SZ/meJhoHRjlVWkNFdQp9flNPBtQS8d0pCSQPVmfp+ZMSRq4KbTjfDV9abTQpClqbeqVDaNhG1CCMCB+bI9OM4MiYeaZurcbWtx5bZUnKSfS2nAACo6T1jCsATPBv5xfDj5u+HXlRZteap0fdrNp28pBo7hUM7WxuAKJJ+QGMYBVIE9OBFoLWTOhNeWq7G1W68NWWuZq1UVa159JcEtqSotVAkFSFhJSoJBGwxOQRuHBo32CntBkGIzk8vimq4XVu1JcVQpddIBS2tYS0qBt3GASpCcjOBMgEmRw3t0ZQFKqVuPVLytx9KQVEDBySVZHfsme/C7Wmr29a67u91oLZb7WLrWO1CaSlQRRW5LjhIZaSpSoaQFQgEnAA492izraPoc8tYMrdWPWgHAzMx2gdMe+B9XcSweVM1k4ZE0w8cavT/AGvlO2UwncQpyIbCoKgAcqPt1gGBjsBlwoLMlKC+75acDYlXqJ/9o/MSe+QM8bqWmaogA3tST1Uo+t0YH2SR9M4PQ8THk7yg1Bzm1O3S2CzV18dbW27Vttfuwwz5qW/Ncz6WzuA3E4BPtPC7LK4+c9eqtSKnp6ZmkkANHIbBRKtuYtiXGk+Y3vO0AuFKlnAzH+EgA+3dvaoqq+ugpGxlAMLUNqQRjI+wPUni4Pis+FZdeTqHL3o9ys1FYFPluoplNbqu3jO1S1pPrY/KVgJKQPV808C/SvLu12F5O8m41iSnzUobJRTr/wC0dSSPzEznAxxyY8HfO3cryKugnj1R7g8gOZ9/YIeaO5UvXctw044nCkkiNxjEf6kiOn3KukuTSGtodipWHAhKG07khX5RtAxJJ6GR2nrxL9L6fuN5bLVNRIaLkJcWr0jckkgpPzK9UAxJGfUI4n1s0Hb7OkG4PVlxeXDrVBRgpNUkemVrUdwbIBJmO5gCeBFZxBS0x8nnd++iJQ2+sqBpcBG305/NRSw6Odvj6qemS6psDzixTEEpn8xWfSnEzv2zuEDqOJKrRNt0VTvLr62nYKAElqleCnFKgbdzsQIJQSEJ/KfaeJZbdM3zV7H4G309NbrY0ZTT0DRCXEKxlzqqAQSUz8sSDMP1h8MlJQu+bc3ad6qWIKnxIIV0Vt3GDgmVqMz9xwk3PieSodiokwOwRiitlLS+wMnuhczqLctw6dtFUopQpAcy2FDcTlZlS8AzMTuyIzw+0HKjUurnQ7cLkqholqCiwykpFQhWAACd6iAUnOFEZCuxitejLLZyip3t1amgmXXwUM08dAlJyU4OBjGcmQ9LulNb0KdbNLTobJC6h5shTmIgDBjESo9cd8r096I/kN+J5ojoLx5QhRpvw7UVkWsUFqLy1IIU++lJedk4mcJzuI7kK/huvXLQ0aCqtdZKtgCdgS6omJGFQkCI+VI+buAeCkzcF3LezQMVtw8knaUU8JUOm3smIMHqMkGeHBnS9bQpUtyno6RtsFe0uJU6omFbiYMEdcDv3jge68VOcyOXjadjDy3QGXyzqbvUlVBZ1LdSA35y2h5aR0O5S9qZEwNoO2B068bKPw8Vcf8APahapEpAAabk7AkCQB6QfboRkZkcGq8WOsvDYapA/VEplKWMCffpAIG7MJjpkSQP9fcsdQ291zzmKSjQEpkurndgQr2MiCPT3nuOJdNdpX7B2lSGwtdtsE1WzkhpZloLXUGsUj8x2qhQJgjMSciZkwB1HC53lfp9tLpSxO+FgthtJIifyjrJnsPtGIXdbddaOoT/AM4wtwpKQVpdkKwkJKpHXpB9z0meG1l65VLqUNugKSQlLagtGyesbsAAGSRESnME8T2id24lJC6/dhyKfNZcsKByjW/TwpyJCSoSFR6QYnHQ49h9QQzrrTn7GacBQwpsp2oVEQnoenQifoR2MEknGgrnUkCtW4Gyn1OOqKkoIxuEE9fsT94HHjVXLWh1wyWgrcpUoKkqkoxBV8og9sR1M9DwQob1NTPAe4kLm+lj6hVJul6cpalQDi2ticBThSFKEAGSBjHQd8Y4Y6u8v1TyQ4QtWOu0FZAwIEkYj/CenBW5nci6zTq3D5Ydp+iXSjy1tpmNoJEQI6E4HtwM+ZNmtdDq+5LsFPd6XT4XNG3cnULq207Ukh0ohJO4Kx7Dix6K6GePUyTIUGSliDgA1JKDUlVbKplcuqStW/blKkmSn0zJ3GO3tHY8FDlq89zBp009KXqyoQkvqbSgb2xHYEEAAdenQEdOBHY6dbq0spSAhSPm2kJUqMAp74xnA+3BZ5RcwLjyXvLF9tzTSq1ttQdC6ZKkOT2OTI29egBgD6NVmnZJI2OqdhnU+iA3mmnigfNRtBeBs07AlPdJdH6GoCVoc8tBle0kAHpEHsFQSSqMj7cE3RWuKDU1OF1KG2q4He4pKSkOZEEAHpJOIg5/QB33mSrV+pauvfZha1HenCQg9jkgdlSZ+kHhzsup3KWubcQ8Q4FQFob+XBMgDqo7hkAwEkzjgu1zIpiYTludvcoLS+aACcYJAyPXqFZHVa2qhht9DmzzEwlUpJQD6SkzJyDEfXvw209GtxpLLqXHEOKSsb25QgEAbhBG0gYx7+4Ewex6se1DaU04UW3G/W3uG4KkEGOmPT2Hv1zMr5Ra1p699NtuLbb7bihtKVb/ACifckbcwO2ZBntw322t14aeSrbim0DQZmc19es7TPnE/vErAW1uJUGnEj07vVBG0wZnak4Bzw7ahv8AceaXL21WjUF0uFw03o1CqO2sOKT5dsD7kqSkH1BK1okROAcpiDOrty7apKFa1O7miCrzIEAQdoVjsfuMdAeog5hsVFlvaXG2VENhRgtDeQRknA2wDO4gDpIPDTLSZjEgGVSM1w8OfQ47o7fDo552zwv8wnn61Ka+ndSlDSSAFGQYgfTPfse3Cj4i/ihtXOHXbFfb6b8KhKdqhAbUpO7IA6lXQYOJiOpFdqYrojvSCslIBaQnzSRJg7cExMwqMp7RmIc5L7VXuyU6hLpZp1FRA3B6MlJ2gzk9zAKI6K2gVV1b44PB07nbPojlNViQblaubLBu9ubqGlOICVKR5jbiSoBIUkLlGSJT8ySBJVG6AQIbvVPPUdKtbe2pYCXG9qtn7xJIxHqSZPeDjsVYO+mtK3Gs0fQpvVBcaOjuyPMtlxqaZSW6ptHpc8pRML2qJlQ3RAk7lhKQ7zT08dJ3StblJcpHgttYmV7idxSlQAB3SATk/aRwrXCkwBUN5Hn70VtVzbK40zuab9M1oermKtCFoc/6dSEJKPKJ3H0JECTEx0CiBOczHWtLU0TLDtU09TuujzWC6VNyDEEThQO4+oAgboGdpI1UtVtuHmNhltSFkbN37pB+SQAVSoR23Yx2gFnmTzkv3N7Qen13y5GvRpi3NWShDoQHqejaJUGSlAG4BThhRkkDr6Y4ZeHq/VCYnKu+MbToqmzxjY80N77tr6CXUkqQ2CmUiUknAnEmMzntAiVHXbXx5b7a4dpnyFuEIcIlJ9XYx6VK7gACSlSZPCpbIcDW0oQEon1SsNAFKyrpuSM5GTgGZ4aKtTVnrGXXAoJfJ/dhuVIJUCpHWd6imPUSMzMDiPdoyCHKTYpsN0lVasNhe1GqlYaQ47U1bqW2QlJU6VJj0R+YkREj8wgz1eeZHKi/8or4uz6psV605eW9rzlDc6JdI8ErTLSti0gkKTkH5epB68b+SfMa5ckuZWnNYWd5K7vpW6MXOhKkB5CX2XUrRKchSSUkGRJAgd4L/wASv4kGufid87WNf8wKawW+5Ulup7YzTWemcYp2Wm/XHrUpZUpxRJKiY3CAAOEpMe6r28nc+6jepCm0kLS5CCCfftiQMRJPtwZuSfi15o6a5S3bk1ZuYN003y01/Vss3q3OOgUC0KWlJW8QgqQgAhTgQYUlJkKiOA44goW6d6QUEqhUpChIyZOO3aTOY4ymSplxCfMJbcWuF7PVEkpUmcIJgd5jrgce5XuB1T3rmztaT1Nc7Y3cLbd/2bWO0qa+hUXKSvSlZQHmlGCtDgRuSoonaoGOg4QJZU47t3gJUghG4hIj04AHfrMR9T3410sLAca2wJW4pUdQPZJ6GT0+okzxY3xI665Ocy+XnKum5W6BvGkb1pywJptbV1ZWJebvlx3JIqGEpUdiI3dkfMkFv0bjqdhldI2F7w1o5oFWKxJqXyspUW0D39SfaD1InqB1j68S7S+uqzReq7dcKdKVO26pbqUsqPoltW4JMYCZ2giZgd+GW4X1FKXGWlEMNIMrSdqRu/KCB8o9x1IEEdeNthtjlWPxFUC02YPkKASJmUqXJz2wZ/WZ45FzQCSUapoHOeIIW5J5rp78Qf8AtCdx8dfgioOWb2jmdO1DjjBvFybqt7dSWQAnyRA2BRGZMwSAB145q01ucvaEpbaU3TI3EnJW4Pm2z7ZnuTHTh0tNjeuS0lxtW8JPqJytOeojAj6gmOwxxvq69ptXk0aZUjLj5UDtHzECIH3iZzGAYFVFY+TZg2HVPto4ep6Nuag4J6dT/pa6OzMUikNJaKtoO3arKBJx09JyQTP29yueeCQaen2peMlIII3E4Kj2H3PuI7njQw0WKVCUqWFqWNxSQVbkiAkZx1z0k/QHjegpp0KZSgPVDp9SU/IgCMlXWYTknrJ/UTI4A6n7lPFLESPDpxpb1PVeyAKjaD57klSCgHcCCDCR2yfc9vaeOhvwaOV34TQertTuMJSLpcGLfTL2ypKGGytyDAkb3QDjqgfc0k5Eck7xzw5j23TNmNO5drutSGEur2NrKGy4sqVEYSgkT3A9+O1fJrk/Z+T+gbZpyxU34a12lnymEE7jJMrUpUepSlEkqP8AIQBDmmL+aE8TyRUlIKVg8z8H4DqfeU92q1+Vn5SBhQOR/X04h+uvBJy95nXN6vrtPsMXKrcL7tTRrUypxcj1LSPQs9zuSZJ4LFstEQQnJ6RiOH+isxaSmEySPaI/24jPYHDS5IUFVNA/XC4tPoqg374a9xonyrT+oaBTClSfxtMpL6Ujp60gpOcZnqYjhpR4KLvoF8OXmnfuaVKK0MWlhb6KpSgDtUojcsk59QxBxEDi9FDb5QkJBKVGQY/jw6U9tIwNyU7cgHEe3AGs4ehmaQwlpPUJrpuOLhGAJSHAfD8lRK6aS1Zp+1MOI09cbHaCAEufhl+YsEdFr6JAgH1EdMcJbToKpr6kD8MtbzagnzFOCEEACZUUgyZHQn1fbi/tXpdi60D9JVspqKeoaUy+2uSl1BBBB+/+/XPAU1d4TK7TrS1aVZo6ppCg40mseU5UJ9wkKwsjAkkE49uE288LTUzA+kGrv3TdZONIJyWVQDHdDvhA5nku8+DVVN1orfSNkOFanAoYGdogCPv79M8NqnOX+k60jz6zU9YTtAWpamlY7AYj0n2iPboYVeD28aieDuqK51tidxQHEq2T12pRICumCZH6Ykdh8P1h0mhTFDanFKAh1bqMmI9RKsn69OuI4Sah01P/APIDgfdj5lMv8apyNLJNX/HkgjQa6cuLATS2k0LAMJKleQFbQREDP+Jzg4B4S1epqK3r3uVVpZKDErVvUiAAQSowIjMiPtA4sHfOSFsv1OWLhUGibWn/AKTCRviI+b8uO2OgMGOGD/8Agw5b3bLi7gl9aRDv4hwKVGR3jBHSB3gZIMeGeBxzJ5fqfqVqb1SsG4PyVfLjzZoqB+V3aocSkb1BpHpGAQcflIgHPfocw8ac5waPuqVU1+bubS1p3qWSCW0j3HUDr2AiOszwRdXfDns1MhVRpm9KfbC1H8NWEHbJKvSomR1z1EcBzWvhKu+lt7iKMlTBDhOxO76kJOFGc4GZg8Eo30Dhp1EKRDcKWoH4bt/kVKta8g9N68sy67S+oqdwqRuU2AjAEGCjqIG3BIj37cV05haAqNJVLia5hhKmhtbW0Skox85IgwR1IBwOh68O1z0rcrBXOKoK6ooqtJkpU75clOYM4GCQIicfQ8R6/wDNu+W6GrykVrSRG9afKdSJwe4nJ9JkCR9uDVvp5mn8J+tvY80SgDmDBdq/NNNkvlTbKpLK1OVFOo7UKdSQsn5QCoHIzHZUqk5jh8qLrUWSrDKtqadtJBMqScHb6REfmEbYJntg8Qq/UlPdnRcLa4UrGNo2gq9UKCgPoOkdes9iTY9OPcytIU9Q0giqZTsUNm0pxk/5de4yOD4oHykFjc91lRVxxbyHAWV1xa1BQ7XlF1hfyArKnEiDHUAK75/7T+oc11yiYu77pp4plAFf7vKVAkR3Ek+5/wC7qBkzI0nddNIUy+y60pfpncpLaCB0HYwIPQxJmIniLuU7i9UN0ppXKvz1JbbQE/vC5JAAzOen16z34yFtRQy45ei5wzxStOgghAGn0KnT1/aU4xtSh8r8tBUVKGfTMBJnPqkqMGBnJQ50an0rcdA0LdsaaYr9sPhBIJxMRHYfxMdOPnOfQlZpCrqqV5ioaXTK8txDjcKZ9Q3NqThUgnKe3Q9+BLdbgp6nXucgIOwkkbkpIO4be8RmBAHt04f7Pe3+CRjOVpJStk0vzsord2VIqFPtuJ3QAoTIUoGSk5/mT/sr07qFbitqt6vMbgEE/MD1nKSDAHXp3HTj5cHUu0ziVktuIVtQqSpBKcCB1n9ScgCTBEXVVOWe5oKlpSkOgfMQFHbG77pyJHQkyZzw2UlSDhB62EtycIr6Q1YLZem1JcDKCEkp7kFI/ugSnKJ6EkiRmOCfp1tylvSPLLhbfh1pCSCDJIMicHMGe8/Wa82+7KqJKCElhZKJBUlUnaowARPqBkCMiAVHBl5La7/HKp6N0uKbCv3RwpSSchJM9SIkkAZMGccNNtqBE8avgku80v3iBxbzwrV6PvbV/wBKOU6iwytLCijITuISmJA9Uz7TOSDwOuY2mlVda+gstKQClxK3mk+iT6UqOBGZkwRjIOOCHyY1jbeV2ubfeLvYEaktLG9FRa1uqaTVJUgwlSo9MEz2JIP6N1XamrvVF/yy21VSsBJINNOQQTmAD83QwInvbVtnEzNB5L5S4oon085lAQqs1mbqmFNL89kKbMJUSPOhcjd0I7ExORnpPA81rS/h3ENupC8khJVsDW0EDGQPSJSSDO6Jk8Hit0s7ZKlTiUv+dTghSdp83bI9QIgTAGQfbgOc27e0ivW1CilUvoG5K5JIUDk9CYPuSTkDHGt0oGhmcILbLo4vwFGrbzi1ffqTT2mrlfLncbNpBT9DZaCpeK6O3JdXvWhoEApCl7VEKJwOogxGuel6eqb/AL0rdd/EMB1RUSCkJV6kwZASk+kqV6Sn3xxso6cP3be2PxRWlRBUoFUhRMnb8w2ySowITIA7oecjaq6ltT6Ggp9wpbwhMtlSlJBR3C4QM9YJJiSrhRroP+2c0JottTiqa/Khte2irpVJX5a2dqElM7d+B6ogbSUj5YJBSBGd3DlYrspyyuNBxCiztbcB6QqdpCR6UiRsCSCUxPvwnoGdtvC9kU6IbUzKUpabKjuABOdxLgCVGSQEmMyi0+ypj8ZbXE+aUBYTLstrG0gb9sJ3BOAE4ntBI4AWmQxTj1TFf4xNTZ7KYWDStZcNM09d5Ly6MrVvfbXtGDGzeRnO7qI3LQCe3EE5huqYfq0haULAXUSjchtxYdG9Se6lbgBCsAgYJk8F7SHPC9v8qkaLFUtdgYfVXtstUyVueYQEKSFTJG3adgmCMRuIAe18FJuKytSQAtL+9I3tLBUJ6AEpEzMKACjkDPDNcz+GCk+2sax/lKrFTVlOhp0qQlJCk7CoblQD7jvABxAyeuON7TKfJTvB3gBLbaiZiMSIGM4IyTHYicoFOvtNhanZdHmK2K/6ihO1RyQQFZ7QBjtx5UtKqTO8qSkgqCZK4MEAE9DPbrkx24Q01LYw1D+9CMwQpMlW9YPUgdZgYPcEwRgbPL3PoTKVbCUQBuUYlKQkjKhETByZ9p4etO8s7zqpFU3bqCrrSwFug0zS3g37rO0EDoO4PrxiZRLtjjdQ4y8ytuoXLat52ISEpASSCCTEEnqPl9+M6ZWNGTgL1aLYbnc0rSUiIdU4FApKSpUjGAFZj6jh3qbs20W6JshpCZUtUgKECSonPsD3mYGeE7zabJRfhmStDygC4fUpRJwN2IBzMH6HhXZ7Om2ufvQH6lR9LIIKWfYqgxu6HqYEj6ccJJA0ZKOUFI97vDh5nmegC3WWylavxNQFBLUBpgblbSBAEd1D6+0mY4mmndOrqgpRKUBhPmKPzJSeoMifUTPyz1kkCSNOjNMv6hvLTbSPxNWrLaAAW2+pEzCcA9/aCPabasqaa0WxVpt7rK0tyurqVE+Wrv5fuZ6FO0Se3EOGF1Qdb9mhOIkhtLGxRNzI75n39lFLjeFv0aqOlQtKHZU6tO390MEweyhPQTAjoRxopLcG07oStQIC90QgY6npP8z6unb2mobcaDoLjVI2r0lXzPqJyEgYPRWAPYdhC+22Z+4pQt1PoaSEIaWZDY6iYwVZnaPvMY4G19XgaWnACarFaHSSCeYZkd07LwilIeQGwHql4QiJlKTIxPRPaSBI7Ykuti0w5WutoUdwyjYlO7cRnalIkwcRI+vXh50zo5dY/wCTTbXHVr/e1a8t9wSo9FGTO0SYBJPQcEPS+lGrS6KWhaU9cFK8pVQ4klW4ylKUIwdpBEjGAeuQVuerJOlnNWJFDHTN1zHkrCfCQ5G1VXzsqtRrpkooLBbXGg44lJWqoqP3aSDMpISHQR1ycAHjpnarOEg7khPeQOn04EXgV8Ob/Inkyigr2FNXq5VK6+4blpW4hRAS22Sk7ZSgR6cAqP1HFiLRYiEhJwD2mAeNmNIHm3JVN8R3JtbWulZ7I2HwSe12fahJKOp9pPbh6o7XPtB6kHH24X0VnWHUHaVJUIWem3Aj9OHyjsmAIPtPsZ43QBNVNZsSEnd14X01oGzKCrufc/bh6pLNGdu1J9skD/fhfT2ny0+rbuH5gM/bjMLEx01qAPqTuVMRH9R/nwqZtGPUkEHJxH6jh5RbfLB9KQUwnrgfT+vfj2UNszJmDMDH8eMwvQSE1NWUAY+UCNyhJ+/+XCbUWiWb5S+WrzGHGzKXBJg/XOR/pw8O3JpslMoBPtmT9fp/twjf1K0ykAqk46dP14j1VDDUxmGZuWldYp3RuD2bEIcVPKO5W+rZbUadxl9YZ81CVHygZyqTIEex7H68aGLVdLa/tYZtgdRM+acNESDIOMmMA5mDnic1+s0JTPmJzgZgGff+fEXu98becWpLjhSvO0qhIECSB9Y7+54r+5/ZzTPYPuR0nPU9Ecj4inO024SKm1fVWupDV1slMUq6lASkgDuPv7Z6DPCq80Nh17av3SQlU7ipSANqiOhPXqTE/bhurX6a+uop3X/KdUqELB9Kcd0k46dR79OMqLDVWujU9DbK0NepaFSFdfcQR9T3PCweFX0s7WSt1DIyMcx70RZVxysL4zocEJObXhlobzQVVW9+EPlDehxtuXAOmd35R1JkwenFU+dfh6XQpqA20mDJDhBVvSRI9pge5M/YkcXpqtWIcp0h4pWogpUlYwqev/jvPAl510tNX2lPkJSUtKUElSjuyJiepGCO+McM164ZFBH97oHaWtGS3JPVGOFuL53ztpKnzajgHbZc3m7G7pTUS2Er2tOShSFwCgSI3ATOREHBj26dPvg6eF7T/NwXa53lH4iltqENfhZKfMUSYK/qAIOfpxzz5y0zNo1GZTuAWCFDASROP4Ht9cmDwW/CF459ReF+5/jLPWIZUtIbdbUmWnkgHqmOgjB4sj7PbnT6tc7g1zm4BPLJ/fNT/tMtNbW20x0IycgkZwSOoC6EfE88J2luX3Ldq/WWkaoCt3ynKdBCQqRO4E9Iif045WajokVd2Q5uSpxhYV5qfSUbT79ZyDMwZ7cHLxQfE61D4o6JukudTToaaSfLZYSUNpB+YxPviZn+PAA0ddTqCufdKpCkLV64hPpzIj3UTiMDvxt9pNTTSNi8Fwe5o3I/fRCPsvtFwoqV4rgW5OzSc4Czm5aF3a2qfecXWVTgBUtwlxxSpmCo5k9ZPYzJOeK9Uuo3uXuuF19PTUF0NMle1qsY81h0KQpG4oJHQq3BROFBJz04srrghmzxKj5be2EJ3zBBPp6YKo69x0xxWzXxpKHUyTUBLrReStxIO3zUbo7fm7CZzu6Dql8NSuLi1WtU4EOB+9lCLgVU6vLhIKkBxa8wTM9QMKIjp/GOGa6oNUNu5SS4pIKtolKicECe5x365yBwbfHDq3k7rDXNre5M2i+2WyptbCK9m6u+Y5+JKR5hEHI3dwYOCAO4PdqYRJGUmVbxEZkkpBAn9IA6GZ4sqNnhP05z6hB4qptXTCXSW56HYhb6B8voS5Li1oICtwUsDMAlU5PToR8oH04KPh/19RaG13YbtcaJu5UFtr2aurplHaKtttY3NqIwCpJAxOf0gXMsgKbKEq9BUjaQTuTGD9ZEziADM9+JHYqlaaQtwYhS29ypJXAIIVOSAQcAnJPcnhs/oY7slQNDnSRu5HIXTPxteNjQHiy1BYn+XukE6XYoraGqtK2EtB9W+EyECCEgQFfMoEiRBHEU5btoqKZhDikHcvcUKG0IEe/QSOncA56cVR5MVqG0KCgnaluAkKEqWVkAEnE5EiO4zMcWF0FeF3KvoWtzbKlwy4t9zchIWQNxJykDeM9hntxZPD8oY0Hovn3jG2sjaYGDOEQNZaYoFUS3adbXmNnaUncUKIJlW4ZTCswIzHeYrxzQs34itrCkLJaG3cpuVMkkEuJgYMJA9UAZxIgnTmO4rSGqq2yuXC23BVqql0ZqKJwO0r6knbuZUSNwMDMEEH3jgQ69T+KQXkrBcdJSC1KScfKMY6mSTmcgEGG6of4se6p9tP4T9TRhBFTBTqBC1eUvzqZIUPMQGx6oJABz6oxgkznuGXmna/J0u2ypa1qZeSoBaZDoKx8/QKSVd4zHcjiZavtLluudM9ChsWfVtlSkhUeoqGcz9fUPc8R3msy28mqpiA0j8OVFJaC9hJneAkScSCnpCRnpwq3CICN3uRy3yuMzUM7NrOsprFX2OmXTppLlUIdd8xlKn0rbUvYApMrSCFOqUEgg7M5EqakuGkujDqW1BKx5ZQkhQgk7ASpJBG0picSJ2yCR9tdsXdb8pLVMmoQMKbV+8RnqCtO0AFW4kekkJQP7o4U6r04u3VC6d9sML8z9wXlJCnE4grBAURjd06LPsJR4AfEBHRPdQ/MBDuqk3K15VZVOqJQYpyqdpUCpJncNvqEFO8FPsme4EP5oUrpRlxSUwv1bZTuKYKUmJgkbQJkYAJztleiypujefd3Nmo2OJL5AUAspJJJkJ/MqAegBMQoFi5tUaW6ZcRvDjbiWwE+YiVbSBuABIABgEYyVCPSzXEk02eqT7cP+5x0Q/wDCRc+TlqqNdL51WvWd2Yq9LVg0sqw1KWnGr6VfuVPk5LEk7jBAIgggxwIm0JStKQ6HHDElJI3wcmOkE9ExPXHt4pGfNAUVO1BVErGSAPaYxOJ7AfWOPaSHUqjylJIDRO4GCdwhAyOpMdsBXbhFTdtldwP7NL4zfDXyA5A6ysvMx/TmntW1D5d/GXlpDorqJKD+4QVAkbSVenEknA45q/EU5iaH1j4steXnlvTCk0tc7s87aW07g21TqWIUP7oKhIG3rEiegFtb/wCz6QPwAkpKPQjcFwMBJgSdsZHY/XhVb6J+uq9xQj8VA2pmQ0ACIA7kgmZ+pnEnHTt0aAOamUduc6TxM7L5R21VCoJQQ5WvKIBM+gnBjtu6/bb1I4lmldIPO1LNNTIU/WOZMK9LckE+qYn1CP8A3g9p4lHhh5Jo5085NO6QbrWqAX6uYonbi8T/AMsh1zaDu7DIJ6Hp1GDdL4qnw3NP/Cgv+n7TQa3TqZvUdtU+oMtoYq2XMp2hKVGEKPqCjmB0jIhtp/FOqQ4aMpvirYqHEMAzK7lty9feqljyOXGnnaVh2jqLpVg/jNyVfu0zBnoAnEEyZJIBHUw1brd1YLrhdXaGVKKSoQqoc6yR0AMRGYn78eq6rcuzqqioJYpVOeYy2JPmECJUokFR6dDGAB78OVqoA7TpuVcCmlQQmnp9sl1XVIgROOmJmO0QOuNyBAiiGGjZN/D/AA3JI81dWcuO+/T1W60afS8Px9yA2iW2WSYmDIKj7kwYEYBmOnE50foKp1c6l1bSk0jJO6USWxmFQr0TAPzSAYJPXj5y55cuaxuLVZWNvuU6BtQlAIQ2Py7lJ6mZyCAJV3Bggts/8TPMWqjATQtqASWk7V1piQkZMyNkn5cQO3ClXveDh53/ACHr6qxaOohYwiDlyz1PuWmio6VmgX+FCmbeB5DjyJK3HMjbuImAUjEQB1meLNfDk5WW658+bLW6hpA008l5FjoXmwgKfSjch5QOSdqCE7hJKPVAA4FY0zTaOZpkP0yKqudBZprayApAUSUpISZKldD0mAR7AWO5Acu2uVmv7LqXW1U8m+PVbTlPR07almkZ3hUwSJUBggdAoz1HC7PeGRFug4BI3PN2/QdvVcLhRPqKZ7XnzEHAHTbquglh08BtEbwYJIPzH34lFnsgcCUlIAHRUY74njfa6ZpDMgpWhQ3gjOCJBB+3fhzZrG2x+UqAiOGgbqjiO6U26zgQqNnYiZBH24cqenQygEmI6T0HDI/qpmmAIXvP/uyof1/hw21uukMNEeYj0Yge/wB/6zxsGrFMl1rVMgA4KACT3BxwlqdQs0424BTJIkZj3+nA6uXMQSRvG7JP6fbpxH7rzLAblLkkdMwD7/fjbGy11BFSt1q2klJXCTOUn6Rww3TmClswFlM9IGR3x7cCm6cygSuHCYGADk8Ru68zIKti4MzGPSOPCcLM55ItXDmL5atwWUp+UAGD/H24Yq/mOHBPmA/mJT+b2j6ffgQ1/MlInIUZnr/HiO3HmSQcukFQ9R9+PC5ejKMVfzKO6AuZO4g4A4ZbjzKkQXJV+YDgN3DmXBIDoz1B6n78R2480S0iA6TOJmT9emMY/nx5qK20o0V3M8oyHSPLGNqsgz7/AOfDTWc4Kl2nW27WVa2Vq3qQXVbSZmSOhPTMcAu7c0UAD1hHvKv4frxHLnzZTlIe94JPQ8czg9FuAcKwFbzX3A7lwlWIBwQe324i+q+Ya7mwGwv0pWmQFQIg9RBgAdzwCrpziDRVLonvnpA4R6d51UlTqdlqqfbabeKpcUjclsgE5EgdoiczwMvEb5aKWNnMgorYyyOvie/kCFA+f+o/xetLgguBpDS0gDBCIAkAzJjrt9+8TxDrfXf/AMrfddDzoCF7wk5ThXpCoIIyD/8AEe08JdWX46j1VUuugpNWohSFSC2gkTMdgIBxmD9YsH4W/A9W84qe01l/eRbdN17pcrksr/5x1tJhICFAhG+CAsGRJVA7rtupn+EyJg3GPkvoC5XCmoofGqXAN/Pbsq5aPrXHK0r8xSWgtTiiFbYgekkdJgTI/NODI4udyA+HdqJXL+qumonTZrnXoSu3W9YCvKSU7pfj5d0gQCSnM+xMvgz+FJZOXfP2t1Vqa5UV5s9rq1uabtqEbm1BRKkLqdwJJaCglKdytxTuJIwboam5et1ZUpptIKpJAIgmev8AHM8HxahLl1R12x+v+FW3EPHLQWw2s9iXfoP1+S5H8++TmotCWB165Wmop6Z0bfNCQplsD0kqIkA/+6CQZEDil3Oa4JTXKlwJW0VIWXFzt3FWM9cdCZkfQjjvvrDlukIfCmvMS8ktuoWJSpJEEK9wQYg+/Ag1jyH0/darzKzTtiqXUjalT9uYWQANoTJT/dgZ7Ace0FpbSyao3bLWm+0V2jTUxZPcFcK3f3zClIVsWleEEgqQYwY9zCgJGO0kSM8tPltnqlPpUQZCRAJI6jB7icn9OOv/ADD8EHLLWtjYoKzRFnpqajW44yLcg0LiFOABZC2ikmYSYVuTKQY4p1z9+FjqXRYqa3RVX/xNbkK8xNA6Es3JtO6QADDTxT2IKVHPpnq1wVDC8ZUui4tpKhpY7yn1/wAqp6QWal1B6lO4zCkwPmM5nM5yMdJjiWaaoQ4umaKvINSQ2pxS52ElICSVCQUx1kHOeuX7T/hD5i6vr61NDo69KNqaCqoP0qqZxSCraEth2A6tIAhKZhMmOh4M3Kv4c2vNW0VM9dlW7TlFUUxdT5xLjrUqMNKaTDiZGREp2kdd0BnbWwNb5nBQZq6njcXOePmmHUGiWOUl9atlNd6W9qShDqKilSspO8boAVkEAqBBHYjr6uCHoh9NZWI8qQS6BAcDhSmSkkGe5AIKpGI6RxG+YnhM1ByGrqBVxdpbpQVbyWU1VGlS0DbkJO4ApcOfpB+b0xxOOVGl6m/31mhpWX6urqH0tJZQkKO4kbAAT6icR1I2nAkRYFkqmSRCWM5Cpnid7Hzuwc+qcq9tFE3BISCQZQobVT0GSACFZOcAYzJ4bNI64e0fqNi5Io7TXPMsuodZuVMKinUpaSlRWhUDcATGZB2HtHC/nZp658ubm9abxRrt1wpv3JaqU7VsL9jMZJj3gexOYRaroF3Bjc4rZuUNpWpIGDiBgJSnJhI+c4JiGiOqbjDlW9VQmQ+VRm+UDd0uzaAHk07yghS/LTC4BBB3TJIKesZJyR0IPxJvB1pnw0aG5d3yxa9tGsHdc2n8VVUtIlC3KRQQmBhZKkkwgSQoqQYBIgibmdranNUVoRT0qA15KlsoAyBCjO4blZ3HYo/QADgX6quTl18tx0+agoSCWxvwT2UNoyPTGe8TuEBLnUh+Qw+9dqG2OifmRK/B5zNsfKXxD6Z1DqKzU2pbHZa1D9ZROpJTUttuJVtJJ2gKHqSI2jYDCSqOCp8VbxJ6K8W3ifOpuWumEaUtNYyhkUaUNMuPuAet0pRKQs9gAZ3JjG6a62qndXWMtrThpsLUAhKUnopslMDaQnYB77TAyTw5Xy3rbQumUlSJ2LSNiWiQoBWBGUmRMn0hJjHAOmo99fVEa2vAaYwt2lFt0mkF1LBStO3YVhG4LUASZ943ScnB6CAQ0cwqn9pW0Ogne0o7lKWgpKQQoKBUTnclRAiMx7Hh/q6lyj02/TJUmoUQh9Xlo3F1UJIicEJMAgRHpwJxEr1VecHada0OI3hKCkH0HcFg+o9CSZHUwIBgQTuB8mhBrYTrLvVCDwy8qRzx57aT0gmp05SPakubVtRU6iq101sp1uApC6l1KklCEiFkg9QgZynhBrHQL+jtfXTTlRV225P2G5P252poXRVUdWth0tqcZXAStCvL3BQGU5AzxHqFhNQWoStYIlYCU7ikqIxJIMk5k9+sdZWxaV26lRTIbK3p2LW0ncBIgNpAMZjMZEqE54QnOxsnikpTI4k8hzSTy0VVwQtxSvyobKAVKSndGBnMwCYPQjMxxJLbZF0G2mKUrqlkOKCQJQIGD/eURB+g+5HGyhtzeicpDC7wv3TvRSAgiT7rBkBJ9QKiemOPbi1Wrc0G1vVbhKlNu5Wsk+olQ64weg9RwTBGaA12l/xRmGQ+GXxjPb3pwtd8esqHWqJ9DRbAUt5vKnFAkglX5SDJET2EgwOHXVOsLrzDu37Qv9yr7s82A2hVVVLddWkKwNyp2gGY7YEATwzW22IbaQ4vaURCCIBcPuB9+8CMf3Y4frfbFXdalur2UtIkF9aUghlBAwEnBUc4OTBzngDcbmXHRHs3srN4W4W0N+8VG7z1PRebRbP2m+9WVm1umZKSRlIWegQBkqBJiPYd+045e6LqNe3lNZV/8tQ0w2NhY3EjBKQB1UEiCQJxiIPHzRehXdZ1qCGzTWplSkq3H1AnBKjgKWcjMyIGAMEporQE223g/IlJT6SKdG7KSRjE9IVmQcmeBbneENR9s8h/aO/vTdUBjhoG0Y54/qPb3L5UIXeG0WKzMeUVIC33lkJab9KQoqVlO1IJAECcR19RH0za2dB21hFG29c73cEhtASDvcKj1QPb1HoBkkz0JbND6c/CoU2zTj8Oh0B5CTCalzqSEkxt6nbugbv04PnLnSFJywZbvVyKXtQV8ONKUd5pkqP5e5WTOMCDknEVzxDeAD4Y3H1cfX0Ryhoi1okeNzyHQD/KU8teWLXI62q1BfG2LrrevSttLIBVT0G8j0JzG6CJUM5EYBHBG0jp1vSFai93tSrtqevUfwlEDu/DpUJEJwJKTG6OxIB4jCLidL1dJdq6ldrb1W+m2W/CnN68+c4kYk/KlPQbQcHIsV4e+RdVohoao1KTcNZV7fmNB8/urWkZhUzJGM5iYnpFd3C4Oa0zTO39Po0dh3Umrljpo8dT8z/gImaF19d9PWa20Wpall25vspJpmGjNIMBPmEfQZUY94GeHWt5iJLJAc75z8vvxBmUi+XGvFH5tbUxNZW7dyFO4hpA95APfA4gmpNcVNlr6imqULp6mmVsW2sFKmlDtnP+szxYnAt5lraVzKh2XtP0KpfiihZBUB0YwHD6+5Fe6czNhUfMAKjhXbhgunM2ZG8qMdjmP9OA9c+ZYQpUOR16nB/r+fEbu3MsBRhyYn6En/GeHjUloNRiuXMpIWoeZ8vUkz3mD/XfhgufM9Klql3AGM9OAzd+ZgEpLpO1Mnt9hxG7rzTk5cA7ACI/n/h9+PCSVsGoy3fmXuUR5pyeyun+3EbufNBJlKVwc9e/1/r24C135qDP73A9zG7iMXbmyEKMPfm98njXJW4wEb7nzRTuUC5GCJ3e+Y4jl25pEqJ82Tu6Dof9+AXeOcCgVISpZHTr8vv9s8R6v5jVFS56Cev5fbj3TlYjfdebkgbXEhIAyVATPcff+fEZufOAbClLwIIIJBIBA/8AHAgqNQ1dcQVLM5k9Y4TP1Li1AqcUVRPWT9+PNgvBlT6782VOqJBBj3P+H68R+s5gVVSowopHsDwO9T83tNaLB/ad6t1KtP5FvAr/AEQmVH+HA21Z46NL2hKhbWLjdXRO0hsMNz/7lZ//AOTxLjpJ5PYYfyXGWphj9t2EfXbzUVGd8Jmcnpxperyj1qego9U7un1+nFONW+OvVN2KkW2lt9pbVMKCS+4AemVemfsngZaq5tak1stRut6uVaFGdi3yGx/8BCf5cEIrFM7+Y4D6qC++QsOYwSfkrzau8TGkNAVK3q66Ww1jagpLVM4F7VT1LaAScY6iI4tr4B/iG6c5uW8N22vdQ/QQy5RPDY8wjokweqY6EdfocccRfN+mPpjh75ecyrxyq1dSXyw1rtvuNEqW3EHBHdKh0Uk9wcHiSOG6ZjD4ezu68qeLq6q0sqXamN5Dsv1U8pec6K1ppYdSUrgJIOQP9f8AXg+aP1/T3elQhawo/UiQZ78cVPh1/EdoeeFibpKp9FDfqNKRWUSl+odB5jZ6qQT+qSYPYnofyr5zpqWEFT0wAJBiZwPsc8ApInxvMb+anxSNe0PZuCra3axMXyiJSgCQZEd+BnrblztBIROSRKf14eNBczBVttodckqiBuwocTsfhtR04Uk7gR6o+ZJ9p45kLpnKrJqPSCmFKPQ5MniG3WzK3GYUPzAjPFndYct/OCihG5MzuTkA/X24EurNFLZWoLbUDJAO3BPXr0/XjxYDhCh6kDh2uAKKDIBlUfXjyEhsHAg9VJPX+PEjutiUytRIIQT3yOGh+gIWRkkGOvHh5L3UVHtWaao9ZWCsttc027T1jfluJUncB/dP0KSAoHrIGeK4WJu9cg+YjNU0tygudidRVU7qCc7SSh5JjbBgnInqO2LUOUaisEH1I+v8eIjzR5QN8wWmXGahdFXUra2UOKTuQpCh6kke05kZx9Tw1cK30UMpinP4bufoe6FXWjM8eWcwqv8AiH5u3vnTra4X/UNZU1VxuLpXUVBGxK1fQfYJAA7HvHAp1NWuWcKU82thflealKyqDuSPLKSRIHQgEmB7dOLC33wWasLo/BO2mqAdCgsVXkqg7QrCkSR83pOACYzHEc5teAbmZqu4MG31FqvC30pZU6agtFiAYSd4yhJ6lMkykBMdLE/6hoD5WSj5oLS26Rpy5pVO9T31166uOpccSuSrcU+WnccBSiQk9QIBk/LIiRwhuDQfp1KVuDlTThBdWkbtsiVKGTkEScfl9I2pBkPNvk5qLk1rlqx6otT9qrqZfn7CWy24kEypLgO1UncAQdoKyJgEcNlLalXCkpkt7nz5u11TRJ2pEQRPqAmT+X1AiMAHpGfEcC3qo9czQHPK2aJ09Uag1GiipqdTrzqQgMMoIcWrcQ2Bt9JKZV8v94GRJPEq52+GzWPhv5k/8OaysVy09eQwKwU9U2lCi0oApdTt9M9iRGAUmTPHzQz7+j3xdaN9dJU09QlVK62opQy4FBYcSsqJEEKVOOoz1HEm8RniQ1p4qeYp1Pr2/PXy/IpW6Np4oQwtplsEJTsSPTElRMdTJMnhhZRuaAen1VZ1NyDpHAHl8kGdZNoZrVUqQ4oObUmYM/MRuQrocTPWZInI4hqqs012YQ6vc6+35rifUVKKSRKt3zCfSVH1SpXsCJhqZpa744sF5DVOr0tp9CFFAJz6oCQYmegiAJPEDrHCmspWqfzUlKVNKXISp4BwLK4BJCyFbSYAA9oJ4E3Ebo3bHYGShbo22olyreKGkMALbWpRgLztJPXckBI7THeMvdNqZKVBdvpmg6oeTTJ2hRkjK4OCB3EGPV3yInT1vn0qWFKCmm0CEAbZnMT0JJg9CEj7kcSLS9sS1bl3CuBCduJV6g2Mwke0nrAk/SeEQS+EfF69FY1LG6pApI9m/wBR936JxbUiy0yqyrK3n1q3NpIILkACVTn9DJHtiONlronCnz6kO71KBKE42qxjHScnBic9eEdmp3dSXVdbUpdRsJCEqSAEiMJiYyD3yJPtmTW63JuDwdP7tpOGxu/KB2/nmO/TgNX1ZY0s6nmrH4ZsjZ5Wz6fw2+z6+vuWy1UDtyV6ihTigAoEEop0bpyegGfc/wA+Jxo3SKtWVrbDQUi20n752pW2f3qpy6uMAxMdkyJyZ4aNMWZd6UostJTTIBQgGf3yyrbHfcB6p7D29zVpfTrWj7KikbR5laSXXNhkbynseoiY9I2mDJzgCMhwIGXdB6qz3PDYy0HA6n0Xtq2ilZbpqVLbNPThLanEtlJUqCIUAJI3EAyTgJ+3Ex0By5RV0qnlOfh6AkIcrIj8QSdwQmI7EkCY6+8HXy20Ku9VTVbUKShDMhKwCC6CQPlBgmdo6zG0iOnBTp6Og0/p9N9uzYTZaKW7fbFZVWvzEEd07gmc+vIzkkRxTUttsQp2u1Tv3d6egUS0H+ISmfGIWbN/8j3W+x0VHpCgpLhV03VR/ZNpAn8UNwhxYidogRIyT3EcT3RlrdpmHr7dAt2r8yGElshDrwAGxAzO3pOc7usSGLlboe5axv69Q6ggVq4TT0illYokdBt6BO1IP93omJji4Phv8P8ATuXOkvdypkKbpcW+lcGGkjq+tMCD12g+844pS73BkOznZPU/oEzVNaymjL3fBJPC74T3bdXI1bqxttV9qpfaYcAUKBs53k9A4R06RtEQBxMNUahq+ceq3dOaUU5S2mnATcrnBJZSMeWgz85yBOJz0E8PGt7xcOZWohpGwOFCFkCsq5grE+qVdgPt798cHLkjygtHLXSrPkUyTTJMoxC7i5OXCTnbPuen24X6SGa4zAuGw5dgO5SHc7x92BqajeR3Idv30XzkhyCoNH2mmdq6UIYalVPTrkpB+bzFyAVLMSZz1gcU8+KPZL1y45xp1PUs25On9Uw1bnabclRcYaQHEvBRy5kHemEFMDBSZ6Sab0i9eXw9WFYClSlsdB9Z9uvFQ/j48htRaj8KdHrG03amp7Jy2dXcbjavw3qqQ9tY/EB7djykqjy9hBDiiTgRffC9kdBTfhsw3n6uPdVHW3l09VqmdnPyC5v3bmkk7j5xhOABn3/r+PEUvnNpFPMOyYIVB4CNZr2srCoblBMxkxP1P14QVF6ef9TjuO/qzwc0jqpek9EUbvzgASUhwkgbQkdv6HfiNV/NR95UIUY+mY4HOoNb2rTCFLuNypKJI6l99KP8TPA+1L4yNHae3JpqipubyZ9NM0Sk/wDyVA/UTxKjppH+w0lcZJo2e24BG2p1RV1i43FPvJMxwkcqnHlfvHCT7SeKsaq8d9zqkrRZ7TS0oPyuVThdUP8A4pgfzPA21R4idY6tCk1N9rG2j/6VMfw6B+iI/wAeCEVlqH7vICgSXmBh8mT9FdjVHMKx6QY33S6UFEEiQH3gkx9E9T+g4GuqfG3pCxBSKH8deHRIHkMltE/+5cY+oB4p86+updUtalLWsyVKMkn9eNfBCKxQN9skofLe5XfywAj1qrx5aguZUi1W2gtrZxveJfcj+SZ/TgZaq546s1nKbhfrg62oyW23PKb/APxRA4ijaCvABJ49qYWlG4pMTBPafb78E4qOGMeRoCGy1k0ntuK8KUoGZyf58YVk8eeM4kKMs6njOM4zjFizjOM4zjFiedEa5uvLvVFJeLNWvUFxoVhbLzRgg9wR3B6EHBHHV/4eXxJaPnLa2rbc3m7fqWibAqaUq9DyRA81qeqD3SZKZ9oPHIfhx0pqmv0VfqW6WurfobhROB1h9lW1bah7f5g4IwccQq6hZUtwdiORU2jrXwOyOXVfp65S8526ttrY+mMTKsH79/6+nFhtA800uoRCxumFJUr5uOIXw9fiPU3NOnp7ReHkW/UtKkF1tJ2t1qRguNz2/vJ6pPuOnRzlZznbraZB84EqgKz1+p4T5oHxPLJBumuGZkrQ9nJXvteoWL1Spykz1B7CPb/Xhq1XoZq4NqcbQMzkD+EjgP8AL/mmQ2g+YrrGff24MOk9bsXWmShUKQrBzn9fpxwXbIQm1foRTC3ISUkHBHUE8D672PyVEEf9u0D+fDl8Rb4oPI/wE25VJrS+PV+sqiiRXUGl7Syai5VbaypKFlRhphslKvU6selMpSvE1C0x8f7w4a/09T1dzqtY6Pr3nfLet9dZF1hYn/1A7TlSVt/UAKnG09eO4pJnN1hpwuJlYDglWSdt5QegmZg9DH+HHluiWophJO8lIKcgkCY+vAkq/ifeHMWqmrjzf0iaety0nc8X05SP3jQb8xvKhhaRjceiSRzh+Jv8Q++K8Ulzc5P8waWm0vcLHRUj9y0xWONmuO3zFeeqdpfbXKUrSlKkoITuOeJFLbpJn6CNPvXKarZGMnddRuaXiZ5b8i1uNav1zpewVLSQpVJVVqfxJJyIZTuXJxAKe89OKbePT4o1Pqe02/S3JzV9tqqG60y/2vcqJuop66nUFjay26oJCAQCSpMrEFKgAQTzR5mc99Wc+dcHUGtb5cdS3lFIih/GVqwX/IbJ2NqKQmdu45OeHDRF0Cq/aFB3bsSUqcVlIUI6RAnBJIwTmeGu2WSGF7ZXnUR06IZLXvkGjp9UaW65bbzzu1dQtSyFvq2lSilMD5pBIxnd2B+XiZaXuy75Y7bSPpo1C3o8qnX5Y3uhR3FRIAKiAO5kSIAJ4gNElnyWVK2ubiA4REkAFRIkH2EqPpP0IJBd5V2hSKNLroceQSSpKnN4JHdJ+5Ewe6vpxY1kpHVFS0AJY4uucdFbnuOxIXrVTZtdC3RqS2hVOgnalzbtUoiQQBkZSMAjMd54YCHGA0FBLTq3Q4lSkOSDOCQD6gNpT7dO44MWoeVlqreR1z1k7rC2s36huzdub06pCjXvtKbKzUAiB5aYIxMkdpHAZ1g6mgWWQnbLinwlKktrR6VE5kbRBAJMn7QRw41jAwEDoqPtlUZsEnc7qFajeP4pp3yAuVKQhIbUPMBP1Akwr0qSmJ6AKTwWPBL8MrXfjw03rzUWjXLamm5f0y6q4LrKgNO1aygrLTeCg7kIiTjbuGJA4Cd+d865BKC8EpRLxcSQdqSVQcf5SJMjG7ibae1rzP8AClQVNtbuOsNAp1na2xU0bRfpDd7e4CtJWgx5jS0KUd3qBAIEEjhMqgXZ0qxKN2ho1KpunrSq6ONo89akFXmLlzATtKSrJ9p+sRkYHE3To2v1Sll1mnU1bET5bykFCFFJkkgCJkSAD39gIa9B0KHrmzSHcpt1SFVCkyRsiI+ij8xTJkdRjjuD4uPE14Rqf4Slp01pi22FOtBQMNUlK3R+XcG6tIAeddWBuKSQT1M4jiunu1Zd25K4qKmfH4cQYXB5GrHbt/lcYWKBtyq/C0+9NG0opJkAknEqPTuenYH24dWqRd2rmLZQFQStQDm1JCW0/UdTjr1I/QEo36v8OyG2m0rqXVny0jJJMnriBmO3Un34KPJvQ/8Awzbxc6gbal1MILkAKJBVM9CIAO0g469RwsBwc8vedh9VekUJhhbFGMHoOwUo0bp1nSjDSadhwO04Hko2jcSTt3EQQe6QCZBEjIjggWLSB8xDLhJfWdzy20hwpJAwoKMJz0GSYg44lWhuTVjHIp3WVVrKz0+omLg3R0+k32XPxj7C0bjWwcEA9J3GQDidpeuXFjpRVGsqUF5KvUy22lRcdA6NCSSZ3ZPUFQPchPGquDbZA6vm3cR5R2+ChSA3CX+H0xw1pw49z7+3qn3TllpdI6dp664FSWlAOUrGE+bAMqUQMJAVBMD2wZKlnL/S9RzHun/Ft/ZcRaaE/h7RT7NocWDG4IP/AKaMZEGSR2PHmzaPrOb2u1UFQshpEP3RxoAt0bKQFBiB3JwexyT24sLy25XjUmrLa0mnbp6ChQkUzCU7UstDCdwEAHrOZOcHHFEXi8uLnTyuzI/6D09eiePDjpohAwYa0bqYcjOUKaekar7iyt2rq2yppsqny2zn1dp/3nvwbNYXpzSWlGaalEXO4thDSU4KEYBP6kgD2+/D9ozlY5abIKmoYXTnKziFuLHfb/c6R9B9ojFfXP6c1c/d6os1ToE0oV/06YwQFlJEkjAAnuTiBwtxcJXetc2pMZDTyH6kJDuHEFI6U63+z+wER+RPJ+n0haFNVe1bxSl25uiZUSJDcD3JMx/HtwfNHWRVa+1WPtoCl+mkZJiBHUj7fy4qfcfGnZtJadZYbstyqKiiQCKdxxJRUvH/AKjq3BkD9CfpngHc2/ic81ajUDlXZrtT6YpkwGaejpWni1tAMFx1ClLBUCT7iE9Bm1OHeGm0mBONhjPcn/A6eqrK81U9a8vB3P7+v5Lq4/W0OkbHVV1fVU1BRUTSqirq6l1LbTLaRKlrUogJSBJkkADilHxzPH/y+5A/Dq5gW9OtbSdS6303U01lpaBCLm5WtOgJcVtQSltpxtS2w+ogJU4FJ3FMcc7vGZ8Qfmd4m6GqtF91C9TadqXQ89Z6BxTNG6sNtoJUmdykEo8wNqJSFOLIHtzt8bl0a07yYujLDbLari41TqCEBO4qWFEmBk+ji2oK6Jz2wxM2OyWHWt7IzNI7cdEGdR+OO5uBSLXaqenR+VdQ4Vq/gmB/M8QHUniT1lqcKS7eaimbV+SlhkD9R6v58QdR3GOPHByOigj3a0IZLXTye08rdV171c6XHnHHnD1UtRUT+pzxq3ZmOPhEcZxJUUlZxk8ZxnGLxfQQONjVKXlhKZJUYECST7cagJPHZz4I3wsbDyT5UWfxHc3LXSXC83ZtVbobT1wbCmaVlEkXSoQrBmCWkqEbfWe3Au83iltdG+urHYY35k9AO5PRTbdb566obTUwy537z7ghL8Pj4BtbeNJ23mJz5aqrLZbghFRadHpdVT3K6IMbXKogb6dkzOxMOKH90Z4rV8XW02LRvitqtMaXt1qs9gsVI01T0VtphT0rS1ISXClIJzuBSVH1EpMyZ47H665jXTxH+bd6y8V9i0tctzlO62tSLtqJrcP3yFDNLRKE7FyHHhJSUIhSuN/xebFbLP4uLkLIltNqDLbCEsiGmXEIR5jQ+qFEgiSZJnM8VpwbxhXXi/SxVPka1hIjA9nce07+8joNhy5qwOJOFILZZY5oxqcXbvPXY7NH9oPXqqp8ZxnGcW+qyWcZxnGcYsWcZxnGATxixZxk44wiOM4xYlljvtXpy6U9dQ1D1JWUiw6y80spW2oGQQe3HSz4evxI0a9cpNN6lfao9StJCWXD6WbikdwOyx1Kf1GJA5j8KKK4P2+safp3nWX2FhxtxtRSttQMhQIyCD0PESso2VDNLufdSqWqfA/U3l1C/SZy05vJqG2jvUptYmMYx29+CXrvxcWfw3cldR671C+8uy6XoF176GSPOqSIShhEwN7iyhA+qp6Djj/8Oz4jatVP0mk9W1TbF7QAilqVkIbuMdvYO+46HqPbg7fFI8TGmWPBHqHSVfdGVah1aKQW+gSSXnA1VNrU8R0DadhEk5JAE8KYo3sqBDIP/SaYqlkkRewrmt4pfEVffFr4hNYcxtS+U1edX3JddUMsrKmqVJAS2wgqzsbbSlA+ieIK2ol0ykBWAPpn/XjR8zipIJVPT7cex6d4/vGT7DPDYxuNguJIxuvlRVKOQopx3PT2H06ceaZYqKxIdK1JSADBlQjqMnHXjRUrB64SBkkD7f0ffjdbWytSRGSk4E9D/l/vxOhbyQioky7ATzUqp0OOKpgtLSjDYdA3SYGf6H+PC7SN8NBeWnVha0AgnGCOmf4/1g8NFTKwlKcmR1GTgwB+nSOHLTlnW9VpUlKVqJhAxBJxGccSsnOy4jnlWDf5g2S6XtpVn/HIo0spbT+LcK3VK2gqH0HYAYyOhAg4cvbszVWdsISVQZM/KowQTAPXB9o6STxXHlByouWp7uinoKZ+seA3lDDaitRAlRIGT7nOBnvxbvwScwXfDfzSterU2KyamNrDqRb7u0t2lf3oKPUIEFM+nr06GZ4s/hOGcAuazJPwVMfaTcIngMdJgD4qO3Lcbe8/uJSCooODCcSeuMZkfT78C7W1wcU+8HipSWygJC9pJVtEp3GTjcOnTd2Hq4L3NO8CoQ/VNsMsJdWpQbaTsbBcUVFCB1gAwJJj08Ae9PfjbkNgSA2N8IlJIyQFEnBgDuB1OZJBjiRxZhnVKXBf4gMp5JmaQ0mpXUPhp1taUtpS4Ey4olRXtPQxEBQ+aBGFcPXNnnXqrnS7bKrWF+uWoamzUFPZLe7WuFw01HTgpYYwR6UjeqBGTJkzwvZ0Clnl1VXdd4tIdZr00JtqlAVh2t7vPKNgCmYBRu3E7ohPQiD6keQWaZpBVuacShxRb3BIISEhQI27kkK+xMCBJ4Spm6G5Ksqmk1nZR/RFmOnLSh5/aupcJWpSkx5ajEEd4TPTB9UmO/qvrXLnXNrcUp0mIClk7ROPuSR/I+3G1Sm2mN7gPloAU2k/+pGJA6YEk+3Qe/HqgonHWzUrBNVVnykJj0mfTt+mNo+3FJVVR4jsDkvtWy2tlJCGAbnclPPK3Ri79qBCnEJGyFrLifQn1RtJOE/rgSJmOD7Y7aw6265UKU1ZLQAt71AeavCtoOMqWZgkYCeoAPEW5Y6SVZLallggPuCHnCcLUQcqKhtAE9+ySe/E6uzNOaJhChutlDucYhSQah0pUreqTIkkkf3gAYAMcRGkOdpdyHNTa+cxx/h+07ktukX1VL69RXFIDlSoClYTgJQRCUAdQcyZmUiAcgA/aS01U6P0UzW1iPOvt7/dUbMlZpkZjb2lMkkjJ+oA4F3ILSyde8yvx9zINptKBVvuAna6RMpIPSIxIwD7iOLa8ldIOa81rR6lujH7nZ5lI1GWacH0Yj80bogf5cVfx5ffOY/6W8/0H6qdYaP7tDrd7RUh5JclqTlZoM1VyTNXVk1FXvPrfJ+VJnp0E/wkA8Wx8KfI4UNtRf7uyE1lw/fNtLRCkDO0qB9gRCT3yfrB+WHKdfMW7t3S6Sm0suhVLTlO41RTiVEYCcY7kz24tDYafyaedigkGEp9zP8AvxA4M4afO/8Aidxbz3YD+ePySZxbxG4g0lO7cnzEfl/lMfMWpDFGSDtIkQex+nFe+ZVbCXfVAJwT1+304N/M+rKEuAH0px0+vFc+Y1atxbkKSczKfyif5ni1FXIQl5g3QKDhgAJnp/rwEeYd12pcAMpj054Kev6ogOKk4z7H78AzmDcCtTkZjqT3PHJ5XVqF+s60OuOmcqJnEHim/wAQvUKW7LZKAFUv1S3znshEf4rHFuNVvlSnDBMyTHFEvH1fxcOaFBQpVKaGj3qHYKcUT/8AtCeCllZqqh6b/ooN3fopj6oErMqPHzj6TPbj5w6pLWcZxnGcYsWEzx6QOvGynoXKp9DbaVrccUEpSlO5SiegAGSTxe3wef2cTxP+LzTVFqJvSlDy+0nXKSW7trGr/ZgcbPVxFPtVULTHQhsA9uOU88cEZlmcGtHMkgAfEr0AuOBzQ1+D34I2fHX42NP6bvSXUaKsLa9Q6peSkwLfTQpTUxAU8vY0P/f36cdoefvMRXPznHV6eNNR02idLULVVeaIjZSqpzuaoLQkAhKULLSnHEj/ANKnIEb+Nnw9PgiK+HHoXWdAxzp0xdr5rpump7hXN6VqF/hqdhTi/IaUatsbFLUlSipOdgEATxKLl8OOrtentS0Vv5/0VHWakuDtyqam56SO5LpbSyhKVorJbS00gJQIVGwkiVE8fPnHPFVuutwZDS10PhxDLQZBvITjJ6eQbjfmrb4FbTUEL5KuJ5e8gHDCcRjcgerjsfRV35p69qub12uNMxWP2/TNteNNdrgxLb1xfSAU0VO7EICJCnVwNghtMFRKOfvxW9MW/UWnrJdrBRU6KDTC1WupVSIKaek3KlLCT0OwlW6CopUYUQTx0o5h/Db5sUOgrVadLag5d6w0pbHTTVDGmrrUWu7KASC4w2Kz9z+IcVBUoveYC4uAFKChTbxaaC1FeLBXctr7o28aVrKKgL7dHXW4UYUoN7UopW1bi6EhJWpbSikhBVmTxO4Mtxt1XFLCdTN8kOByDzc8jmTnYdB9bK4krrXfLJUU8D2iQDIBGDlu4YxpwcY5nqfpy0ODxnG+uoF0NU8ysjeytSDBkSDB40kQBxfS+WiML5xnGDhx07pW4avvtHa7VRVdyuNwdSxTUtKyp56ocUYShCEglSiegA49wvEgbQVnAmOJZyf5Ea08QutKfTmhtL33Vt9qiA3Q2micqnzmJKUA7U/VUD68dRvBB/Zw6DQ2k6PmH4tdQP6MsSgh5rRVsqEIu76VAlH4x8ymlCgFHyk7nSBHpOB0k0FcLRyC0mvSnKDl/ZOT2hEIBOxtTFwu6QDDj0TUufd5SBk4IGUbibj63WiM/wD2PG2kd/UnYe7cpq4f4Pr7rIGxjS09T+g5n8vVcTtS/Ao5j8i+Ttw1xzj1LpDlfbbe1u/Zj1V+1Ly66RIZ8inlCF++9wQNxjEcUivKKZu5PponHnaQLUGVvICHFpnBUkEgEjsCeOjHx1vG7U8w9ZtcurTWqVabOomr2xtq3TBLpIJB67B7BKuhUocc4HFb1EnhgsNZVVdCyqq2aHPGdPYHl8cblCr1RQUlY+mgfrDNie5HPHpnZeOM4zjDwYQpbaR5dM8lxtSkLbIUlSTBSRkEHseLaae0fzT8enhgRdbdpq9apq+WVxat71VRMB96vbqk+mUj1lxoto3ESkpdBMKyqpVOje4AJJ7ACSeP0jfCN8Ko8Kng50lp99jyL3cmDeL0oABf4upSlakE/wD20Btof/4ziSeB1yqGwxh5GT0RO1Me+bDTt1X58dUaVueitR3Cz3igrLVdbW8ulq6KraUy9TPIJCm1oUJSoHqDwgqF7EnOVDMZKozx+jvxzfBy5ZfEErWb5cDWaR12ldKh7UVtbS4utpG1Qpl5lRDbqi1KEOK9SClv5kJKDQHnL/ZY+clHri5J0BqvRWo9Mbql23OXKtcoK/YmCzTvpLZbL6wop3IJbGwqJSCBxypLnA8ZccH1Rioie0YauWyh5iz0hXqHDnaqQ78pBSkAYHQ9BxcDxafAc8Qfg2pP2teNNN6r023TGpfvOlnF3OntxEFSahKUhxraT85SUKAkKxHAO0pylUUsrU2VJcG7zNnoEQSQZyZI6/z4PUZbLuw5CCyMcDhyiNo0uuqILgO6BuKkwACIAnEf114JfLrQRqbgwA0JWZBIIUFYJ/SYz7zE54X0el2mKkU7cpJwFggR8xIHsIEHv1M5jgj8q9KKdqmlhoBTywSIGxSfmAx6k9ukGftwettvM1Q1iiXeoZR0bp3lGXwt6+1F4Y9Zsaj0hWotN0ZpF0qXFtpfaWh1lSFoCVdZQSBgYIzw8uU5oKUbAnzVqJKlD1b8de39Dpwh0haUN1A/dmGUyVdAcAGZSIMgHE5jv03alqvwNvcdI2pXKQkpkpEyev1x3n2zxe1otwpoDJgDbn7l8jcVXd1bVCFpzv8AVDXnBe01QeZbw0UrEEnchMGR0yemPqSex4h9osL13Qp0JCl1J80nyEFSEKwkCBJO3b/dj6GCSFZuaNxsVBqbT9Hb7DcEawtIttS7WUKKp6kaC/MDlI6qVMOSE+tPXoQQAeNlLp9Fqs4QUwUo3kEJTj/ICe3XJ4V3Uz66pe93IbJvhrI7XRxws9oob6gojbU7Qhtax8hbE7FbsEDHtnAziR04gtyQt9xDadqEpUEhIML2kTtIHpgkDKgBKVTGRxP9enakhBUz5qISSU+nOEkmAOhP2TGMkRzTGkqmsvNoVUUdSlFWpDjClwkubfzflMyOhAPfInhbuNN+MIWpytVX/wBsahyhdFbheHypY2UjaiBPUAASo+56AdckDvxLuW9sGp9UCoS3NJRo2tAo3pMTIIESrBJJI6HPEXuaRRUrVuYIQViFKC/TnvPaf8T32jg1cttHJ01YmGXWiHEALcQU4WEAEqAI7E/b7iRx87FwDTIei/QdrcYj6nmplpS0bKNSCQ29VKKA2gAFInr9BA/UDsInLtVm56gfZpPSim2tBSQU+c4D6UqIkynuDnBOCkcO1KHbLpolDZXU1I8tpsYUpe2TA6kAwDIj0xEcHvwj+AC73ty33vVITbaBfqTT7VCrdUn8ym1DYnduOTkgkxkcDXzPEDvD3cULuNTBFOH1DsNH19wTRyp5S1qrfZtJUVKtf7UeadvNWyySlQXuWGln5cpCiUzMJETMHoNy55UU1F5P4lve2AlIQkelYCQACABjEwf5xx50HoC16OolU1uoKejS+Qt0to/ePK7FxRysj3JOOkcT+w0yUdABHXMx9uFqPhaKSZlRWHUW5OOhJ6nvjolS78YSzh0VMNLeWeuFM9IUiWktIQhIS2AAEgekdgO3QcT+kSGaIqCSkFM9MDHWPfiG6TZO5CAQUqjv/LiW3B38JbCoKG7aRPeOG1oAGyRiSdyhnzRr5W4ZPfI6cV85i1JLrkFUZk46e/8AX/g08z7lt8z1lAEkn6cALmDVkqcgSTI67v148WBCDmJWJhwbiBHUA8AbmBVS6uST2gmMf68GbmJVEFzKsDHXgF8wamQvsJwO/wBvpxxdzXdgQ31NUFCVSYVBP2P+/wDnxzq8Vd9/b3Pa/LCwtundTTI+gQhII/jPHQDW9d5VM8oq27ZlRxA/045n62u5v+r7pXHrWVbr323LJ4YuHo/M9/ogd/f+G1nqmvjOM4fOW/Ly7c1deWrTljpFVt2vNSmlpmUmNy1dyeiUgSSo4ABJIA4aErpkQncY6T3PEw5J8h9V+InmfZNG6NslfftRahqkUdDSUzSlqdcWoJE4gASJJwOP0F/2dz4Yng8unK67asOqNHc7eY+j1/8A9Ru1CVqtWnZC9pbbfbQHGiG1nzjuQdhgnqbg8vV6M0rdqrxC1GmqOxUl4oTaOXOnbZY0MOUltecCG6lLDTaVu3C5r2FCBlLKmGht3uqI273WC3Uzqmc7DYAc3E8mgdSTsFvDG+V4Y0IC/C5+B9yw+F/oi36o1jb7NrvnU6yH6i5VyA9btPT/AOlSoVKZT0U8RKjO2EwTZ7W/M6o1nXzV3FyolSdqfMKYGRhIzE+5Huew4kWnfCLzH521pu/MDWddoSnfV5iNPaYTSv1zIIEfirk824C6O6aRttCCIDrkbis1X8LSx3m1uJs/NLnJYrgElSH16iTdmVL7FynrWnWlp6ykBOFGCDBHzbxv9mHG/GUpqq2qZBF/RDl2w/8ALSCC7vzTlZ7xabc4DQXu6u2+mUF7jSsXVJpnWW10DoKahLid341JwQruUQSCJgzHvxDeXaqSt1rebRcUuKRpu7NUDXnTscpFtsPsrUT85CHQ2pUgSyQYgkxDxEo114FdZUdi5pJsT1s1C441pnmDa6ZdLbaytSJapK6lJWaOocHpIQtbTqd22CClMIrdSal0xaqXVdDS1NDX827wu33G1vrFWLSqtpVtWqoJQoplgUzfm7SUrFSpUEpkUZL9ltytE76G5uDM40Ho5wOolvcFgcPfgHfCtOjvcNSxs9Hu3fPcZ23+OPgkdFea5nkDyp1ZU1blJcL5U2+zXWpXITc6iqbX5D6uxdRUBYKiJIdWn+4UyPU3MVetdUXfTGqbNbtXaPbttK67p25tIeaD1Q68S82tQ3tOJQjYnYtJxJ6g8D+6cqaWh1DYeTbNW8NGaXbt2r7UtTZ81inYS/R+SFx/1G65IqgokH951O3iOWzmLXWS43fmPfnk0ptFQ1pK/sqaCG6ZFE+4yutAgwBVOKVAx5SzjATxY1OZo5BVW2bDsEtxts92YxjuQHDHTAHZH6Olimj8CujBbq3z6AB2T6HBVFPHZ8JOj1DRVvMbke5WXGz3AKuDmmqtRNW2wokhylUfU4kAZaX+8T7q7c7q+jcoX1NOtracbJSpKwQpJHUEcdqr5qXUOguWq9K3J543K2Uf7Ysqiv8AfW5xDoXUUylDbuKPM2DP7xLoTBhR4rZ8QrwhWbnzWM6u0ZTsUGqq1lVRU0ySltF8bSJLiEYIeQgoKxHRRGSmVfR3DHFb5AKevcCSSGvHIgcs+vy3CQeMvs1dDCbhbGnAALmds7HT/j12VE+RHIjVHiS5oWbRujbPV3zUF9qUUtLTU6CpRUpUSY6JHUk9geO9vw+fh66J+FxQUluslstvNXxLXmnC7tcS6lu26Jplg9H4V+Hax6nkAuvRtaABB4DXw5OSGn/A7yLeqdP1tqc19f7YHdTave2KotJUi0qJS0pR2uuqSYGxQSSnapQR5hJ9tdZp+1aGW3XVVzs2lnAK+uonaxVPedTvKSD+Lu1X6VU7aoSQygpcKYEsNlLZh8Z8XOhY6npSOxJzt8tzno0bnrgbpY4X4RmqXiRzSSeQx9Tnb4nYeqnlTcm7hrV69Uda5za15bXlJrtTVjotumdJudTT0c7kskSQfLD9YpJ9W0GEwDxD6vZ5Kcn9QcwNV301wtdOp1hNPR/hKU1SlHyVBC1Kce/ed3FpAAPoHTh9t2v9Q67s9E/Z9PPUtkoGtltRT0jVrtTDKiClNN5m0BsyVb2W1zkkkgnjnz8eTnrX2LSVj5dPV7VZUXB5N2rFNhakJayhgJWsAqO1LhKgEiVkQQNxpyw0tVfL9FRkaWA6nF2NZa3c+X+gHlsAd+ZVtXSWOw2WWYS6pMaQG5Lcnbd39RHPY425LmtzN1tVcxda3K81jrrz9e8pwqcUVKIJ6knuTJP1J4jyu3HwmTxnH1kAAMBfMhOTkrOM4zj1u3K4xeKynwlvDEPFL429KWurpjUWKwO/t67AiUKYpyFJbViIcdLSCDEhRjPH6UdH20sMJkSVdf8AuUcmf1zxzK/s6HhXGgvD9deYtfTFNy17VeXSKUnKbfTKUhMewce81X18pHcTx1V0pQFK2pE9usfrwo3mo8SfQOTfzTVaINEOsjcqUaUoClaQRJjaRHX6/wBe3E4tdAPKSkjdtEYHzD69+GTTVv8AIZSSANuPUeg6/wCPEpYbKWikAgDuR0+n04E9coq5an/QVAbkkgjcR0kRBHt7g9QY44c/G08PeifDx4pG2tG25u3J1Pa03i425ojyaOtcee9bSD6kIcA37EQhKjiAYT3EqiYJ9+x7/wBTwJPE/wCDLlv4vLEKPXGmqS51DLP4elujMsXO3J37/wBw+n1oG8TtykyQQQTJiyXIUdR4j8lpGCotTTCRuOq/OXp3TYulfsU6hCSNsnJjcEzgZBJ6xJjpBHFh7PyqpdOt0rrN0oruiupW6pZpyT+EcUsgsrSUg+YA2Cdu4bVIyRIB18YPwravwaagYvdkertTcv6l5tlFbVLbTWUVQ4CnyH0tpTO4oG1aEwcJO08C63WlilrENvuM0wUsIUokhttEiVKCASR3hIJABgHpx9M8CW+nqacV8bgQfz7e8L5l+1ziuop6r+G6SMDP+/il1kpaa16dfK2Q+66kpb2kpO8nqRBnEiP+76Dgfa7qnKt3yQQsKMBrIHuAQBn6D3PXHEn1TdVgNqlSW0wWxEkjdg/rjqR/KOCF4fPBpzF59MuOWe2V9bb6RxT6mQgusturSncokQCtQSiYgwlIMxw/XKqi0CHUAOpVG2aklEjqx4LnZyBzQ0sPh2csXKCz68dvViq3L7c6i3JtaagKr6ZTISrzVskHa2qYBkyRnrmOajo/w6AtqNxEpkhQVn8wz6ZSQPf9ODHzL5FXXlLf37fd6CooKpBKVpW1sUpJ+/8Al0n3OYFzR10y/wArmdOm3WptdvrHqtu5IpCmuqA4hKS0t0mPLTtCkgJBlRyePRb/ALvSgghwJ5jHL4c+2V2jvBrK3SQQRtj1/T3Ku/M15Lz7yd26kBKQp1X5PUcqJkSYMTJPTrhn0/qyoeuzFTUun8SyEtJRvO9RwrcsdeoGDIgEAxx55lXMtLdO5IWk+b8gWFEEQ2nESoSCYgJBHeOJX4OfC9qnxk85qDl7o/yarUl2feXSuVLqqdlBS3veUqZhITu6ndEEdSOKhr6sCtc89Ff9uoj/AA5sY6qE8mtMDUd+duDyZbbUtayrA2piY6eqSBJMkkwCeDbpKiVd7gpa0hLTbMgBBAS3JVuHSFYUYPecjpxB9L25rS2j6Gkwkup854LTEJE4PYerdERP3PBD0Zbi5YyFpC11SkKcEKKtoAhMATMgJJAnJyYnj5mr58N0jkv0Up4iG63cypZpDVCrBzEst4/CM1ot9Y2pukdTvbeCSFKKhHQHb75STMZ46o2DZXUzb7S0PIfSlxDqTuDgVlKp+oIP68cpdM2n9u6lpkKAhs/u9qASIUComJI3QAZ+sdBHT/k7zMs/MOzIVQXGjrq+mShFe02TvpntvqSUnIEhUdcDrngfRVjNZjccHZIXHVE8COZoyNwUQbVTkkAyT3SQMcSayMbiBBBEwY6fT68MVtp9wSY3IAg+/wBOJTYGyXQVBKQCNuevt/Offguq3Dt91NdK0x3oOFGYBnh21bUCmtcEqUQI+/1406YaG1BMgDrGccItf3A+StPQA4J6Rx7nZYQg3zJqpKzu2ggid0gjgEcwKmCrqCCTE/Tgw8xq5ThWT6SJMdZHAJ5gVilKXI2yTgfw/XvxqV4EHuYlf1ncpJ7/AF/y4CWtnt5WI/Q5A68F3mO8ApRII7Ge/twFtZPhClhWQkHA7jPHE81JYEE/EZfzpzlffqyRLNC6UmYklJA/mRxzoeI3QMxxeDx56gNr5L1lOkwqtqGWBB91bz/JBH68UdJJOeG+wMxTl/c/kla/PzMGdh+a+cWf+FzyQvHPbmxrS1acc8nUw0XcxZlpUUuCrcQlptKCBIUvcUSP7+MxxWDiw/wwvEsjwu+Lqw32pqGqO23BDlprah1oOop2XoHmFPfaQlUDOMZg8GznGyBrr18EP4cGtPCtZrhprmOw5ZNUc5WA5qexJfgWHR1tWl18vhrAqK+qcaowmY8pT3UyB1G5M64oudPjXulBX0bb6tFacpr/AESSr93Su1dTU0rSkIGJQzTuAE/L5x6EAisfgc1I5qblNqvmnUKIc5m3T9l2ICR5Wm7UpdLShufUE1FQKqpM5Jd9UkTxLvDtzcoeSnxIrA7fHk0dp5r6WXpJqqWYaau1FUqraRlap2p85h+sCSeqmkpGVDihrlxvBW/aLScOE+SEOcexlLSWj3tH1KcaOySNsU1xxucAf8cjKJ/xxvibPfCr8C1drmzW5i5awvde1YdOtVDZXTN1bqFuF50CJQ222o7Z9SikdJ45H/Bn/tM/PbmB4/ND6D5x6lpNbaR5oXpuww5aqekfsVXVOeXSrp1sISVN+cW0KQ4FelZIMpk9u/iWfDr0Z8ULwq3PlhrR+ut9NUVDVytl1otpqLTWtBXl1CEq9CxtWtCkKwpC1CUnaoUF+F9/ZatE/Dv8RNu5p685gq5k6l0s+qt0zQU9t/ZtFRupSQipdQtxxbrqJlGUoQqFGSBF+gZKSshrd1dX4wmibNzN8CGqdP31DaaK+LYoDUnL9A4txPlvs9/NS6GymCM98QeIfh78Y991tyo5fs17qVv6MurNxvrCDn8Hb0OIqnyfTvdWHhtJ9P7tWSQTxfr4x/j7tV00/W0NruS3KaxsVDC3mKgJpaBQlTlQpZw4sNpbSIAA8+ErO8AcafDNq1y16NcQ4HwjVqXaS4hKyV0lG84POdSZClFtK1AAAlSlxhU8I3HtqgqqVnisDnA7Z7Eb/RWX9mUrzPM3VhmnJ94O31XSO485UWzmzcdVvJpm9K1tkYs1nuTS1rqK55l52pWtpE7Q09+IUGNol38KDtPmI3ijVnMGsHha5nvXyiqKmvrbhcUVtnpnBUfjXayvbqaUpWcElDqIHq3KC0+4Dhz55k2i6VvL/Tool0IoLsLwloKS8KNFJTqbpAJPd1xIHpElHaDwOqu8ruviIFOXwmkobPTvVNMXUJcerC482w4Oo/dtKfBAwkutkkEA8UJZqKMxsmdFpwGuH/GJ2lnxIzkjrghfRFLZtUbhr8xdo77vGXH4HBHpkJz8QGu6Gt1ly9TS17VQ5clXK5sVBpFKbfS6yktpSRAglSVpJ9J8oRkRwL71q2psGqr/AHCjqi2jRz7tRQNqdAcO9LT1W0TO1KTG3CoClnsAOI5cr26nQlS9StU6Ljbq1lq1OO1CVMtUdPVksKnuHCt6E90JQIIE8DbmFzEqWKS+W2nrKakoa9hhyscdIcefPlhBaT3UpxQWpcGc56DixLTZ9DfCiOdPl3/5ZJ+TsKRV3ZrY9Uw2dh2239JAHzbq92FZfkpq4AVQpUsKVYrj5tGuqrEIttv8weY26pMldU8EKSEOEKSNpiDkG3Q2qbXqStbr6Wkd13d0K81C0U5qKSle3kqWnzFClQoKzvcU656vnJEmk3JvVS6jXFxLVPZBTN2ylpUIultFc0ylsqI9E70IG4wtcjcPUoGJunyT1BqvUtmpqpik0XeKVgSPwtzqKNKiQSRhh1B3RJAUPr1EL/GWumj1OOB6u08+hOM/IhCrPDDNKSwDGf7cj34yAfiCixRDVl4X+Iut5ore6oeaW6ZZuNUMTCnFgNJIJgwlQ9Ig8cf/AIz+r3tReNu/0iy+W7ShqlAWsqAUhtKDEmJISCYAEnoOOwFk11qGzvJJ0ZVOFIBCqWvoHyFEdNqnkq987MyffjlZ8cnk7c9O+Im365eoqygoNbUSXE09SEpcpn2oQ6ISSNqlDcCMeqB0ngJ9i0zGX+USaGl7DpDSD1BO+S4nHdLP2zNe60RBm7WvGdsdCBtgAD/KoxxnGdTxnH1QvmNZxKuS3Km5c7+a+nNH2dHmXPUtxZttPIwlTiwncfokEqP0B4ioE8dHv7Ov4XTzH5/3vmRXUwXb9FU/4GgUpMg1tSkhSgfdDAXPt5w9xxxqJhFGZD0XanhMsgYOq7L+HXlJbOUnLLT2mbM0GLVYKBi30ifZplsIST7zBUZySSTmeDfpe3FW0hJATAH0+n6jiJaTtstoAEiNsEdRE/19+CZpW2zskAnocZPCE5xcS49U8MaGgAKS2ei2sRmD0MZ4WrUWxG0wnAEzjv142UiA22UkwYEE9DxoqHdoOFJAke0f1/PjxZndJauo3AkZgAmfvjhMpYUuB6ldJ7/+ONlQRtCphIO0YzGP8+3CZxwECQZVggdgR/XTjFshF46+SNf4hPDVfLFafPcvLTjNxt7LJSDVvsqw0dxSAFJUrJPpMHPTjk3zs5Z6q5D379k6pslZYq5KDsS+namrSkgktLHpcSD1UkkHpPy8dvC6CqCQVewgkf19uKu/GB5f2PVfgd1Hfrt5rVx0SW7lZ3UO7CipddbYU0o/3HUKKT3BSkiCOLP4A45ntj221zQY3u59QTt8lUf2k/Z5T3cOuzTiWNvLoQN9/Xnhcr2OYRRZ0Wo01IaVdWm4B9VMDVgttqGxLnXyhO4pn5gk9Rx3i+Edq/S938Idkp7S5Ror6TcLghJCXC6TJUR1gg4J7cfnas9yXXXkJSlRB2rEgJKjMJCgZ6iYH/dAAweLBcq+d170Cyr9jXevt6lgb1U7ykFQP97aYEGQB/rxeNTZxeoDA6TQc5B/RfPcd6dw7UMqGRCTYgjOPkV0B+N7qDTV35g2elo6mkTcGmFfjHm0hZbnoDGScf1HHJ/mtdgt1bbTiVNg7FGQBnqP8sfp2kic0ucFbqB916uqah59353XnPMWoSJMkxGZkmPr7ibRXiJq+RHOTTesqGit9zq9NXNq501JcqcuUj62iTseRAJSYGAQRAMzEm6yWO02plCx+ssbjPLJ/RA7HSTXe7vuUrNHiOzgdB+qEVv1s5p3mdZbq3S26uds9e1VttVtKKqkeWlSVpadZV6XGiQNwO2QVAnMj5o/mhe6bmOjU1juFZpi6P3J2pD9n30QtwcUp0+VsJDaUBSgET6QkGew2c9OcNbz35yao17cqG2W+5auuNTdaiioKfyaWlecUSUspP5QrEZJ2kzKiAi5e2ZSqlLiQ0IAQS0BJkhRUTJzBJKhmAfuKiDXTz6upX0GZGU9NjOwRBdV+0ri6SC8XloZQpOCR0In8pxtzJ65HXgo0FEinoqSmklBBcU4AfUSRhQED5eoJJGBMEcQPSNnLtT5pQ4l2llwyNqWyEkAgxMAGZ+8dI4LPL60J1FrunZKXFpQhCSUpgLSVEgKkEgfL0EkE9ySfmS6zCOMuX6GFupwx0TzouzK0xUs1TiUtOMtqdcC1BMFIJAwOgO0wR3yMji1vw89NKt2ibheXWwhNbVKe8wABS0N4iR1AUVqE9QQOkcV/wCYun10FYhhhohx9SUn92FFRJkiCPoY9uLjeHfTqNJ8nKOnRAWpjarakeoqnMdes/ftxVF+vD4YWVEftagR/wDnf81Du8DJKYxH+rZHGzOprGkLSc9Du6p79P17e/EusLMFIAgq65z9/txENNPlavLTtKgoJ67ZIAwP66jia6eaC1HB2kiYESf/ABxbtguRr7fFVuxlwGff1VAXGmFPUvhHIHbKnenGSGU/Yyev9DiMcyq7c2sKSSlSYKZifrxKrQrZSlRwAnH3/qP48D3mZXeVuC1bZkhSRk8GCMKGgrzIr1FazIjJKZ6n6/b+p4CWvK30uDBnMESBwWuYdZuU5hQGQATAOP5cBPXlZCT1iSBBkfYcakrdiEev6oDeScqERmf04DGqKhT61ykpiZEzGeCvzAqT6oCUk9ft/hwIdWvyXOwEzHHAkZUgclTr4jV+AobBbwoy5UPPqHYhISkfzVxVU9Z4O/j8v5ufNqjogvciioQoj+6payT/ACCeAQcAcPlqj0UrB8fmkm6Sa6lx/ey+cFXwfc+qfw5c97Pqep0xo7VzNOFsLt+p7eK23KS4NqlLbIVkJKgFQSndIE8Crj0lzaoEdeJssTZGOjfyIx+8KFE/Q4O7LvHyK+Ira6mistt0vRL0Y1T06U0eg7hWJqbQ+yBO3T9xnbIklNI8qCSQkMmdx05jVmmPGbyPXSitqG2qsIqKSsYJYqrbVsqlt5Ewpt5l1JHZQUkgjJ4/Pbyi8SeoeVja6Jl1Fxsb42v2ysSHqZ0T/dUCAfqM8dBvCV8SK2DTxpWXmGk7T/ySoaekJTvUB+ZwjO7cokNgkHCR80/aL9l07Z23m0gioY4ODwd9v7jzPoTk8wSdsfQH2f3e21zDQTvDQRjQ7YfDpjuPoraaY+NXzu8KVwRpHmje65X7PTsor/Q25usbuKAISp+nCkrbWIBUW/MSds7EZBZOfXxnLTqzRtczduct3uzD7BXUUdnsVdR3J8mISpTjSW4SZG4ux+8VDaYHA85l89NI839HK/GP0FRSOIEpLYedWYmUAkyUxhSyJzEJTxWbUfIPTd0uaHbay9TWxQUadK3AfPVuPrCykySCPlG1PuZ28WJwvx7WS0obdIS2QbE42Pr6E/JRb59jdO6ozb5vKenPHoot4hPFHffFFX09ls9Cuw6OafH4el3y9cHC55gW+sHapQWpS8EgboJgRw+8vtPv6dondMUzqXr7dqFpoOhcppi688HFbkHcGkojqQVLHX1SG+7Wqh0s7VKtb7XnMMLqKl5tnc1bkJz+HVuO5biwEpOQUhU4hIU7VXM5rQ9jq6G3viovlxShb9SgBaEq2QVelIXIIhDYgI27oElSp1zuE9cWkDIPIfLn6DnjqjVj4corNA6NjvP/AFO235jA7nPy+iI2oNein0lYKqlfqqvVmpn7fVioq0fvSlsoKkpQ3hLDSCoJiSQ4clSyox/Ule/pLWFPbrdVebX3C3u/tCuckqbSXmgmpd6HdIcCEAAAFCQNqVECms1G/pu1UiaAIbqyG6eqqqxakIYDaUqSnsIjbtQJCtyxJUkkQ2+8wnHb1UqbuDrpUhbD9fUNesjqNiB9JAzABPWTMKmsrg4gct+fryHuHQIx/G4YWBxGCC3ryxz95PUp95l8wUpstdYbMkOUNPU7w4k7UoQh0LayoZ9YWP1JMdwVzW5khddVUdIpC3XPRU1KZyO7bfYJnJV1V75jhPr7mgl2lfoLctwsrqC+pZcncqACon8yj79oEcQF14uEkzJ78PlttrYmhxHr/tUrxVxY+peYYHbctuQGTsPmjL4bOe7WltYUKLrWP0TjTqfw11bcUh6njolagDKO5BSQYEgieOj3Jbm3alM01VfWnbWdgKNS2N00fm7xhx/y5CAr/wC8lbS4gED08ce0uwODn4X/ABpXTkjVs0F1TVXbT6VSGkvlL1J9Wz7dik4IJHATirhZtyhJj2djpsfhtj4HI9EX4B46ioJBSXL+WeTsZx7+uPUEELtNYL1qVmnbdor5p7VNC+PMQqqSbdW7SAU/vGUqYcGevloEdjwMvHbyHuvi88MN3047pWtF8tSnLvZ3GKynqQy82j942SNiyHWwQAEH1oRgdeB/4ZOYNu5l6ddv+gtQPs0aFE1SaJAU2ysgqUH6NYLSP/c35U/9xO7g/aS5vX+3NJLzFvr2krDf4y0O7XoIwfIeVInodjqhEfXj5hntdbY7syrp2s8SN2dwY3eowCGnI22znsvo26W6ku9rfEx7nxSDYtIkb6dC4Y9QML8/lxtz1qrXWH21NvMqKFoUNpSR7jtwnJ2k/Xi6Hxi/DHT8uudC9dWChqaXTuryHnkLpVMppa8gGoZB+VULJUCkkQtOSZ4petMAcfYNpuLK+kjq4tg8A47Ht8F8V3a3SUFXJSS82Ej/AH8QsQjcZM/px+kP4S/hXT4XPBzpDT79MKe9V7H7YvJ2wv8AF1AC1IViZbR5bWRjyo7ccUfhLeGMeKTxraXtdVTfibHp9Rv12BTKVMU6klLZkQfMeLSI9lHtPH6S9DWg01O2CMkSSnqo9eB99qMNELeu5Uyy0+SZj02CmulrZvIMbUpMgDJ6TwR9PUgQhO35QPv+v+XEX0pQEKG7H5ukjPE8trAp0hIHUYxB/wB+FtMHILetSUMkgGMGPcz2+vCOrcSggqjeD06lXbpxuqHcQcycfXhvefBUgFQIEp3A+x/yIP8AA8YvQFrfX5aFkSpWcpyeEjziUkqAhJOfr7fy/wAOPZqSQTJCgqDGRMx/p/HhKpYVux0zE9/6/wATxi9WLcHmTKSYGU8Cjxpch7h4p/DrfOXlDdaWyJ1K/SMV1yfphVKo6VFQh54tNEgKeKUBKJKQlSgqcQSiPSTgp6SOPO6SqUn2gZgxn+fv7cdIZnRSNlZzBBHwXKeBk0ZikGx2PuXFXxifDt1R4I7tR19TU0t70neKhVJRXemSsJp3sqSzVIIlDqkgqEShUKCVSOD18IHwJWHx1aw1AdQ3SoprPptltZYplDzqlS1EAiRCR7kDrMTPF+fEvyYpfEV4f9XaKrC0gaitzjFK478tJVJG+mensUPJQZ7Z+vHFblD4huZvgQ5n3hm11tz0nqu2VDluutPuO5hxEhaFAylSZIIOQU7SDBB4vbhTimrulA+DWGTN68tu6+fOLuD6O2XCOofGXwO6c9x0z+SNHxivCPZ/BjzzVYLJdDcKOqoWq1lpR2uttqO0IWBgqESCBnHfpz61/eUOIUGNi0oAHykIXgqCoEAZExnqSPSCSWvEt4kdT8+tT1uoNUXGpvV5qv8ArVDrkl8g7AMDoJERAxnoeAlbrHW621RRWi1sfjbjcKlqipmicvOuqDbYkwJUTAnoTkTjg1cquVsQjmfqIG/v7ofZqGB0pmhZpaScDsFqs1uVeHGmgylaEvwd6VQQVfMRIwJ6D+8ehzwRLXZW7VZnahSQinWmCJA80bjuCo6HP16+2OF2n+QF45Z8wrnpXUtvct1/0zWv2y50anEOGlfZXtWjehWyApJGJBKSSTAIkOuaM0tChmnn1KKvMQnO4RCoA9zJ/U8E7NbT93NU74IVxBeAKltFH8U9aZpk0dkqn1OFAqXyhhs+oqhQTuzncAe5OT9RwQ/D80A6/Vl1QDK0pSFAjcJ2lQx9Osk/URHA6vDiKHSbCioAs0pWifQJ37cyCRlPTESZ6iSlyOYFFpM71FDiUhKgo7QMggDsEnORnMY4+Kr44mmPyX6gwnz4PVF/mJQN12t9P71BtAp/PUsyNhiMx7TJMjp79bSaCq00XLyhSClKFRvJUBtAAP8AH/HiruplqpqHT9d5a2iwgMK2+k7SDKZMgjAIEY+gODbozVoodE06QQqHxOzoQoHA+n6+/txTd2jMkLB2yudbEXNbhHi23FDtuSoYU8oFtREbTM/+P58E3lxdkXylQsEhxPpcBwQsdf8ACfrxWzSuu/MtRaW6FOU7oStROQJHqOfePoJA4NPJav3a1LaV+isbUtJMhKhAUQPbI9uivrxP+z27zUFy+5u3ZKQPce4/JV5xTZw+nM3Jzd0b/N8q1FQlICQCdsGf9OBPzLrlBS4MD/t78FC9Pli1kQNxGQO/68BjmDWpU4sGREgGc/19OPoRyq5B/X1VtDhMQcHBnrwENe1SklYiD37n78F/XtV6FwQZ7JBmOAjrtz5xt2pggGf1Mx/hxzet2IS67dBK8z+nvwJdWvShZHQiff7cEzXdQS8sEgyPfPAo1lVCmpXVqI2hM/aDxxA82FIacBc6fFHfDfeemoHBBRTvppkx7NpCf8uB6oQBw5awvQ1Dqi51wn/nKt18A9YUsn/Phs4smFmiNrewVezP1PLu5WcZxnGcdFyX1BjhVbbvU2erQ/SvOMPo6LQYUPp9uEnGTHGEZGCvWuLTqacFFjR/ipvdlfQLhNalIgvNlCKjqCCFqSoAyJmOueueJbUeKFnUTDIdWwyVKR5ilsKD6EJIJQlU7BMGIAAJJiTxXrjOB0tqpXnVpwUz0vGF0gZ4fiFw9VYu58+v2m/SU7VdTfgaVW9tlDAbb3jMuL7jcd0ASSJP1fOTN+a1XzKRb2Kxq5VQbcqhTBPrqwhJLzKASPMUtsLKQTKlNgfmg1Z3H3PEr5Xaoe0/qWhqaWrcoq+hqEVNHUoVtUy6kyFT9CB/rxxfaI/DLWbHBRCLjSrfKC8dR1PToF0dofCjatXUFtRUVhZtt4Q2vTte2wX2KlJBWaV5HymckAJlxJlKt4PDFzY+Fc47ZH019dU6SbSUpRc22V1dnKsk+Y+lJcpxOQh9LYAVIUoAKVKfCp4rdN6otNVar8xSClvXrudqed2NsPk7lVFCkn1NqMOKZQQ6yrLczAuDy25iX/ltSUj9rr16206EBluo/GNtXVtHUoLqlBFUAn8qy06IyHD6lUtcuIbraajw3jkds7B3ucdg7uHbHm3nhXjHabdeqPxof6huN8j0IBzjsR8eS4j+JjwZa38Ltxb/AOIKBuptFZ6qG70LoqKCuT1BbdT6TgiR2kT1HAjXx+iHXlLyp8UOlrnYtUaXsi665BSHGallVoubTzgIDu1SUeY4BmVJWSIExxxE8cvhFu/g85z1Onq1Lr9rfBftlcUFLdYyfzJkA46QQDg8WPwdxnHeQ6CVjo5mDcOGMjuOhVG8XcHTWhwmaQ6Jx2IOce//AGgvx6AgzPHkiOM4eUkJ70XzAvfLm/M3Ww3a4We406gpupo31MuoIMjKSD1gx04t5yV+M7rHS9O1S64sds1cylJSqvYQmiuKiZ9bhALTy+mVICj3UTnilHGcQa+2Ulazw6uMPHqP3hF7Vf7jbJPFoJnRn0P5jkfiul3Pv4hnJrxa+G3UGlq643nTdTUNirpae6281CzVNhQSUOsgoTM9PSPVJOMc262mTT1S2krS6ELKQtOUqAPUcJZPvxNfDvyaufiG52aY0TZ0FVw1NcWaFCwJDKVH1un/ALUICln6IPHC0WeltcJgpBhmc4znHuXa+8QVl5qRVVxDpMAZAxnHLOOq7C/2dfwrK5feHGv17X0paufMCrCqcrEKRQU6lIaiegW6XV/UBByCOOqGk7aEAK29hIOYHv8Az4GHIHljbeWGgrPp6y06KS1WSiaoaRsCNjTTYQge07QCfqTwctKWslQwIPWRPaIj24WayczTOf6o3SQ+FC1ilGmaANpHeI6p6/78SQJ2NQSn9ex9/wCXCG1UvkMiO6Ynv7cb3FHO2M5A6j7cRl3O+y01dQQYB6HP14b3VAiI+Yjr278b6pwLRIJAwQY6cIKp2FEHKTj/AF4xerUp4lpUo2haiSJ6fc8aFwXDJAMRMYOO/Ht1SvLJ+YgbQZ+b7+3Glbgnd27mOnT9eMWLx5hO0x95MRj2/wA+PpHZU4yZwM4/hx9bhMQlIIJSABGf6HG1tuVoCZWVkAEZJVgAfr9OMWLW7ChEgk9onuc/bjmF8erkto/R+odNcxqa9Wq26xvxTRXOylyay7MIQQ1XpQJIS0EpbcUQlKgluCpTZB3fEc/tENh5I37UGhuT1tY1NquzV5t9dqK4tpesbBQIeFMhC0qqXEuDy95KWpSo+sQDyB5leI3UfOfWtz1Fqi+Vt61Be6g1Fbcq1zzXqhcRKowEJGEoSAlASAkAAcPvCNBUU1Q2te7QO3Uj17D6pB4tq6esp3UTG6ievQe71Ul1Lqv9suKIX5mwFIMBaIEEd8wAE9I3Rk4lnt77rV4SptflpWkoBPpCiEgNrIBkK3QoREKJ+U5EeRqRtdOpDjnluODJjcQIkiQZ3YH1yf1eNCuivrVVLzag2QIBABSD0k4HeJgkgACSOHaSd1RLg9UnxUzKWIkDkj1ywcXQ0UVClF/aFuLUsqWJG8kqyonr16kmck8fNc6oF1fSy0ltI8xW1S46TtVuMwrqnBPUEDBkR1vU6qC1ja6t1apWSVYHy+r5pPcGcglJnsNVvq1vOB8mdiQoAwkgRjp9BBiB1kiSRYLK3RTNpo+WFWD7YZKt1XN15IoX9aEP/h3VKCHmCyoBQbSmDhIUJISPT7kp29eC5ycbcc0r5flqUVgZ8sQralIwfynvBiCQOAvzGKkMVDiTsdYIMyYQCYz9YUOsRHYjJj8O1zTdbW6IQpTyErATtIKtqZAkQcj7kQfrx8RXuJxoXPHQhfqiHATAHsjfrCnbPLQqT5inKZttYKjs2pOYicDHbr9usl5Z1hvGkU04dAdCN2xH5T+YE9zg9v8AEjhhu7Ip6FFMobjVUwSlREbBtgdO5wCc9enDDyN5gtaWrmqOuUlCWXw2ggE7wOuVdITHt/gOKpdA6Wmdp3IOV0kHkIG6JbN+f01ekKWFuU7raGHVQJP93+G769Rnix3hX5hsV2pre0/UJLjKi1uWQB6kqnJ6dsH37Y4AHMSzsVd5aYQG/wAPcGw5TlWRkT9ZPTB6fYE8RG2avvugqw/hy+8hjdsQqSomflgmZA9p7Y6ngdRksnjqWDzMIOOWcKDW0La2mdDnBIIXTbU90bqLckNOtrABnaoK657cBrXlblwzAJMgmM+/+314B3KDx4MW+qprTfWiyw6pDRqSDsQpRA3E9SnM+wEx14GvOT4ha6zUFYzY2KY0LClpbqFNqWt4JkBZkgAH2A6d5ni6qDieKen8WRpa7kW8z7/cqjl4GuQqDAwZGM55BEzmFc2mKd9xxZSGwSo/3T9/1A/24CmuLuzV+aW1ev8AOnoR7fX/AF4GeqfE5cda14bqXXA3JQlCUwlSgPlKB3MxP8O3EXr+dSKepARVtthJkgmUrAHQHBInA+ggHHGG9yOnA8P8P6593ZFv+gXtpS50g8XmAOXuynHW7+1xUiDJx1g+3AF8TWpxpblVqCskpW1ROxHUEp2pj9TxNtceIG10bSVfu1KcI3IKpJSRMgDMR/DuBHFbvFbzVHMzlJerfa6aqW4hxlQCGypS2A4NylRMGR0HUH6Hhkt7RNO0DlkJKuduq6WmfLLGQACqXnrxhOOPTiCgkEQRg4yOPPFjKsFnGcZxnGLFnGcZxnGLFnGcZxnGLFnHtt0oWDnHtjjxxnGLESuW3Oz9ksijuiPNpwU7XdoWREABaSDuSPtPT24Odj8Zus+W+l/2roDUwYqKUeXVWyoKK1lTZHzMhwFQjumYxgY4qIhRT9uPSKhTRlCiD9OB09rp5j52gjqCMg+8HZHabiCshi8NriMciDhw9xCudZvjmc1qK1N0lwsmh7yhhBS3+Lt69qSRClbUrCTPcRHsOBV4r/iBX7xgWihpNRaW0bak2tJFIuy0SqNTZJkk+ohQPtAzngBEcZx5S2ehpna6eJrT6ABRKu8V1U3TUyuePUk/mvvU+3HzjOM4JIas4zjACo4BPG6konK2obaabW864oJShsFS1GegAzxixatu6O3HTf8As5PhVVqrmdqTmncKYKpLA3+xrQtQ61LqQqocT7FDO1E+7+PlPFH7N4KuaFy0/wDtmo0ZebNZQf8A+43lr9mUny7iQt/YFCM+me3Hen4P+kNCaH8IWjtL6Q1Pp3UlfbLczW3tVtrEvrRV1X751SwPUn1HYncBhtIMHHAu61OiA6OZ2RS10xfMHOGw3VvdF2bY03AkmBHY9z/lwT9MW/y20zA/KcjrHEY0pa0nbA3RhWPt/lwQbLR/h2oI3kZBT+ccJw9U2FLkNeSJyFEZUR0+3CWtd8segH0jcAO/fA4VuKSlKCJgGUn2PDdVukpmRvBII7q78erUd0jrHvmUEqMgnb34SOLIMzhWevQ/XhQ+AlaYVOZ65/8AHCdTQKldNgUADH1/1/z4xbLWlSnGh8xBkHt79OPi2FdBk5jEAjjalpSnEhCN6yQkJTn1GIH8eOdGuv7R3yp5Xc/OZ2kb7p++XC16NuRttgvGnlorBqBbZDdQVBxSEMpS6HNqwpQUlHQmJk01HNUEthbkhR6irigAMrsZVqPH542tMeAHw43DX2pGl19QpxNustpac8t69V60qU3ThUHYgJSpbi4OxCVEAmAePPMH+0ic9NfckNY6Yet2jbNdtSuFuh1DZWHqSssdIuQ6w0grWhSyg7UvqPmIlRyogpA/xIfiYaz+JFzbaut5bFl0hY3njprTqChaLS26ltK1OOpSkvvrDSSpw4EbUBKetbKt/wA0bUjpgD+97/r/AL8PtosMUMAfUty8/RIV3v0s0+imcQxIn6gMogJ2hOEgdEgDoJ+3CJFQpo9MpwOk4mfv2+g+vHuve3O4kJmR6ZIHCo6tYotDv2gWi3OPO1qalNxWhRq2kpQU+SnO3YSdxkTI4Kyu3wFBhj6lahfVbg2dq21TtBJJ69T1zj/DgpaQqbezYqCqZr1VNa+t1NVSrZWgUTYUA2SsHavzEk4T02wT24DdJUulBaOW1ncv0gkkHGevvAnOOH2zV5p0ADYUpEKJGTt/h/A9f8OlLKWyalrVwB7NKONPcnL1/wAs2VPKcSAkyVKV2Bk9gO3YIP8AeMze76Xq9Ns0jNdTKp1VSQ6A60tr8QJwRuj0jaqCTEhWe3An0RdQ0EKc3gEgLK1KzkCAYJ3SSQP+2PqbO+KLx7am8ZFr5eUWpLZY6Oj5cWVFioP2VSfh1VjIWlRW/mN6gEAhJ2g7iACs8NtHVEuAxz+iR7hRkdeXLbmtOomP2oghxZ3Lp1QEwRKfUPSYA9+v5YAJIPD14cNWmyXKlClLW0GhuTuJ2x2mI7xHuTPQDjbzU0dW8udb3Gx1CQupttSunK0Kn1pxIJ27pI7jGBAgcQ/SledNauK1pmmfcIW3MhWDuyIxjAI2jHeOPnGSjD4paSQbkL9FJ5CdFRGfLsr+8yU2u86JtddbXVqZLDfmLWkHaYTKDBgxMDuR1PAV1fYnWKxypQlYZSoOHae4BhU56YknIiT7F35f8xVN6JqrSSlbBG5JIBJYMK6GVAiUz0JHfM8PWir3T3NP4erbSW3xKFqGPpHbt9cR24pZ9PJQvcCM4P0ROFp8Pmkeieai7vbjp+5uF1umlykqAqHKZQ6ImdsdPuBHfEtptXXCrtATXUQujaAUpfGV7ff6jAAIzk/TgRcwbIrSGrlrSHF0yiogEA+naAcnvAgE4A29+CFyv5yU1FTKZqlNOsuelC42rI/lPc/m7kwM8eVtM17BNC3K3azGzQh7zB1wq1uS1bS827O9mFEqT1kGDPT26/zGWo+a9lFOUfs2tYUgSElRO2T3EHvnMmIJnizmq6ax6yZWhmrS2ogwjzDLhPqMdEyCD7DA6xwJNdco6Zxai4+lTPRCUKkIJPUCIzJ+5nABjgraquHAEoIKyaTAGyAep+dSKphTdHSvrCgdpLREAA9DmB74xnIkkwO66su92cUWkKbBn1DcCZ985iDiYE9COhzvvKxqlcLiWzuBIG9RSSInEnHT9Jx14gGrdFqYQreXApaQAkmVHE5nt26f4cPFA+B/sfVA6yqDG4IwhJdLLcLkrY9Ur9UpUW1SAScwZ6ZgiT1ni2/w1/hm3HxFaOu+t62oUzpkVSrWwyEhX4upbCFLcn8oQF7enqKsYB4qnqewVVKqP3hUYSElwBSSTAIgzkxA/wBJ47ofBRs1Xyt8D+mdIampX6S6rqK64VNLVwHKXz6lakojsdgQqDkFaunThkjb5cgqt+J7lmn8P+49eyplzV+BDoDW7DqqxmuoKxYxVUKw27PvkFKvsoHimHiX+AjzH5Zofr9B19Lrq3NAq/BqAobmkewQo+W73+VYUeySccfpB1lynZrGy7ToS4yoH/4+36fXgSax5XBhbifJlIkQof4HiXBc54tmnI7FVrNb4Zdy3B9Nl+TrXHL2+ct9Rv2jUFouVjulMdrtHX0y6d9BEjKFgGMHPThnUmAOvH6b/ED4OdGeIHTy7VrLS9q1JSIEsisZlxjt+7dEONn6pUn2yMcc5PFT/Z5WXnKi48qtSG3nJFmvyitoHsluqSJT9nEEZ+cdODlPe4X7SeU/RB6izyx7x+YLlSTgcZwSOf3hJ5ieGG+qoNb6Uu1iO6Gql1ndSVP1bfTLa/0VORIBxwOVt7FRwXa4OGppyEKc0tOHDBXnjOMjj6lO7j1ar5xnGxLClKAAJKsAdz+nBL5S+C3mvzzdR/wpy91beGXDtFS1bnEUoP1eWEtj9VDjwuAGSV61pdsAhgI7zx9E9uL18pv7Plzv142y7fajSujmXQCUVlaqrqWz7FunSsA/dUfXiz/J/wDs1WlLQtpesdd6hvziSFqYtlI1b2lDuklXmrP3BSfpxCkudMzYvUyO31D+TVx28sgTHD9onlhqPmVXfhtO2G8X2oKgny7fROVSpPQHYDH68foN5S/Bp5DcrFMrpeW1kulQ16kv3pblzdWfch1RbP8A+A4snpDlJb9I0bdLbaCloKVsBKWaRlNO0lIGAEpAEfTiBLfYx/LaT9FPisrzu92Pqvz1cpfgweIjmq62o6Ef01SuH/6i/wBUigCfqWjL0fUN8Wg5Tf2ay61xZc1rzHpaQKSN9PZLap8hXcea8pA/UJP247KUuiEhPyCB2gpn647cOtLo6QMYQesAQOIEl7qHbNwP36qbHZ4G+1krn5yZ+AdyH5dBh6usN31hUNq3JdvdxWttR9i0z5aCPun9eLS8p/CRoXk2yhvSmjdM6cATsBtlrZpnI9i4B5ip9io8HSn0hHpCNsnEjp9OHSg0iS4nYkIcUoJBmNqpwZ4gSVc0ntuJU+OliZ7LQF+eT4w/iPf5o+LzV+nqC7VNZYNNVhtNMkqltstbQ6huCQUh7fK+qinPQcRX4dHjpvngB5hXvUthslFqX9vWlVudoq2pXTsLX5qHUuqKElR2luIEfPJONphvxILtYbr47ubj2l7Suw6fGqq5umoC15Pk7HNi4bHybnErWB7LBgdOB7ou4Nlrasp81RIAyCfv3j7frw5U9MySNsWNsD9+9R2Frcl5X6IfCX8bDkT4hecWm+XdHdbrZ9V6noaV2mNZSBNtNc6wHV28VO7/AK7aj5fqSAtcJSZ6k34mfxU9EfDF5d2+qvdC5qnWN9qPLtWl6Otbp6p1tMKdqX1KkssISRClJJWtSUpxuI/NmNHLqXP3KFOKCgpJA2hBBEEHqD0gzIjtE8auZD911VqC4Xa8XO5Xu+3FXnVdyrKtdVV1Ss+px5xSlqMT1VPb248PDMQlDgfL1CjOrZSw5591130//a2dLVmqaGnvnJG/W2zPJZTV1dJqJqqqKZReh5SGS2gOISz6kpC0qUsFJKRB4u54bPiveHzxbM0LWj+Z+nmrzcGHXk2S9OC03RkNgFwLaeITIGfStQKQopJCTH5ZbuwqjcAgpVlJSn+Ufb+uvCYsF2G1jzBJKW1epCfrnjlPYad5wzyqK25yR891+x1FxYulIzU0r9PUU9Q15rDzDqXmqhM4UhaSUqTI+YE9OuOIvzq566J8OujRqLXuqrFo6wKqEUSa67VaaZhx9YJSykn53CErO1MmEqPQE8fmV8G3xUeeHgQtH7H5d6xVT6cVVGtXYblRtXC2vOlO1R8twbkiACQ0tAJSCQTwn8bPxGua/wAQa/Wet5kXtmopbAwlmhtltZNHbWXCCF1Ip0qKA84CApfUgQIEjiHHwzKZMF3l+q9ff4wzIHm+i62/Fy+PBpXkjywb0pyM1VprW+t9T07iXr5bKpNfb9NUqkqQpSXG1QqsVMtJkhtMuGTsSeEtO3KQFbiBBE4/U5/onjy0n8Q93IKiBPY9Zx/lj24UtN5PUqM9RBOOG612yKkbhnXmUoXW5SVL8v8AkvoOxrAjuf8AtH+nbhw0Rox/mTrOgsVLU0NHU3NzyUvV1QllhBEn1LVhPTv3j34aax4JISCOwn3zwidcCT1A3kAiPlnrI7j/AF4nyvA2Q2GM5ytN4YNuq32lLacWy4psqbO5BIJSSkjqMcNbx8xRKgBGVQZjt+v368KahwrJkqjuOsDuQP6/lx7tlvNU6CsScHb3Of8AHiAfO7ZFWkMZkrZbaFSyoKG0rMwpM9/bv/X6yG0aeqHXElxkpE+neCAOm0QY7SZ6e46xIOVGlmHdUW6orqf8XQtPoW80MF9IIJQB23dP19uLu/Fn8TXIfxN1/Lxnkny4Xoqt0/avIvZ/DoYFY8dpbbCUE+aUlKyXDBV0zJHE9lE9pBI2KGyXCNxc0O3CpzpynW0gNs+altO3a55Y24g9T1E/pCZ69ZPfLodPaWdcdSpLtShYQhRy6SdsxEqk7xGJUkgZAHHuisTNtbK1gvuqQElSCBux2OZMmJyTgdYmDc0NS/i3kBDgbYUZZCnAUgJCSRk5EDbAIB3HJKhwUncaOAvPtEbIRTNFdOGD2RzKuFqrUFVcbuLpWVDj71a4S+oKHmLUVesqx1KlEYjJGO3DPfqZNqriVIaUhMCEZBnMBXQESB9gY7g7n3vO0o+0n5qR4lKSr1BJKkgSBnHYQP4jh6asX7b0nT1aUtqQCEJJ9Ux+WN3pjOOs4VmeKlrKEuInjX3vRVTSDTvCedDanqKa2teS84XWFKRhPpJjdEYI9p7wPYy702unLU+p1CFJadPmjbkBXZM5kdMkmJ7cDWw1ZtNyW0pQAIKcIwcgnABE9J6/6yVyrWGxUNoJplubnPUCXATEDB9R9Q7jr0BgI94tOpxk05yi8L2t8pRIe5mWrXFpSxWPhdQhJKlElQT2B3Gc4P8AH643s8r2LolC6Cvf3rAkT0USIJ64BAgGcHIxwKmaS2XF1S1OoYIKlqKnCOkE+pPWZAOcQSOoBcbJVV2mlLVR17yKYepxK3ISkE9ZzEj/AAOcjhPmtpjGIiR6HkpxcCNkRqnkvfklKmKlSzuSE7mNu5IBOfVOBBkgjPbB42s8l9bUoU429TVJQCkE75X6oChnImTHTPeSOGex+JC6WB9LDzm9xMGT68dp7T1zGZJx3ea/xeXQ04SLhSpQoFKMphQjrA6jAkECI6dwPfDXZ0hoKiSmRo2TVqjSGo9NNPLq6NohMqSralIGSYP6xH/tEdeBRr/WztK04PI2KEz6AQDBETkZ75Mex4lWv/FZX1TIDlSxUhM7vVu2Kic9BETjH0nPAD11zbrb3ULUtlCgRuSENiB2ASf4Y7k8Mdnt85IMg3SncppSTqCb7pqL9t39igbo3krqnkMpVkwVKAHpT1iRhOZ+vHUzwCVWstHFmmu6qhbKFbWnFkkiMDr+nXioXwjfDXWeJvxI097q7S0/pDRhNfWvOD0Gs2j8MymCASFkOdClIQJHqE9aqDl9SadCfJbSlKTkhIjHDnGPDZpAVU8ROL5w3sjny718LhRMtvyVEQcDOOHrUGkKa+MhbYAKhBHAd07dVW2oTtUZSSBnKuCjo7V/mNJQVgpH5T3HsONkGOyg+qOXKqZxY2SRgdo4g990SmfU3EHI29v0/wDGeLL1drYvtJvRG5QgE9fsf69+IRqnQ6WStQbQPLHvtAHWST0Ed+w4wrNXVU38UuoND8j+SV91Dr922N6WpmSH6OsbRUN3JwpV5dO20oEOuuEFKU7ZzOACeOBfPPUHKzntq6uuGn9I0XKmprFjyKWjq3qi2trn/wBQL3FAVJktQhO2QiMcXI/tBnjksXiB5uad5faGvVtv+k9EtLrq6ut9QH6WtujyShSEOJO1YYaARIJ9TjkEjjnGXCqpKpSIzPvwyWynMTNWSCVCqoWyjDhlXu8LH9nzufPjQ9l1ZX809Mq09e2vxNKvT1C/XqqW9ykkBbwZShQUkggpMEHi3nKX+zo8k9HJbXfU6r1i+k5/H3H8Iyuf/tU4Sof/AOw9OHD+zl+KrS3NPkvUcmWdNsad1No1td0NZTuKUzqVt1076he8yipSQlKkj0qQgKSBtUOOoNFy7Hl/9MJIjHc54iV1fUtkLNWAuNPboAMlu6qnyd+Htyu5JhP/AAxy90pZ3EAJL7NsaW+QOhLqwpaj9Zngx0fL6QkLSpxKR6d+dvt+n04LTPL8BCjskJTtPsTPQ/y4cKbQ4Tkp3d54Evkc45ccogI2NGGjCE9LofaTsR2gyIAHDnT6I2lKdgMeoADA9v6+/twUafR6Ns7NpAAInBHCyn0gltcEAZkyPrPGi2x6oaUmjdx3QoAZwP8AP/Thwp9Ipj5UwDn6D7fw4ITGl0oVgfNhX1HCpNgQiRgjAP2j/fjFnJQKn0eCSdgA6H2iev34WM6XLYHowBiBk/TibM2ZDQEgEZyf6zx7XbG0JCjCzmPqOMXhKirGmkbAoJC0+wEkcK2LEEJTAJ2HOM/pw/pZbQSARMY7ZH0/rvxji0MjcMdp/wA/48YsBXIn+04/D205dvD7/wDr/Y7caDV+mq6ktmoHKSnSlF1oahZaRUVBEfvGXPLSFxJS7CiQkRxB0c7trUTuUNwgDoPp/n+nH7CedHLSx89+Uup9Eakp01OntW2uptNwbUmf3LrZQVAf3kkhQ+qU8flT8SHgm174M+Z2pbHqy0umm0zqBzTi7uyPMt9fUpp26gJacgBSzTusuqTG5PmQQCI4cbBWa2eG7mPyQeujLX6uhTvpxlL9vbDYSErSVKUsjJAPSBOZPQjuRjht1NZ0O0y005bKDIRtSSrpPpAzkgZV7z78L9C07lRQmSvy0gjbvIKsJ7jIxt69zjpiSOULTZLSEhbiyP8Ap+ohI6EIOQRBIJ7oyMGXuOEuZqfsEIrrk4uEMIyUC9Q6HXSVDgUhRWCQUmQEH2+sfWftieI5V2w0Z2QOskx8w/y7ffPB61BpYISfM2B9UhaCoEmMEz1HykZHbHXiF1nLh15LjimzBMbts+YfpnOfb/bjm23ucctCXq64xw7SO3Q6qEtKccLDbjLMnZvVuUMDE9/fH09uPiWd8JJIAgerpnM/+OJXdtB1DB3EEdxHbt1/UfXhmetC6VSiQdoI6ffjY0zmHcIQK6OX2SkLI2JHQH36E/1PGxx7yhEiMD0yTPt/X8eMcSULJGBGMYP9f58LdO6qa0yzc0vWiiuxrqJdK2agE/gVk/8AWRH5hGJ9xnjnJIGjAUmKEvOSmKqqEtgbiAI7d/8AX2/qeENS7O6em7aCeg69Y+38+NlUVTEdOh9x7f5/x4yho11Dg24HWB3ziP67cQXOLtgiLYwwZK+UFrXXOpwoARgCI6fy4nektDrdI8xMpyTHUjb/AOP0+3HrQ+k/xDwJbhKQSSkSBPTHtP8AHi3PgW8AWtvG3rZWntDWxusq6OmNQ+48sJYQgKAKlEwAkqEde4AHB630DQ3XJsB3Sxd7o/V4UO5PZBmz2Cm03Z1vuJRtAPRUFWZOe/369B2y20FAwbi7V1SYacG5tBTt3JGCfuAJMxjPq4KPiu5MX3w8c0blo3VVGKG5WlzyKunUQApcjsZ9OZGSIjIyOBJUA1tRDQQl2pQts+qcCB6QcmCBG49hAyCDU8zGAMaNggVBTyOLnvO7uf8Aha9QXRdXSkqKypSStwpSZ9kyeoBBORJ7ZxA2vV182vWtT+Gx5oWBjZCVBUhQBTtJyB6sHaT0leurl/yiUpCUCoTkOLKkhKgVbVhI6EBPYA9SYBPEDcU3WVyQWvxAeT5QSp7KVeXCQVJBCinbuIxuHsI4TLvOXyaScqwLPTtjjy0K6ml7+liqcQ44ksVAU0szKlE9CTA9UZ6mARHE75Jupo1XCyuhMj10zjySSQmSEZ7GFAAnOfYcQ7lXybvHMnQmv9U2+ts1NR6AoW7tcGqytSw+6hbwaT5CFZdXKspTkDHdIKvSFX+JubdY2t5LzCIdWlsrSlQJBMgj6SAeqfpHAqztEkPhvX1Neasw1PjRHONjj3A/kU8a40240866lDiCAdiiQAnP8VTB6EdMyclkpdd1WmKxKiFLS8Av0wSEpkESQMyAkq9gAJngzt6aReqVtatnneUFkLSSESJBj2A+mZycRwNeaXK521KIaTuQpRWpM4Mdz1nv79+0DjW58O5aXjkplBxLDKQ0ndJHrzR6kCXaN5DS3CS60pfloZImQIOZ9j3CZM7uPIuryHgFKJH/AKe4BS/mIwPuE9ROfmEcQWkoKq11DikJdO31KS50cVgJ9IyICsx+hmTw+Iurjill9KFHfkbNwCumCDBAB7nEEDhDqbH3TXDVNIzlO1U+i5IUhTzp8s+lM+kHGSR7wZOQZiY6R6u0g9dcmrXChCo9H0knoSZnGJnI7LqZ9mpQA064N235U9ukBJ69R1kAH+G01gbUEkuqIQHNqwAqJxGQQCTIxn7SeBTrJURguaFK+8Ru2BUSf0aKR8qcDjrQMBSFA7T3GAMZBjBj34ZrjpVVW3LQfCIkbsIAjp16A4gfY9uCM66iCp5CkpUCVHYVKKZghREScH+sceHKpFMyspS00tCSkJMbgYz/AAMg/wAI7CO0Tx82qPUUjJBqWeDHm5d/B14kdP64oGqi4U9KtVPX0CXik3GidGx5nPpST6VJVBhSUzMcdw+Vuv7bzy5Saf1laGK1m2aloEV1O3Vo2OoSvG1Y6YIIkYIgjrxwoqdu9QCUoSkBAWsylJjp7erAiB0meo47W/Deor9VeBPlw7qR2oernLe4qm8+N4ovPcFIOkx5AbInO0pHE2OR7x5gq140tccLWVDeZOPf/lTVykUw6SoylWMCCOHiw3pVO4gAgqiCmMK9uN90s4DihAVH8R34ph8TP4ht58Gt509YNGUdruWo6+ncuFeitaU43RMFSU00gEZdUl0Rn0QZEg8d42lx0hItPSyVDxFEMkq/NdztsfKvST1+1PfbPp2xUZCamvudW3SUzJV0SpxwhO49hMnMA8cwPi//ANoI5X80fDvrvlHymYv2qa3V1tXaH9VNf8hbqRpak+cllLiS9UBaAWydrYIckKxxTv4rvjia8dnMLSj1ot1ztWnNN2pLJo618bV1zii4+6EAlHo/6SFkbilJmJCRTC729TaiJCSCCo5nMmM5n+X04J0tMzIceaKfwJ7GB8/Psosthurqdj76KRsJUQrYSEwDCQB0kwB7cNPmT7yQMDrw53ZoNLJJnHv1nrw0vvAHtPbGBwwM3QKpwHbLpr/ZYGV1vxH7i2thx+nb0XcnFKFKhxDK/NpwlRWQVNEytIKSCokpJiRx+iM21pLZIIxII44Jf2TC/u2vn9zmp0IpQ27pehfU4pP74FNbtCEnptO4lXeUo47l/wDFxqQF/wAZ6jhdun83C4xuySQpEaZrzUylBUFSk/3evT9D/Dj6otNpj0iI/SMf19+IurVZxC8ggifbhO7qgqUsEhIKvTnGR/X+3A7ZdFLjVNoAgpEGInAPTjwbmhsgn1T3HSOIkrUsOKMnasyRA9PtxrXqMkbh6k9579OnGYXoUuRc0gkCB6h1PHpy5pLiVBRSDII9vrwNNf8AOGxcotA3jVWprtT2TT9gpl1dwuFTu8qkZHVagATAkYAJ+/A3uPxO+QFt5W1Gs186OXK9O0ydzlQ3eG1vLyQEpph/zClkgwkNkn7CeN2QyPGWgn3LV0jG80R/F34vNNeCzw/3zmJq1bgtFicp2iwyQl2qdffbZQ2gHqr1lX2QrhJ4UPGdoXxp8lbRr7QN2auNhu29vZUFLFZQutkJdYeaJlDiCpO7qmFoIJCgT+fz433xi3/iGa+Z0foOru1Byf007vYbdKqdWpqwK/8Ar3mSApCEglDLapKRKyApRAotbNU3CzMPtUlbV07dQ0ph5LTym0uNqKFKSYMQotokd9iZ6DhggsgfCC84cfp6Iaa0+IdPsr9S2p/jV+GHSWtq/T9bzj0wu4W2uat1T+GQ/UU7T7igmEvNoLTjaSYW42pSEQdxHB90pzKsfMS2m4advdn1Db/MWwuqtle1WsJcQtSFJ8xtShuCkqSRPVCh1B4/HrVa7uVRpijsiXUJt1vfW+w15aQpKlYMqAkiB07SeDR8Pvx5aq8BniGsmtrI5UVtJTti33O1uVK0sXG2qJ305SPSFJ3FbaiPQ5CsyrjrJYWaMsO68bXPzuF+rI1gU7ukeqN0nEjHHNn+0p82eX1P4S7Nom9XY/8AHlTemL7YLfS+WtxgNbkPvVXqCmmXGnFJSoetxaYSFJSspG3Ov+04afuPhBau2g7C/aub90rlUP7IuSfxVJZWkAKVXF0AJdSvclLbfzbt+4QgFXIPm3zk1Hz25nXjWmsbzUXvUuoapVZX1zxBcqHD0iBCEJACUoSNqEgJSABHGlptcrZfFl2AW1VUtIwN0TtEXHzmfLQCPLbCAdgEkmJI6bYkx/GI4le8UtOunpQtTz6AXVD1BQgjqCSREgDIJgyDwGNCauLA8rdCGwsEkCIgSRjt26R/IGmx01Ezb6Sqcfpq124IU44ht4l+mg7Ql3cAEqMBQifT3PTizraxsxDT0SNea4UsRewblKrRpf8AbJDjwKwtJJjptAyCegSNvQGOo+vE95N6e0baeZForddWm43vSNFUJXdqO3ueTUVjUEFttwwEq6HqPliRjhttC37mhDTaUIkCEoaO0+oDM5EHABwD2mOJ5yn5iWzkVzBs+oK7Ttm1Y3aKs1K7PdEKXR10JILbkdpIPQ524IkcOsdPGyIgD9Cfiqjmnnqqge/3oG8ydN2pd5rl2+ndorUt9xVOy64HHadpSz5aVqj5gmCSBnJ6cDO96IXWKCtnlpMmVQAM9QPacR9IHBu1zczrDU9wuDFAzQoq3nKgU7DZ8imQVqUG0SZCE4AIyPSPpxELjZ/wrhO9tK4CCSASPzHrJj0RPfAxKeA9W0P2TTQ0/gDPVB6t0d5bSitojqEpGCIGZHvkA9e3DFX6cU88EpbSqVdQMKJjP+P8uCzdLQt7e4pBPkYG4AFcxCVQcQMkzI/WRGF2wVr0NNoUmZEAAqKfyzORGMzPbrwvVUIB0t5php5dtTioAnTa3FICE7wqQmFyFmYGe3frMj3zxIbFoRaUNfukqLyQoH3BJGD9ZH6QZI4nlj5fOUbfnKStVQ4CjdtymACECY98gduCdyF5aaZvnMC1o1ld6iwaaqalLNyudPTmpet7KgoechoH17fT6UnAPEqhsxzrk/JBrrxCxrdMZ9ENbBpj9mttpUElQxPXZHSYGD9O/frxazwAePHWXw+9U1mpdJCkqXrjSimqKB9nzG6hoGU+YAQEkKBjacZ6HHAMGmWbpfax5Ky5bad1TTZKQ2pxtMknr0IxkmFGM9eG2/3E17pU4FinQs7UIQf3YBjKQIEmAMEgdJggnJYG6Cxw27d0txyFzw9rvN37KVeMPxNai8XvPi7631O9SqvV2cLz/lbUJ2hP7tCBAIbCIgnsR9eBnarXUOUq3mE1inW9qEfuzucCUmD7hM9FT6YUdxMDiWaT5U12obTU1cOspYaipWUhJQFKgqV3MqgGPoOp4uP8GPwsaG8Vfi5tOk9XNpXZWWHq9VCl0oFctsAhuTmFdSnrASOo4gXKP7pSmplG3Qd+yK2yVlbVikidv1PbG5XMTXdoradSi4yWfwju5W/OSlSyVEYMwlMxtPqM+mRGax1TD8uLWpSW5G9RDkeakp3QohMwckR6e8if0K/2i74a3JXk14Q6TXWlrFa9G3uirUW9LVKShuubUlRMozuUnYk4GQADHH58r22ikujTRbbbceWg+pwgpIASU4OVBJSAQIBB6ySK8dK6Rxc/mrNbEyNgDOStk7pOot9/rrPW079LVKaJLTjam1gSSAoKKYE9AqIwMGOHXR7irXeGFOJUoPN+W7Hq2GIJAOAPYwIIgR14d+dHiJvviB551Wu9VKp6m/3mpLta8wyltp5cJgACIT2KQTgmPfh35ycxaXmnrk3hmx2rTbNQlDS6GhbhDRQhKSsTkFZyT23ZxI44wSCF+tp6r6fbGaqmEb24ON9+R7eqsF4ZtG/8aaauQ/aVBTfsSnRUIpnH9jlYS4EwykYUYztGOpnB4W849F0zlBTgDZ5o2kJAGxUgGTMxP2n9eBz4eb04zXU5Q6ht5kgt7HPSR3CcH056kmVdTwYNe3KlqqaHBKakBQXkhC+gBg94VjrgdAOLIonRzw5VN3KKooapxaUDrbyhTc72djJeUFiBClQD+YxkEQTu3YxmOJVe/CY3dbck06gl1KAlKkypCpwEp9oMffqYPSXaUQhirc8wJekENK3lKR8sCPeIzA6nGJ4l3/Ei2G2yydkrCTEqJMwD9TPbE/y40ksdPKCCFtBxlWRuBB2CqJzD8Ntz0sXalhsutA+ptRKSg5xumCf9sRgj20Wm/XW5tW+jp362tUS0ilbZLi3zMGEpzk/cycdJ46BMst6vpFoqUuB1aMqURCZPdWAn8xgkjPUCBxH+Xmjr34f+a9o1vpJqmpr9p+oFTTLfpEPIQSCn1oPUkE+0YiDwu1vDOn+WSQmqj44MkZJHmHLoFSKuuz1kuamK2hXR1TDhbUghaFoUkGUZMpIOCD0gn5shM5WsvtBRP7o/lKtoG4dOsQcgAD7zxZ7m7ykuvOjmJetT3ZlK77qKpcq615hISgPLJUopSkSPUOsCAIkzwNtY+G+rs9OVtUj7IUCjrt6iFfKBP8wTERJ4ES8LTvGprdkwUfGtN5WSOAceYyhPT0H7WcLKNzqnAWA0hsrW+tXpgYg5URAEnoOp471+B3Tev7L4VtI23mUyhGq6CmNO4ylfmPMUzYCadD5GC+lsDcU4+UEbgeOa/wAKXkZYH/EDdNba6RTUemeVtr/4gerKp0CmpqlLqU0/miCVlK1FSUpglSERu+Utvjx+KVrbxF62uNq0dfLvpfl4y4yaSlp1inrK0sqUU1DjrZ8wFaxuDYUANiJkgyqVVpkMpYBhc73M66ytpIAAG7lx9eg+HNWe+Lt44m+Ta9NaR0jqZyl1QxcBdrwLbVAqp6ZoQildUk+lTiiSpBztTkdOOYPOvmzeedWua7UeqLs9dbxXqBfcc3Qlv/02UemQ2iVAAACBJkk8MVXRrStSkoWSSqXPNlaiSSSSATvmCcwSep4a6yzvhaN+5ZTnegFM5yAY+43dBPuY41bbHjyhFLXb6aiiAG7up6qOagdRtUVKAgElIhJUJxBV6oyevv3xwPtULTBSnoSQdvUfr9B7x29uCRe7QhhtQUlPpG8pTChEEH0jAzg5kwfpwP8AVSUpAMCRCUkwe8kiBHt0+v0HE+ChcweZDrtOwtOEN76j1kwR3AjGc/58NCmt7hzw+35I3rhJBOMQIz/t/XZrZYU4/EZkp+hPEwNxsqxqW5k2XWH+y7cl9TUPMnmJzFbfoGdKizosDjSqgKq6qqU8h9tQaSfQ2lLa/WsQoqARJSuOyKLqSJBycj69JB4/OH8HDnvU+Gzx4aHublwrqHTl9rRZL61T1IbYqGalKmWi+D6XENPradjr6PSQTx+iV5a2S42sKStpRQtA6gg5/hnhbuzSJQ48isEJYMHqnRd1UFGTuSoEAzmPtx4VefMUUpWN6oOD/P8AThv84qWQTJ6gz04+CFLnuoZnr+vAtbBqcRcySMkdlDrxsTX7YAUrarBHb7/z4bWG3KhQSlLi1kQQASSOv+/EF8RHia0Z4XuU1+1hq29W2koLFROVRpE1rYq61YBCKZludynXFgIAjBJJgJPGzGF5AbutXEDmVRH+04+J6z6b8Pej+V1NcaZ/VF/vDeoKm3SSuloGGX2m3nABAC33VBAJBJZKoITPHEBxP7zfsQFp/NthX34JHih8R2p/F9z41JzE1Y6h696jqA6ttrFPQspSEM0zQJ9LTTYShPUmJJJUeB082tIAAGcDM/x/hw+0NJ93gEZ59UtzzeI8uXlf7sAH1CZ+o48IQpxQJTEdc8bmqIrKsn0CfofcwOFztjeoassPJ8tSYmFBQAIB6j6EYnvniYAtQeiTUzG3b2kyN2P4nrHCxhlLRCnCpSk9j16dPrP+XGM7W1Ao6yMz/wDt7fT+saHKoIAKd0kelSjJP6Dj1bZAStyrCI3ET7R8oye2ONKVqq3gEpSRPTt+o40NsqqCCsKCTlKT39v8e3DpRMBCJMJEQBjpx2jiLyh1ZWhgx1S+x1KretsoUvcj5FBW1QwYiPafr/hwUOWuqEtrSStIQgAq3kpEE+wPuev/AI4F9MhTywkJMq+k8SCw3E0DqYXtkpx7/wBR269ODtC7wSMpQryajIKuXzGXo7l27ZBpjVStTprLG1W3N1ujLX7OrXEr8ylUlWCUmAVjIkkAdOB6m9LrHj5soITMOboUdw++4YA7Z+3A2suoVLphC3VKCt5ROCRISE4nImOkR1PQF7kbyH1RzkrbizpbTl41NU2ikdudY3RMuKFJTsplx1wpMwJEq6FS8EzAOtr3OGMofR2+KJ2rC6DeDj4RulvEd4JbxzCuWp00NxpmnnGGUbFoY2IBlwnA/rpxzi1/aGLRX1DSw0tNIqAtDgAcKiNuxMndAiSACAqSenE5s3ii1do3l3Vadst3rKSxVwmop26hXluABIVuSD8smNqhJj3BPEFprVV6vr3amq3eUV9VLVK1E4CTJTAnp7HpMnjvBBK7LSckntyU6+Xei8Nojbp081EqyyP6jrFuwkUnslIdLqQVYEjpHeJGJPs/WPQSLczTrISl1YSQkqlaI7gzMxtHWSB7DibWjSDYfU4SXX0GQ4tuFJ+iQYgYjABEe2RJrNpFLSEuuypRKlrEgYAEkxnqR16wr2gMNDw6NQfJzVV3jjA6SyM7LX4ZvDdT86ebumdNXG9WrTLF8q00qq+uhFLR4MFZxJMiOmSn9N3N3kqxonmne9OUNxobxTWGufoW6+lWPw9YGlFPmtLPRJgkdR7ZyXJu2mueKVN7mxuKkbDtg9x39sT7E9+GTX2qGdJ06vJUPOIKEBIKcZzjpicgxAPuYYJbfHTxmR58uPr+aSIrtPVTiKP2j9AkHOnkjqrltoG03q4WC423Tl8So2+vU0UM3HaU7g0uJUd2T9B0iQA9SMftCrbUGg4htaVJQlsALMGClOITJEjrAOc4NvO7x98x/EFyM0ny11Fc2rno/RhAs1L+CQh0CFN7nVp9T0AqQOkD368Bu6PI0vZlmohwIO4Heny1L9XynsSMzB/mTwlySOlcXyDAGfl0VnU8AhjbHGck4+fVK9QcyHbFY2aBrym1g+gh0n09ZicwARiCPb5QYty28RWouS+u6W+aduj9ruiXErQ+w8pDiV7UwoqHdMjIiSRISCRxGLzcF1fnvrS6VPAqCAknYArAwD1KcCcRiAMsValVTVKX+/KFrUv5inaVCAo+lI2gqVIMSpYyY4VL1cXVX4Tt2jkE6WC1spPxWDDjzKM3it+I1za8ZlLbmNf6vu+oKa2ILlM2/Vq2toAhS0JQkx6gApUnEzAI4gPIbwxM8/dMcyLw9rPRmk6rl/ppV/pqO8rLb+oQlYCaWlQgq3vwR6fUEbfUOpEBulUuqpUIQ6vy3WULUkJIIMn8o2mAdwIjJSqCCUkxq4bnHWGtop23tym1yQVJkq3iBCv7uD2TJESFF7Gt9lOoe52NSti60uqs6XGSVPU5MJnDQATKIIEYIPScdo4eqe/oqLNQ1O0NGfKWsSopUI6f3ZHZMkQT78NFpS48hTPlmpgggmSJwIOR1J/XcehJ4+K3W+oXSOKJSXEraWsFPl5jp+VPpEyZxPXji+IBxaV9D2q5amCUFHHkPqiltOoVmsqhTIJ3JKAFyqYg9iTnPbcQB1JO2pta2HUWolm3IrWrW4pCtlQ4HakNSgORgCcqInoMSDJ4pnoy8uN1Bplr3ONyWvmAcA6TkmYk+8k9uCbpvWpNQyipdVJbhveohxaTPzAHpA6yMDIABgzaa7wsRu5LnxLZW1sf3iPnhWYTo2lpa41NsW+7Yn3lPUFXUsJS48yJSArJAUUwCkEpBA691DrbCCUggAOl0grHpGE5Bge3XOCPpwM9Mc3nWbSzSJr/ADKRClbGFrISwV+o7RgAHoSB2mOFl25kOPU61BAcWgSSr0oKhgxIGSRjPsIEwXWOraQCqentrmZaQjto7VmnrRoA21yzuft9Nal5Fx/EENt04RlotQQVlXfrtPcEcSGzU9Nd9pUykJScFO1cqMmQTgARP+WRxWGk5hor2FrDqnVjKSHAU4k+3pyCYHuBjiTaZ5+v2UJQXHGlpABWrcUAexHWMdQDiTExMiKSFx3KDVEE8LC5i6JeCrweaf5v11bVXhpTlDSJT5iEHYt1R6SeoBAnBmBHbiS+PD4fOj7Dyiqr7YqZdGq3gFxrzCUlJOYnoeKyeDj4lyOQ99dcqkC4UVdDdUwp1KVgjoQRgHJwe3fideMX4vFo5v8ALxdisducoaOr9dQ888lSlAdRAwB1mT+o4kupa91wY6F4+77Z3GMdcg9VWUkkzQ4uY81OrynfvtvyxjmufnNfSVwp7BdLBQrUxZ7pXM1ldTCA3VutBYZ3EkBYbLjigBPrIOYECh7lwmnKvU8W4WslLcl8g4UEomRkZPWMkEYJmteblJerk6ufKgglK1AKWnP5Vd+pEiZAwOvEI1RzIapVSlX40pCVp/eQmQQdxz0gRBxAkdchbhRUonc9gG5X0LYrtcTTta8b4UZq9AMstrGxKFIlKwSncAAdogHcmMCMYMe44jGqLZS2fDigysnosBSlEEdhmJJgR7Zxx41ZzLqnEoQ3UtwmFtJY+VB2hKSCD6gAYMenPUHPA3vd3qLsSEqU4D6tqDlafcGCCAfbpGTHC5UzUsR0tCb6c1kgy8rVrC9U6gpDSQuDgtqkdYBlRkqOBgjP8COrxNWsqbb80rUcqyTJ9x1wPc9veOJ0dD1t2dUU0617VbRCTtT8wMrIABmJyfyxnHGxXLIJcPnPNrWRuIEq2n82foc4Hb6cAZp3PPkC3npzp85QXuVgVUICgkyepEGfv279Ppwjt2ng5UhO3G4gbcnt0+33/T3LWo9OMUowkubfSDv+UDp1wP1iJGeGG22eKlISkpgSYMQY6T1H8uMjpnPO6VawRRyZB5J65d6I8xTaxClBSVDPQ9ck/YfwxMcdqPhG+Je+c6OUV405qm7Vd8vej321s3GufU9V1lE/u2eYpXqUWloWgE9UKR145H6Hs7qw2UI8wyFbidpRMRnpAkd+/wDC53wpufdn5B8+69eprixa7Fqu2G3uVzy4pad9LqXW1rVPpbkFG8zt3ZIkx3uVkL6Jxa3LgM+qF1V4pyQ3K6tJSU9R16KAyT0/x4a9f66tnK7QV51NfKlNHZdP0Ttxrn9m4ssto3LUEjJOIAHUqHA85reOTl1yr1HbbAq6r1HqS7+SaO22cIfU8HQC2VPKUhhtKkncNzmE56QeKceOn4vGj+bfhm1loqwaf1Zbr9eA1Qtv1IpnKbamoSt07krMyhsgQFAhXWRPCbR2Ctm0uMZDCcZ5e9RKi8UsZcwSAuA5Ku3j/wDi98w/EZWXqyaSq6nRXLSvpzSGgQhCLncmgQpbj9SkkoKikfu2VAITKSVlU8c+L5TNrfU88sqd3KUSVFTiyeuVGZOcnJ6kk8EnVlPdb+8pbqS2Y3BCpjaREkQTtwfYYJnIBg14tlPbaglxzzFqwkBXb2+v6Dh3jo4qduiMYS6KmSRxc481GXaZTii22ghAVjqCfof58fRZShne4JEjamMAdvr9uHI16W5DKEIn8wP17d+3CKuqgidyj9ACSY65Pt/pxjgOa7brQpCkLGyYSc4AGcRHfhMt0JG1IPpx6TP8+nvxrqazzMJISB7Hr/DhTZLu1Qpr/OoKevTVUymWlOqI/CqUcOpiPUAI9uOWcnAXrnhoyUicdU7BBTkYAEgzxtYodz0ncFHvPT6E8bqelhcJgrOJmRPsP5dOHO327ecgjd1AzxLgpi85KEV1wEY2K0U1HsJTECOh6DHX78KW2ZEyBAyomeNymwlWY9/m4Q1tb+72pkDqRuM/rj69+nBAhsbUFYXzuyVvfuQpyQgmDjr1/wAuv+PCm01ZXUNZ2K3RMYT2M/TIGfY+x4YWlFdTJVAhQJ7Z/o/5cPtlShtLfllCirJlM9ziD0EE/rP3EXxi4qcIGsaiFpZ8W5tDQTUbkr2kj0qwDIQesSRIj78GTkb4j9bcha25VmjNS3XT1fd7cu1V7tE5sU9TOJAcZXIKYJIIMggztjrwDrI8VF5ISIdcRvSckzJHt/eInoSB168ErRNrVd/KcWVOF3BBUpWf+4kkmZA7D6nHDFbmiTAS5dZvBaXZwpFpzT/7Rq4KSloOGG0EKAglIknCgJAwcT1HE+sNgLTrLbIQXm8l2ApM+wJ+u72Gf4J9L21TqErMJBSYCfT+VOfftEdY6RJHE4tVm8gNpPoLchRgREA56Z956Y4tG0WwNbqwqO4hv7nuLCUmt1gSy00qJVEJkSCZ3Awe0R1yY9uHFu1l59SQColUkYCcd56dCPoP8HJdEmip0egyEgbSCCnrjpiD7e/XHCC6XhFgoy+VtoViCTlJyAf1k9epmIyeGcxRxMy/kkHx5ah+iMZJSLWOqaXSenV+tKFuDzFrkAIGRHbB6yDn7cCnmXobUgoLHcLjS+TQaop1VdtW3UJWaxlK1NFxLaSSIWmBIhUpiRu4Rax1XV6xva2WFK2NEhSk5nMGAYwQDkwMEDhsrLoq1Ujeyrc8xoBTSt0hlICjuBAmRP8AAyADHCFebi6qk0tPkCtjhyyR0EWs7yuSMOJtdN5ywpLphotFpSgtAwoZkxIj2xkg8QbU9bVXmC3O1SJTuJUpCAqYEEKB6kqwkCehVl9TUN3q+0tOpYRSpcKEQoJSgAyT0Pqgz3gk9+p7+KRyq5D8oLNoRPJXWdfqw3SzIqL8h4pP4WrMIIV6QUrUSRsztgZPCrXSF0RIOB+qdrcAyoaxw8x+n+VULU9yDtQ4tpDiEOSUBKZiDsO2QSrCU4EEArBIHVpSUs1CXyny22kLWlTQDhlIOE9QrrkmUknMAceKx4PPvFLiHVuqELKvMWgLAUpJgQomTMDMQSkE8aq65KZp3yhXzrSspDigt3bJKgIOEJbBM+qTkkdEuZ+2VYkEeCAm281CWlklJQlhYkFQ3IlJDfqAHoBAkHHzn8ySWNttNS7Spa2pcQ6ASEYUsEJ3ExBImSeme/QrawrFclCylpDywle9ClKbT6UlYGFKnIxOERmJKFa1NtbgyyXKn0KSVFQb3DcCBJSYgFWARtIwZ4FOdlFGq19GULo2Slxhbji1llrchSpGxEQMqjpBJBxKogiRrsabpscT+K8kvKAUYbIhRmVZSOsyJA78RfRN8S7TN07j9RteLcjzAougH+4ImAFQYgncVEgRxbr4fXgvu3jI5u09gs79PSKqWhVuVji1OhoJHqUkmQYJEHJmc+8mVoJVgWG6lrC0nAAVYam1P2dpipdQpryklQIWQnacmRAO0ycf/jjiQ2y6JqqJSAotOpBS4JTCUkQQRIAEHOCOkzPFo/Hx4Eb34O9eKsl8eRcRXNmpYrky4p9BMq64CiQQVKySoRxTK/tvaIvK0ZU15hTvTkoMA9YgRuKgR6v048FMfaarCor7EIQHnIPI9FPLPfaj5FvFK0YeAWFKViRnuYJxHeM9OJBqbVzabxVN2x2tqKMHaz+KWhNTG0TIk5yQB/dH3HA5pb05W+elRAcpWEnadxGAIiSQPcDp8sYE8SO06iU4wW6gJS6WikPEFSXRuSNpzhJV2Ez7HEFoZZGswhFcaSV+chOy9bCpeC1oJUVSkKA2ORtOFJzCZwMJO7oMDh6ptbI1A3RttNUDQp0JbCkKBVVrKioLWZIKlExJ27toG0HPCO3N2q6PIKvLaUuf3bioCjlRIJTIwR0gyTHsH1nQ9luiN4rmm3FwVNpJdKIk/LM7oExOBu7jiSyQu5HdBJqBrTvuEhGrTS0u1K3m/SpCkpTvUUyCkKVHcpmCYkdY4Zr7zGfdfV1DhUCtRQDKwQBlWAUzkgEkDpIw/wBfoS0UTCUftTzNxPpCxKQRlQITtAmE/afoSxXfT9pokEIeYc2wSlgnd0lQBGEgyrEY3cd9VQ4YD1xaKCJwcWclELxq2pUpNOULKEynYYlK49SiCDnb7SQSc5A4jdfeaivIHmKLa1nYtCtqVKBMlSlYBHbEdepPE9XpW1vEJWhLbY9SUxtwFRCVLHuDJ2/MP1H2k0vahUbkOU7jajv85DYEkFMbSogQZEEAmQQMTEF9vmfsXIoziajhadLAhqxZKm7PrKW1FKlpVJ3KkHr82SEzGI+4jh9t+i6ilaUpSEKTO1tRUlKSoJMiCAlUSoz6sj3PE2rnKG0UYCEMhUeW+FhLhBhM5nvlUTiehHEO1DWVl5q3kp3SswtBhITBIiJBIBSSQrHUn341/hMMQydyoc3GT5Dpi2HuX2kt6bxeqa3suGrq33UstNhyBuMJQkrUAkCTGfb9eGXXdur9MXeutdYymnrLW84y8224HEU7qFbVgbRCvUCJJORjoOFFBp119bbaVqcSte5DSBswBknqCSJPUwSJMRMhsPJ1qutyqmsXVoDbIcZKGVKC1ASEzAgkkgkYkdenHpgYwbDCiurKipOS5BW60jr7xSrcAlJCQMewkYg9T/Gfrw46OsbCKtKqtZCSsDYAqSR1iSAIPbrkZHXjZrK21FFWODyQwEuQlsNgLT0wB1GB/nPbhrttmr6twlhsulBASveClAORk4GJzMe3vxzZOyN+ojOFBq7W9zCC7BVibzetKU2j6NVA35dWw0ZbCwJPYZODj37HoCeIxcubv4BhSqVxlhMy0WhCdu5MnfAMSfy9MicZjunOVd1uOz8R5gZZKWlhCQNqiFdFrhKCMdZEicKE8Smm5SN6baTUVrlMisUNohAUpJyZEkDp0k5HU4I4OV96mrJQ6JgYAANvRI1Dw1T2+ItnlLySTlx33OfkOiZGK+46gU6KJh8oX6gYLaEpMYUTlSfUkQY94JEhlvDVPaadT1Q+XXylKUICpeWraJCQMAHtEjBg44ll4ti31hlpC6dBVKVrWlbpXgewAkAdNvzdcwY9dNKsUdMXVu+W7ghRQoOQTkEnpEScgAjv2FzSycnlSYqWJ7vwW7d0M9X3x6qQUNJdQyYKEGEmI646kD7x2zkDm/ocaCifVkhQ/vmepPTJ9z7cE3WtdT2111LJRugFaSfUknOQSBJye+MnJngVX1T1xdUPLIJUQoBIJBHacdgR0/04DSPLnbKbI2OJuCd0z1dyS0kgKBOQVbsf0ONf7Hqaiwm5hKPwfn/hiouDcF7d0FPzEQOsRx9fsrzpUkqWokwYPy/fHGpNuU04ScAyoR269D7Z+/Ggie7oh0tW1oSRthSwYT37YHf+v04X0lt2rJGVHIkEdv44+n+/CiioiYSlIBBgR2/zP68PlBYlKbCyN3ZIGZPtjp1j+PE+npN0Fq7h2SCjoNpSYmehERw6Jb/DsmY3EEgEwTH+PC6jsYp/UspKUpJkyEgAT/D/AB4QXVRW4pKQkkTAIyn+GOh+vQcEyxsbMlA8unkGybLjVlaylJICvQAcDp0PX/zHvwo0fqShsH7T/HWC2ag/aVseoKYVq3EmgeVGyqb2qEutjoFSn3B4TOUS33QI2lQI6fMABIHvj+OI+u2ntvp3JCvKWSn0rk4kn9I9v4Znga8l5RuLTG3CR0dKWfKnaAFASVxI6wYyT+nY8PFrpv3KJOwuKCJ29dwEGfuekZk+/C2zaddqllhhtBUoRAiSoZgHp16nv/LiX0nLattT6nHmKimG/elC1SsddqnI/n0kjHfjpFSvfyC4z10TdnFadH2NTq1FaSGwtKCmSAuUqT3ndEdSoRJAHUA1aD0+XgQsrBGXdxG87v730JIAyTHuOInpDRooyhIbSFBfqGwkkxOCInGOvX9eDVpLUdksXLq4W+pspqLvW1DJpbgH1b6RCCsLR5eQvzCQN3UbcdTw+8PUAZ55Oiq7iy8Of+FD1Sq33NFqAS0hK1IykhQMQmSe5IjvPtjsH2wXRx2uZStzaCrcgIXhIMAkQTBEASf7o6xxBrG887ucdSFvvrG1sNJKnSo9NoJKs/TqQMDiY21//hukNW+pZqCZSVLAQADJJwdyuhj1RuyDMh9irQ0ZJ2VWVVqfISGjfuuufw6/hqcsecHhNpNQ6iYRcbxekuFxwOwbeQSAI7KHUz78ckfGzY2NKc8L7pm31jL1vs9a7T+aj1JWhCoUspAgfKPlESesHiT2rxicwuVOn6+0WjUd4sNK+Ciopm6ksI3QYBBPQpjvMH3jiunMPmE5U3RdSuoZffW6XFKdTA3T7L6gkT9CcjB4WKqaoYZnzzFzXnIB/pHYfkm210EB8GOCANdGCHOByXHueyn+n/CfzB1jyNvet9OaTu1forTJT+1rq2gq8lfpEJGZ2pJMAqCRBJMjiuV81S9cXnEIKoVO5sBKipUmYSM5VECIwT8pjiyXJr4wXNnw7+FPUfJnTjtic0tqtp9p9yspCqvpi6CHShcpELbHqJSqAcGTitVgs6E0z1S+gimQ6FMoIiRvA6j0jPyqAGcwCY4UZK2WokMYGGg7e71Vjw0ENLGJHHLj+foldO83p+nFc6Q+4lKtiJSCmTJndEmZkgZnCRIJgOobs5WPONrceDgBVsSolSlFKkHaMiUz3lXYdZDtrTUQr9zXmJbC0ltkAypcHaencj8wGCe8nhitWlrpq4VdbbqKtrRRMqfrk0TKyzRU4MFwqR8qQoknaNo7kTt4DXOqyfDYdgmC0URaPGkG5SOhWmoDqxKfJRCkyVNrCikGDJAAyrPcTkiOGW73NDtI6gKeU6rYlajJwEK9QJAzComJAChkcLq2qabom0s7CsKGwlG/zITnB9MJO07QDBM4IHEcK2qirbLSQd6wFKcCnCkmcDochPuJPWIgA55PLgJhhZg5K8tVaXqcFwsBxQ29JPpklUGSpJJSQfoYzwnfrWm22UOJUolIbUlJTlJgwVZn0xmB1VkRn29dSEpWhb6kpSGwCqCiSYAPQJnIgY68JH0ht55AdZbCHTsG4EJO4BRmDMdZnHUexglS1Z+mqjaUbypwIUSpKg8tsKCZAAyqPyhGcFKgojcscWQ8Gni71H4U9f2++6dudLRXWlchJC/NSsKSApsj/wBRJkSB7/ljgGNWlTrLW1twJeUtLZSolYUAgByff0wFA5XJMJ+ZvYZrLLtSlDrexRJW2YjcVESYJgbgokzO36JPDHPQOaMrLbeA04CuR4uvG7qzxW67/wCIdZ1TQfQwGGSxLDTTeSNqSNwAJTMiYVPcEA/U1pZ1FTrlTcQpLpMrS0okgEmAVbSCqeg6SIjiMcwOfd65x3Rusv1ciormKVm2pU2whr8O1TJS0hASjb6wlABVgkkrBKscI9L8wm7bdAKh0KQstqcQFD1gKBKFJTvMmQAJOdvckiEZPD2TnS3ESRhp2C9nQ7lFVoVSPq2bFJR6wlZ7AQpXy/XEwDtOYIGgNLPXhva+CpYQnGw7QkErKvUJjI7EAA4kmZHy+uemNbWKrqf20TfKYNCmpfJSoVExvBUkwjan5ZGZEwRPE65Va3tvLXU7NzrrHZdS25lCgbY8tSaWqCmylKiU7T6FKQcADAmQSOJ8Dm+1nK0mkcMlqgOoOXbdjcDDTyUIW4krCVlxJkwEgwkKIJA3AxITt68Ru43Fy2sqVDzLAKgUqUN6IkkSDtKon05mJxOSNcL7RVBIUo7Q6skEgt7lEJ3ZgDdBECZVtOR8zRVaFotRMJS04PLUpxsLYAVtBBMyBkJG7MyANuYHEvwwd2ryOrkaMZQxuWs2kLKU/im2JhYUQorG4xO1RmUyJmevSeGw6wZYfSlSqplQTI3oKNhEnA6kSIJM/oDHE/u/htuaWEPp3uMLJCiCdySZxBlMyoCZG0Z9iYpceSz9AFKWxWMqeKzJUSEAABAidylYJED7Y6do9bNsIHXvmc4kOSKj1tQobOXXlLCVje3lJk42kSfdIAxCczkSHTd/OoKlKGXXnVFKVJKQlPlyc7TKQcqiUxIIz1HDVYOV5VUswyEOpKXN+5aRuJABURtEkJ2pVOd4wOoNHJ/l5bdMUrYddUmoUvchS0klO1JI27uqc9ComBJPXgtbKZtRMIn+UHqlG9V1RSQOmblxHQcymuzcp6nUVG665TVSA8lKqby1EpJk/M4pUpAkxsGTOEkjhwo/C9XlgrqEIQwAhCQ4VJTsG4wCpPp6YA29THQ8GG2a7sui6XznXGFLSkOEIdBKslJgAhWPf2IBgQDDNd+KJNwqagISy8XdywhsJHf1EkZMKBx1iMQDBG822mpX+Gx2pReF75VVkHjTxlh7FR5PI2ktFKX6hbGzzIO9SR2AUIGRCdpI7AHHqE9Ofh6aa8OT/g+vCdX/APCa7sF1Bua7mUoqUtx+72boUAE9CMzMknjlLdOZlfe6tTg8wqUN6FJfK1OKiPV6QQN0GMbQFCPZwtd3fTSjzzVKQYUoEk+mQFbcEyDPQDM/qoVUHiYA2Vhw3EOj0OJHLcc1Fue+i9Pr5lXYWdDtVa231uUyluAJCPMMEgpJJg/ZJH68Ri32UUDi/LRTJcA8kKKQ44PUUlCt2E9VKiM5T2AE3vd7tqHVOvVFKhCnNihuLu0kJ+nqVgSZIjGBwq0NetI3S2X5Vw1NTWJ222pyqt5cpis3qpb2g0yFI+UqCyRuMYIPXjyOiYN3lZdb65zNMXNP3hY5d6Z1Pzf0+xrnUruk9PtvFqovD1MKgUaUI3pV5WT6lBAP5fUJEQVQbmFR2+l1Rdm6K5Lr6ZmuWG6hLSg4tALgQsFIlCSACU7knIGSOIDqXm4urU47vrHN6hMCQ4QjJlUx17FPzI6CIi72ta68PJRSvVIp24Lfl7lLcgRMEjatRxmCIVlREK8lqo2jRFuUuQxSF3i1B2Uu1BrCksTADSUpWlG1IQkFwjduMJBGyTJyT0x2BgGp77W6geS0t3y1Pq2BlB3rRBIOI3bZCsjJj8wnhyt2lap5JePniUpWVEJAgKBz9jsMDbE4gSOHeh5fihUsLStbjhLZO1JXGATBwe0iDBg9N3HkVpqandwwFyq+KqalboYc+5C3/hAV9G2t5wKDilLiNziQZUJBEwSSMxGZj8zbU6SQpRS2hKhugQqEnbkyCM+0GB/Lg1q0Gq4+S08ryt6kjy5UPKBOwKCeqiOwPsY6ji2nh9+B1zB8TfIir19ZmEN28oWumYeqEocrggGQgH5ozHST+o4kzWuOlALygAvctYSGBcy7np4qDiS2dpJCEnsJ+0DH8YM+3EarbEfMWopED5s4n+vpweuZ2hHNDXmpoHabynqVam6gLSIQpKoOf0VPTp2zxDtM6Ho9Q6ytlLeK5Vpt9XUIZqazyS+aJoqhbvlJ9S9o9UdTHvjjnLTBq5MncUPrbZ8oSlKyVAwQY3exMxg8GjkHz61D4ebNrKmsdHYqlGuLG9YLgbjQN1SqdhwgqUwpf/ScBAheD34ilTbbbZ7nUIo3/wAcy08tLNSoBsPpCvSoggwNoBg9N4HGmofcq1qaBUlskBIIMKkCAR1VP0/w47RRNaMqJM9z3YTNdkOOqUC6hKimTKjIIEgGYjPvgduGty1CQlQDhWnfsg5T1BGe8AicfwB4kX7Dfq9oSHBuSSEmVAnInGTkHqInEcOlLoV1sphEKW6pCTnKZ7jGP1zH6caGmfIchdBWxwjBKgqLPtBU4kGSApK0EoSJggRjt2z/AIcOFHYgppBQkqIkSBAGYAgTkwc9ZBAAjiY0OhgKlpKw4qFbwABvT6pKiMSraMDvuxmOHG1acYolFSkLSUgKQoHaE7e5mB7dTH0njvDbTncKNPdgRscqQ+GZdu5ac0rFeb1bG7tbrZXNVVbQOK2IqmkupK0KMwQtMJmeihxcv4uXi/5YeNzmfpS6ctNH/wDC1NaLb+Gq3HKRqlXULUQoBSWzG1tIKQr/AE4plpehbu1XTNLUmmYWr1vJRv8AJABlQSYO0TkSCMntl5FWGGypltG0jY1CQpSjIIMdTJBzkkkxjbwepaOEFr3ZyP3ySxWVU79QaR5v0/fxUw5SafsdVrixUF3vKLHb6+sZpqq4lpTrdqZcdCVulIyoNo9UgmUgjHaUc6OWFg0Vzz1JYtHaiRrXTVtqVtW29tNbGri3tBKwAduCFCRg7ZE44GdDSppW3Ha0wlKUjy43z/8AEyJ9HzHtgAdONrWp6ispSzQn8FQEhK1KIH4lUwIT3j5gEiQYMgxwabKWbM7cv1KXXUeo6nHOTz/QKcV17t2hmkJQ4hypebP735incBCW0pCcT2AG4pncEkAsdw1Mupfcq64OM7UoV5O2EgfMoiIOTJATgE4joYpUXxizredSFv1LeHXFwAVBJAnOMmf06cNPMjmDZXdF0VLRpqG9RNVjrdXUIfSqlcaVCadDKQNylT5ilqIgEpiAJ4jTVohGuU7qZTW505EcIwDzPVbde8zn7hV1C0uBThlW70kJjad2O8GDGY+44gztS7dSqpdD61OKAS0JK90DaBnHUTImDPc8Rw3tC3TUrUhakIWEhbwKwrrIgdBn8xAmJnPF3vge+GnRXio8ZFksvMCqpkWQtVB/CuOhCnnEJBbZCiBlZmTGYMRPCvU3UzAud0TjS2ZtOGtZzKqTZtGOVDwqKoFTSE7w33fCgr1pBPqSSAJwOwzE/NVPvGiQltva03IASIUCqMjdJk9AfqImAB1w+Pf4GOU/hWpNJ1Whaduy193LgqremoKw2hMBLmTuAVJEf9v345Ya2oWG6FLrSw3uQdqkp8vYAARmTA749jHaJFFC2WlMse2QodXVmKtbDJvgoTVlCp1Zd3tsMuOOJSVStG6VAkiUgSMgq29Mg9OJ54dPHHr/AMIlt17TaDuVtoUcyNPuaXvyn6NNQ5UUbiVFwMqXJS4QpSt6SQNwiYBA+1RWtU9xdbQ+2ryX/MDiD5beEkbkKE+mCSInuT3mKVVxcS3uKVkJSlpxJTLaiVZIPQgbgAlXcjuSAlVbhqIVi0bSWgrTVVSVtJSGkObdq0ArKwpCUhQEDpuEGTklMCIJGtuxu1d3dt1CuouhL5ZZ/DMEOVCSIKdplad2PSATgkgTBQ1z6U7vKBKyPJTKxuV+UAQSZOSe24xPfh/5M84NQeHrmhY9aaYqzaNRafqzW2+saOKR4IKUrAUkjaRIgzuJ7dOBxKIAKPsuOW+6OI2JbcCvKIU0FhCyTIUnJ+UEJ25Bjpw3u1RaaR6i2W0oASViU5MSB1gT1yMdwTwoutTUXq6OVDzrztU+tTjriQVKU6pQK5KcZKiSR2P24SfiF1oZ85SkturK1An05hMwcAnAn2A6wZ0Kz1XT3x9+B+7/AA/ufC9B3i+W+9LRQs1Ka+lSoSl1EbFGNycJI9WAAekiAdXpbr6Mo8unUtpIHlpaKEtBRkJSMnpJABgwDKtxHD4ddVevWU1V3qX6p14FtbtQ+HnSoN7Q2A5J2gHA3TCSokbEgtVfTpTWhLTjJp9wH4cLAQ6CsBaACTIkpJ69IAJE8WDE7U3D90rVFOY362bBRS5MOUoJQ8EUyFbVIUJlM4BHqgQUwYGR3JMIKy4ipS8yVKS/6/JKCoBBUZIM4iSQCCFSJkpB4k9PTKeo0pab85lCSpaXQkkwJCcYACRMAYBUenDfUaVUXgFNoUVuKaKUO5CVeiEk+lUkYxB2me5Aiut59toRi3XQtIa8poqb7UtqU6HFOFsJcQqJgJQAApBkbSUCEkhJ3AJTEw40XNG50byqeqXvSkHakqV5iwYSoJXtMEwCZ9MA9zw1CzPOpaQW1IUhzalQktnamSd0GVbjG4iMx1HpS/sU+WoBThp2l7mkqUk7TJKk5BAXG0bT1gkiIhf8KRvspmiuDe6k55x1LzjPlvVCCpSQk7pKFFJhZJkiAE9twkQZHD3p/n+7RvB92oWghRCm0oLi0K/LshQCSAIzBhGI6AXP2MIBCP3LmUBQchvOIJEYMmEiTkGIBlN+x1lzY4hKVESjzDKNpAEzMT9gAM5kYz71LHyUqORsmDlWd0V4ylaatpaefBYXgjaE7IHUyYCZ2gT7kdJPDoeeln1PUOLcqXApa9iwl1ptCyBISteYylcqnG38wA4CV48NGttHcsqHWNVa7hbtOXFXlIqmWFFDK0kHMqyodj0//KeID+9aSChbYcbIhPmrSUjACdxBkxnEYSTI2xxJjus4ADlu9oI55CvTyncteu01q3rmxTtopws7lqQREH0oT+ZJJJhMgnrIkxrUevk2CqXT0qm0KYX5aVIaUreJQBB6boiZmeh6EGplr1VXCrCqd8tIcT/1QA2kpKoRtCSJEpUTnqZJ6wutdZXXBKDv/dqhspMrUuT8oKiQkkqJntkDrgrFepS3DBuhVXDTkZcOSOt45oU1Y8lTtZU1K94UttpQndIEiJO6SSQBMkwesxuu5p01O2tLBR5iUStJCYQdwweyDJSZwqScSeItTaWdvDLanAypLY3O1BJcSNqOoBJAGDnr6QCT1Hi4cunUKaR5LLK0FSAkErcaPTZuIBWqYmIOFEyAeMlqqlx1HdQ2eByZhOV45xVyKZxJcNOlO8k7FHCSd0JMCTJM5BLsCD0YrxzZr6h5QVU53BtsqJI3xhKBPqTAgbfSO27B4dnfC1qWm5dta1ftdYxpsXAWxNzCUltqt27koJPyqEqJgCQCR8oAhzWji04lNHTEu+UtBUoEqTtmVKxIAAVtEYKR6sZjGeY7LdzGtGcpS3rKqutXDZeUtwBPr9RUqD3TgmffqQYMHj5+KqrygtNuKdStahtQk5B9iqRHo/L7xuE8OVl5cVVwec8xUlW9S2RkdYTu2EFQgHCj+WYJIkh6K5NUvmoNQ6K1xJCyhO4oKQBAEAST1BIBH96eJ9FZqqreNkGuN+paNmXHdQrS2jm7ptqAyp9zeHSt9BAAEQcSYEgQFQSTAjgjWTQYpaBK1JbJWhMQkBpRIG70gZJBgACDAIz0L/KDkhp/Utm1M9XaksWnntN2Z2606LhUFtV2W0oJFKykAfvlQQBP5cTmB9cOY9FTIBp0NvLd3ALWRCxnAUDJ9KQCEnqSI4daGz0dICZCNQ/VVpcOIq2uOmAYaUpoNLCjdDiQ2Gqckh95SUhCRISRgkExEEjAAjvw7aeotFq05flXTVSbLW2m3Kq7ewihW4LvUhaQKZKgSEFQUsgr6BJmOnA3vGsay8kuuv7Q2gLQpRAbWvegle0TIkKzIUQmOs8MNUlKyhRadeJAUHcpDaYPqQkndGTJPXAIzB3nrmnaMYXCmtrtnVLs+ieKzVKqi4FSEtBQUtKoSPUYn0jBkwnJyYVIwAb6eFT432rPDP4alaBatdnrkU9O4igfckLogo+ojbAWmVAgDIM9ccc5F3sUrCg2xSMlTkKCmgVISEqKjBVA6xCifpA6OXKHSi+b/M6wWCqvFLZaW81rFE5X1rxbapm3nAkKWs/KhJUZk9jkxPASsLZR+MM4R+mzDnwjhe+b/MVjUeqrlXuFl6prqlx5SgkDbkkuSrO0kkwBIkZE4YdN8jtVc1ba9XWyzVVwpGE+a8aZpag2M/N3iBGR24JfxHvB/Q+BbxIVmhrdq21a1pKWmZqkVlEUjaVoK/JUEqgKEdAYIInPF6Pgj/Fb5R+Dbkhq7T+v6ByjuNwqE1TD7FKahVchLYQWFEiYkSJ/vHvwNqpHOGYm6s4RODS04kOFycv+n16ZuQp3kKdqW0pGxKSBuIg7Se5xJEAmdoAgcJKO2ru1UhsNEpcWCEIAUpeBEbesx0j6nJwUvFPru2c8ef2oNR2W3G02+63B55mjaAKadtTpKUo2kDoY2nAM+5ATaT0ApTqXHVpNQ6opASkLIgABUYgkd5z+s8EKC3vmeGgIJc7tHTtLiQpd4ZvDVWc5dW26z2+mFXX3F9FPTtoTAqFLMgfQDPTp17Rxfnm38AXWfKjlK7qPfQXD8HT+fVUtLhTQgFRAjMZmOvfgOfD95jU3h2506Y1JXMB6ms9Wl19JysJMhX2I3GO478deufHxVuU6eRFyftt7buVyudCtqnoW0HzCpSYhQ7DPDNWUdwo3wR0dPra/2jjPX6bdVWtPdbfX/en1tWY3MHkAOM7fXfovzx8wtIv2S6+TsKvKWUKlMBJzmT0Oc5MfXA4i9DpOpu1YEpaWjyyQtcFKUAfLIPc9P4ZweD/zBoWL9qh6qO1G98vKKsATJEDsAJz37e3EbvVytWnrK4EIUrYFbpwlZKSDuI75Eye0z6eC9RZACXyENaotu4kcWNijaXOKg9LokUTLUK8pttsqQt0wlSgQSe5JEiCDORJmeEVXdqS1vw1tdcKlNFwNlSAgETtAJUT2+p9xI4W6juFw1Ql5Sj+FpEggNghGZCJBGQIWemOmZnaz195pbKk+Ult1xKlOIVI3LEiJVHSEnA/7pJJkhZjE04ZyHVNVO2Vzcy7nsP1WxNmNaymruZUimIKW0AfMTMIxgYMnMxuJ6EjRW3ldzqHktKShptMvHyyskZ9IPUAZBznvPThhcvbt7fW4VhIKStb64KWkwJSPr1BzjiL655koprYKWl3oo2kFRQVDJ7E46n0iDI6AlJE8RJayGFhceX1KmxW6eolDBufoP9pXrLWdNbWjToCkra2oXvPrWSCBtI9x9gBkkdRBa6sqa4oeqAGE7fM8wKiIAgj2EKmB6vVkAAwjrKx78Q8p9SkrQVl0mQSZBKZx1UQJPUkSIAPG6lt4tri3Nhc/DuqbKXUh3zFgHaYxE5IBONoKpHCDX3B9TMTyCsy22qKkhxjLuqdqOiCHVObS0mnDTTqAlXpAIykHqkekCQJMDHeXaa5qVvL64U1Ta3naaqQUvMqaUptwGEhOxSTKiAMiYwcxgD126ob8kqhKVPqc9RCggSFKVuIABKSTMdgCO/H2kqg3TF55xlRXueLJUd7YETAIiNpAA7dZSN3G9POGclzqKXxOYRh1hz71DzJqUVupL3X3N0qS0l+sqlvLkpxC1yI6GBGCDgSOBrrrmA/W1DiGtwLSf3ZCCQIPuAYE5yIxI3do9e9Rfgmi048ErUCVo3StQBhRKpPZSSI/VOI4i1Zc6httaAlbKlpkpCQ2CoiJIkZxIPaE53CeNq2+O8IwxnmuVDYImzCd45Jdcn3rhUPpLakBLa1OiFGUneohSVKjckAHMRBMZVLReFr/ABLjKXijzVo8zcd/ykwdwJOSATEjIHYHgweA/Tug9eeLHQdn5o3den+XlyurVNeaxDiYZYKVbVKgEJBVsTChCdyjETNqv7RP4a/Dj4bOcWkKPkBe6CtZu1p82/UduuouVHblBSUtOBxKjtccQDuQqRASY+UBaOT5imgYBAC5x1G5x5RcVKTlaZICTgEkKEKyDj6gdIHHSDwh+BDwwc2PhNc0+Y+seZJsPNSw1FS1arZ+PaS6HGm9zLYpzKny+Soen5ds8c5326r8SmrSy8lCyIWBtTuT6iEGACQDgAYCUnthE/dnkUzO6EhnaiVD1FMkgAg5BxMZxOO+rTg5Wzhnkpbr7mBbdcaO0nbqfS1utN4sjD7Vxu1E8pTt9StaVM+Y38jamW5SCkkqCgSeEnPTS1i0DzJuVq03q2k1xp+kDCqa90lvco26vzGmnFANPBLiFtrUplU9S0YJTxEFn8QraooCm4bG31JjPTMDoAD0j9Txo9T0ubdnZXqyepjrnH+A9+MJysyrsczfEvaOcHMav1DatHaf0Ja7jSMUyrPaajyaVry2EoW4CtSzvcHqcELkqA2y4Bw4ou72pqNbVS4relDrqA4yE7nV7IKVhJVuCl/KpKgMbVEE7a6aW1Ahyrp3SpQRUtlCoQUNGBBQCDjC/USTPpySsgE+vtSNGaT0veXLtpm4q1JRPOop6GtS9V24NOqaWKxmUeS47s3hoqUFhZODkscFW7n2UJ7G8ip7bLax+3aQVLiGwry1Ph353uikrKcbpTtMHMe3Bg1Fyz0zedLKdRUodUtMHykIHRCQobzhQAMEkHrMQSDXzT9+WpgslldQlL/lOB5twqB2j0yCCdwXjPzKRABnbOKjm4bly8t2naez0VA7YXqmoVXMU5FbWodDY2PSY2t7DAMBJdUCZJh0snEVPA18U0evUPkkfiPhuoqpI5aaYx6TnA6+iIvOnwCcxOSfLDT2t9WaVq7dpjWIQmjuKsl1WxSkJVA3pCh6hv2lXq9QngTVegEVIUp5a3nIUttJQklwHqorBBiSB+bpH5uCx4h/iR82PEp4fdJ8uNS3+gqdL6R2LoVt04Q44lCNjbjqmgoubWVKCYSkggmM8Aml1BcLCkofSpKvMPqQ5LYViTMztGBuB2wrHvwPgNO/aZmN+i71JqogDE75pxqOUzyWQGfMZ2JTiVgGSem7cP8AtJAA7yTB400XK1yjeSUISg7/ADAjeQCBE+n2lAKsCRuwDHE95Aau0TqHU71HzCut80/Zk0FSW6m3UYq3V1nlqNO2tE/9NTiUoUfyzMpT6uG2jvpqnEOPKbYcSsKKkJMqJI/MkCYIjqJ6j34kNs1HLks9FBHEddD7aInMrxe61154TLPylrW0N6YstZ+JZWlKXDG6QnHSDt6K7bsDHFdajlgKdCXVIDOQWlOjZ5UZ2gbTkkyOvQAk9OCFW1i0KS22ttalpOTASkSCoqgSAEiCOvXAxw3UGoaCxakp66so2qyjYeQ5UMqUsofQIUpO/sCAUkyFRmSRPEaexQMy7CIU/FVTJ5cqMUmjNjKFlaghKgpSA4dqVknaQFerqBMAnIEz0lul9Aodd2rbSmG0ndMg9BEd8AnBBkBXRIgicw9RNeLfxDX1/lLyzqrLT1AcraTTdtLlW7RsMU8vrxBUJDqzEABXzTgOPh61JYk3FX49sJWhxKFbVbiIGAmIBSEwAPvOOk6yWqCombANie6HcR8RVFFTvqXAuDeg5plodIooqR5RKw6zMJU4CVbk9ciCogpSDuT1SACY4ba6wuVbnUFS1QFMoCd53JgknI/LBRlI6nqeC1zJvFjav77lndFRSqCgkqZ2K2QEpXGIVJmZMT78Rmr1S21TLh9xaCVJVtAAdG443dDI6iTIjIM8H6nhtjHlpPJLdBxxJJE1wB3781B1aVvFz08baaypRbVLFT+DL6gwl4D/AKgRO3zClJzM/N177aDkq0w6pS3VKbSopU4kJkCIVG8JEqAmIIjcO44fKzWdCl4JWoguQ4oU5IU3KfmGJPXoTjaDBOSyXbUNRdg6UMU6AILinlFtpsp9KgqfUScnsRuVEZ4iG2U0Z5ZKmniWslHtYCckUdnsjSE/MtKgQ0D5ra0gnBmMED7iSmTPDbeuZaUUv4WnpVAiQVuulanTtMkZGCE9OhIEqGOGt2203nbqm4gtok7W/SYxClKx0ke84k9uNC9Z2jTrS26SnogYkIEOEHAAkjpP92AYBnIjsC5o2OgIbNUCR4JaXu+iT3mvvWsStbqXVNbhmPQ0DnuAJCY/KN0caG9KKSy47VVakectSpEEOgHHX1KIG3MSYUJMbuGi981Kisd3NNhtIJ3VDxJISZEJQI9IAkTkQfbhLqyx6itOkbPqCrdZVb9Qh80ZYqWlv7miG3C41uLjY3KTAWhO5MlMgSB01RTt5kuKIwwVbwAAGD6p5uNVTUdSVIaUQJhdSvZB3dSSSRInIJMHBE4YL3qAKWtJdCHQsNKQlA2LV/dM4lWMEHOfdPDPT0z90c8xwvKQg+YlO+VqH1OYMdDAJOI6J4dKbRRQlC1+YzTwdqXG43JEhR6QTux9SJI7cRw58v8ALauznxQfzXZKa13AMuIUhourKlQVLEuJCjncRtAOe3QTk8am6qrr/MCFKebUoqcUBO8kzJHyicH69568ShvS9tpadYJW69O1aHQAhgHaUGAI/vEg4iPm4V1KGrXTpKklTu1C/LCTAB+UlShJmTnJg/fjYUJzl68fdWYwwKM1Gj6hawalCgNkpLiyklEY6diM47YzMl0suiXq4LlBQ0mCpayUBXQ4yEnJGZzEAGOJHY7Ebw75r6EBCyklJgKcEzBIMzgZxOegJ4mFo0+0H3F1I2q3iCvcFSTAMYkmT0jr24K0llD3ZIwPVL9z4j8JuAd/T9FHNJ8tUpZBUgBxKtylLJBQTM4jHvhMwD0kcEiz0NDpy3pS6yhbjigpZBA2pzJ+oyM46TnoH/RnJ+5aq5aas1LbBQVFs0Wyy7cQ7UtNPoQ6vYkpbJCljcIJSDtj7HgXXjUKHEEVD4bpwobokFZIzmDIPt1jMcNFP4NID4YyRsfRItQ+ouTsOJ0/mpQNXVFSSGUFKSsALSoIC4zHvP8A3fSCOnE35Cc+9OcteZ7FdqjRrevbKhh5ldncqlU6X1qbKUuFaQSCgkKg9oMjB4DaKev1KyNjAo6FtJ8x4pgARClDsrp3Kj6pMdvtRqe36etqkMrWt11vzHSuApQCyEpVgERBJAI3bs9xxCqb2NDmHJz6kfUbhFaPhUPcHEAY9M/PKfrle6u8PbWB5QSvcYWkpBzAKp9xBI9u8kcRO+XujsxK11FHV1TcBkBY8hoA/UErJHsOgSSep4jGseb9RVOJpKVQNIQAIbW0l5sFSZ2rKlAjPXcTOZMAQ+vrCismoKnKhTaZUtsDaT0ABEgQSQJzEnsCs1l4mndlyd7fw/T0rQ1v/tPV/wBf1lY44lsuLWAStSVGDEAkGCTnbg4x2nEZW+usfacfLjhXKdrhBSoASIyBsAnqQOo77uNdK6h5QK0hpICU7lrCUpQSYPtjBJIkwB7wjuFf57bhDaPIQrzl78hcLgFQ+aDunqRJABgmA8tTpGuQpjp6MuIZGEmump3a8ww4UsFQKwkkAp3EeoiNxicDunIAJPEVHlqblAWlZaClbUNkNgpPcknJwcA/NOYHC66ViHLgoy3UshxI3rbWTUtJ/IITuSFETJ27SoCAUcM1UVuLaZQSp9W2nQC0loOLBUFLBCgEeooBJEQTKsAlQr690rtinW325kDduaVOsihaCnUrZQEqWUrBKS3sSUqAUTBVuc9RTHyx88hHX3NxRcpw2ZbXKEJRKUOGUyNpAwUExEQ2MQI4SVr6nN623AG3CopLbYKXCAolyPYSlIPcJkiAY1vLUKJ5e90tpIQBEpG5BAWkKTJK5UoElJwkEdJGiXCI6M7JYh5TCg88pxPkblpb8zzFJVu3+YU4KmiASNpyUJz04SXS6O07boSahJaP76HE7Wdx2qSpSTCjCkgHPUpPRQKC5VqXANxSkJCH/wDo4dkJ3Ix1CSRAxuCZx3bbm+l6pUt9aklSNiFJWHFEDYCd0dQEnviAYHfwzu6LBCMp0oLZW3x5VPQMVVatphbpaoWFvOIS0hS1OKCc7EphaiISkbj2KeGRTrTTbLhbS4pSDKSUFtQBKj83ypmT0g7o6Agyrlhzu1RyNvNzuej75WadrLza6vT9Y5TqG6ooKtHl1DCpGQ62SVGJGIAUBxDK15KEtqa/d7AXQVJgxO1PXB2AACIwD7ccMldgMJVR3B2ibDoUlgtKMlELQDGSqAPTBSIknr7gceaqpaq1uLDyVBLhg7yd0jr2OcSoAElPQmeEVe8tK1rG6Hwpzes/NBIEKAz0juCQT7cfKja1WLCiFlFQoJJAEwCJKuv6TGCep48WKa3fxAapv/Ja08tnryv/AIO07d6q/wBBb1tso/DV1Uhtp51LoHmHclCRtUspGyREq4YtIctr1r+k1BUWWyV17p9OW1d2ujlHSKWi10aVoaVUPFKf3bYU42lSyAAVYIJkMjZkZS2D+Yx8qRGenYDMZMx7jj7b71WW2iqksVD9MiuQaZ4NOKbTUtSFKaUUwCkHaSk+nCcSAeMXmegW+0WOr1FWoYpqZysqXHEBLbTe8vAdCAnqPYAdBGSDCW+WOssFc5S1yV0dUw8ULQ6Cktmckp/rp9eJFR1N95K6xttexUs092pm6e5UjrLiXsOJStsSCU7gDlJEgjj7q6/ap8S3My9XutS9fNS3RL10r1ssoQVJbRvdXAACdqETjsDkk8YsIX//2Q==" />
These are GGUF quantized versions of [Envoid/Fish-8x7B](https://huggingface.co/Envoid/Fish-8x7B).
The importance matrix was trained for 100K tokens (200 batches of 512 tokens) using `wiki.train.raw`.
The IQ2_XXS and IQ2_XS versions are compatible with llama.cpp, version `147b17a` or later. The IQ3_XXS requires version `f4d7e54` or later.
Some model files above 50GB are split into smaller files. To concatenate them, use the `cat` command (on Windows, use PowerShell): `cat foo-Q6_K.gguf.* > foo-Q6_K.gguf` | {"language": ["en"], "license": "cc-by-nc-4.0", "tags": ["not-for-all-audiences"]} | null | Artefact2/Fish-8x7B-GGUF | [
"gguf",
"not-for-all-audiences",
"en",
"license:cc-by-nc-4.0",
"region:us"
] | 2024-02-08T18:23:39+00:00 | [] | [
"en"
] | TAGS
#gguf #not-for-all-audiences #en #license-cc-by-nc-4.0 #region-us
|
<img src="data:image/jpg;base64,/9j/4AAQSkZJRgABAQEAYABgAAD/6zLqSlAwAAAAAAAAADLganVtYgAAAB5qdW1kYzJwYQARABCAAACqADibcQNjMnBhAAAAMrpqdW1iAAAAR2p1bWRjMm1hABEAEIAAAKoAOJtxA3Vybjp1dWlkOmI5NWU5NWE4LTM5MDQtNDBiZS1hMWZhLTAyMjdiZmFmMjU2NwAAAAQianVtYgAAAClqdW1kYzJhcwARABCAAACqADibcQNjMnBhLmFzc2VydGlvbnMAAAADIGp1bWIAAABBanVtZGNib3IAEQAQgAAAqgA4m3ETYzJwYS5oYXNoLmJveGVzAAAAABhjMnNoKLEwSV8v4AnyIFxkhZX2gQAAAtdjYm9yomNhbGdmc2hhMjU2ZWJveGVzjaNlbmFtZXOBY1NPSWRoYXNoWCBxVjrYAGFAft6cbzFoNihL03EKUgxaeSte2hy3A2kIFWNwYWRAo2VuYW1lc4FkQzJQQWRoYXNoQQBjcGFkQKNlbmFtZXOBZEFQUDBkaGFzaFggpbugkv/qGbiwV/poyU2sCfAca0V95uZnpd4l6rlbtBpjcGFkQKNlbmFtZXOBZEFQUDFkaGFzaFggow7IX4ji/mamhqg8tPHBxtfwMx9F36Y1MdJKk/zMChNjcGFkQKNlbmFtZXOBY0RRVGRoYXNoWCCz6roXFIfVkeIcYArL/vOFvekosqTDT8EkyxRHkqwE22NwYWRAo2VuYW1lc4FjRFFUZGhhc2hYIAriPEsLiEpOgbUZmp9jrt497yq0OR7gZ2AX0+76cthNY3BhZECjZW5hbWVzgWRTT0YwZGhhc2hYIEOZIfOKtsMxQseaNacrnLcTTB8dbyaLAxbzRV5+xedVY3BhZECjZW5hbWVzgWNESFRkaGFzaFggi33gSmKpOgwEasSxD71IHV762BXSHCZWJT7LoAa233BjcGFkQKNlbmFtZXOBY0RIVGRoYXNoWCAqabdjzjoE8U/M1CiRaSX4mnCr/X6/ak691R9jR4ft42NwYWRAo2VuYW1lc4FjREhUZGhhc2hYIPNXdg8p1I1tkgB5cC7JtIgNxzQTJvEup/Xr9nXv0aHsY3BhZECjZW5hbWVzgWNESFRkaGFzaFggYn9oeDtfUFm8zmTdUOIdC564oIMFHxBdaoCnYM0qVk5jcGFkQKNlbmFtZXOBY1NPU2RoYXNoWCAXkRMmLxrYvN72JWzsXSxSv9k1fn1wmYdPXtSZa7Q1u2NwYWRAo2VuYW1lc4FjRU9JZGhhc2hYIM3mbnjlQZ3qdN989D2aqHa4xmnUAGeZLnGc75CsXz/gY3BhZEAAAADRanVtYgAAAD5qdW1kY2JvcgARABCAAACqADibcRNjMnBhLmFjdGlvbnMAAAAAGGMyc2gosTBJXy/gCfIgXGSFlfaBAAAAi2Nib3KhZ2FjdGlvbnOBpGtkZXNjcmlwdGlvbnJBSSBHZW5lcmF0ZWQgSW1hZ2VmYWN0aW9ubGMycGEuY3JlYXRlZG1zb2Z0d2FyZUFnZW50eBtJbWFnZSBDcmVhdG9yIGZyb20gRGVzaWduZXJkd2hlbnQyMDI0LTAyLTA4VDA4OjE0OjMyWgAAAnJqdW1iAAAAJGp1bWRjMmNsABEAEIAAAKoAOJtxA2MycGEuY2xhaW0AAAACRmNib3KnY2FsZ2ZzaGEyNTZpZGM6Zm9ybWF0amltYWdlL2pwZWdpc2lnbmF0dXJleExzZWxmI2p1bWJmPWMycGEvdXJuOnV1aWQ6Yjk1ZTk1YTgtMzkwNC00MGJlLWExZmEtMDIyN2JmYWYyNTY3L2MycGEuc2lnbmF0dXJlamluc3RhbmNlSURjMS4wb2NsYWltX2dlbmVyYXRvcngcTWljcm9zb2Z0X1Jlc3BvbnNpYmxlX0FJLzEuMHRjbGFpbV9nZW5lcmF0b3JfaW5mb4GiZG5hbWV4KU1pY3Jvc29mdCBSZXNwb25zaWJsZSBBSSBJbWFnZSBQcm92ZW5hbmNlZ3ZlcnNpb25jMS4wamFzc2VydGlvbnOCo2NhbGdmc2hhMjU2Y3VybHhdc2VsZiNqdW1iZj1jMnBhL3Vybjp1dWlkOmI5NWU5NWE4LTM5MDQtNDBiZS1hMWZhLTAyMjdiZmFmMjU2Ny9jMnBhLmFzc2VydGlvbnMvYzJwYS5oYXNoLmJveGVzZGhhc2hYIL09f8NxAMKjRnFb5mueViPD2DHpJo7kNW97D8OtpO4lo2NhbGdmc2hhMjU2Y3VybHhac2VsZiNqdW1iZj1jMnBhL3Vybjp1dWlkOmI5NWU5NWE4LTM5MDQtNDBiZS1hMWZhLTAyMjdiZmFmMjU2Ny9jMnBhLmFzc2VydGlvbnMvYzJwYS5hY3Rpb25zZGhhc2hYIH2f3m5nJE+74zAePnjKB4CcYrGk61TKEyO4iyg+X1xRAAAr12p1bWIAAAAoanVtZGMyY3MAEQAQgAAAqgA4m3EDYzJwYS5zaWduYXR1cmUAAAArp2Nib3LShEShATgkomd4NWNoYWlug1kGKTCCBiUwggQNoAMCAQICEzMAAAAaw97CUKNYKBAAAAAAABowDQYJKoZIhvcNAQEMBQAwVjELMAkGA1UEBhMCVVMxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEnMCUGA1UEAxMeTWljcm9zb2Z0IFNDRCBDbGFpbWFudHMgUlNBIENBMB4XDTIzMDgwMzE4MTUzNVoXDTI0MDMzMTE4MTUzNVowdDELMAkGA1UEBhMCVVMxEzARBgNVBAgTCldhc2hpbmd0b24xEDAOBgNVBAcTB1JlZG1vbmQxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEeMBwGA1UEAxMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMIIBojANBgkqhkiG9w0BAQEFAAOCAY8AMIIBigKCAYEAhx0Xi7XUYeyCzx60eBBWBhULK6pknzhPA91nLOZ1/GjHUN+4urt+GT47vXNbxAhOBtKt/I4dNkrC8PHIvnRxWuiVS95KAzVxW8Eoa2R/aQqwHCIDT7cRVG4lelfwuBcP556ytmc3N4GZ1oytR7a4VesIGricmIKa75A1E7657GDIuLib2D9Mplj2wgYIvWhmlJ3AB2LuKpjCFJsa3es6CeQhegEgGcZSZOKmYMbnh7ZxgjCjuLxx/nCyjjpWXdS5j9yDFm6aep1eYCIUaoB24DsqTXbpJUho7h6NLMSvpny8Zwou+IGDp4zcuAmnOFSbFmRj5ZHEFh4Ad6qM9nwZER430VtD9mZkYCoEYEpbawhDXiDcaaa1VhY9DPmZGnGu22AVxZQ3DgZQagp2lqe0nv/mDszpdtbvJ49B9GKD30CinK9WJnSKxi9wQwMMjxr6nYRKWArHudzvptaagAR+qQ4UcMbEo3+FmI9im8Sbi7bdha77OPG4yRSJe8rLIv9zAgMBAAGjggFMMIIBSDAZBgNVHSUBAf8EDzANBgsrBgEEAYI3TDsBCTAOBgNVHQ8BAf8EBAMCAMAwHQYDVR0OBBYEFO+OY+GF1Ob/ECtYIpG37FLbs7MfMB8GA1UdIwQYMBaAFIutmvyPdc3ODY1prXxMpGBW2bREMF8GA1UdHwRYMFYwVKBSoFCGTmh0dHA6Ly93d3cubWljcm9zb2Z0LmNvbS9wa2lvcHMvY3JsL01pY3Jvc29mdCUyMFNDRCUyMENsYWltYW50cyUyMFJTQSUyMENBLmNybDBsBggrBgEFBQcBAQRgMF4wXAYIKwYBBQUHMAKGUGh0dHA6Ly93d3cubWljcm9zb2Z0LmNvbS9wa2lvcHMvY2VydHMvTWljcm9zb2Z0JTIwU0NEJTIwQ2xhaW1hbnRzJTIwUlNBJTIwQ0EuY3J0MAwGA1UdEwEB/wQCMAAwDQYJKoZIhvcNAQEMBQADggIBANTqew+p+AT2aoc5q/OvQmtFhFPDvuq/dwAiE3hr/fCr4R+HKLnHi0YWhuHIYwe4VoeVnfArs2zPY6rGy5xzt/WV4A7brdV2rjEJFy8Cym1Ou8lalKvelcC4RqBbcn5k75NKtqsRBz+AaEWkKANSPXjfB4C5V5PCLWpAmBjWiZWSG8Q2//mqCOLVnlRynYrh7n5lpN7qOLstO10VddcHgBwLklAkU3xvHSGyioid6mCZxJ+1xJjtmN3m3Kvd074edvPOhs65zmgRZEqfjf8l8RrwSzpZWm4SRqcPjdh5zz06GpKuHb/6CvaiXg+ggV8iA4/dAoxdTyUXgN3xWfJkOuUR95OiQFjveIvLzFbNShdB1XI2L39p1XZXy0d+KP3mcBHCQInxrbPI8BYYU2EOwf7uoAZz3CFE8hHVoaB0wJEYe/MV4bMLXZf6YWIOT/44CabfWyvEBkPKfkTs4XXLPKXAonRjgOnTL+6E42w/XqeOWnVLEmKAe5Qw4VvBIMZfJKiTA4ddOIfdzCOaOQziIzV1zZfnGXZKd2Wtp60xKpwhpRn4QA09p1ay05urDGZn06bnJnfxAC+wFMJmnFctVQRJ9Lf/L50sg/zgbtOAfS/PMPkiOTGVs6lCXGKFwGBLpPqmaLVkgebCnIUUKjWBqnae1G2JLCms6PI5ik1Cz/JtWQbWMIIG0jCCBLqgAwIBAgITMwAAAATR1uF6CiJiDwAAAAAABDANBgkqhkiG9w0BAQwFADBfMQswCQYDVQQGEwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMTAwLgYDVQQDEydNaWNyb3NvZnQgU3VwcGx5IENoYWluIFJTQSBSb290IENBIDIwMjIwHhcNMjIwMjE3MDA0NTI2WhcNNDIwMjE3MDA1NTI2WjBWMQswCQYDVQQGEwJVUzEeMBwGA1UEChMVTWljcm9zb2Z0IENvcnBvcmF0aW9uMScwJQYDVQQDEx5NaWNyb3NvZnQgU0NEIENsYWltYW50cyBSU0EgQ0EwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDXJeLb8443bNoExaat+SmT0DUBXhFfqUaWfR6k7SCD6AEWd5EAq0w3Ue+rrNM0kanwhMIOjG41cUIdInoLdbgaHVBeBccEvAJNgir6KJpZo/H8VqNbFi+cQ9h4JLGZHNGFAKZ/rGaQ4kSzZ+zecVwYl9Z0w+ipTO/B412lryhgfZhaJdbifP89TEAaEqE1LDUVMPlISMjvDPoHyA6i/CMw0L+QB0u+ziGT5H7dhbBUXWf7vN369iZS0+Dxt7FfXowzHmnNG5ll9wZinCP6EUyVDdBX3iZMHYCa66kKQMICuyplN2jmRfjy0kV3nGVJC/sUvV262AdgZ1t914LzPeAsrCQRQid5M2aIc6wv/fX24tjor+OVpHxrqraMglsxWXQPuOui1x8HmMaeMLB/TOaHVc4iMrVLikoE9wcYGb0V9mnCpBFaD0S9/pzo6sx9xpl3g1O1L0WU0kIx/SxdnMQ9epxGDtJcjxCgZHHM56MdlmclOybk2DQO4cOO8lp0Dj+FfK2eZ7Bd+t/ZXC0C1RgIGKK3kWotVyv0nUtDPXOu0vwTcd/ckeOtRuBjgLHXPjJM/7rTXF2BN7HnwOtPIG5yniRaxfOUvrtmOfhOO61cWvylcMLNaYnHzwTsbtL2sFxIOV0gSWGJAlbmv2FQSgWXK1cWohnLQiktiGWBNVX1aQIDAQABo4IBjjCCAYowDgYDVR0PAQH/BAQDAgGGMBAGCSsGAQQBgjcVAQQDAgEAMB0GA1UdDgQWBBSLrZr8j3XNzg2Naa18TKRgVtm0RDARBgNVHSAECjAIMAYGBFUdIAAwGQYJKwYBBAGCNxQCBAweCgBTAHUAYgBDAEEwDwYDVR0TAQH/BAUwAwEB/zAfBgNVHSMEGDAWgBQLs2g7r9qv7nCldtkh98xEFgfQ+DBsBgNVHR8EZTBjMGGgX6BdhltodHRwOi8vd3d3Lm1pY3Jvc29mdC5jb20vcGtpb3BzL2NybC9NaWNyb3NvZnQlMjBTdXBwbHklMjBDaGFpbiUyMFJTQSUyMFJvb3QlMjBDQSUyMDIwMjIuY3JsMHkGCCsGAQUFBwEBBG0wazBpBggrBgEFBQcwAoZdaHR0cDovL3d3dy5taWNyb3NvZnQuY29tL3BraW9wcy9jZXJ0cy9NaWNyb3NvZnQlMjBTdXBwbHklMjBDaGFpbiUyMFJTQSUyMFJvb3QlMjBDQSUyMDIwMjIuY3J0MA0GCSqGSIb3DQEBDAUAA4ICAQBpxEcsFAQ9oJ+nX90jxdm0zaRrpzJVb7sSue3McYCfUjf3mnVTtU/AsWpQAWg6C/z65YH0AEeBV8qlDb+QoGDWjAc5Q9JKHzoIdJxJA07Q0GwJHAA8mPCkG9OJDMATcdXHe5Rr3gyWFWADZfcBEPdExeaQrAlXycwDlRtYpI86tjoqzb/9OKs4hl8FfK8kjm+9XeWCdiwlmPBMl6GdH/otPRkHzxtWuFv2ZPfVsIzsA04/QwhmUmY8OrCeKMTD4rY4aOrmgjR7MghQRfDoDNAueUDs5yYVdfkb5z6u3kpXPP5H/AsGrY5U3hvjmTQjGKvqc9vaSTsb9tHXk5g+6EQRK+OC8UE5K/lE+bhBiExQTCfwJeWgahGxTeXy807rrE3KZUg4j80lnaMt0DNbMSxhlPF9zLzMK8edCPctFwKvfWMOoE9mTf9giYJ3V2g45mQKOZfk93VcUkcNazLd+iiUzFlYTB8NLmu4Sc1Lqgr507Wtip3UEANCuVZCt/KyK3xupM40vubUyWHk3QxwPwaXy5/3kGxDtKzy7hVTCc5ILiHBnNQyvjiNYU8zOt2Fs/JkIMpPy7sqAvurpoGsSxv/0od0ns4p3Zg2ZIskuaI66ccB/6qLeAp1AAwTbV7lOgtiqhl9vKS9CyFLH4Sd44M+ZiGKGlunx26nOprkYnZSxVkFszCCBa8wggOXoAMCAQICEGgo1Ux+XNq9QzmuDMFaKjUwDQYJKoZIhvcNAQEMBQAwXzELMAkGA1UEBhMCVVMxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEwMC4GA1UEAxMnTWljcm9zb2Z0IFN1cHBseSBDaGFpbiBSU0EgUm9vdCBDQSAyMDIyMB4XDTIyMDIxNzAwMTIzNloXDTQ3MDIxNzAwMjEwOVowXzELMAkGA1UEBhMCVVMxHjAcBgNVBAoTFU1pY3Jvc29mdCBDb3Jwb3JhdGlvbjEwMC4GA1UEAxMnTWljcm9zb2Z0IFN1cHBseSBDaGFpbiBSU0EgUm9vdCBDQSAyMDIyMIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAniUBZhkfZDTBnTkjYh1xi1bqJdKbH+8nAYK/d9iUM4MYSJtQnnuWZMLQw19F/zKc6BcXvXjtdZhfOgYIKxw3m0ZKkAqwr0aSPjOJKvq45zJj8yPHbtIU+yZY7v4GhFT6wR83qtvU7FYqv0m9zOsC7cZO/KwZtRI1aRWJF02jaOpsHimaCfPOeiHGCdEZ6o8wRmk7aAQrfIot1mNd6m3WOZ69Bj5b7i8RWyhrp1KkaF5MpOquziO/TDZx2oFFUI7Khs7/U8O4Q7Mk7gd6orT6xwode8ZSNTHsCB+EgJJb+LHaOdbJ5+WJBH5Rf/TmamRHSer47Kb2oENT/trDIyTYJdoTLCq3P5TedxxMeBxq+ZqP62oVd3etSYTOEEDHmUgP1ZYegJxzoTihA2/TTSDQtUPk9y54D073vL9l2m2QC1u/3uonJ5lk+Dl8cz3WIdLu1vNTES5Vw9zq8SlX3lGheHOQCy/1yXU2643SbY55XboaOP/fGQGo0sjR1vLrivUu0cyTE5uckHhlY3kExPGen4w682QM/pgdk+KPVqVjUyO4bnMWRRq293sPzaQy/1r+lo3hh3jbcIOoJIVpIMJtEg3lefYqWc/Wq+eB5qCxiC0IjAuxz9dsNq+e+QNn2UFzqatFuHFgWBjUFixlutEF3pLFUBARkM5HzPuvvyPAnwUCAwEAAaNnMGUwDgYDVR0PAQH/BAQDAgGGMA8GA1UdEwEB/wQFMAMBAf8wHQYDVR0OBBYEFAuzaDuv2q/ucKV22SH3zEQWB9D4MBAGCSsGAQQBgjcVAQQDAgEAMBEGA1UdIAQKMAgwBgYEVR0gADANBgkqhkiG9w0BAQwFAAOCAgEASMc3///BaFfXi0NmRjomay/o+t5ooY9H8T00lXraVTH0ldI4Xyy6j6WNUTFqiVVobCtCZGqFJKBRA8fd0XJY7WwejNiRxedJEZ0ZejdYHndE+8IImELETeObig7PQEVPG4BwWYyTgegP1cgmlan3H3cGuMrvnPvoZtrlOeDS0mLDp9S2GJonmyZQSnKX1bNbKqT9Xy9+5mKjJ2YM+tkZzBEdMagBUgjmVAyZYyvq2ITUtAgW775KW4hY9AYWoOt6XeHHRNa7L1VWJfCeDOQPEtvAf69WXcaJDnGpVhLkuZyoZB61R5WSrtBwyJN9fFpY8QXxSrhschiprh9XmSZ0ZvUdD99d8Oc3W1+68LTv5GMHfh8yGGmpcFqS+XmcWNR+v3JdU0YrbqOZYNaFjGZ3Fnav4sUYW+JdCDbWZjcXZfAuz6HlvOaNDWW0VlNdn8ivTm5Rz4i+kuow+yzndT9CYMRx55efc8efytG4bCPqUCgdDkPM9akbQOummOXlD8WSL6WWx9f6PBjuHRthA/2G5yRBM73Y87ZgfPMcggPVYK/f9CCk5IEGIlrMhTN9ZPjkuL+AF9T7IT9jruePtxdE7HIuNckL0IEd6XIDCUHZ3wlI5s23shxgJRlS8z0SSe2dlCKOcSj4wQdUc904CLSFjxRsqgCvQKu1h862OVxz+ZBmc2lnVHN0oWl0c3RUb2tlbnOBoWN2YWxZFzQwghcwMAMCAQAwghcnBgkqhkiG9w0BBwKgghcYMIIXFAIBAzEPMA0GCWCGSAFlAwQCAQUAMHcGCyqGSIb3DQEJEAEEoGgEZjBkAgEBBglghkgBhv1sBwEwMTANBglghkgBZQMEAgEFAAQgtbgZ8oIWmS5XENkuY1VBNvFYGj1GMTUpNuIyT0vMtCACECZn80XjVP6AQSxxVKEpiNkYDzIwMjQwMjA4MDgxNDMxWqCCEwkwggbCMIIEqqADAgECAhAFRK/zlJ0IOaa/2z9f5WEWMA0GCSqGSIb3DQEBCwUAMGMxCzAJBgNVBAYTAlVTMRcwFQYDVQQKEw5EaWdpQ2VydCwgSW5jLjE7MDkGA1UEAxMyRGlnaUNlcnQgVHJ1c3RlZCBHNCBSU0E0MDk2IFNIQTI1NiBUaW1lU3RhbXBpbmcgQ0EwHhcNMjMwNzE0MDAwMDAwWhcNMzQxMDEzMjM1OTU5WjBIMQswCQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQsIEluYy4xIDAeBgNVBAMTF0RpZ2lDZXJ0IFRpbWVzdGFtcCAyMDIzMIICIjANBgkqhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAo1NFhx2DjlusPlSzI+DPn9fl0uddoQ4J3C9Io5d6OyqcZ9xiFVjBqZMRp82qsmrdECmKHmJjadNYnDVxvzqX65RQjxwg6seaOy+WZuNp52n+W8PWKyAcwZeUtKVQgfLPywemMGjKg0La/H8JJJSkghraarrYO8pd3hkYhftF6g1hbJ3+cV7EBpo88MUueQ8bZlLjyNY+X9pD04T10Mf2SC1eRXWWdf7dEKEbg8G45lKVtUfXeCk5a+B4WZfjRCtK1ZXO7wgX6oJkTf8j48qG7rSkIWRw69XloNpjsy7pBe6q9iT1HbybHLK3X9/w7nZ9MZllR1WdSiQvrCuXvp/k/XtzPjLuUjT71Lvr1KAsNJvj3m5kGQc3AZEPHLVRzapMZoOIaGK7vEEbeBlt5NkP4FhB+9ixLOFRr7StFQYU6mIIE9NpHnxkTZ0P387RXoyqq1AVybPKvNfEO2hEo6U7Qv1zfe7dCv95NBB+plwKWEwAPoVpdceDZNZ1zY8SdlalJPrXxGshuugfNJgvOuprAbD3+yqG7HtSOKmYCaFxsmxxrz64b5bV4RAT/mFHCoz+8LbH1cfebCTwv0KCyqBxPZySkwS0aXAnDU+3tTbRyV8IpHCj7ArxES5k4MsiK8rxKBMhSVF+BmbTO77665E42FEHypS34lCh8zrTioPLQHsCAwEAAaOCAYswggGHMA4GA1UdDwEB/wQEAwIHgDAMBgNVHRMBAf8EAjAAMBYGA1UdJQEB/wQMMAoGCCsGAQUFBwMIMCAGA1UdIAQZMBcwCAYGZ4EMAQQCMAsGCWCGSAGG/WwHATAfBgNVHSMEGDAWgBS6FtltTYUvcyl2mi91jGogj57IbzAdBgNVHQ4EFgQUpbbvE+fvzdBkodVWqWUxo97V40kwWgYDVR0fBFMwUTBPoE2gS4ZJaHR0cDovL2NybDMuZGlnaWNlcnQuY29tL0RpZ2lDZXJ0VHJ1c3RlZEc0UlNBNDA5NlNIQTI1NlRpbWVTdGFtcGluZ0NBLmNybDCBkAYIKwYBBQUHAQEEgYMwgYAwJAYIKwYBBQUHMAGGGGh0dHA6Ly9vY3NwLmRpZ2ljZXJ0LmNvbTBYBggrBgEFBQcwAoZMaHR0cDovL2NhY2VydHMuZGlnaWNlcnQuY29tL0RpZ2lDZXJ0VHJ1c3RlZEc0UlNBNDA5NlNIQTI1NlRpbWVTdGFtcGluZ0NBLmNydDANBgkqhkiG9w0BAQsFAAOCAgEAgRrW3qCptZgXvHCNT4o8aJzYJf/LLOTN6l0ikuyMIgKpuM+AqNnn48XtJoKKcS8Y3U623mzX4WCcK+3tPUiOuGu6fF29wmE3aEl3o+uQqhLXJ4Xzjh6S2sJAOJ9dyKAuJXglnSoFeoQpmLZXeY/bJlYrsPOnvTcM2Jh2T1a5UsK2nTipgedtQVyMadG5K8TGe8+c+njikxp2oml101DkRBK+IA2eqUTQ+OVJdwhaIcW0z5iVGlS6ubzBaRm6zxbygzc0brBBJt3eWpdPM43UjXd9dUWhpVgmagNF3tlQtVCMr1a9TMXhRsUo063nQwBw3syYnhmJA+rUkTfvTVLzyWAhxFZH7doRS4wyw4jmWOK22z75X7BC1o/jF5HRqsBV44a/rCcsQdCaM0qoNtS5cpZ+l3k4SF/Kwtw9Mt911jZnWon49qfH5U81PAC9vpwqbHkB3NpE5jreODsHXjlY9HxzMVWggBHLFAx+rrz+pOt5Zapo1iLKO+uagjVXKBbLafIymrLS2Dq4sUaGa7oX/cR3bBVsrquvczroSUa31X/MtjjA2Owc9bahuEMs305MfR5ocMB3CtQC4Fxguyj/OOVSWtasFyIjTvTs0xf7UGv/B3cfcZdEQcm4RtNsMnxYL2dHZeUbc7aZ+WssBkbvQR7w8F/g29mtkIBEr4AQQYowggauMIIElqADAgECAhAHNje3JFR82Ees/ShmKl5bMA0GCSqGSIb3DQEBCwUAMGIxCzAJBgNVBAYTAlVTMRUwEwYDVQQKEwxEaWdpQ2VydCBJbmMxGTAXBgNVBAsTEHd3dy5kaWdpY2VydC5jb20xITAfBgNVBAMTGERpZ2lDZXJ0IFRydXN0ZWQgUm9vdCBHNDAeFw0yMjAzMjMwMDAwMDBaFw0zNzAzMjIyMzU5NTlaMGMxCzAJBgNVBAYTAlVTMRcwFQYDVQQKEw5EaWdpQ2VydCwgSW5jLjE7MDkGA1UEAxMyRGlnaUNlcnQgVHJ1c3RlZCBHNCBSU0E0MDk2IFNIQTI1NiBUaW1lU3RhbXBpbmcgQ0EwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQDGhjUGSbPBPXJJUVXHJQPE8pE3qZdRodbSg9GeTKJtoLDMg/la9hGhRBVCX6SI82j6ffOciQt/nR+eDzMfUBMLJnOWbfhXqAJ9/UO0hNoR8XOxs+4rgISKIhjf69o9xBd/qxkrPkLcZ47qUT3w1lbU5ygt69OxtXXnHwZljZQp09nsad/ZkIdGAHvbREGJ3HxqV3rwN3mfXazL6IRktFLydkf3YYMZ3V+0VAshaG43IbtArF+y3kp9zvU5EmfvDqVjbOSmxR3NNg1c1eYbqMFkdECnwHLFuk4fsbVYTXn+149zk6wsOeKlSNbwsDETqVcplicu9Yemj052FVUmcJgmf6AaRyBD40NjgHt1biclkJg6OBGz9vae5jtb7IHeIhTZgirHkr+g3uM+onP65x9abJTyUpURK1h0QCirc0PO30qhHGs4xSnzyqqWc0Jon7ZGs506o9UD4L/wojzKQtwYSH8UNM/STKvvmz3+DrhkKvp1KCRB7UK/BZxmSVJQ9FHzNklNiyDSLFc1eSuo80VgvCONWPfcYd6T/jnA+bIwpUzX6ZhKWD7TA4j+s4/TXkt2ElGTyYwMO1uKIqjBJgj5FBASA31fI7tk42PgpuE+9sJ0sj8eCXbsq11GdeJgo1gJASgADoRU7s7pXcheMBK9Rp6103a50g5rmQzSM7TNsQIDAQABo4IBXTCCAVkwEgYDVR0TAQH/BAgwBgEB/wIBADAdBgNVHQ4EFgQUuhbZbU2FL3MpdpovdYxqII+eyG8wHwYDVR0jBBgwFoAU7NfjgtJxXWRM3y5nP+e6mK4cD08wDgYDVR0PAQH/BAQDAgGGMBMGA1UdJQQMMAoGCCsGAQUFBwMIMHcGCCsGAQUFBwEBBGswaTAkBggrBgEFBQcwAYYYaHR0cDovL29jc3AuZGlnaWNlcnQuY29tMEEGCCsGAQUFBzAChjVodHRwOi8vY2FjZXJ0cy5kaWdpY2VydC5jb20vRGlnaUNlcnRUcnVzdGVkUm9vdEc0LmNydDBDBgNVHR8EPDA6MDigNqA0hjJodHRwOi8vY3JsMy5kaWdpY2VydC5jb20vRGlnaUNlcnRUcnVzdGVkUm9vdEc0LmNybDAgBgNVHSAEGTAXMAgGBmeBDAEEAjALBglghkgBhv1sBwEwDQYJKoZIhvcNAQELBQADggIBAH1ZjsCTtm+YqUQiAX5m1tghQuGwGC4QTRPPMFPOvxj7x1Bd4ksp+3CKDaopafxpwc8dB+k+YMjYC+VcW9dth/qEICU0MWfNthKWb8RQTGIdDAiCqBa9qVbPFXONASIlzpVpP0d3+3J0FNf/q0+KLHqrhc1DX+1gtqpPkWaeLJ7giqzl/Yy8ZCaHbJK9nXzQcAp876i8dU+6WvepELJd6f8oVInw1YpxdmXazPByoyP6wCeCRK6ZJxurJB4mwbfeKuv2nrF5mYGjVoarCkXJ38SNoOeY+/umnXKvxMfBwWpx2cYTgAnEtp/Nh4cku0+jSbl3ZpHxcpzpSwJSpzd+k1OsOx0ISQ+UzTl63f8lY5knLD0/a6fxZsNBzU+2QJshIUDQtxMkzdwdeDrknq3lNHGS1yZr5Dhzq6YBT70/O3itTK37xJV77QpfMzmHQXh6OOmc4d0j/R0o08f56PGYX/sr2H7yRp11LB4nLCbbbxV7HhmLNriT1ObyF5lZynDwN7+YAN8gFk8n+2BnFqFmut1VwDophrCYoCvtlUG3OtUVmDG0YgkPCr2B2RP+v6TR81fZvAT6gt4y3wSJ8ADNXcL50CN/AAvkdgIm2fBldkKmKYcJRyvmfxqkhQ/8mJb2VVQrH4D6wPIOK+XW+6kvRBVK5xMOHds3OBqhK/bt1nz8MIIFjTCCBHWgAwIBAgIQDpsYjvnQLefv21DiCEAYWjANBgkqhkiG9w0BAQwFADBlMQswCQYDVQQGEwJVUzEVMBMGA1UEChMMRGlnaUNlcnQgSW5jMRkwFwYDVQQLExB3d3cuZGlnaWNlcnQuY29tMSQwIgYDVQQDExtEaWdpQ2VydCBBc3N1cmVkIElEIFJvb3QgQ0EwHhcNMjIwODAxMDAwMDAwWhcNMzExMTA5MjM1OTU5WjBiMQswCQYDVQQGEwJVUzEVMBMGA1UEChMMRGlnaUNlcnQgSW5jMRkwFwYDVQQLExB3d3cuZGlnaWNlcnQuY29tMSEwHwYDVQQDExhEaWdpQ2VydCBUcnVzdGVkIFJvb3QgRzQwggIiMA0GCSqGSIb3DQEBAQUAA4ICDwAwggIKAoICAQC/5pBzaN675F1KPDAiMGkz7MKnJS7JIT3yithZwuEppz1Yq3aaza57G4QNxDAf8xukOBbrVsaXbR2rsnnyyhHS5F/WBTxSD1Ifxp4VpX6+n6lXFllVcq9ok3DCsrp1mWpzMpTREEQQLt+C8weE5nQ7bXHiLQwb7iDVySAdYyktzuxeTsiT+CFhmzTrBcZe7FsavOvJz82sNEBfsXpm7nfISKhmV1efVFiODCu3T6cw2Vbuyntd463JT17lNecxy9qTXtyOj4DatpGYQJB5w3jHtrHEtWoYOAMQjdjUN6QuBX2I9YI+EJFwq1WCQTLX2wRzKm6RAXwhTNS8rhsDdV14Ztk6MUSaM0C/CNdaSaTC5qmgZ92kJ7yhTzm1EVgX9yRcRo9k98FpiHaYdj1ZXUJ2h4mXaXpI8OCiEhtmmnTK3kse5w5jrubU75KSOp493ADkRSWJtppEGSt+wJS00mFt6zPZxd9LBADMfRyVw4/3IbKyEbe7f/LVjHAsQWCqsWMYRJUadmJ+9oCw++hkpjPRiQfhvbfmQ6QYuKZ3AeEPlAwhHbJUKSWJbOUOUlFHdL4mrLZBdd56rF+NP8m800ERElvlEFDrMcXKchYiCd98THU/Y+whX8QgUWtvsauGi0/C1kVfnSD8oR7FwI+isX4KJpn15GkvmB0t9dmpsh3lGwIDAQABo4IBOjCCATYwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQU7NfjgtJxXWRM3y5nP+e6mK4cD08wHwYDVR0jBBgwFoAUReuir/SSy4IxLVGLp6chnfNtyA8wDgYDVR0PAQH/BAQDAgGGMHkGCCsGAQUFBwEBBG0wazAkBggrBgEFBQcwAYYYaHR0cDovL29jc3AuZGlnaWNlcnQuY29tMEMGCCsGAQUFBzAChjdodHRwOi8vY2FjZXJ0cy5kaWdpY2VydC5jb20vRGlnaUNlcnRBc3N1cmVkSURSb290Q0EuY3J0MEUGA1UdHwQ+MDwwOqA4oDaGNGh0dHA6Ly9jcmwzLmRpZ2ljZXJ0LmNvbS9EaWdpQ2VydEFzc3VyZWRJRFJvb3RDQS5jcmwwEQYDVR0gBAowCDAGBgRVHSAAMA0GCSqGSIb3DQEBDAUAA4IBAQBwoL9DXFXnOF+go3QbPbYW1/e/Vwe9mqyhhyzshV6pGrsi+IcaaVQi7aSId229GhT0E0p6Ly23OO/0/4C5+KH38nLeJLxSA8hO0Cre+i1Wz/n096wwepqLsl7Uz9FDRJtDIeuWcqFItJnLnU+nBgMTdydE1Od/6Fmo8L8vC6bp8jQ87PcDx4eo0kxAGTVGamlUsLihVo7spNU96LHc/RzY9HdaXFSMb++hUD38dglohJ9vytsgjTVgHAIDyyCwrFigDkBjxZgiwbJZ9VVrzyerbHbObyMt9H5xaiNrIv8SuFQtJ37YOtnwtoeW/VvRXKwYw02fc7cBqZ9Xql4o4rmUMYIDdjCCA3ICAQEwdzBjMQswCQYDVQQGEwJVUzEXMBUGA1UEChMORGlnaUNlcnQsIEluYy4xOzA5BgNVBAMTMkRpZ2lDZXJ0IFRydXN0ZWQgRzQgUlNBNDA5NiBTSEEyNTYgVGltZVN0YW1waW5nIENBAhAFRK/zlJ0IOaa/2z9f5WEWMA0GCWCGSAFlAwQCAQUAoIHRMBoGCSqGSIb3DQEJAzENBgsqhkiG9w0BCRABBDAcBgkqhkiG9w0BCQUxDxcNMjQwMjA4MDgxNDMxWjArBgsqhkiG9w0BCRACDDEcMBowGDAWBBRm8CsywsLJD4JdzqqKycZPGZzPQDAvBgkqhkiG9w0BCQQxIgQgtjbcUw8JQ+4BJRYa3i6d02c3ponVYOJR1O+/tvEoOcIwNwYLKoZIhvcNAQkQAi8xKDAmMCQwIgQg0vbkbe10IszR1EBXaEE2b4KK2lWarjMWr00amtQMeCgwDQYJKoZIhvcNAQEBBQAEggIASj+0yT/ruWTV+BtWc7aN6dgh5DBlH8jj8pAijzf8u23eMA8nWRDXHCHEa+VMg9yq+mBSeO8dGY9y7T4CJc3C3BLG9FPrMv0a9h8QAoFBqtGak5HrkZ85VTUUbmDpjWdVBtuiY5Y6Rj979Gm0jJGr7HNtFL/kXBy1C0kmP1iKMsymeF/+0oZw1Cgj3wY+JBqdTiDj8sCyDe2bPLCUl9LOMoVOPzRWjVK8jiZGKty5RbnirqLzXtfu9dk82EkFZyvTVOXg3lJ8kyqBfpgmB7m8TM/RemZ4cHwpaiKP6pB3FjQvzC1DPsiD1N+ngD+Zw2PCRWPTaIJ9VLZkFQA/CepE6WV9qiz7lhzfXK+bG/HC737MZrwDvK2wSF6m0Cb+dRDfBVfleX0FCRKcEa8s/qqstFDCG/gm7Mzrsdns4Mw/0SKVqN1BvX9JJlOqhlzlNahdg9heLB8U4uxnvbasHNcVauiLqAU3GG9l/NkZyWXd14qaYpvNLMpg/6KIs6MxSTttNg1EMgA7ir9cv1zsdA2gyJAlf0SIMjESFEA3/di4RJ512rzQ4sCeEiu0mDW1zUgU7WSjki4QvkWpNnnYASQ6SShPr2n6azwe3rtWE12pTFyY84t+6/0aeIjjTc+0xDos491Kx8R8WqZG/Emc74+QN7s+hRmEXD9txmXz2GQbqDr2WQGAUq8+ncyKpuNTAm/GKfZBOpMH8QVsBUuEstuKIB+vUps3glqhtZ99pu3/Pxq78wHQ6/zZBxEV26sgrZ1QQ4zMAwL79Ch/1eAcIMsN6dJjz3iRprWj+yOkT6wruMU00BJvQTYRsefPJsz0p62rbYBwMWmNWwdvRj2S1+uT8E6VeZRy6NwPABXIga5yuYOK7xephG5Y+LmXTzB7JZVXVuTRmsf2kUHuOmSn6SB0US5YMsEkUxO70HHh+aWbLEALx7ZbuDdPLaflymzkB5EFEQR5BpPLPNABhJRPXalbscliue9zkwH+GkrdwZUbLSj6784QHizhcbjrN83RyBBKnD2cXcXFHtwcpU5TZH0om9JBoUG29oimvDklsR7pfxazB1dePj0lhgrPGkMOscXZMloQtxXYUXrujm2+TDSEayNN2ML/FuduuyA5xNXuzk1/hXk26bB3pjZI45eRJZsGKAHmp6l+m025WywniuyzIC6L3iEMBpLXoF8PKv9uU4OAHKKJ/+EARkV4aWYAAE1NACoAAAAIAAQBEgADAAAAAQABAABREAABAAAAAQEAAABREQAEAAAAAQAAAABREgAEAAAAAQAAAAAAAAAA/9sAQwACAQECAQECAgICAgICAgMFAwMDAwMGBAQDBQcGBwcHBgcHCAkLCQgICggHBwoNCgoLDAwMDAcJDg8NDA4LDAwM/9sAQwECAgIDAwMGAwMGDAgHCAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwMDAwM/8AAEQgBgAGAAwEiAAIRAQMRAf/EAB4AAAEEAwEBAQAAAAAAAAAAAAcEBQYIAAMJAgEK/8QAShAAAQMDAgQEBAMFBgUDAwIHAQIDEQQFIQYSAAcxQQgTIlEJMmFxFEKBFSORofAKUrHB0eEWJDNi8UNygjRjkheyGCUmNXOTov/EAB0BAAICAwEBAQAAAAAAAAAAAAUGBAcAAgMBCAn/xAA8EQABAwMCBAMFBgYBBAMAAAABAAIDBAUREiEGMUFREyJhMnGBkaEHFCOxwfAVM0JS0eFyFiRi8TRDgv/aAAwDAQACEQMRAD8A4H09NtZlaQWyRJ/KMT26cb0tq2oWN6QpJInGAJx3mD/LjzCwyNoIjoZwSIMR0nJMHhUwz+JlCEqUFTuSJ2wOhM4EGf5A8YvcLUyFNvKQoNpbLZThJ2mMds5xMxnrjjammV+8CUpCQFBRSsbSQDiZzGMA/wCPChmhIdb2lbRcbITkjaFHdux+UDMf6yLbVfwrrjbfhXNeKI670a7bam+GzJ08p8i4pIeDI+U7VObklflFMlG1fSeMG5ws2HNVIWyEMbEuEhtAQJVIbThSs4GZJkZxPSSVFutNTd7ghNI0p2qqCUttNq3uKJIThIzmQBJA6kA9tNW0k1bxQgK80TOTtE5MgSBmegwIjJ4lvh+54Xnw7859Ia5sKbeu/aPudPdKJq4MJqKdT7KgQhxCyApJMbhjqciOPCt1FqqiNnfdZqEqp3UOLQpK0H0LkDaQQNpAKusdDgTx5bbKU72kuJQopSFNq2CMgAnp1AOc578TjxDeIC++J/nxqnmFqk0SdQ6zrnblcG6GnTTsB50ZKGx8qIAx1J6knJhbCvN8tbgcPlkDdv6AiDJESIImft9OPVim3LPlzo2/aR1pVag1kzpi6WO0Jq7BQrtjtWdRV34hCBSJWjDCvLWpe907BsIycCGsraQ5tB3ISZCUyhKgkzujoTG7qMAHjxT0riqcpHnOBI3uKbTvAgnJIOIKhnHQdONrLKW9ydrZbQN4I3FOU9szmInBET9eMWYPNamWi+ny0oWFYwFbiI6CTO5JI3YmJEYMcOFDaFqrEoBaSdqgSJM/3s98T8pO3qCOPlv/ABDlS5TtpO5agNskFw4BTAMHvJOYERmOLH+KjxBVHjk1fo5Vj5caR5dI0zpqk085T6cpvLTdXGpH4l0CNzy5jaJOACVEDjXBcdLdypcMIx4smzRzQKoBQ2J5MNF6sc9IQhCZKjgJ2wQkTJEeomJAnglctOVGtOeV/boqChqairq2lQxRslyrqPLSVqXgblbUBSiTGASRHSVWfwqXDl3oBOpbtY7oxa6x1dC3cH6Yhj8UgJK2ypQCdwCkyO24GAFZQp17qOgq1P2S5XDTLCWnGjU0rymXglSdriSUkH1IJlMDCyOnzT/uIhAdU8+y2huclQDHQnDRzPT5pAzqjT/LTSt9tLdgtN8vt2NOmmu9Qp41VpdbUpbpp9iocU7OwlY9KUgpAUTxD3bDWXVhutvtQaOmQmWmHCAsY2wEgbUKJR7SYJMYlydqqHSr6gyyHKlTeHHEhbpJ6EJMmTP0iJ7glirkVV/qvNW6tJWYC1EzI6wR1HUykfrkDiO+WR5wB8OgUiOlhaNUjs/mT6LRdtUt20KZtqPKUpBJf+eqWDiCuJiPykRj6cR9GmlXSqU68A6+4ZWpaip1RJIlUmIx1JOD2McTOl0s1SHaUbAoztgAoIiMAwf/AJGehM9ndm2/hmYQhunQUEBWwbtogyTAHY5wcdiJ4FVdcyLYnJ9E12fheesIeW6GevNRG16FYoktufg2kK3BKlurjHulMdMHIj29jw70FgpwMArDacwkJCgATugTkZkn+MmOJDT6eLiSt8EiRKSQQn2UojGBmJ7/AK8bBTBK0bFNLJO6CQkHtMGJJgZAHf2wDnrHyeisi3cOUtJggZPc/omlmjW69KtjCQQcqIASSenWIGfaB1OOFNNRIUryxuU6QQVJEg5AGSMA5k/6cL6WhFepCZdrJXtQE7giCQBhIk5MCTOffh6p9Opo2UpuNQWkLI2UrABcBIyITIMjvk4MweIfiNbuSj7IZHZAGE1ULL7lak06lB0kLbLa9uwSBuDnpxkRtBycA8PFJpR2qQaisecWVwrc4grWfTgJ6AAAGCT1AxBI4kto0VWvUwXT0LdspSQhSn8uL6dJOQr2GBnAzDsnQrpfQlbqnnFnJdcJ3SSYjtuxI2g9sHjk6qc7ZuyltpYmfzDkqNt21q3FMBE/KVbt7y1A4BBMDM9CMZgjr7fcVUNpT+9d2gJMqJQM/NMACD1AjH/u4m9HyyLeahUpWUpIJLSGup9SiBJjMyQN2ZiQ40OlKFhgLdq2UuFtUBJLmw9d3sSZJyokkGJnjeKn1nLyskr44hpjCHtBYqmt2bKbaFdCdoySJHQ+oAGDtMEAz1hxouV9W8w2XXvMDZKClAhsH+6CSZ79ATjBzAnzVEyHHW7XQV9UtRUQ6GcmB1hUA9Ae4wJ6QfSNFXevXL7KGGkEgh55C9mdoAAmFmQJBAwczEmaeKhj3eclBZ66skOIxgKFv8urdblDzqlKXEAqaS49ucyTO4JQYA7hRn5faApGmbYwhfk2x+rLajsWphxTgUEypJKyUJAETMkbjIBMif0HL0opAk1Tz3REsJ8tKTtMoMCZkZj2PcyF9Dy1pnVpKmGycICFQSgAQlJgEgmRCQCPlPUmStPU0PJoQ6aluJ8xchqiGnNtNZaJtDw3JHnh1J2gGAYyn1HoY6SZzx8coa95tUsU7AdSFjctYWiQJ+VO8iCTKpAiBPY3WXkzdK8rNsoapllaipZaSQtwjAlUQR6R0Ix0ESOHK2+EK8XVlLtS1VO9AFLISASmPnMEGAQRHuAevBinbSybtG6GTC4M2yFW1+g8+RuqXGnipxJp2ylTm2SFDIUqRtkmEiB1kDhHU2pCH0rKa1pPVC1BIKyDG6JiSSOpPVX1ItNcPC01p1pa6tt/yvNBkVKVlZ6JnIlUScdYGO/EduPh+t1zcWilcDq/Una2tCVpEDCQkgAR1Ix1gSTJVlsiePZQievrIz5noAptX4Yuf8yd4CoKXd8wCrfKMKCYEFUDG6IIl0p6N1LDxAZVsQNgKpS3khZhRB9QM9hMkCeJ5qvk87YVLSrzXElKCsIdCWUlO0JJ+Ug5UIOcmY3cQOvVU6ceQXFuqidrqfQRClQoYkHGY9OJkxxxko3xHHRdYKqCbyzDdPNPVqZRsU2PKb8xfypSQkSreoLMBICspIO7ESrI9uimXTz5LjwS1sb8xJQreoEpSVAKkzuEJSnaVAJJJALSzqNl5lrz0hQQYS4y2VrRHdRmCcGCVKJkfYfBdKF6rU1TlttZAStRJC1n3BBMH6ghOBAIIPHPxnMGFL/hdPIOhC83+gFtJVS4QnclaPPQrcADsBTgTuVJUTuAkbk44Yfx7bdQlBU++wg7UbylCRG5KFAidhUmO4kySolW3h6dubbwSA40HGR09OFgq7iAB0ESSTmdoHEa1JRLQ2tTZSph1K1gqTvcCgDtVHclJ7jpIGJ3eGoaeaXrnYXwjXFuFovGkUB1yoYD29LimlNEq3IWVmSd6CrJIMmSIUQUgAG1/wAMf4hWhfB9S64puYnLyk1kvUFoVQUTrjLSnGnJWT8yYCF7huUnMgnuriptnuiC35Lqwwy635jZHq2gCAdzmFAzORuhEBU44W3SxLvCC+2pzzwkjCyoMkkqGVZBEA5MDeDAMgcWz6HZSbW2tk7CCN1IblcGrzcal5tX4Yu1DikNNna3TFREIiYCgCB0O336kSzlhW6KRZ9S0OrtNXG4uXO1OMWh2jqvJbt1VvTDziNqi42Etx5ZOBJyBgTW6/uUbbjdQ46H2lek7ilVQkH5ljChAIVGDKwQDuJ4mFoLFzQW14cQnaosqS2VkoSYiRPpkykCdpAJzw50NS2pb5TuqxudI6ifh/JMWvdPNsXRuo/fho5b2LEgESQUwNqe/VOUjpBmO1LJ8ko/+oS6oICSna4tQgjESidwBKTKj6RuiOCbW0jyGXaZ5FQp1n0hbidqXNgJS2qBiDIwABJMKk8RW425lFMC5tQiSFLgIM7SgkkpPp2tgZE5BABniNX0RwcIlZ7i1zgCU58tuVeqdW6Hvt8syGa+i0q2h+rXtQ4Utrc8uQlRB2iVDbIBUYUDuIDaKug1bZlvNFtxLHoqKZx1LikM99ywUlaIWkqXBJUCkAAyN7ddX6fttfSU1RV0tFXiKqnTvK9iin0FI6hJk7VAdZyDJg1nuT2ndR1IBYUyVB1TTjW5CkFRGO8jPqSesAH1AcLUjSOafGVDRgKrLISrZucSoKwB0BJAEfQfX6Hrji9fwIdLeHHVfi6LHiVuFHS6V/ZrrtuFc+ult7tcgphFQ42QUNbCrbkJKh168Und05WW610VbU09S3RVwWlh9Tag06pBG8BREKiUzGRjHGuhqPKRBUVoJCUysJ3IJ6KAk5n+BPvBFjnugQ3GFYf4n1n5QaY8cWtKXkRV/j+WdPXE2uoQ4txtYLY8wNqVClNpcKwg4JSBnM8AWludU20GEVLrlO1K1Nq9KFEp2pUUkgExPqIBgAdJ49WOxV+qq+noaSlqK2teO1qlYZUt1z/tCEg7iCCekE5zGElTRqp3nGVtIbfaWUKCkjr0IBxI/iAOnWePXHJyF61uBgr5gKVkDYZ2kj5ZJk9zETJA6jOQOPrjCFLSt5JdOyTuUPlMGDjKiD1x3zGePqArytwTMynCgZBmf1JAInHv24+fgy75yk+ajcduVEAKSEqVJIJwIzgDB6Y41Wy3lSENrRLYKN0ygnqACD+UnAgED5hMjj40ve6y+SqW1eYpRj0ztEmczg+o/SPpJeZ/J7UvI/WYsGq7FW6evKqZm4fgK5sNuhipbD7S9g6JU0tKk9ilU4HRBoi101/1HaaGquNLaWq+qZpl1j6VLZoEKWlCqhSW5UpttOTtBnarBPHhKwKZjxCUo8LCOWStF6TU9/xN/wASJ1R+HJvoQGC3+BDswaaVBZTAO8AnrxB6ChqLg80huUIQPLS4tMGN24dP0jp0mRB4k+uuX9FovmRe7RZ79R6os1srnKaivdOwunYvLKFEIqW23QHUoWmFBK8gKEgzwusNkSEpW4lxxpEQFhUKOIgD3PTsCYg540LnOOGopTUTD55dgvGjdCN1q22UIacClpbUBg5k7U4gCAVdD8pGCRJq0RdRyuVSVlpeeoq2mM079M2WngQqQUg53JiAobYGIn1JH9Jf02WnCaZpt2rCJ3YDKMgqkDA+sYjp9fNbqI0b3n1S3HnypSkJdXjptATIKoPsE9EmIyQaoZWU27Bl/foFBuERqvLL5YxyHUokXnmbqLmJam9Ou3O6V1pVU/imbOxufRU1akBHm+WSf3isJmNygAPYAX6t1LV01SaYU7tArcoN0oSUPMhJAIUoiQN0CEgeoK7weFegucGqeXvMOx6q03cqqxXnTtY3W2usYBbXQPNxtcQgzuKTHzDb3z0OrVGprnrHVFxvl7rai73661DlZWVLh/e1Dq1lS1rKY+ZaiSBiT9OINwuMbDrkOXH9/BE7HYKms8lOzTGOvRM9tsq3ypSm0sJUtSy4lUBJkEK9UdAJgndIntw60lKzSOhbaVuLlJUsY9P3MRJjH2/TVS0j1SpMoKvLMBAAhIk4wDEQegxP1PDwlhq3HzKoh12EkoC5UjBypR6CPoTjHQErNVcppfw2HS399Vatn4XoqH8WUeI/5/IL1QW0oZlKPw7LYMuBO51Ak9AqMn2xMdT2d9M2Fq73qjomBRsvXKoTRoeuDvkU7bqykSt5e0IRMkqIAGJMcRS56pLxLTQDhSTASkhIVkyAclWCZVnqRHb5QWisury5CnlLELgSEzJgnpBkjJzkfcW8MbvzTfTySvO2B2A/VPN8ffp7pUUhLVU7SPLaUttaVs7kqIJSoDaoYJkdeuZMprTb03l5CFioqVgE+RTNlWCcndnp75gfTPEn0nydqbwFOPtF1TKd/lIEhInBMfljvAHTMxM6tnLgtN+Q2hnytqVLRO0EFQOUIKirJ7qHePbjmY3lusjA9f8AakGpiiOjVv8AM/RRK0aYcUPLWGaZLYJUinT5iwDG4FZhLc4mMn744mWmtJmjUTR0rKEPN7vxMALB2/KXFSpSvqmB6fbiVW3QKqOm3PpYbT5nloCUh56okY2IHpBIHaevYZMoouVVyvAUC63Z6dZSUPVKw/UuQSYS2J2CFkDcZABH14FTV9DFvI7V7uXzXV332UYiGkeu5+X+VC3GQwz5wWCXEpQp0J8pG/cYlSyVHviY7YBgKba1VVq1U9tQupWiGgtgOBtBkR6sGOgxgkiAdsCdL03ovlwr8RdK43KtUYUKypK1Ff5gltA9MkCE4PywcCd45mvVVOpuwWGrca/6ZWG00rY6BJCjJIj6ie2M8QH3USkCBmG/L811ZRPYCZ35P77KKUHLGrrktqqm6pUFCYa/doaMCSColWYEQB8w6Azw9UHLmnpdw20TCgU5SDUFZBg+o4nBzJgwAeNdwvl+q0OKcTbaZskifMWsogTG4qCR1j5sSMRPDLc7hdKyWxXklxZCvJSC0nEk7SnpEH0jogGQJI7Nkc/YkD4rGtib6/BTb9mW2gC0KrKYkD0hcgge23APQASZ9RgYke6essbK2wblRgJSoN7ErMGJGf1UcbYgduIExpOuqXUOVT5JCSvy1O7IHRRUEdc4+/c9OFNPpOqbP7tVKzuJLBNSRIjqfQVdzB9jg54lxQU5Pnl+S3Mrx7MaMmjbHpjUSwmtvVtZS56pdKhtSMEQJET7RI6Zjgu6WsugtO04dYq7ZXKWAUCmgboJmOpkGepmR3nioCdI11SwEpuYfbwAhNQAnfOT2AjHZQz3yeG6psPkqUlC/Ic7KZqFbiR12wQNoxjAEjrJHBWntkDjmOYrhI6V3tN2VyNR+Ke3aaBZprfbggApG50AnbAMgdZj+U/Xgb618VLmoQUtOmkB9IAn0me5yYx3iO/1rk9cK+jQECreUW0eYlIeUpSQkbkxGRIJjdPQZgYQp1umld8ioShCEkqKgsuBMpEqEplRGOoUYB7DDLSB1OPKcqPJTwP2e1Ei+81K91TynEIUtzcFnapQB7Y6EHPbKsRwz1HNV3cUllaUOKUpQ3KlRAAAg5IgTtAz9ZJ4Y6S6t1ZG9R8tSUqUpLmAowAOhTgj1dQADkHjWKFis/duH0q9UISV7UeknMyJ9cYPf6AFReKiPBPJQJrBSTDDTupzauZzt42pqQthIKWQFrK0pIwQoiVR6QduSQU5PCa8aH/byEI8pDqXk7liVKEqKkgewJAgRAGDiSBmgKLTNu5e6ht1ys9bW6kqXKZdlujNeG6eiQFq/EB1qD5hWlMJ/ulJxmeJXoKjS5XooaxtUrIUyogTugBQCSYjt0kYmCQOD9HcmVAAPVJV0sM9Ll7OQQjvnLe6WBa1hlTjZMJKkgpa6CYSSAO8xkgfKQOInqKwvU4adbbWh5lYQ2EJI3yeqlHcQcntMfX1G3l8at7VCmlcp2WnmEiS7hPsIBBGCDmO5mTB4F+t9BU9wbcLZ8hodF7e4VEjMgTOZTOZmABLmtrXeZiXHXeWLyHbCAVf+/p/MZVuUVQptDcEE9JgFJMx1EY6AyOPDVM5d2ko8k1AUoMnH4iVbySZGSokkx1zOCY4m995UuwssL8t9e8nIASvaREqO04URMgg5mThjunL26aTWzV1FDVU9JVJK2lOMnynEoISqJAlMhQgHE5JieBE9A8cgitLxHFIzw5Xc1G7houronAlXmMu0z8pcUlaA0CdkZSRAASSZlISoAnHBc5DcmblzOU1T0jDrymQViEuK2FO3aACNqkztMEfQ54nPIvw/wBZzopEsW6ldqapgraUlpkykgg+YpM4MkAZIj+JtD8P6tpPAl4i6N3VlA4advzCWHkbTvM5g+kEQrb2MduPI7VO9viuadPdArlUxxsdJCQSqSc+PCNdeXVNUViqR2kS0sJqGlpU25S7p2QDlUBM+3pyJIkMaavNVaakhZLJZPljakgoVIj2JAEZTAgAeoxx0t+MD4tdLeIzmm1dLFRJom22m6Z1p5fluVK5j1pR1klIwT0H145tavs4RfatbRSX0tytyQsqbAKR0EgGRP0I9G7HHtLL9zqcA9shK91oTX0Ilc3BR2qeUl2oeSdj1y+3TfsDU9ZUW2hcarkuLQ8wE7g40DKAJQQcAwM/lIvvdtVSvuJBCgo+WlGFKxgIBCSBBUev/biCOPHKu/Oqo6hl5biUEHzSUHf0T7YAyDug4Ik5jh11NStqXUtKTKVkqKULB2naRMgfbMdiYERxYxayenDlSkTpaWqLHnkVH7NVpYuAQHUtE7mlOOGPLWVdYB65T6hmFE9AE8NvOPkVqCxcpaTmGikpxpM3T9g+Yl9LjqH1t+aELbJCtxCHDJEEglIECNrla7R3ZtbUuKCkk7AFgplRJHypQjcQCJgkZHzHj1qQp1HpN6gR5raS835jaTu8xwgHfs/KrJQkbQR6j2PCNXQhhLVa9uqTLEDnkq4aq8UWtOY/InRnLK7XyorNG6DdrKnT9uU20G6FypWFVCypI3LUqCJcJhOAIPEERUqWhtSVhuTKilUDBAnMdSBj69gOE9PSKShtwqb2qMkHqARicdJA6HEn6nhbb2UpqFlbZWoLQlQCsznI+mBgnv8AXhaJyu4GOSJng18W2qfA34jNPc0NFKtKtT6YLxo0XOlFXSFTjSmylTaiCYCgPSQpJSCkgieIjzE5h1vOHmFftVXkJevF9uD11r3GW0stuVFQ8t10oQnCQVLwlPRI9hAj7TZqKZHlILq1J2JgKyI9RB94nr2H3ixHKn4ZXNvnV4OtY88tN6a/aXLrQFQ7SXer85tp4oQG1urbbnc4GkKQVkTtmZPTjwDKwnCryw0HSinQpKglsAq3AhU5UB1G337yPoeNiKZTziCtLzyOu1YzmQMdZJmBPf3k8bHEkupR5qvMYWImTtUUjOCTOP1gdJx6dZ8wy/8AuS6lKj5gKVEESDke+T7gmJ7YsSysra3V1Y6/Vvv17qkwuoq3S44oISEJlSiTtSlITGdohIwI4c6CjZpaqAp1+rXCAEJlZO0QkADH5oz9Zzw1jz31oaTuBUCodA2kREe394SZ/meJhoHRjlVWkNFdQp9flNPBtQS8d0pCSQPVmfp+ZMSRq4KbTjfDV9abTQpClqbeqVDaNhG1CCMCB+bI9OM4MiYeaZurcbWtx5bZUnKSfS2nAACo6T1jCsATPBv5xfDj5u+HXlRZteap0fdrNp28pBo7hUM7WxuAKJJ+QGMYBVIE9OBFoLWTOhNeWq7G1W68NWWuZq1UVa159JcEtqSotVAkFSFhJSoJBGwxOQRuHBo32CntBkGIzk8vimq4XVu1JcVQpddIBS2tYS0qBt3GASpCcjOBMgEmRw3t0ZQFKqVuPVLytx9KQVEDBySVZHfsme/C7Wmr29a67u91oLZb7WLrWO1CaSlQRRW5LjhIZaSpSoaQFQgEnAA492izraPoc8tYMrdWPWgHAzMx2gdMe+B9XcSweVM1k4ZE0w8cavT/AGvlO2UwncQpyIbCoKgAcqPt1gGBjsBlwoLMlKC+75acDYlXqJ/9o/MSe+QM8bqWmaogA3tST1Uo+t0YH2SR9M4PQ8THk7yg1Bzm1O3S2CzV18dbW27Vttfuwwz5qW/Ncz6WzuA3E4BPtPC7LK4+c9eqtSKnp6ZmkkANHIbBRKtuYtiXGk+Y3vO0AuFKlnAzH+EgA+3dvaoqq+ugpGxlAMLUNqQRjI+wPUni4Pis+FZdeTqHL3o9ys1FYFPluoplNbqu3jO1S1pPrY/KVgJKQPV808C/SvLu12F5O8m41iSnzUobJRTr/wC0dSSPzEznAxxyY8HfO3cryKugnj1R7g8gOZ9/YIeaO5UvXctw044nCkkiNxjEf6kiOn3KukuTSGtodipWHAhKG07khX5RtAxJJ6GR2nrxL9L6fuN5bLVNRIaLkJcWr0jckkgpPzK9UAxJGfUI4n1s0Hb7OkG4PVlxeXDrVBRgpNUkemVrUdwbIBJmO5gCeBFZxBS0x8nnd++iJQ2+sqBpcBG305/NRSw6Odvj6qemS6psDzixTEEpn8xWfSnEzv2zuEDqOJKrRNt0VTvLr62nYKAElqleCnFKgbdzsQIJQSEJ/KfaeJZbdM3zV7H4G309NbrY0ZTT0DRCXEKxlzqqAQSUz8sSDMP1h8MlJQu+bc3ad6qWIKnxIIV0Vt3GDgmVqMz9xwk3PieSodiokwOwRiitlLS+wMnuhczqLctw6dtFUopQpAcy2FDcTlZlS8AzMTuyIzw+0HKjUurnQ7cLkqholqCiwykpFQhWAACd6iAUnOFEZCuxitejLLZyip3t1amgmXXwUM08dAlJyU4OBjGcmQ9LulNb0KdbNLTobJC6h5shTmIgDBjESo9cd8r096I/kN+J5ojoLx5QhRpvw7UVkWsUFqLy1IIU++lJedk4mcJzuI7kK/huvXLQ0aCqtdZKtgCdgS6omJGFQkCI+VI+buAeCkzcF3LezQMVtw8knaUU8JUOm3smIMHqMkGeHBnS9bQpUtyno6RtsFe0uJU6omFbiYMEdcDv3jge68VOcyOXjadjDy3QGXyzqbvUlVBZ1LdSA35y2h5aR0O5S9qZEwNoO2B068bKPw8Vcf8APahapEpAAabk7AkCQB6QfboRkZkcGq8WOsvDYapA/VEplKWMCffpAIG7MJjpkSQP9fcsdQ291zzmKSjQEpkurndgQr2MiCPT3nuOJdNdpX7B2lSGwtdtsE1WzkhpZloLXUGsUj8x2qhQJgjMSciZkwB1HC53lfp9tLpSxO+FgthtJIifyjrJnsPtGIXdbddaOoT/AM4wtwpKQVpdkKwkJKpHXpB9z0meG1l65VLqUNugKSQlLagtGyesbsAAGSRESnME8T2id24lJC6/dhyKfNZcsKByjW/TwpyJCSoSFR6QYnHQ49h9QQzrrTn7GacBQwpsp2oVEQnoenQifoR2MEknGgrnUkCtW4Gyn1OOqKkoIxuEE9fsT94HHjVXLWh1wyWgrcpUoKkqkoxBV8og9sR1M9DwQob1NTPAe4kLm+lj6hVJul6cpalQDi2ticBThSFKEAGSBjHQd8Y4Y6u8v1TyQ4QtWOu0FZAwIEkYj/CenBW5nci6zTq3D5Ydp+iXSjy1tpmNoJEQI6E4HtwM+ZNmtdDq+5LsFPd6XT4XNG3cnULq207Ukh0ohJO4Kx7Dix6K6GePUyTIUGSliDgA1JKDUlVbKplcuqStW/blKkmSn0zJ3GO3tHY8FDlq89zBp009KXqyoQkvqbSgb2xHYEEAAdenQEdOBHY6dbq0spSAhSPm2kJUqMAp74xnA+3BZ5RcwLjyXvLF9tzTSq1ttQdC6ZKkOT2OTI29egBgD6NVmnZJI2OqdhnU+iA3mmnigfNRtBeBs07AlPdJdH6GoCVoc8tBle0kAHpEHsFQSSqMj7cE3RWuKDU1OF1KG2q4He4pKSkOZEEAHpJOIg5/QB33mSrV+pauvfZha1HenCQg9jkgdlSZ+kHhzsup3KWubcQ8Q4FQFob+XBMgDqo7hkAwEkzjgu1zIpiYTludvcoLS+aACcYJAyPXqFZHVa2qhht9DmzzEwlUpJQD6SkzJyDEfXvw209GtxpLLqXHEOKSsb25QgEAbhBG0gYx7+4Ewex6se1DaU04UW3G/W3uG4KkEGOmPT2Hv1zMr5Ra1p699NtuLbb7bihtKVb/ACifckbcwO2ZBntw322t14aeSrbim0DQZmc19es7TPnE/vErAW1uJUGnEj07vVBG0wZnak4Bzw7ahv8AceaXL21WjUF0uFw03o1CqO2sOKT5dsD7kqSkH1BK1okROAcpiDOrty7apKFa1O7miCrzIEAQdoVjsfuMdAeog5hsVFlvaXG2VENhRgtDeQRknA2wDO4gDpIPDTLSZjEgGVSM1w8OfQ47o7fDo552zwv8wnn61Ka+ndSlDSSAFGQYgfTPfse3Cj4i/ihtXOHXbFfb6b8KhKdqhAbUpO7IA6lXQYOJiOpFdqYrojvSCslIBaQnzSRJg7cExMwqMp7RmIc5L7VXuyU6hLpZp1FRA3B6MlJ2gzk9zAKI6K2gVV1b44PB07nbPojlNViQblaubLBu9ubqGlOICVKR5jbiSoBIUkLlGSJT8ySBJVG6AQIbvVPPUdKtbe2pYCXG9qtn7xJIxHqSZPeDjsVYO+mtK3Gs0fQpvVBcaOjuyPMtlxqaZSW6ptHpc8pRML2qJlQ3RAk7lhKQ7zT08dJ3StblJcpHgttYmV7idxSlQAB3SATk/aRwrXCkwBUN5Hn70VtVzbK40zuab9M1oermKtCFoc/6dSEJKPKJ3H0JECTEx0CiBOczHWtLU0TLDtU09TuujzWC6VNyDEEThQO4+oAgboGdpI1UtVtuHmNhltSFkbN37pB+SQAVSoR23Yx2gFnmTzkv3N7Qen13y5GvRpi3NWShDoQHqejaJUGSlAG4BThhRkkDr6Y4ZeHq/VCYnKu+MbToqmzxjY80N77tr6CXUkqQ2CmUiUknAnEmMzntAiVHXbXx5b7a4dpnyFuEIcIlJ9XYx6VK7gACSlSZPCpbIcDW0oQEon1SsNAFKyrpuSM5GTgGZ4aKtTVnrGXXAoJfJ/dhuVIJUCpHWd6imPUSMzMDiPdoyCHKTYpsN0lVasNhe1GqlYaQ47U1bqW2QlJU6VJj0R+YkREj8wgz1eeZHKi/8or4uz6psV605eW9rzlDc6JdI8ErTLSti0gkKTkH5epB68b+SfMa5ckuZWnNYWd5K7vpW6MXOhKkB5CX2XUrRKchSSUkGRJAgd4L/wASv4kGufid87WNf8wKawW+5Ulup7YzTWemcYp2Wm/XHrUpZUpxRJKiY3CAAOEpMe6r28nc+6jepCm0kLS5CCCfftiQMRJPtwZuSfi15o6a5S3bk1ZuYN003y01/Vss3q3OOgUC0KWlJW8QgqQgAhTgQYUlJkKiOA44goW6d6QUEqhUpChIyZOO3aTOY4ymSplxCfMJbcWuF7PVEkpUmcIJgd5jrgce5XuB1T3rmztaT1Nc7Y3cLbd/2bWO0qa+hUXKSvSlZQHmlGCtDgRuSoonaoGOg4QJZU47t3gJUghG4hIj04AHfrMR9T3410sLAca2wJW4pUdQPZJ6GT0+okzxY3xI665Ocy+XnKum5W6BvGkb1pywJptbV1ZWJebvlx3JIqGEpUdiI3dkfMkFv0bjqdhldI2F7w1o5oFWKxJqXyspUW0D39SfaD1InqB1j68S7S+uqzReq7dcKdKVO26pbqUsqPoltW4JMYCZ2giZgd+GW4X1FKXGWlEMNIMrSdqRu/KCB8o9x1IEEdeNthtjlWPxFUC02YPkKASJmUqXJz2wZ/WZ45FzQCSUapoHOeIIW5J5rp78Qf8AtCdx8dfgioOWb2jmdO1DjjBvFybqt7dSWQAnyRA2BRGZMwSAB145q01ucvaEpbaU3TI3EnJW4Pm2z7ZnuTHTh0tNjeuS0lxtW8JPqJytOeojAj6gmOwxxvq69ptXk0aZUjLj5UDtHzECIH3iZzGAYFVFY+TZg2HVPto4ep6Nuag4J6dT/pa6OzMUikNJaKtoO3arKBJx09JyQTP29yueeCQaen2peMlIII3E4Kj2H3PuI7njQw0WKVCUqWFqWNxSQVbkiAkZx1z0k/QHjegpp0KZSgPVDp9SU/IgCMlXWYTknrJ/UTI4A6n7lPFLESPDpxpb1PVeyAKjaD57klSCgHcCCDCR2yfc9vaeOhvwaOV34TQertTuMJSLpcGLfTL2ypKGGytyDAkb3QDjqgfc0k5Eck7xzw5j23TNmNO5drutSGEur2NrKGy4sqVEYSgkT3A9+O1fJrk/Z+T+gbZpyxU34a12lnymEE7jJMrUpUepSlEkqP8AIQBDmmL+aE8TyRUlIKVg8z8H4DqfeU92q1+Vn5SBhQOR/X04h+uvBJy95nXN6vrtPsMXKrcL7tTRrUypxcj1LSPQs9zuSZJ4LFstEQQnJ6RiOH+isxaSmEySPaI/24jPYHDS5IUFVNA/XC4tPoqg374a9xonyrT+oaBTClSfxtMpL6Ujp60gpOcZnqYjhpR4KLvoF8OXmnfuaVKK0MWlhb6KpSgDtUojcsk59QxBxEDi9FDb5QkJBKVGQY/jw6U9tIwNyU7cgHEe3AGs4ehmaQwlpPUJrpuOLhGAJSHAfD8lRK6aS1Zp+1MOI09cbHaCAEufhl+YsEdFr6JAgH1EdMcJbToKpr6kD8MtbzagnzFOCEEACZUUgyZHQn1fbi/tXpdi60D9JVspqKeoaUy+2uSl1BBBB+/+/XPAU1d4TK7TrS1aVZo6ppCg40mseU5UJ9wkKwsjAkkE49uE288LTUzA+kGrv3TdZONIJyWVQDHdDvhA5nku8+DVVN1orfSNkOFanAoYGdogCPv79M8NqnOX+k60jz6zU9YTtAWpamlY7AYj0n2iPboYVeD28aieDuqK51tidxQHEq2T12pRICumCZH6Ykdh8P1h0mhTFDanFKAh1bqMmI9RKsn69OuI4Sah01P/APIDgfdj5lMv8apyNLJNX/HkgjQa6cuLATS2k0LAMJKleQFbQREDP+Jzg4B4S1epqK3r3uVVpZKDErVvUiAAQSowIjMiPtA4sHfOSFsv1OWLhUGibWn/AKTCRviI+b8uO2OgMGOGD/8Agw5b3bLi7gl9aRDv4hwKVGR3jBHSB3gZIMeGeBxzJ5fqfqVqb1SsG4PyVfLjzZoqB+V3aocSkb1BpHpGAQcflIgHPfocw8ac5waPuqVU1+bubS1p3qWSCW0j3HUDr2AiOszwRdXfDns1MhVRpm9KfbC1H8NWEHbJKvSomR1z1EcBzWvhKu+lt7iKMlTBDhOxO76kJOFGc4GZg8Eo30Dhp1EKRDcKWoH4bt/kVKta8g9N68sy67S+oqdwqRuU2AjAEGCjqIG3BIj37cV05haAqNJVLia5hhKmhtbW0Skox85IgwR1IBwOh68O1z0rcrBXOKoK6ooqtJkpU75clOYM4GCQIicfQ8R6/wDNu+W6GrykVrSRG9afKdSJwe4nJ9JkCR9uDVvp5mn8J+tvY80SgDmDBdq/NNNkvlTbKpLK1OVFOo7UKdSQsn5QCoHIzHZUqk5jh8qLrUWSrDKtqadtJBMqScHb6REfmEbYJntg8Qq/UlPdnRcLa4UrGNo2gq9UKCgPoOkdes9iTY9OPcytIU9Q0giqZTsUNm0pxk/5de4yOD4oHykFjc91lRVxxbyHAWV1xa1BQ7XlF1hfyArKnEiDHUAK75/7T+oc11yiYu77pp4plAFf7vKVAkR3Ek+5/wC7qBkzI0nddNIUy+y60pfpncpLaCB0HYwIPQxJmIniLuU7i9UN0ppXKvz1JbbQE/vC5JAAzOen16z34yFtRQy45ei5wzxStOgghAGn0KnT1/aU4xtSh8r8tBUVKGfTMBJnPqkqMGBnJQ50an0rcdA0LdsaaYr9sPhBIJxMRHYfxMdOPnOfQlZpCrqqV5ioaXTK8txDjcKZ9Q3NqThUgnKe3Q9+BLdbgp6nXucgIOwkkbkpIO4be8RmBAHt04f7Pe3+CRjOVpJStk0vzsord2VIqFPtuJ3QAoTIUoGSk5/mT/sr07qFbitqt6vMbgEE/MD1nKSDAHXp3HTj5cHUu0ziVktuIVtQqSpBKcCB1n9ScgCTBEXVVOWe5oKlpSkOgfMQFHbG77pyJHQkyZzw2UlSDhB62EtycIr6Q1YLZem1JcDKCEkp7kFI/ugSnKJ6EkiRmOCfp1tylvSPLLhbfh1pCSCDJIMicHMGe8/Wa82+7KqJKCElhZKJBUlUnaowARPqBkCMiAVHBl5La7/HKp6N0uKbCv3RwpSSchJM9SIkkAZMGccNNtqBE8avgku80v3iBxbzwrV6PvbV/wBKOU6iwytLCijITuISmJA9Uz7TOSDwOuY2mlVda+gstKQClxK3mk+iT6UqOBGZkwRjIOOCHyY1jbeV2ubfeLvYEaktLG9FRa1uqaTVJUgwlSo9MEz2JIP6N1XamrvVF/yy21VSsBJINNOQQTmAD83QwInvbVtnEzNB5L5S4oon085lAQqs1mbqmFNL89kKbMJUSPOhcjd0I7ExORnpPA81rS/h3ENupC8khJVsDW0EDGQPSJSSDO6Jk8Hit0s7ZKlTiUv+dTghSdp83bI9QIgTAGQfbgOc27e0ivW1CilUvoG5K5JIUDk9CYPuSTkDHGt0oGhmcILbLo4vwFGrbzi1ffqTT2mrlfLncbNpBT9DZaCpeK6O3JdXvWhoEApCl7VEKJwOogxGuel6eqb/AL0rdd/EMB1RUSCkJV6kwZASk+kqV6Sn3xxso6cP3be2PxRWlRBUoFUhRMnb8w2ySowITIA7oecjaq6ltT6Ggp9wpbwhMtlSlJBR3C4QM9YJJiSrhRroP+2c0JottTiqa/Khte2irpVJX5a2dqElM7d+B6ogbSUj5YJBSBGd3DlYrspyyuNBxCiztbcB6QqdpCR6UiRsCSCUxPvwnoGdtvC9kU6IbUzKUpabKjuABOdxLgCVGSQEmMyi0+ypj8ZbXE+aUBYTLstrG0gb9sJ3BOAE4ntBI4AWmQxTj1TFf4xNTZ7KYWDStZcNM09d5Ly6MrVvfbXtGDGzeRnO7qI3LQCe3EE5huqYfq0haULAXUSjchtxYdG9Se6lbgBCsAgYJk8F7SHPC9v8qkaLFUtdgYfVXtstUyVueYQEKSFTJG3adgmCMRuIAe18FJuKytSQAtL+9I3tLBUJ6AEpEzMKACjkDPDNcz+GCk+2sax/lKrFTVlOhp0qQlJCk7CoblQD7jvABxAyeuON7TKfJTvB3gBLbaiZiMSIGM4IyTHYicoFOvtNhanZdHmK2K/6ihO1RyQQFZ7QBjtx5UtKqTO8qSkgqCZK4MEAE9DPbrkx24Q01LYw1D+9CMwQpMlW9YPUgdZgYPcEwRgbPL3PoTKVbCUQBuUYlKQkjKhETByZ9p4etO8s7zqpFU3bqCrrSwFug0zS3g37rO0EDoO4PrxiZRLtjjdQ4y8ytuoXLat52ISEpASSCCTEEnqPl9+M6ZWNGTgL1aLYbnc0rSUiIdU4FApKSpUjGAFZj6jh3qbs20W6JshpCZUtUgKECSonPsD3mYGeE7zabJRfhmStDygC4fUpRJwN2IBzMH6HhXZ7Om2ufvQH6lR9LIIKWfYqgxu6HqYEj6ccJJA0ZKOUFI97vDh5nmegC3WWylavxNQFBLUBpgblbSBAEd1D6+0mY4mmndOrqgpRKUBhPmKPzJSeoMifUTPyz1kkCSNOjNMv6hvLTbSPxNWrLaAAW2+pEzCcA9/aCPabasqaa0WxVpt7rK0tyurqVE+Wrv5fuZ6FO0Se3EOGF1Qdb9mhOIkhtLGxRNzI75n39lFLjeFv0aqOlQtKHZU6tO390MEweyhPQTAjoRxopLcG07oStQIC90QgY6npP8z6unb2mobcaDoLjVI2r0lXzPqJyEgYPRWAPYdhC+22Z+4pQt1PoaSEIaWZDY6iYwVZnaPvMY4G19XgaWnACarFaHSSCeYZkd07LwilIeQGwHql4QiJlKTIxPRPaSBI7Ykuti0w5WutoUdwyjYlO7cRnalIkwcRI+vXh50zo5dY/wCTTbXHVr/e1a8t9wSo9FGTO0SYBJPQcEPS+lGrS6KWhaU9cFK8pVQ4klW4ylKUIwdpBEjGAeuQVuerJOlnNWJFDHTN1zHkrCfCQ5G1VXzsqtRrpkooLBbXGg44lJWqoqP3aSDMpISHQR1ycAHjpnarOEg7khPeQOn04EXgV8Ob/Inkyigr2FNXq5VK6+4blpW4hRAS22Sk7ZSgR6cAqP1HFiLRYiEhJwD2mAeNmNIHm3JVN8R3JtbWulZ7I2HwSe12fahJKOp9pPbh6o7XPtB6kHH24X0VnWHUHaVJUIWem3Aj9OHyjsmAIPtPsZ43QBNVNZsSEnd14X01oGzKCrufc/bh6pLNGdu1J9skD/fhfT2ny0+rbuH5gM/bjMLEx01qAPqTuVMRH9R/nwqZtGPUkEHJxH6jh5RbfLB9KQUwnrgfT+vfj2UNszJmDMDH8eMwvQSE1NWUAY+UCNyhJ+/+XCbUWiWb5S+WrzGHGzKXBJg/XOR/pw8O3JpslMoBPtmT9fp/twjf1K0ykAqk46dP14j1VDDUxmGZuWldYp3RuD2bEIcVPKO5W+rZbUadxl9YZ81CVHygZyqTIEex7H68aGLVdLa/tYZtgdRM+acNESDIOMmMA5mDnic1+s0JTPmJzgZgGff+fEXu98becWpLjhSvO0qhIECSB9Y7+54r+5/ZzTPYPuR0nPU9Ecj4inO024SKm1fVWupDV1slMUq6lASkgDuPv7Z6DPCq80Nh17av3SQlU7ipSANqiOhPXqTE/bhurX6a+uop3X/KdUqELB9Kcd0k46dR79OMqLDVWujU9DbK0NepaFSFdfcQR9T3PCweFX0s7WSt1DIyMcx70RZVxysL4zocEJObXhlobzQVVW9+EPlDehxtuXAOmd35R1JkwenFU+dfh6XQpqA20mDJDhBVvSRI9pge5M/YkcXpqtWIcp0h4pWogpUlYwqev/jvPAl510tNX2lPkJSUtKUElSjuyJiepGCO+McM164ZFBH97oHaWtGS3JPVGOFuL53ztpKnzajgHbZc3m7G7pTUS2Er2tOShSFwCgSI3ATOREHBj26dPvg6eF7T/NwXa53lH4iltqENfhZKfMUSYK/qAIOfpxzz5y0zNo1GZTuAWCFDASROP4Ht9cmDwW/CF459ReF+5/jLPWIZUtIbdbUmWnkgHqmOgjB4sj7PbnT6tc7g1zm4BPLJ/fNT/tMtNbW20x0IycgkZwSOoC6EfE88J2luX3Ldq/WWkaoCt3ynKdBCQqRO4E9Iif045WajokVd2Q5uSpxhYV5qfSUbT79ZyDMwZ7cHLxQfE61D4o6JukudTToaaSfLZYSUNpB+YxPviZn+PAA0ddTqCufdKpCkLV64hPpzIj3UTiMDvxt9pNTTSNi8Fwe5o3I/fRCPsvtFwoqV4rgW5OzSc4Czm5aF3a2qfecXWVTgBUtwlxxSpmCo5k9ZPYzJOeK9Uuo3uXuuF19PTUF0NMle1qsY81h0KQpG4oJHQq3BROFBJz04srrghmzxKj5be2EJ3zBBPp6YKo69x0xxWzXxpKHUyTUBLrReStxIO3zUbo7fm7CZzu6Dql8NSuLi1WtU4EOB+9lCLgVU6vLhIKkBxa8wTM9QMKIjp/GOGa6oNUNu5SS4pIKtolKicECe5x365yBwbfHDq3k7rDXNre5M2i+2WyptbCK9m6u+Y5+JKR5hEHI3dwYOCAO4PdqYRJGUmVbxEZkkpBAn9IA6GZ4sqNnhP05z6hB4qptXTCXSW56HYhb6B8voS5Li1oICtwUsDMAlU5PToR8oH04KPh/19RaG13YbtcaJu5UFtr2aurplHaKtttY3NqIwCpJAxOf0gXMsgKbKEq9BUjaQTuTGD9ZEziADM9+JHYqlaaQtwYhS29ypJXAIIVOSAQcAnJPcnhs/oY7slQNDnSRu5HIXTPxteNjQHiy1BYn+XukE6XYoraGqtK2EtB9W+EyECCEgQFfMoEiRBHEU5btoqKZhDikHcvcUKG0IEe/QSOncA56cVR5MVqG0KCgnaluAkKEqWVkAEnE5EiO4zMcWF0FeF3KvoWtzbKlwy4t9zchIWQNxJykDeM9hntxZPD8oY0Hovn3jG2sjaYGDOEQNZaYoFUS3adbXmNnaUncUKIJlW4ZTCswIzHeYrxzQs34itrCkLJaG3cpuVMkkEuJgYMJA9UAZxIgnTmO4rSGqq2yuXC23BVqql0ZqKJwO0r6knbuZUSNwMDMEEH3jgQ69T+KQXkrBcdJSC1KScfKMY6mSTmcgEGG6of4se6p9tP4T9TRhBFTBTqBC1eUvzqZIUPMQGx6oJABz6oxgkznuGXmna/J0u2ypa1qZeSoBaZDoKx8/QKSVd4zHcjiZavtLluudM9ChsWfVtlSkhUeoqGcz9fUPc8R3msy28mqpiA0j8OVFJaC9hJneAkScSCnpCRnpwq3CICN3uRy3yuMzUM7NrOsprFX2OmXTppLlUIdd8xlKn0rbUvYApMrSCFOqUEgg7M5EqakuGkujDqW1BKx5ZQkhQgk7ASpJBG0picSJ2yCR9tdsXdb8pLVMmoQMKbV+8RnqCtO0AFW4kekkJQP7o4U6r04u3VC6d9sML8z9wXlJCnE4grBAURjd06LPsJR4AfEBHRPdQ/MBDuqk3K15VZVOqJQYpyqdpUCpJncNvqEFO8FPsme4EP5oUrpRlxSUwv1bZTuKYKUmJgkbQJkYAJztleiypujefd3Nmo2OJL5AUAspJJJkJ/MqAegBMQoFi5tUaW6ZcRvDjbiWwE+YiVbSBuABIABgEYyVCPSzXEk02eqT7cP+5x0Q/wDCRc+TlqqNdL51WvWd2Yq9LVg0sqw1KWnGr6VfuVPk5LEk7jBAIgggxwIm0JStKQ6HHDElJI3wcmOkE9ExPXHt4pGfNAUVO1BVErGSAPaYxOJ7AfWOPaSHUqjylJIDRO4GCdwhAyOpMdsBXbhFTdtldwP7NL4zfDXyA5A6ysvMx/TmntW1D5d/GXlpDorqJKD+4QVAkbSVenEknA45q/EU5iaH1j4steXnlvTCk0tc7s87aW07g21TqWIUP7oKhIG3rEiegFtb/wCz6QPwAkpKPQjcFwMBJgSdsZHY/XhVb6J+uq9xQj8VA2pmQ0ACIA7kgmZ+pnEnHTt0aAOamUduc6TxM7L5R21VCoJQQ5WvKIBM+gnBjtu6/bb1I4lmldIPO1LNNTIU/WOZMK9LckE+qYn1CP8A3g9p4lHhh5Jo5085NO6QbrWqAX6uYonbi8T/AMsh1zaDu7DIJ6Hp1GDdL4qnw3NP/Cgv+n7TQa3TqZvUdtU+oMtoYq2XMp2hKVGEKPqCjmB0jIhtp/FOqQ4aMpvirYqHEMAzK7lty9feqljyOXGnnaVh2jqLpVg/jNyVfu0zBnoAnEEyZJIBHUw1brd1YLrhdXaGVKKSoQqoc6yR0AMRGYn78eq6rcuzqqioJYpVOeYy2JPmECJUokFR6dDGAB78OVqoA7TpuVcCmlQQmnp9sl1XVIgROOmJmO0QOuNyBAiiGGjZN/D/AA3JI81dWcuO+/T1W60afS8Px9yA2iW2WSYmDIKj7kwYEYBmOnE50foKp1c6l1bSk0jJO6USWxmFQr0TAPzSAYJPXj5y55cuaxuLVZWNvuU6BtQlAIQ2Py7lJ6mZyCAJV3Bggts/8TPMWqjATQtqASWk7V1piQkZMyNkn5cQO3ClXveDh53/ACHr6qxaOohYwiDlyz1PuWmio6VmgX+FCmbeB5DjyJK3HMjbuImAUjEQB1meLNfDk5WW658+bLW6hpA008l5FjoXmwgKfSjch5QOSdqCE7hJKPVAA4FY0zTaOZpkP0yKqudBZprayApAUSUpISZKldD0mAR7AWO5Acu2uVmv7LqXW1U8m+PVbTlPR07almkZ3hUwSJUBggdAoz1HC7PeGRFug4BI3PN2/QdvVcLhRPqKZ7XnzEHAHTbquglh08BtEbwYJIPzH34lFnsgcCUlIAHRUY74njfa6ZpDMgpWhQ3gjOCJBB+3fhzZrG2x+UqAiOGgbqjiO6U26zgQqNnYiZBH24cqenQygEmI6T0HDI/qpmmAIXvP/uyof1/hw21uukMNEeYj0Yge/wB/6zxsGrFMl1rVMgA4KACT3BxwlqdQs0424BTJIkZj3+nA6uXMQSRvG7JP6fbpxH7rzLAblLkkdMwD7/fjbGy11BFSt1q2klJXCTOUn6Rww3TmClswFlM9IGR3x7cCm6cygSuHCYGADk8Ru68zIKti4MzGPSOPCcLM55ItXDmL5atwWUp+UAGD/H24Yq/mOHBPmA/mJT+b2j6ffgQ1/MlInIUZnr/HiO3HmSQcukFQ9R9+PC5ejKMVfzKO6AuZO4g4A4ZbjzKkQXJV+YDgN3DmXBIDoz1B6n78R2480S0iA6TOJmT9emMY/nx5qK20o0V3M8oyHSPLGNqsgz7/AOfDTWc4Kl2nW27WVa2Vq3qQXVbSZmSOhPTMcAu7c0UAD1hHvKv4frxHLnzZTlIe94JPQ8czg9FuAcKwFbzX3A7lwlWIBwQe324i+q+Ya7mwGwv0pWmQFQIg9RBgAdzwCrpziDRVLonvnpA4R6d51UlTqdlqqfbabeKpcUjclsgE5EgdoiczwMvEb5aKWNnMgorYyyOvie/kCFA+f+o/xetLgguBpDS0gDBCIAkAzJjrt9+8TxDrfXf/AMrfddDzoCF7wk5ThXpCoIIyD/8AEe08JdWX46j1VUuugpNWohSFSC2gkTMdgIBxmD9YsH4W/A9W84qe01l/eRbdN17pcrksr/5x1tJhICFAhG+CAsGRJVA7rtupn+EyJg3GPkvoC5XCmoofGqXAN/Pbsq5aPrXHK0r8xSWgtTiiFbYgekkdJgTI/NODI4udyA+HdqJXL+qumonTZrnXoSu3W9YCvKSU7pfj5d0gQCSnM+xMvgz+FJZOXfP2t1Vqa5UV5s9rq1uabtqEbm1BRKkLqdwJJaCglKdytxTuJIwboam5et1ZUpptIKpJAIgmev8AHM8HxahLl1R12x+v+FW3EPHLQWw2s9iXfoP1+S5H8++TmotCWB165Wmop6Z0bfNCQplsD0kqIkA/+6CQZEDil3Oa4JTXKlwJW0VIWXFzt3FWM9cdCZkfQjjvvrDlukIfCmvMS8ktuoWJSpJEEK9wQYg+/Ag1jyH0/darzKzTtiqXUjalT9uYWQANoTJT/dgZ7Ace0FpbSyao3bLWm+0V2jTUxZPcFcK3f3zClIVsWleEEgqQYwY9zCgJGO0kSM8tPltnqlPpUQZCRAJI6jB7icn9OOv/ADD8EHLLWtjYoKzRFnpqajW44yLcg0LiFOABZC2ikmYSYVuTKQY4p1z9+FjqXRYqa3RVX/xNbkK8xNA6Es3JtO6QADDTxT2IKVHPpnq1wVDC8ZUui4tpKhpY7yn1/wAqp6QWal1B6lO4zCkwPmM5nM5yMdJjiWaaoQ4umaKvINSQ2pxS52ElICSVCQUx1kHOeuX7T/hD5i6vr61NDo69KNqaCqoP0qqZxSCraEth2A6tIAhKZhMmOh4M3Kv4c2vNW0VM9dlW7TlFUUxdT5xLjrUqMNKaTDiZGREp2kdd0BnbWwNb5nBQZq6njcXOePmmHUGiWOUl9atlNd6W9qShDqKilSspO8boAVkEAqBBHYjr6uCHoh9NZWI8qQS6BAcDhSmSkkGe5AIKpGI6RxG+YnhM1ByGrqBVxdpbpQVbyWU1VGlS0DbkJO4ApcOfpB+b0xxOOVGl6m/31mhpWX6urqH0tJZQkKO4kbAAT6icR1I2nAkRYFkqmSRCWM5Cpnid7Hzuwc+qcq9tFE3BISCQZQobVT0GSACFZOcAYzJ4bNI64e0fqNi5Io7TXPMsuodZuVMKinUpaSlRWhUDcATGZB2HtHC/nZp658ubm9abxRrt1wpv3JaqU7VsL9jMZJj3gexOYRaroF3Bjc4rZuUNpWpIGDiBgJSnJhI+c4JiGiOqbjDlW9VQmQ+VRm+UDd0uzaAHk07yghS/LTC4BBB3TJIKesZJyR0IPxJvB1pnw0aG5d3yxa9tGsHdc2n8VVUtIlC3KRQQmBhZKkkwgSQoqQYBIgibmdranNUVoRT0qA15KlsoAyBCjO4blZ3HYo/QADgX6quTl18tx0+agoSCWxvwT2UNoyPTGe8TuEBLnUh+Qw+9dqG2OifmRK/B5zNsfKXxD6Z1DqKzU2pbHZa1D9ZROpJTUttuJVtJJ2gKHqSI2jYDCSqOCp8VbxJ6K8W3ifOpuWumEaUtNYyhkUaUNMuPuAet0pRKQs9gAZ3JjG6a62qndXWMtrThpsLUAhKUnopslMDaQnYB77TAyTw5Xy3rbQumUlSJ2LSNiWiQoBWBGUmRMn0hJjHAOmo99fVEa2vAaYwt2lFt0mkF1LBStO3YVhG4LUASZ943ScnB6CAQ0cwqn9pW0Ogne0o7lKWgpKQQoKBUTnclRAiMx7Hh/q6lyj02/TJUmoUQh9Xlo3F1UJIicEJMAgRHpwJxEr1VecHada0OI3hKCkH0HcFg+o9CSZHUwIBgQTuB8mhBrYTrLvVCDwy8qRzx57aT0gmp05SPakubVtRU6iq101sp1uApC6l1KklCEiFkg9QgZynhBrHQL+jtfXTTlRV225P2G5P252poXRVUdWth0tqcZXAStCvL3BQGU5AzxHqFhNQWoStYIlYCU7ikqIxJIMk5k9+sdZWxaV26lRTIbK3p2LW0ncBIgNpAMZjMZEqE54QnOxsnikpTI4k8hzSTy0VVwQtxSvyobKAVKSndGBnMwCYPQjMxxJLbZF0G2mKUrqlkOKCQJQIGD/eURB+g+5HGyhtzeicpDC7wv3TvRSAgiT7rBkBJ9QKiemOPbi1Wrc0G1vVbhKlNu5Wsk+olQ64weg9RwTBGaA12l/xRmGQ+GXxjPb3pwtd8esqHWqJ9DRbAUt5vKnFAkglX5SDJET2EgwOHXVOsLrzDu37Qv9yr7s82A2hVVVLddWkKwNyp2gGY7YEATwzW22IbaQ4vaURCCIBcPuB9+8CMf3Y4frfbFXdalur2UtIkF9aUghlBAwEnBUc4OTBzngDcbmXHRHs3srN4W4W0N+8VG7z1PRebRbP2m+9WVm1umZKSRlIWegQBkqBJiPYd+045e6LqNe3lNZV/8tQ0w2NhY3EjBKQB1UEiCQJxiIPHzRehXdZ1qCGzTWplSkq3H1AnBKjgKWcjMyIGAMEporQE223g/IlJT6SKdG7KSRjE9IVmQcmeBbneENR9s8h/aO/vTdUBjhoG0Y54/qPb3L5UIXeG0WKzMeUVIC33lkJab9KQoqVlO1IJAECcR19RH0za2dB21hFG29c73cEhtASDvcKj1QPb1HoBkkz0JbND6c/CoU2zTj8Oh0B5CTCalzqSEkxt6nbugbv04PnLnSFJywZbvVyKXtQV8ONKUd5pkqP5e5WTOMCDknEVzxDeAD4Y3H1cfX0Ryhoi1okeNzyHQD/KU8teWLXI62q1BfG2LrrevSttLIBVT0G8j0JzG6CJUM5EYBHBG0jp1vSFai93tSrtqevUfwlEDu/DpUJEJwJKTG6OxIB4jCLidL1dJdq6ldrb1W+m2W/CnN68+c4kYk/KlPQbQcHIsV4e+RdVohoao1KTcNZV7fmNB8/urWkZhUzJGM5iYnpFd3C4Oa0zTO39Po0dh3Umrljpo8dT8z/gImaF19d9PWa20Wpall25vspJpmGjNIMBPmEfQZUY94GeHWt5iJLJAc75z8vvxBmUi+XGvFH5tbUxNZW7dyFO4hpA95APfA4gmpNcVNlr6imqULp6mmVsW2sFKmlDtnP+szxYnAt5lraVzKh2XtP0KpfiihZBUB0YwHD6+5Fe6czNhUfMAKjhXbhgunM2ZG8qMdjmP9OA9c+ZYQpUOR16nB/r+fEbu3MsBRhyYn6En/GeHjUloNRiuXMpIWoeZ8vUkz3mD/XfhgufM9Klql3AGM9OAzd+ZgEpLpO1Mnt9hxG7rzTk5cA7ACI/n/h9+PCSVsGoy3fmXuUR5pyeyun+3EbufNBJlKVwc9e/1/r24C135qDP73A9zG7iMXbmyEKMPfm98njXJW4wEb7nzRTuUC5GCJ3e+Y4jl25pEqJ82Tu6Dof9+AXeOcCgVISpZHTr8vv9s8R6v5jVFS56Cev5fbj3TlYjfdebkgbXEhIAyVATPcff+fEZufOAbClLwIIIJBIBA/8AHAgqNQ1dcQVLM5k9Y4TP1Li1AqcUVRPWT9+PNgvBlT6782VOqJBBj3P+H68R+s5gVVSowopHsDwO9T83tNaLB/ad6t1KtP5FvAr/AEQmVH+HA21Z46NL2hKhbWLjdXRO0hsMNz/7lZ//AOTxLjpJ5PYYfyXGWphj9t2EfXbzUVGd8Jmcnpxperyj1qego9U7un1+nFONW+OvVN2KkW2lt9pbVMKCS+4AemVemfsngZaq5tak1stRut6uVaFGdi3yGx/8BCf5cEIrFM7+Y4D6qC++QsOYwSfkrzau8TGkNAVK3q66Ww1jagpLVM4F7VT1LaAScY6iI4tr4B/iG6c5uW8N22vdQ/QQy5RPDY8wjokweqY6EdfocccRfN+mPpjh75ecyrxyq1dSXyw1rtvuNEqW3EHBHdKh0Uk9wcHiSOG6ZjD4ezu68qeLq6q0sqXamN5Dsv1U8pec6K1ppYdSUrgJIOQP9f8AXg+aP1/T3elQhawo/UiQZ78cVPh1/EdoeeFibpKp9FDfqNKRWUSl+odB5jZ6qQT+qSYPYnofyr5zpqWEFT0wAJBiZwPsc8ApInxvMb+anxSNe0PZuCra3axMXyiJSgCQZEd+BnrblztBIROSRKf14eNBczBVttodckqiBuwocTsfhtR04Uk7gR6o+ZJ9p45kLpnKrJqPSCmFKPQ5MniG3WzK3GYUPzAjPFndYct/OCihG5MzuTkA/X24EurNFLZWoLbUDJAO3BPXr0/XjxYDhCh6kDh2uAKKDIBlUfXjyEhsHAg9VJPX+PEjutiUytRIIQT3yOGh+gIWRkkGOvHh5L3UVHtWaao9ZWCsttc027T1jfluJUncB/dP0KSAoHrIGeK4WJu9cg+YjNU0tygudidRVU7qCc7SSh5JjbBgnInqO2LUOUaisEH1I+v8eIjzR5QN8wWmXGahdFXUra2UOKTuQpCh6kke05kZx9Tw1cK30UMpinP4bufoe6FXWjM8eWcwqv8AiH5u3vnTra4X/UNZU1VxuLpXUVBGxK1fQfYJAA7HvHAp1NWuWcKU82thflealKyqDuSPLKSRIHQgEmB7dOLC33wWasLo/BO2mqAdCgsVXkqg7QrCkSR83pOACYzHEc5teAbmZqu4MG31FqvC30pZU6agtFiAYSd4yhJ6lMkykBMdLE/6hoD5WSj5oLS26Rpy5pVO9T31166uOpccSuSrcU+WnccBSiQk9QIBk/LIiRwhuDQfp1KVuDlTThBdWkbtsiVKGTkEScfl9I2pBkPNvk5qLk1rlqx6otT9qrqZfn7CWy24kEypLgO1UncAQdoKyJgEcNlLalXCkpkt7nz5u11TRJ2pEQRPqAmT+X1AiMAHpGfEcC3qo9czQHPK2aJ09Uag1GiipqdTrzqQgMMoIcWrcQ2Bt9JKZV8v94GRJPEq52+GzWPhv5k/8OaysVy09eQwKwU9U2lCi0oApdTt9M9iRGAUmTPHzQz7+j3xdaN9dJU09QlVK62opQy4FBYcSsqJEEKVOOoz1HEm8RniQ1p4qeYp1Pr2/PXy/IpW6Np4oQwtplsEJTsSPTElRMdTJMnhhZRuaAen1VZ1NyDpHAHl8kGdZNoZrVUqQ4oObUmYM/MRuQrocTPWZInI4hqqs012YQ6vc6+35rifUVKKSRKt3zCfSVH1SpXsCJhqZpa744sF5DVOr0tp9CFFAJz6oCQYmegiAJPEDrHCmspWqfzUlKVNKXISp4BwLK4BJCyFbSYAA9oJ4E3Ebo3bHYGShbo22olyreKGkMALbWpRgLztJPXckBI7THeMvdNqZKVBdvpmg6oeTTJ2hRkjK4OCB3EGPV3yInT1vn0qWFKCmm0CEAbZnMT0JJg9CEj7kcSLS9sS1bl3CuBCduJV6g2Mwke0nrAk/SeEQS+EfF69FY1LG6pApI9m/wBR936JxbUiy0yqyrK3n1q3NpIILkACVTn9DJHtiONlronCnz6kO71KBKE42qxjHScnBic9eEdmp3dSXVdbUpdRsJCEqSAEiMJiYyD3yJPtmTW63JuDwdP7tpOGxu/KB2/nmO/TgNX1ZY0s6nmrH4ZsjZ5Wz6fw2+z6+vuWy1UDtyV6ihTigAoEEop0bpyegGfc/wA+Jxo3SKtWVrbDQUi20n752pW2f3qpy6uMAxMdkyJyZ4aNMWZd6UostJTTIBQgGf3yyrbHfcB6p7D29zVpfTrWj7KikbR5laSXXNhkbynseoiY9I2mDJzgCMhwIGXdB6qz3PDYy0HA6n0Xtq2ilZbpqVLbNPThLanEtlJUqCIUAJI3EAyTgJ+3Ex0By5RV0qnlOfh6AkIcrIj8QSdwQmI7EkCY6+8HXy20Ku9VTVbUKShDMhKwCC6CQPlBgmdo6zG0iOnBTp6Og0/p9N9uzYTZaKW7fbFZVWvzEEd07gmc+vIzkkRxTUttsQp2u1Tv3d6egUS0H+ISmfGIWbN/8j3W+x0VHpCgpLhV03VR/ZNpAn8UNwhxYidogRIyT3EcT3RlrdpmHr7dAt2r8yGElshDrwAGxAzO3pOc7usSGLlboe5axv69Q6ggVq4TT0illYokdBt6BO1IP93omJji4Phv8P8ATuXOkvdypkKbpcW+lcGGkjq+tMCD12g+844pS73BkOznZPU/oEzVNaymjL3fBJPC74T3bdXI1bqxttV9qpfaYcAUKBs53k9A4R06RtEQBxMNUahq+ceq3dOaUU5S2mnATcrnBJZSMeWgz85yBOJz0E8PGt7xcOZWohpGwOFCFkCsq5grE+qVdgPt798cHLkjygtHLXSrPkUyTTJMoxC7i5OXCTnbPuen24X6SGa4zAuGw5dgO5SHc7x92BqajeR3Idv30XzkhyCoNH2mmdq6UIYalVPTrkpB+bzFyAVLMSZz1gcU8+KPZL1y45xp1PUs25On9Uw1bnabclRcYaQHEvBRy5kHemEFMDBSZ6Sab0i9eXw9WFYClSlsdB9Z9uvFQ/j48htRaj8KdHrG03amp7Jy2dXcbjavw3qqQ9tY/EB7djykqjy9hBDiiTgRffC9kdBTfhsw3n6uPdVHW3l09VqmdnPyC5v3bmkk7j5xhOABn3/r+PEUvnNpFPMOyYIVB4CNZr2srCoblBMxkxP1P14QVF6ef9TjuO/qzwc0jqpek9EUbvzgASUhwkgbQkdv6HfiNV/NR95UIUY+mY4HOoNb2rTCFLuNypKJI6l99KP8TPA+1L4yNHae3JpqipubyZ9NM0Sk/wDyVA/UTxKjppH+w0lcZJo2e24BG2p1RV1i43FPvJMxwkcqnHlfvHCT7SeKsaq8d9zqkrRZ7TS0oPyuVThdUP8A4pgfzPA21R4idY6tCk1N9rG2j/6VMfw6B+iI/wAeCEVlqH7vICgSXmBh8mT9FdjVHMKx6QY33S6UFEEiQH3gkx9E9T+g4GuqfG3pCxBSKH8deHRIHkMltE/+5cY+oB4p86+updUtalLWsyVKMkn9eNfBCKxQN9skofLe5XfywAj1qrx5aguZUi1W2gtrZxveJfcj+SZ/TgZaq546s1nKbhfrg62oyW23PKb/APxRA4ijaCvABJ49qYWlG4pMTBPafb78E4qOGMeRoCGy1k0ntuK8KUoGZyf58YVk8eeM4kKMs6njOM4zjFizjOM4zjFiedEa5uvLvVFJeLNWvUFxoVhbLzRgg9wR3B6EHBHHV/4eXxJaPnLa2rbc3m7fqWibAqaUq9DyRA81qeqD3SZKZ9oPHIfhx0pqmv0VfqW6WurfobhROB1h9lW1bah7f5g4IwccQq6hZUtwdiORU2jrXwOyOXVfp65S8526ttrY+mMTKsH79/6+nFhtA800uoRCxumFJUr5uOIXw9fiPU3NOnp7ReHkW/UtKkF1tJ2t1qRguNz2/vJ6pPuOnRzlZznbraZB84EqgKz1+p4T5oHxPLJBumuGZkrQ9nJXvteoWL1Spykz1B7CPb/Xhq1XoZq4NqcbQMzkD+EjgP8AL/mmQ2g+YrrGff24MOk9bsXWmShUKQrBzn9fpxwXbIQm1foRTC3ISUkHBHUE8D672PyVEEf9u0D+fDl8Rb4oPI/wE25VJrS+PV+sqiiRXUGl7Syai5VbaypKFlRhphslKvU6selMpSvE1C0x8f7w4a/09T1dzqtY6Pr3nfLet9dZF1hYn/1A7TlSVt/UAKnG09eO4pJnN1hpwuJlYDglWSdt5QegmZg9DH+HHluiWophJO8lIKcgkCY+vAkq/ifeHMWqmrjzf0iaety0nc8X05SP3jQb8xvKhhaRjceiSRzh+Jv8Q++K8Ulzc5P8waWm0vcLHRUj9y0xWONmuO3zFeeqdpfbXKUrSlKkoITuOeJFLbpJn6CNPvXKarZGMnddRuaXiZ5b8i1uNav1zpewVLSQpVJVVqfxJJyIZTuXJxAKe89OKbePT4o1Pqe02/S3JzV9tqqG60y/2vcqJuop66nUFjay26oJCAQCSpMrEFKgAQTzR5mc99Wc+dcHUGtb5cdS3lFIih/GVqwX/IbJ2NqKQmdu45OeHDRF0Cq/aFB3bsSUqcVlIUI6RAnBJIwTmeGu2WSGF7ZXnUR06IZLXvkGjp9UaW65bbzzu1dQtSyFvq2lSilMD5pBIxnd2B+XiZaXuy75Y7bSPpo1C3o8qnX5Y3uhR3FRIAKiAO5kSIAJ4gNElnyWVK2ubiA4REkAFRIkH2EqPpP0IJBd5V2hSKNLroceQSSpKnN4JHdJ+5Ewe6vpxY1kpHVFS0AJY4uucdFbnuOxIXrVTZtdC3RqS2hVOgnalzbtUoiQQBkZSMAjMd54YCHGA0FBLTq3Q4lSkOSDOCQD6gNpT7dO44MWoeVlqreR1z1k7rC2s36huzdub06pCjXvtKbKzUAiB5aYIxMkdpHAZ1g6mgWWQnbLinwlKktrR6VE5kbRBAJMn7QRw41jAwEDoqPtlUZsEnc7qFajeP4pp3yAuVKQhIbUPMBP1Akwr0qSmJ6AKTwWPBL8MrXfjw03rzUWjXLamm5f0y6q4LrKgNO1aygrLTeCg7kIiTjbuGJA4Cd+d865BKC8EpRLxcSQdqSVQcf5SJMjG7ibae1rzP8AClQVNtbuOsNAp1na2xU0bRfpDd7e4CtJWgx5jS0KUd3qBAIEEjhMqgXZ0qxKN2ho1KpunrSq6ONo89akFXmLlzATtKSrJ9p+sRkYHE3To2v1Sll1mnU1bET5bykFCFFJkkgCJkSAD39gIa9B0KHrmzSHcpt1SFVCkyRsiI+ij8xTJkdRjjuD4uPE14Rqf4Slp01pi22FOtBQMNUlK3R+XcG6tIAeddWBuKSQT1M4jiunu1Zd25K4qKmfH4cQYXB5GrHbt/lcYWKBtyq/C0+9NG0opJkAknEqPTuenYH24dWqRd2rmLZQFQStQDm1JCW0/UdTjr1I/QEo36v8OyG2m0rqXVny0jJJMnriBmO3Un34KPJvQ/8Awzbxc6gbal1MILkAKJBVM9CIAO0g469RwsBwc8vedh9VekUJhhbFGMHoOwUo0bp1nSjDSadhwO04Hko2jcSTt3EQQe6QCZBEjIjggWLSB8xDLhJfWdzy20hwpJAwoKMJz0GSYg44lWhuTVjHIp3WVVrKz0+omLg3R0+k32XPxj7C0bjWwcEA9J3GQDidpeuXFjpRVGsqUF5KvUy22lRcdA6NCSSZ3ZPUFQPchPGquDbZA6vm3cR5R2+ChSA3CX+H0xw1pw49z7+3qn3TllpdI6dp664FSWlAOUrGE+bAMqUQMJAVBMD2wZKlnL/S9RzHun/Ft/ZcRaaE/h7RT7NocWDG4IP/AKaMZEGSR2PHmzaPrOb2u1UFQshpEP3RxoAt0bKQFBiB3JwexyT24sLy25XjUmrLa0mnbp6ChQkUzCU7UstDCdwEAHrOZOcHHFEXi8uLnTyuzI/6D09eiePDjpohAwYa0bqYcjOUKaekar7iyt2rq2yppsqny2zn1dp/3nvwbNYXpzSWlGaalEXO4thDSU4KEYBP6kgD2+/D9ozlY5abIKmoYXTnKziFuLHfb/c6R9B9ojFfXP6c1c/d6os1ToE0oV/06YwQFlJEkjAAnuTiBwtxcJXetc2pMZDTyH6kJDuHEFI6U63+z+wER+RPJ+n0haFNVe1bxSl25uiZUSJDcD3JMx/HtwfNHWRVa+1WPtoCl+mkZJiBHUj7fy4qfcfGnZtJadZYbstyqKiiQCKdxxJRUvH/AKjq3BkD9CfpngHc2/ic81ajUDlXZrtT6YpkwGaejpWni1tAMFx1ClLBUCT7iE9Bm1OHeGm0mBONhjPcn/A6eqrK81U9a8vB3P7+v5Lq4/W0OkbHVV1fVU1BRUTSqirq6l1LbTLaRKlrUogJSBJkkADilHxzPH/y+5A/Dq5gW9OtbSdS6303U01lpaBCLm5WtOgJcVtQSltpxtS2w+ogJU4FJ3FMcc7vGZ8Qfmd4m6GqtF91C9TadqXQ89Z6BxTNG6sNtoJUmdykEo8wNqJSFOLIHtzt8bl0a07yYujLDbLari41TqCEBO4qWFEmBk+ji2oK6Jz2wxM2OyWHWt7IzNI7cdEGdR+OO5uBSLXaqenR+VdQ4Vq/gmB/M8QHUniT1lqcKS7eaimbV+SlhkD9R6v58QdR3GOPHByOigj3a0IZLXTye08rdV171c6XHnHHnD1UtRUT+pzxq3ZmOPhEcZxJUUlZxk8ZxnGLxfQQONjVKXlhKZJUYECST7cagJPHZz4I3wsbDyT5UWfxHc3LXSXC83ZtVbobT1wbCmaVlEkXSoQrBmCWkqEbfWe3Au83iltdG+urHYY35k9AO5PRTbdb566obTUwy537z7ghL8Pj4BtbeNJ23mJz5aqrLZbghFRadHpdVT3K6IMbXKogb6dkzOxMOKH90Z4rV8XW02LRvitqtMaXt1qs9gsVI01T0VtphT0rS1ISXClIJzuBSVH1EpMyZ47H665jXTxH+bd6y8V9i0tctzlO62tSLtqJrcP3yFDNLRKE7FyHHhJSUIhSuN/xebFbLP4uLkLIltNqDLbCEsiGmXEIR5jQ+qFEgiSZJnM8VpwbxhXXi/SxVPka1hIjA9nce07+8joNhy5qwOJOFILZZY5oxqcXbvPXY7NH9oPXqqp8ZxnGcW+qyWcZxnGcYsWcZxnGATxixZxk44wiOM4xYlljvtXpy6U9dQ1D1JWUiw6y80spW2oGQQe3HSz4evxI0a9cpNN6lfao9StJCWXD6WbikdwOyx1Kf1GJA5j8KKK4P2+safp3nWX2FhxtxtRSttQMhQIyCD0PESso2VDNLufdSqWqfA/U3l1C/SZy05vJqG2jvUptYmMYx29+CXrvxcWfw3cldR671C+8uy6XoF176GSPOqSIShhEwN7iyhA+qp6Djj/8Oz4jatVP0mk9W1TbF7QAilqVkIbuMdvYO+46HqPbg7fFI8TGmWPBHqHSVfdGVah1aKQW+gSSXnA1VNrU8R0DadhEk5JAE8KYo3sqBDIP/SaYqlkkRewrmt4pfEVffFr4hNYcxtS+U1edX3JddUMsrKmqVJAS2wgqzsbbSlA+ieIK2ol0ykBWAPpn/XjR8zipIJVPT7cex6d4/vGT7DPDYxuNguJIxuvlRVKOQopx3PT2H06ceaZYqKxIdK1JSADBlQjqMnHXjRUrB64SBkkD7f0ffjdbWytSRGSk4E9D/l/vxOhbyQioky7ATzUqp0OOKpgtLSjDYdA3SYGf6H+PC7SN8NBeWnVha0AgnGCOmf4/1g8NFTKwlKcmR1GTgwB+nSOHLTlnW9VpUlKVqJhAxBJxGccSsnOy4jnlWDf5g2S6XtpVn/HIo0spbT+LcK3VK2gqH0HYAYyOhAg4cvbszVWdsISVQZM/KowQTAPXB9o6STxXHlByouWp7uinoKZ+seA3lDDaitRAlRIGT7nOBnvxbvwScwXfDfzSterU2KyamNrDqRb7u0t2lf3oKPUIEFM+nr06GZ4s/hOGcAuazJPwVMfaTcIngMdJgD4qO3Lcbe8/uJSCooODCcSeuMZkfT78C7W1wcU+8HipSWygJC9pJVtEp3GTjcOnTd2Hq4L3NO8CoQ/VNsMsJdWpQbaTsbBcUVFCB1gAwJJj08Ae9PfjbkNgSA2N8IlJIyQFEnBgDuB1OZJBjiRxZhnVKXBf4gMp5JmaQ0mpXUPhp1taUtpS4Ey4olRXtPQxEBQ+aBGFcPXNnnXqrnS7bKrWF+uWoamzUFPZLe7WuFw01HTgpYYwR6UjeqBGTJkzwvZ0Clnl1VXdd4tIdZr00JtqlAVh2t7vPKNgCmYBRu3E7ohPQiD6keQWaZpBVuacShxRb3BIISEhQI27kkK+xMCBJ4Spm6G5Ksqmk1nZR/RFmOnLSh5/aupcJWpSkx5ajEEd4TPTB9UmO/qvrXLnXNrcUp0mIClk7ROPuSR/I+3G1Sm2mN7gPloAU2k/+pGJA6YEk+3Qe/HqgonHWzUrBNVVnykJj0mfTt+mNo+3FJVVR4jsDkvtWy2tlJCGAbnclPPK3Ri79qBCnEJGyFrLifQn1RtJOE/rgSJmOD7Y7aw6265UKU1ZLQAt71AeavCtoOMqWZgkYCeoAPEW5Y6SVZLallggPuCHnCcLUQcqKhtAE9+ySe/E6uzNOaJhChutlDucYhSQah0pUreqTIkkkf3gAYAMcRGkOdpdyHNTa+cxx/h+07ktukX1VL69RXFIDlSoClYTgJQRCUAdQcyZmUiAcgA/aS01U6P0UzW1iPOvt7/dUbMlZpkZjb2lMkkjJ+oA4F3ILSyde8yvx9zINptKBVvuAna6RMpIPSIxIwD7iOLa8ldIOa81rR6lujH7nZ5lI1GWacH0Yj80bogf5cVfx5ffOY/6W8/0H6qdYaP7tDrd7RUh5JclqTlZoM1VyTNXVk1FXvPrfJ+VJnp0E/wkA8Wx8KfI4UNtRf7uyE1lw/fNtLRCkDO0qB9gRCT3yfrB+WHKdfMW7t3S6Sm0suhVLTlO41RTiVEYCcY7kz24tDYafyaedigkGEp9zP8AvxA4M4afO/8Aidxbz3YD+ePySZxbxG4g0lO7cnzEfl/lMfMWpDFGSDtIkQex+nFe+ZVbCXfVAJwT1+304N/M+rKEuAH0px0+vFc+Y1atxbkKSczKfyif5ni1FXIQl5g3QKDhgAJnp/rwEeYd12pcAMpj054Kev6ogOKk4z7H78AzmDcCtTkZjqT3PHJ5XVqF+s60OuOmcqJnEHim/wAQvUKW7LZKAFUv1S3znshEf4rHFuNVvlSnDBMyTHFEvH1fxcOaFBQpVKaGj3qHYKcUT/8AtCeCllZqqh6b/ooN3fopj6oErMqPHzj6TPbj5w6pLWcZxnGcYsWEzx6QOvGynoXKp9DbaVrccUEpSlO5SiegAGSTxe3wef2cTxP+LzTVFqJvSlDy+0nXKSW7trGr/ZgcbPVxFPtVULTHQhsA9uOU88cEZlmcGtHMkgAfEr0AuOBzQ1+D34I2fHX42NP6bvSXUaKsLa9Q6peSkwLfTQpTUxAU8vY0P/f36cdoefvMRXPznHV6eNNR02idLULVVeaIjZSqpzuaoLQkAhKULLSnHEj/ANKnIEb+Nnw9PgiK+HHoXWdAxzp0xdr5rpump7hXN6VqF/hqdhTi/IaUatsbFLUlSipOdgEATxKLl8OOrtentS0Vv5/0VHWakuDtyqam56SO5LpbSyhKVorJbS00gJQIVGwkiVE8fPnHPFVuutwZDS10PhxDLQZBvITjJ6eQbjfmrb4FbTUEL5KuJ5e8gHDCcRjcgerjsfRV35p69qub12uNMxWP2/TNteNNdrgxLb1xfSAU0VO7EICJCnVwNghtMFRKOfvxW9MW/UWnrJdrBRU6KDTC1WupVSIKaek3KlLCT0OwlW6CopUYUQTx0o5h/Db5sUOgrVadLag5d6w0pbHTTVDGmrrUWu7KASC4w2Kz9z+IcVBUoveYC4uAFKChTbxaaC1FeLBXctr7o28aVrKKgL7dHXW4UYUoN7UopW1bi6EhJWpbSikhBVmTxO4Mtxt1XFLCdTN8kOByDzc8jmTnYdB9bK4krrXfLJUU8D2iQDIBGDlu4YxpwcY5nqfpy0ODxnG+uoF0NU8ysjeytSDBkSDB40kQBxfS+WiML5xnGDhx07pW4avvtHa7VRVdyuNwdSxTUtKyp56ocUYShCEglSiegA49wvEgbQVnAmOJZyf5Ea08QutKfTmhtL33Vt9qiA3Q2micqnzmJKUA7U/VUD68dRvBB/Zw6DQ2k6PmH4tdQP6MsSgh5rRVsqEIu76VAlH4x8ymlCgFHyk7nSBHpOB0k0FcLRyC0mvSnKDl/ZOT2hEIBOxtTFwu6QDDj0TUufd5SBk4IGUbibj63WiM/wD2PG2kd/UnYe7cpq4f4Pr7rIGxjS09T+g5n8vVcTtS/Ao5j8i+Ttw1xzj1LpDlfbbe1u/Zj1V+1Ly66RIZ8inlCF++9wQNxjEcUivKKZu5PponHnaQLUGVvICHFpnBUkEgEjsCeOjHx1vG7U8w9ZtcurTWqVabOomr2xtq3TBLpIJB67B7BKuhUocc4HFb1EnhgsNZVVdCyqq2aHPGdPYHl8cblCr1RQUlY+mgfrDNie5HPHpnZeOM4zjDwYQpbaR5dM8lxtSkLbIUlSTBSRkEHseLaae0fzT8enhgRdbdpq9apq+WVxat71VRMB96vbqk+mUj1lxoto3ESkpdBMKyqpVOje4AJJ7ACSeP0jfCN8Ko8Kng50lp99jyL3cmDeL0oABf4upSlakE/wD20Btof/4ziSeB1yqGwxh5GT0RO1Me+bDTt1X58dUaVueitR3Cz3igrLVdbW8ulq6KraUy9TPIJCm1oUJSoHqDwgqF7EnOVDMZKozx+jvxzfBy5ZfEErWb5cDWaR12ldKh7UVtbS4utpG1Qpl5lRDbqi1KEOK9SClv5kJKDQHnL/ZY+clHri5J0BqvRWo9Mbql23OXKtcoK/YmCzTvpLZbL6wop3IJbGwqJSCBxypLnA8ZccH1Rioie0YauWyh5iz0hXqHDnaqQ78pBSkAYHQ9BxcDxafAc8Qfg2pP2teNNN6r023TGpfvOlnF3OntxEFSahKUhxraT85SUKAkKxHAO0pylUUsrU2VJcG7zNnoEQSQZyZI6/z4PUZbLuw5CCyMcDhyiNo0uuqILgO6BuKkwACIAnEf114JfLrQRqbgwA0JWZBIIUFYJ/SYz7zE54X0el2mKkU7cpJwFggR8xIHsIEHv1M5jgj8q9KKdqmlhoBTywSIGxSfmAx6k9ukGftwettvM1Q1iiXeoZR0bp3lGXwt6+1F4Y9Zsaj0hWotN0ZpF0qXFtpfaWh1lSFoCVdZQSBgYIzw8uU5oKUbAnzVqJKlD1b8de39Dpwh0haUN1A/dmGUyVdAcAGZSIMgHE5jv03alqvwNvcdI2pXKQkpkpEyev1x3n2zxe1otwpoDJgDbn7l8jcVXd1bVCFpzv8AVDXnBe01QeZbw0UrEEnchMGR0yemPqSex4h9osL13Qp0JCl1J80nyEFSEKwkCBJO3b/dj6GCSFZuaNxsVBqbT9Hb7DcEawtIttS7WUKKp6kaC/MDlI6qVMOSE+tPXoQQAeNlLp9Fqs4QUwUo3kEJTj/ICe3XJ4V3Uz66pe93IbJvhrI7XRxws9oob6gojbU7Qhtax8hbE7FbsEDHtnAziR04gtyQt9xDadqEpUEhIML2kTtIHpgkDKgBKVTGRxP9enakhBUz5qISSU+nOEkmAOhP2TGMkRzTGkqmsvNoVUUdSlFWpDjClwkubfzflMyOhAPfInhbuNN+MIWpytVX/wBsahyhdFbheHypY2UjaiBPUAASo+56AdckDvxLuW9sGp9UCoS3NJRo2tAo3pMTIIESrBJJI6HPEXuaRRUrVuYIQViFKC/TnvPaf8T32jg1cttHJ01YmGXWiHEALcQU4WEAEqAI7E/b7iRx87FwDTIei/QdrcYj6nmplpS0bKNSCQ29VKKA2gAFInr9BA/UDsInLtVm56gfZpPSim2tBSQU+c4D6UqIkynuDnBOCkcO1KHbLpolDZXU1I8tpsYUpe2TA6kAwDIj0xEcHvwj+AC73ty33vVITbaBfqTT7VCrdUn8ym1DYnduOTkgkxkcDXzPEDvD3cULuNTBFOH1DsNH19wTRyp5S1qrfZtJUVKtf7UeadvNWyySlQXuWGln5cpCiUzMJETMHoNy55UU1F5P4lve2AlIQkelYCQACABjEwf5xx50HoC16OolU1uoKejS+Qt0to/ePK7FxRysj3JOOkcT+w0yUdABHXMx9uFqPhaKSZlRWHUW5OOhJ6nvjolS78YSzh0VMNLeWeuFM9IUiWktIQhIS2AAEgekdgO3QcT+kSGaIqCSkFM9MDHWPfiG6TZO5CAQUqjv/LiW3B38JbCoKG7aRPeOG1oAGyRiSdyhnzRr5W4ZPfI6cV85i1JLrkFUZk46e/8AX/g08z7lt8z1lAEkn6cALmDVkqcgSTI67v148WBCDmJWJhwbiBHUA8AbmBVS6uST2gmMf68GbmJVEFzKsDHXgF8wamQvsJwO/wBvpxxdzXdgQ31NUFCVSYVBP2P+/wDnxzq8Vd9/b3Pa/LCwtundTTI+gQhII/jPHQDW9d5VM8oq27ZlRxA/045n62u5v+r7pXHrWVbr323LJ4YuHo/M9/ogd/f+G1nqmvjOM4fOW/Ly7c1deWrTljpFVt2vNSmlpmUmNy1dyeiUgSSo4ABJIA4aErpkQncY6T3PEw5J8h9V+InmfZNG6NslfftRahqkUdDSUzSlqdcWoJE4gASJJwOP0F/2dz4Yng8unK67asOqNHc7eY+j1/8A9Ru1CVqtWnZC9pbbfbQHGiG1nzjuQdhgnqbg8vV6M0rdqrxC1GmqOxUl4oTaOXOnbZY0MOUltecCG6lLDTaVu3C5r2FCBlLKmGht3uqI273WC3Uzqmc7DYAc3E8mgdSTsFvDG+V4Y0IC/C5+B9yw+F/oi36o1jb7NrvnU6yH6i5VyA9btPT/AOlSoVKZT0U8RKjO2EwTZ7W/M6o1nXzV3FyolSdqfMKYGRhIzE+5Huew4kWnfCLzH521pu/MDWddoSnfV5iNPaYTSv1zIIEfirk824C6O6aRttCCIDrkbis1X8LSx3m1uJs/NLnJYrgElSH16iTdmVL7FynrWnWlp6ykBOFGCDBHzbxv9mHG/GUpqq2qZBF/RDl2w/8ALSCC7vzTlZ7xabc4DQXu6u2+mUF7jSsXVJpnWW10DoKahLid341JwQruUQSCJgzHvxDeXaqSt1rebRcUuKRpu7NUDXnTscpFtsPsrUT85CHQ2pUgSyQYgkxDxEo114FdZUdi5pJsT1s1C441pnmDa6ZdLbaytSJapK6lJWaOocHpIQtbTqd22CClMIrdSal0xaqXVdDS1NDX827wu33G1vrFWLSqtpVtWqoJQoplgUzfm7SUrFSpUEpkUZL9ltytE76G5uDM40Ho5wOolvcFgcPfgHfCtOjvcNSxs9Hu3fPcZ23+OPgkdFea5nkDyp1ZU1blJcL5U2+zXWpXITc6iqbX5D6uxdRUBYKiJIdWn+4UyPU3MVetdUXfTGqbNbtXaPbttK67p25tIeaD1Q68S82tQ3tOJQjYnYtJxJ6g8D+6cqaWh1DYeTbNW8NGaXbt2r7UtTZ81inYS/R+SFx/1G65IqgokH951O3iOWzmLXWS43fmPfnk0ptFQ1pK/sqaCG6ZFE+4yutAgwBVOKVAx5SzjATxY1OZo5BVW2bDsEtxts92YxjuQHDHTAHZH6Olimj8CujBbq3z6AB2T6HBVFPHZ8JOj1DRVvMbke5WXGz3AKuDmmqtRNW2wokhylUfU4kAZaX+8T7q7c7q+jcoX1NOtracbJSpKwQpJHUEcdqr5qXUOguWq9K3J543K2Uf7Ysqiv8AfW5xDoXUUylDbuKPM2DP7xLoTBhR4rZ8QrwhWbnzWM6u0ZTsUGqq1lVRU0ySltF8bSJLiEYIeQgoKxHRRGSmVfR3DHFb5AKevcCSSGvHIgcs+vy3CQeMvs1dDCbhbGnAALmds7HT/j12VE+RHIjVHiS5oWbRujbPV3zUF9qUUtLTU6CpRUpUSY6JHUk9geO9vw+fh66J+FxQUluslstvNXxLXmnC7tcS6lu26Jplg9H4V+Hax6nkAuvRtaABB4DXw5OSGn/A7yLeqdP1tqc19f7YHdTave2KotJUi0qJS0pR2uuqSYGxQSSnapQR5hJ9tdZp+1aGW3XVVzs2lnAK+uonaxVPedTvKSD+Lu1X6VU7aoSQygpcKYEsNlLZh8Z8XOhY6npSOxJzt8tzno0bnrgbpY4X4RmqXiRzSSeQx9Tnb4nYeqnlTcm7hrV69Uda5za15bXlJrtTVjotumdJudTT0c7kskSQfLD9YpJ9W0GEwDxD6vZ5Kcn9QcwNV301wtdOp1hNPR/hKU1SlHyVBC1Kce/ed3FpAAPoHTh9t2v9Q67s9E/Z9PPUtkoGtltRT0jVrtTDKiClNN5m0BsyVb2W1zkkkgnjnz8eTnrX2LSVj5dPV7VZUXB5N2rFNhakJayhgJWsAqO1LhKgEiVkQQNxpyw0tVfL9FRkaWA6nF2NZa3c+X+gHlsAd+ZVtXSWOw2WWYS6pMaQG5Lcnbd39RHPY425LmtzN1tVcxda3K81jrrz9e8pwqcUVKIJ6knuTJP1J4jyu3HwmTxnH1kAAMBfMhOTkrOM4zj1u3K4xeKynwlvDEPFL429KWurpjUWKwO/t67AiUKYpyFJbViIcdLSCDEhRjPH6UdH20sMJkSVdf8AuUcmf1zxzK/s6HhXGgvD9deYtfTFNy17VeXSKUnKbfTKUhMewce81X18pHcTx1V0pQFK2pE9usfrwo3mo8SfQOTfzTVaINEOsjcqUaUoClaQRJjaRHX6/wBe3E4tdAPKSkjdtEYHzD69+GTTVv8AIZSSANuPUeg6/wCPEpYbKWikAgDuR0+n04E9coq5an/QVAbkkgjcR0kRBHt7g9QY44c/G08PeifDx4pG2tG25u3J1Pa03i425ojyaOtcee9bSD6kIcA37EQhKjiAYT3EqiYJ9+x7/wBTwJPE/wCDLlv4vLEKPXGmqS51DLP4elujMsXO3J37/wBw+n1oG8TtykyQQQTJiyXIUdR4j8lpGCotTTCRuOq/OXp3TYulfsU6hCSNsnJjcEzgZBJ6xJjpBHFh7PyqpdOt0rrN0oruiupW6pZpyT+EcUsgsrSUg+YA2Cdu4bVIyRIB18YPwravwaagYvdkertTcv6l5tlFbVLbTWUVQ4CnyH0tpTO4oG1aEwcJO08C63WlilrENvuM0wUsIUokhttEiVKCASR3hIJABgHpx9M8CW+nqacV8bgQfz7e8L5l+1ziuop6r+G6SMDP+/il1kpaa16dfK2Q+66kpb2kpO8nqRBnEiP+76Dgfa7qnKt3yQQsKMBrIHuAQBn6D3PXHEn1TdVgNqlSW0wWxEkjdg/rjqR/KOCF4fPBpzF59MuOWe2V9bb6RxT6mQgusturSncokQCtQSiYgwlIMxw/XKqi0CHUAOpVG2aklEjqx4LnZyBzQ0sPh2csXKCz68dvViq3L7c6i3JtaagKr6ZTISrzVskHa2qYBkyRnrmOajo/w6AtqNxEpkhQVn8wz6ZSQPf9ODHzL5FXXlLf37fd6CooKpBKVpW1sUpJ+/8Al0n3OYFzR10y/wArmdOm3WptdvrHqtu5IpCmuqA4hKS0t0mPLTtCkgJBlRyePRb/ALvSgghwJ5jHL4c+2V2jvBrK3SQQRtj1/T3Ku/M15Lz7yd26kBKQp1X5PUcqJkSYMTJPTrhn0/qyoeuzFTUun8SyEtJRvO9RwrcsdeoGDIgEAxx55lXMtLdO5IWk+b8gWFEEQ2nESoSCYgJBHeOJX4OfC9qnxk85qDl7o/yarUl2feXSuVLqqdlBS3veUqZhITu6ndEEdSOKhr6sCtc89Ff9uoj/AA5sY6qE8mtMDUd+duDyZbbUtayrA2piY6eqSBJMkkwCeDbpKiVd7gpa0hLTbMgBBAS3JVuHSFYUYPecjpxB9L25rS2j6Gkwkup854LTEJE4PYerdERP3PBD0Zbi5YyFpC11SkKcEKKtoAhMATMgJJAnJyYnj5mr58N0jkv0Up4iG63cypZpDVCrBzEst4/CM1ot9Y2pukdTvbeCSFKKhHQHb75STMZ46o2DZXUzb7S0PIfSlxDqTuDgVlKp+oIP68cpdM2n9u6lpkKAhs/u9qASIUComJI3QAZ+sdBHT/k7zMs/MOzIVQXGjrq+mShFe02TvpntvqSUnIEhUdcDrngfRVjNZjccHZIXHVE8COZoyNwUQbVTkkAyT3SQMcSayMbiBBBEwY6fT68MVtp9wSY3IAg+/wBOJTYGyXQVBKQCNuevt/Offguq3Dt91NdK0x3oOFGYBnh21bUCmtcEqUQI+/1406YaG1BMgDrGccItf3A+StPQA4J6Rx7nZYQg3zJqpKzu2ggid0gjgEcwKmCrqCCTE/Tgw8xq5ThWT6SJMdZHAJ5gVilKXI2yTgfw/XvxqV4EHuYlf1ncpJ7/AF/y4CWtnt5WI/Q5A68F3mO8ApRII7Ge/twFtZPhClhWQkHA7jPHE81JYEE/EZfzpzlffqyRLNC6UmYklJA/mRxzoeI3QMxxeDx56gNr5L1lOkwqtqGWBB91bz/JBH68UdJJOeG+wMxTl/c/kla/PzMGdh+a+cWf+FzyQvHPbmxrS1acc8nUw0XcxZlpUUuCrcQlptKCBIUvcUSP7+MxxWDiw/wwvEsjwu+Lqw32pqGqO23BDlprah1oOop2XoHmFPfaQlUDOMZg8GznGyBrr18EP4cGtPCtZrhprmOw5ZNUc5WA5qexJfgWHR1tWl18vhrAqK+qcaowmY8pT3UyB1G5M64oudPjXulBX0bb6tFacpr/AESSr93Su1dTU0rSkIGJQzTuAE/L5x6EAisfgc1I5qblNqvmnUKIc5m3T9l2ICR5Wm7UpdLShufUE1FQKqpM5Jd9UkTxLvDtzcoeSnxIrA7fHk0dp5r6WXpJqqWYaau1FUqraRlap2p85h+sCSeqmkpGVDihrlxvBW/aLScOE+SEOcexlLSWj3tH1KcaOySNsU1xxucAf8cjKJ/xxvibPfCr8C1drmzW5i5awvde1YdOtVDZXTN1bqFuF50CJQ222o7Z9SikdJ45H/Bn/tM/PbmB4/ND6D5x6lpNbaR5oXpuww5aqekfsVXVOeXSrp1sISVN+cW0KQ4FelZIMpk9u/iWfDr0Z8ULwq3PlhrR+ut9NUVDVytl1otpqLTWtBXl1CEq9CxtWtCkKwpC1CUnaoUF+F9/ZatE/Dv8RNu5p685gq5k6l0s+qt0zQU9t/ZtFRupSQipdQtxxbrqJlGUoQqFGSBF+gZKSshrd1dX4wmibNzN8CGqdP31DaaK+LYoDUnL9A4txPlvs9/NS6GymCM98QeIfh78Y991tyo5fs17qVv6MurNxvrCDn8Hb0OIqnyfTvdWHhtJ9P7tWSQTxfr4x/j7tV00/W0NruS3KaxsVDC3mKgJpaBQlTlQpZw4sNpbSIAA8+ErO8AcafDNq1y16NcQ4HwjVqXaS4hKyV0lG84POdSZClFtK1AAAlSlxhU8I3HtqgqqVnisDnA7Z7Eb/RWX9mUrzPM3VhmnJ94O31XSO485UWzmzcdVvJpm9K1tkYs1nuTS1rqK55l52pWtpE7Q09+IUGNol38KDtPmI3ijVnMGsHha5nvXyiqKmvrbhcUVtnpnBUfjXayvbqaUpWcElDqIHq3KC0+4Dhz55k2i6VvL/Tool0IoLsLwloKS8KNFJTqbpAJPd1xIHpElHaDwOqu8ruviIFOXwmkobPTvVNMXUJcerC482w4Oo/dtKfBAwkutkkEA8UJZqKMxsmdFpwGuH/GJ2lnxIzkjrghfRFLZtUbhr8xdo77vGXH4HBHpkJz8QGu6Gt1ly9TS17VQ5clXK5sVBpFKbfS6yktpSRAglSVpJ9J8oRkRwL71q2psGqr/AHCjqi2jRz7tRQNqdAcO9LT1W0TO1KTG3CoClnsAOI5cr26nQlS9StU6Ljbq1lq1OO1CVMtUdPVksKnuHCt6E90JQIIE8DbmFzEqWKS+W2nrKakoa9hhyscdIcefPlhBaT3UpxQWpcGc56DixLTZ9DfCiOdPl3/5ZJ+TsKRV3ZrY9Uw2dh2239JAHzbq92FZfkpq4AVQpUsKVYrj5tGuqrEIttv8weY26pMldU8EKSEOEKSNpiDkG3Q2qbXqStbr6Wkd13d0K81C0U5qKSle3kqWnzFClQoKzvcU656vnJEmk3JvVS6jXFxLVPZBTN2ylpUIultFc0ylsqI9E70IG4wtcjcPUoGJunyT1BqvUtmpqpik0XeKVgSPwtzqKNKiQSRhh1B3RJAUPr1EL/GWumj1OOB6u08+hOM/IhCrPDDNKSwDGf7cj34yAfiCixRDVl4X+Iut5ore6oeaW6ZZuNUMTCnFgNJIJgwlQ9Ig8cf/AIz+r3tReNu/0iy+W7ShqlAWsqAUhtKDEmJISCYAEnoOOwFk11qGzvJJ0ZVOFIBCqWvoHyFEdNqnkq987MyffjlZ8cnk7c9O+Im365eoqygoNbUSXE09SEpcpn2oQ6ISSNqlDcCMeqB0ngJ9i0zGX+USaGl7DpDSD1BO+S4nHdLP2zNe60RBm7WvGdsdCBtgAD/KoxxnGdTxnH1QvmNZxKuS3Km5c7+a+nNH2dHmXPUtxZttPIwlTiwncfokEqP0B4ioE8dHv7Ov4XTzH5/3vmRXUwXb9FU/4GgUpMg1tSkhSgfdDAXPt5w9xxxqJhFGZD0XanhMsgYOq7L+HXlJbOUnLLT2mbM0GLVYKBi30ifZplsIST7zBUZySSTmeDfpe3FW0hJATAH0+n6jiJaTtstoAEiNsEdRE/19+CZpW2zskAnocZPCE5xcS49U8MaGgAKS2ei2sRmD0MZ4WrUWxG0wnAEzjv142UiA22UkwYEE9DxoqHdoOFJAke0f1/PjxZndJauo3AkZgAmfvjhMpYUuB6ldJ7/+ONlQRtCphIO0YzGP8+3CZxwECQZVggdgR/XTjFshF46+SNf4hPDVfLFafPcvLTjNxt7LJSDVvsqw0dxSAFJUrJPpMHPTjk3zs5Z6q5D379k6pslZYq5KDsS+namrSkgktLHpcSD1UkkHpPy8dvC6CqCQVewgkf19uKu/GB5f2PVfgd1Hfrt5rVx0SW7lZ3UO7CipddbYU0o/3HUKKT3BSkiCOLP4A45ntj221zQY3u59QTt8lUf2k/Z5T3cOuzTiWNvLoQN9/Xnhcr2OYRRZ0Wo01IaVdWm4B9VMDVgttqGxLnXyhO4pn5gk9Rx3i+Edq/S938Idkp7S5Ror6TcLghJCXC6TJUR1gg4J7cfnas9yXXXkJSlRB2rEgJKjMJCgZ6iYH/dAAweLBcq+d170Cyr9jXevt6lgb1U7ykFQP97aYEGQB/rxeNTZxeoDA6TQc5B/RfPcd6dw7UMqGRCTYgjOPkV0B+N7qDTV35g2elo6mkTcGmFfjHm0hZbnoDGScf1HHJ/mtdgt1bbTiVNg7FGQBnqP8sfp2kic0ucFbqB916uqah59353XnPMWoSJMkxGZkmPr7ibRXiJq+RHOTTesqGit9zq9NXNq501JcqcuUj62iTseRAJSYGAQRAMzEm6yWO02plCx+ssbjPLJ/RA7HSTXe7vuUrNHiOzgdB+qEVv1s5p3mdZbq3S26uds9e1VttVtKKqkeWlSVpadZV6XGiQNwO2QVAnMj5o/mhe6bmOjU1juFZpi6P3J2pD9n30QtwcUp0+VsJDaUBSgET6QkGew2c9OcNbz35yao17cqG2W+5auuNTdaiioKfyaWlecUSUspP5QrEZJ2kzKiAi5e2ZSqlLiQ0IAQS0BJkhRUTJzBJKhmAfuKiDXTz6upX0GZGU9NjOwRBdV+0ri6SC8XloZQpOCR0In8pxtzJ65HXgo0FEinoqSmklBBcU4AfUSRhQED5eoJJGBMEcQPSNnLtT5pQ4l2llwyNqWyEkAgxMAGZ+8dI4LPL60J1FrunZKXFpQhCSUpgLSVEgKkEgfL0EkE9ySfmS6zCOMuX6GFupwx0TzouzK0xUs1TiUtOMtqdcC1BMFIJAwOgO0wR3yMji1vw89NKt2ibheXWwhNbVKe8wABS0N4iR1AUVqE9QQOkcV/wCYun10FYhhhohx9SUn92FFRJkiCPoY9uLjeHfTqNJ8nKOnRAWpjarakeoqnMdes/ftxVF+vD4YWVEftagR/wDnf81Du8DJKYxH+rZHGzOprGkLSc9Du6p79P17e/EusLMFIAgq65z9/txENNPlavLTtKgoJ67ZIAwP66jia6eaC1HB2kiYESf/ABxbtguRr7fFVuxlwGff1VAXGmFPUvhHIHbKnenGSGU/Yyev9DiMcyq7c2sKSSlSYKZifrxKrQrZSlRwAnH3/qP48D3mZXeVuC1bZkhSRk8GCMKGgrzIr1FazIjJKZ6n6/b+p4CWvK30uDBnMESBwWuYdZuU5hQGQATAOP5cBPXlZCT1iSBBkfYcakrdiEev6oDeScqERmf04DGqKhT61ykpiZEzGeCvzAqT6oCUk9ft/hwIdWvyXOwEzHHAkZUgclTr4jV+AobBbwoy5UPPqHYhISkfzVxVU9Z4O/j8v5ufNqjogvciioQoj+6payT/ACCeAQcAcPlqj0UrB8fmkm6Sa6lx/ey+cFXwfc+qfw5c97Pqep0xo7VzNOFsLt+p7eK23KS4NqlLbIVkJKgFQSndIE8Crj0lzaoEdeJssTZGOjfyIx+8KFE/Q4O7LvHyK+Ira6mistt0vRL0Y1T06U0eg7hWJqbQ+yBO3T9xnbIklNI8qCSQkMmdx05jVmmPGbyPXSitqG2qsIqKSsYJYqrbVsqlt5Ewpt5l1JHZQUkgjJ4/Pbyi8SeoeVja6Jl1Fxsb42v2ysSHqZ0T/dUCAfqM8dBvCV8SK2DTxpWXmGk7T/ySoaekJTvUB+ZwjO7cokNgkHCR80/aL9l07Z23m0gioY4ODwd9v7jzPoTk8wSdsfQH2f3e21zDQTvDQRjQ7YfDpjuPoraaY+NXzu8KVwRpHmje65X7PTsor/Q25usbuKAISp+nCkrbWIBUW/MSds7EZBZOfXxnLTqzRtczduct3uzD7BXUUdnsVdR3J8mISpTjSW4SZG4ux+8VDaYHA85l89NI839HK/GP0FRSOIEpLYedWYmUAkyUxhSyJzEJTxWbUfIPTd0uaHbay9TWxQUadK3AfPVuPrCykySCPlG1PuZ28WJwvx7WS0obdIS2QbE42Pr6E/JRb59jdO6ozb5vKenPHoot4hPFHffFFX09ls9Cuw6OafH4el3y9cHC55gW+sHapQWpS8EgboJgRw+8vtPv6dondMUzqXr7dqFpoOhcppi688HFbkHcGkojqQVLHX1SG+7Wqh0s7VKtb7XnMMLqKl5tnc1bkJz+HVuO5biwEpOQUhU4hIU7VXM5rQ9jq6G3viovlxShb9SgBaEq2QVelIXIIhDYgI27oElSp1zuE9cWkDIPIfLn6DnjqjVj4corNA6NjvP/AFO235jA7nPy+iI2oNein0lYKqlfqqvVmpn7fVioq0fvSlsoKkpQ3hLDSCoJiSQ4clSyox/Ule/pLWFPbrdVebX3C3u/tCuckqbSXmgmpd6HdIcCEAAAFCQNqVECms1G/pu1UiaAIbqyG6eqqqxakIYDaUqSnsIjbtQJCtyxJUkkQ2+8wnHb1UqbuDrpUhbD9fUNesjqNiB9JAzABPWTMKmsrg4gct+fryHuHQIx/G4YWBxGCC3ryxz95PUp95l8wUpstdYbMkOUNPU7w4k7UoQh0LayoZ9YWP1JMdwVzW5khddVUdIpC3XPRU1KZyO7bfYJnJV1V75jhPr7mgl2lfoLctwsrqC+pZcncqACon8yj79oEcQF14uEkzJ78PlttrYmhxHr/tUrxVxY+peYYHbctuQGTsPmjL4bOe7WltYUKLrWP0TjTqfw11bcUh6njolagDKO5BSQYEgieOj3Jbm3alM01VfWnbWdgKNS2N00fm7xhx/y5CAr/wC8lbS4gED08ce0uwODn4X/ABpXTkjVs0F1TVXbT6VSGkvlL1J9Wz7dik4IJHATirhZtyhJj2djpsfhtj4HI9EX4B46ioJBSXL+WeTsZx7+uPUEELtNYL1qVmnbdor5p7VNC+PMQqqSbdW7SAU/vGUqYcGevloEdjwMvHbyHuvi88MN3047pWtF8tSnLvZ3GKynqQy82j942SNiyHWwQAEH1oRgdeB/4ZOYNu5l6ddv+gtQPs0aFE1SaJAU2ysgqUH6NYLSP/c35U/9xO7g/aS5vX+3NJLzFvr2krDf4y0O7XoIwfIeVInodjqhEfXj5hntdbY7syrp2s8SN2dwY3eowCGnI22znsvo26W6ku9rfEx7nxSDYtIkb6dC4Y9QML8/lxtz1qrXWH21NvMqKFoUNpSR7jtwnJ2k/Xi6Hxi/DHT8uudC9dWChqaXTuryHnkLpVMppa8gGoZB+VULJUCkkQtOSZ4petMAcfYNpuLK+kjq4tg8A47Ht8F8V3a3SUFXJSS82Ej/AH8QsQjcZM/px+kP4S/hXT4XPBzpDT79MKe9V7H7YvJ2wv8AF1AC1IViZbR5bWRjyo7ccUfhLeGMeKTxraXtdVTfibHp9Rv12BTKVMU6klLZkQfMeLSI9lHtPH6S9DWg01O2CMkSSnqo9eB99qMNELeu5Uyy0+SZj02CmulrZvIMbUpMgDJ6TwR9PUgQhO35QPv+v+XEX0pQEKG7H5ukjPE8trAp0hIHUYxB/wB+FtMHILetSUMkgGMGPcz2+vCOrcSggqjeD06lXbpxuqHcQcycfXhvefBUgFQIEp3A+x/yIP8AA8YvQFrfX5aFkSpWcpyeEjziUkqAhJOfr7fy/wAOPZqSQTJCgqDGRMx/p/HhKpYVux0zE9/6/wATxi9WLcHmTKSYGU8Cjxpch7h4p/DrfOXlDdaWyJ1K/SMV1yfphVKo6VFQh54tNEgKeKUBKJKQlSgqcQSiPSTgp6SOPO6SqUn2gZgxn+fv7cdIZnRSNlZzBBHwXKeBk0ZikGx2PuXFXxifDt1R4I7tR19TU0t70neKhVJRXemSsJp3sqSzVIIlDqkgqEShUKCVSOD18IHwJWHx1aw1AdQ3SoprPptltZYplDzqlS1EAiRCR7kDrMTPF+fEvyYpfEV4f9XaKrC0gaitzjFK478tJVJG+mensUPJQZ7Z+vHFblD4huZvgQ5n3hm11tz0nqu2VDluutPuO5hxEhaFAylSZIIOQU7SDBB4vbhTimrulA+DWGTN68tu6+fOLuD6O2XCOofGXwO6c9x0z+SNHxivCPZ/BjzzVYLJdDcKOqoWq1lpR2uttqO0IWBgqESCBnHfpz61/eUOIUGNi0oAHykIXgqCoEAZExnqSPSCSWvEt4kdT8+tT1uoNUXGpvV5qv8ArVDrkl8g7AMDoJERAxnoeAlbrHW621RRWi1sfjbjcKlqipmicvOuqDbYkwJUTAnoTkTjg1cquVsQjmfqIG/v7ofZqGB0pmhZpaScDsFqs1uVeHGmgylaEvwd6VQQVfMRIwJ6D+8ehzwRLXZW7VZnahSQinWmCJA80bjuCo6HP16+2OF2n+QF45Z8wrnpXUtvct1/0zWv2y50anEOGlfZXtWjehWyApJGJBKSSTAIkOuaM0tChmnn1KKvMQnO4RCoA9zJ/U8E7NbT93NU74IVxBeAKltFH8U9aZpk0dkqn1OFAqXyhhs+oqhQTuzncAe5OT9RwQ/D80A6/Vl1QDK0pSFAjcJ2lQx9Osk/URHA6vDiKHSbCioAs0pWifQJ37cyCRlPTESZ6iSlyOYFFpM71FDiUhKgo7QMggDsEnORnMY4+Kr44mmPyX6gwnz4PVF/mJQN12t9P71BtAp/PUsyNhiMx7TJMjp79bSaCq00XLyhSClKFRvJUBtAAP8AH/HiruplqpqHT9d5a2iwgMK2+k7SDKZMgjAIEY+gODbozVoodE06QQqHxOzoQoHA+n6+/txTd2jMkLB2yudbEXNbhHi23FDtuSoYU8oFtREbTM/+P58E3lxdkXylQsEhxPpcBwQsdf8ACfrxWzSuu/MtRaW6FOU7oStROQJHqOfePoJA4NPJav3a1LaV+isbUtJMhKhAUQPbI9uivrxP+z27zUFy+5u3ZKQPce4/JV5xTZw+nM3Jzd0b/N8q1FQlICQCdsGf9OBPzLrlBS4MD/t78FC9Pli1kQNxGQO/68BjmDWpU4sGREgGc/19OPoRyq5B/X1VtDhMQcHBnrwENe1SklYiD37n78F/XtV6FwQZ7JBmOAjrtz5xt2pggGf1Mx/hxzet2IS67dBK8z+nvwJdWvShZHQiff7cEzXdQS8sEgyPfPAo1lVCmpXVqI2hM/aDxxA82FIacBc6fFHfDfeemoHBBRTvppkx7NpCf8uB6oQBw5awvQ1Dqi51wn/nKt18A9YUsn/Phs4smFmiNrewVezP1PLu5WcZxnGcdFyX1BjhVbbvU2erQ/SvOMPo6LQYUPp9uEnGTHGEZGCvWuLTqacFFjR/ipvdlfQLhNalIgvNlCKjqCCFqSoAyJmOueueJbUeKFnUTDIdWwyVKR5ilsKD6EJIJQlU7BMGIAAJJiTxXrjOB0tqpXnVpwUz0vGF0gZ4fiFw9VYu58+v2m/SU7VdTfgaVW9tlDAbb3jMuL7jcd0ASSJP1fOTN+a1XzKRb2Kxq5VQbcqhTBPrqwhJLzKASPMUtsLKQTKlNgfmg1Z3H3PEr5Xaoe0/qWhqaWrcoq+hqEVNHUoVtUy6kyFT9CB/rxxfaI/DLWbHBRCLjSrfKC8dR1PToF0dofCjatXUFtRUVhZtt4Q2vTte2wX2KlJBWaV5HymckAJlxJlKt4PDFzY+Fc47ZH019dU6SbSUpRc22V1dnKsk+Y+lJcpxOQh9LYAVIUoAKVKfCp4rdN6otNVar8xSClvXrudqed2NsPk7lVFCkn1NqMOKZQQ6yrLczAuDy25iX/ltSUj9rr16206EBluo/GNtXVtHUoLqlBFUAn8qy06IyHD6lUtcuIbraajw3jkds7B3ucdg7uHbHm3nhXjHabdeqPxof6huN8j0IBzjsR8eS4j+JjwZa38Ltxb/AOIKBuptFZ6qG70LoqKCuT1BbdT6TgiR2kT1HAjXx+iHXlLyp8UOlrnYtUaXsi665BSHGallVoubTzgIDu1SUeY4BmVJWSIExxxE8cvhFu/g85z1Onq1Lr9rfBftlcUFLdYyfzJkA46QQDg8WPwdxnHeQ6CVjo5mDcOGMjuOhVG8XcHTWhwmaQ6Jx2IOce//AGgvx6AgzPHkiOM4eUkJ70XzAvfLm/M3Ww3a4We406gpupo31MuoIMjKSD1gx04t5yV+M7rHS9O1S64sds1cylJSqvYQmiuKiZ9bhALTy+mVICj3UTnilHGcQa+2Ulazw6uMPHqP3hF7Vf7jbJPFoJnRn0P5jkfiul3Pv4hnJrxa+G3UGlq643nTdTUNirpae6281CzVNhQSUOsgoTM9PSPVJOMc262mTT1S2krS6ELKQtOUqAPUcJZPvxNfDvyaufiG52aY0TZ0FVw1NcWaFCwJDKVH1un/ALUICln6IPHC0WeltcJgpBhmc4znHuXa+8QVl5qRVVxDpMAZAxnHLOOq7C/2dfwrK5feHGv17X0paufMCrCqcrEKRQU6lIaiegW6XV/UBByCOOqGk7aEAK29hIOYHv8Az4GHIHljbeWGgrPp6y06KS1WSiaoaRsCNjTTYQge07QCfqTwctKWslQwIPWRPaIj24WayczTOf6o3SQ+FC1ilGmaANpHeI6p6/78SQJ2NQSn9ex9/wCXCG1UvkMiO6Ynv7cb3FHO2M5A6j7cRl3O+y01dQQYB6HP14b3VAiI+Yjr278b6pwLRIJAwQY6cIKp2FEHKTj/AF4xerUp4lpUo2haiSJ6fc8aFwXDJAMRMYOO/Ht1SvLJ+YgbQZ+b7+3Glbgnd27mOnT9eMWLx5hO0x95MRj2/wA+PpHZU4yZwM4/hx9bhMQlIIJSABGf6HG1tuVoCZWVkAEZJVgAfr9OMWLW7ChEgk9onuc/bjmF8erkto/R+odNcxqa9Wq26xvxTRXOylyay7MIQQ1XpQJIS0EpbcUQlKgluCpTZB3fEc/tENh5I37UGhuT1tY1NquzV5t9dqK4tpesbBQIeFMhC0qqXEuDy95KWpSo+sQDyB5leI3UfOfWtz1Fqi+Vt61Be6g1Fbcq1zzXqhcRKowEJGEoSAlASAkAAcPvCNBUU1Q2te7QO3Uj17D6pB4tq6esp3UTG6ievQe71Ul1Lqv9suKIX5mwFIMBaIEEd8wAE9I3Rk4lnt77rV4SptflpWkoBPpCiEgNrIBkK3QoREKJ+U5EeRqRtdOpDjnluODJjcQIkiQZ3YH1yf1eNCuivrVVLzag2QIBABSD0k4HeJgkgACSOHaSd1RLg9UnxUzKWIkDkj1ywcXQ0UVClF/aFuLUsqWJG8kqyonr16kmck8fNc6oF1fSy0ltI8xW1S46TtVuMwrqnBPUEDBkR1vU6qC1ja6t1apWSVYHy+r5pPcGcglJnsNVvq1vOB8mdiQoAwkgRjp9BBiB1kiSRYLK3RTNpo+WFWD7YZKt1XN15IoX9aEP/h3VKCHmCyoBQbSmDhIUJISPT7kp29eC5ycbcc0r5flqUVgZ8sQralIwfynvBiCQOAvzGKkMVDiTsdYIMyYQCYz9YUOsRHYjJj8O1zTdbW6IQpTyErATtIKtqZAkQcj7kQfrx8RXuJxoXPHQhfqiHATAHsjfrCnbPLQqT5inKZttYKjs2pOYicDHbr9usl5Z1hvGkU04dAdCN2xH5T+YE9zg9v8AEjhhu7Ip6FFMobjVUwSlREbBtgdO5wCc9enDDyN5gtaWrmqOuUlCWXw2ggE7wOuVdITHt/gOKpdA6Wmdp3IOV0kHkIG6JbN+f01ekKWFuU7raGHVQJP93+G769Rnix3hX5hsV2pre0/UJLjKi1uWQB6kqnJ6dsH37Y4AHMSzsVd5aYQG/wAPcGw5TlWRkT9ZPTB6fYE8RG2avvugqw/hy+8hjdsQqSomflgmZA9p7Y6ngdRksnjqWDzMIOOWcKDW0La2mdDnBIIXTbU90bqLckNOtrABnaoK657cBrXlblwzAJMgmM+/+314B3KDx4MW+qprTfWiyw6pDRqSDsQpRA3E9SnM+wEx14GvOT4ha6zUFYzY2KY0LClpbqFNqWt4JkBZkgAH2A6d5ni6qDieKen8WRpa7kW8z7/cqjl4GuQqDAwZGM55BEzmFc2mKd9xxZSGwSo/3T9/1A/24CmuLuzV+aW1ev8AOnoR7fX/AF4GeqfE5cda14bqXXA3JQlCUwlSgPlKB3MxP8O3EXr+dSKepARVtthJkgmUrAHQHBInA+ggHHGG9yOnA8P8P6593ZFv+gXtpS50g8XmAOXuynHW7+1xUiDJx1g+3AF8TWpxpblVqCskpW1ROxHUEp2pj9TxNtceIG10bSVfu1KcI3IKpJSRMgDMR/DuBHFbvFbzVHMzlJerfa6aqW4hxlQCGypS2A4NylRMGR0HUH6Hhkt7RNO0DlkJKuduq6WmfLLGQACqXnrxhOOPTiCgkEQRg4yOPPFjKsFnGcZxnGLFnGcZxnGLFnGcZxnGLFnHtt0oWDnHtjjxxnGLESuW3Oz9ksijuiPNpwU7XdoWREABaSDuSPtPT24Odj8Zus+W+l/2roDUwYqKUeXVWyoKK1lTZHzMhwFQjumYxgY4qIhRT9uPSKhTRlCiD9OB09rp5j52gjqCMg+8HZHabiCshi8NriMciDhw9xCudZvjmc1qK1N0lwsmh7yhhBS3+Lt69qSRClbUrCTPcRHsOBV4r/iBX7xgWihpNRaW0bak2tJFIuy0SqNTZJkk+ohQPtAzngBEcZx5S2ehpna6eJrT6ABRKu8V1U3TUyuePUk/mvvU+3HzjOM4JIas4zjACo4BPG6konK2obaabW864oJShsFS1GegAzxixatu6O3HTf8As5PhVVqrmdqTmncKYKpLA3+xrQtQ61LqQqocT7FDO1E+7+PlPFH7N4KuaFy0/wDtmo0ZebNZQf8A+43lr9mUny7iQt/YFCM+me3Hen4P+kNCaH8IWjtL6Q1Pp3UlfbLczW3tVtrEvrRV1X751SwPUn1HYncBhtIMHHAu61OiA6OZ2RS10xfMHOGw3VvdF2bY03AkmBHY9z/lwT9MW/y20zA/KcjrHEY0pa0nbA3RhWPt/lwQbLR/h2oI3kZBT+ccJw9U2FLkNeSJyFEZUR0+3CWtd8segH0jcAO/fA4VuKSlKCJgGUn2PDdVukpmRvBII7q78erUd0jrHvmUEqMgnb34SOLIMzhWevQ/XhQ+AlaYVOZ65/8AHCdTQKldNgUADH1/1/z4xbLWlSnGh8xBkHt79OPi2FdBk5jEAjjalpSnEhCN6yQkJTn1GIH8eOdGuv7R3yp5Xc/OZ2kb7p++XC16NuRttgvGnlorBqBbZDdQVBxSEMpS6HNqwpQUlHQmJk01HNUEthbkhR6irigAMrsZVqPH542tMeAHw43DX2pGl19QpxNustpac8t69V60qU3ThUHYgJSpbi4OxCVEAmAePPMH+0ic9NfckNY6Yet2jbNdtSuFuh1DZWHqSssdIuQ6w0grWhSyg7UvqPmIlRyogpA/xIfiYaz+JFzbaut5bFl0hY3njprTqChaLS26ltK1OOpSkvvrDSSpw4EbUBKetbKt/wA0bUjpgD+97/r/AL8PtosMUMAfUty8/RIV3v0s0+imcQxIn6gMogJ2hOEgdEgDoJ+3CJFQpo9MpwOk4mfv2+g+vHuve3O4kJmR6ZIHCo6tYotDv2gWi3OPO1qalNxWhRq2kpQU+SnO3YSdxkTI4Kyu3wFBhj6lahfVbg2dq21TtBJJ69T1zj/DgpaQqbezYqCqZr1VNa+t1NVSrZWgUTYUA2SsHavzEk4T02wT24DdJUulBaOW1ncv0gkkHGevvAnOOH2zV5p0ADYUpEKJGTt/h/A9f8OlLKWyalrVwB7NKONPcnL1/wAs2VPKcSAkyVKV2Bk9gO3YIP8AeMze76Xq9Ns0jNdTKp1VSQ6A60tr8QJwRuj0jaqCTEhWe3An0RdQ0EKc3gEgLK1KzkCAYJ3SSQP+2PqbO+KLx7am8ZFr5eUWpLZY6Oj5cWVFioP2VSfh1VjIWlRW/mN6gEAhJ2g7iACs8NtHVEuAxz+iR7hRkdeXLbmtOomP2oghxZ3Lp1QEwRKfUPSYA9+v5YAJIPD14cNWmyXKlClLW0GhuTuJ2x2mI7xHuTPQDjbzU0dW8udb3Gx1CQupttSunK0Kn1pxIJ27pI7jGBAgcQ/SledNauK1pmmfcIW3MhWDuyIxjAI2jHeOPnGSjD4paSQbkL9FJ5CdFRGfLsr+8yU2u86JtddbXVqZLDfmLWkHaYTKDBgxMDuR1PAV1fYnWKxypQlYZSoOHae4BhU56YknIiT7F35f8xVN6JqrSSlbBG5JIBJYMK6GVAiUz0JHfM8PWir3T3NP4erbSW3xKFqGPpHbt9cR24pZ9PJQvcCM4P0ROFp8Pmkeieai7vbjp+5uF1umlykqAqHKZQ6ImdsdPuBHfEtptXXCrtATXUQujaAUpfGV7ff6jAAIzk/TgRcwbIrSGrlrSHF0yiogEA+naAcnvAgE4A29+CFyv5yU1FTKZqlNOsuelC42rI/lPc/m7kwM8eVtM17BNC3K3azGzQh7zB1wq1uS1bS827O9mFEqT1kGDPT26/zGWo+a9lFOUfs2tYUgSElRO2T3EHvnMmIJnizmq6ax6yZWhmrS2ogwjzDLhPqMdEyCD7DA6xwJNdco6Zxai4+lTPRCUKkIJPUCIzJ+5nABjgraquHAEoIKyaTAGyAep+dSKphTdHSvrCgdpLREAA9DmB74xnIkkwO66su92cUWkKbBn1DcCZ985iDiYE9COhzvvKxqlcLiWzuBIG9RSSInEnHT9Jx14gGrdFqYQreXApaQAkmVHE5nt26f4cPFA+B/sfVA6yqDG4IwhJdLLcLkrY9Ur9UpUW1SAScwZ6ZgiT1ni2/w1/hm3HxFaOu+t62oUzpkVSrWwyEhX4upbCFLcn8oQF7enqKsYB4qnqewVVKqP3hUYSElwBSSTAIgzkxA/wBJ47ofBRs1Xyt8D+mdIampX6S6rqK64VNLVwHKXz6lakojsdgQqDkFaunThkjb5cgqt+J7lmn8P+49eyplzV+BDoDW7DqqxmuoKxYxVUKw27PvkFKvsoHimHiX+AjzH5Zofr9B19Lrq3NAq/BqAobmkewQo+W73+VYUeySccfpB1lynZrGy7ToS4yoH/4+36fXgSax5XBhbifJlIkQof4HiXBc54tmnI7FVrNb4Zdy3B9Nl+TrXHL2+ct9Rv2jUFouVjulMdrtHX0y6d9BEjKFgGMHPThnUmAOvH6b/ED4OdGeIHTy7VrLS9q1JSIEsisZlxjt+7dEONn6pUn2yMcc5PFT/Z5WXnKi48qtSG3nJFmvyitoHsluqSJT9nEEZ+cdODlPe4X7SeU/RB6izyx7x+YLlSTgcZwSOf3hJ5ieGG+qoNb6Uu1iO6Gql1ndSVP1bfTLa/0VORIBxwOVt7FRwXa4OGppyEKc0tOHDBXnjOMjj6lO7j1ar5xnGxLClKAAJKsAdz+nBL5S+C3mvzzdR/wpy91beGXDtFS1bnEUoP1eWEtj9VDjwuAGSV61pdsAhgI7zx9E9uL18pv7Plzv142y7fajSujmXQCUVlaqrqWz7FunSsA/dUfXiz/J/wDs1WlLQtpesdd6hvziSFqYtlI1b2lDuklXmrP3BSfpxCkudMzYvUyO31D+TVx28sgTHD9onlhqPmVXfhtO2G8X2oKgny7fROVSpPQHYDH68foN5S/Bp5DcrFMrpeW1kulQ16kv3pblzdWfch1RbP8A+A4snpDlJb9I0bdLbaCloKVsBKWaRlNO0lIGAEpAEfTiBLfYx/LaT9FPisrzu92Pqvz1cpfgweIjmq62o6Ef01SuH/6i/wBUigCfqWjL0fUN8Wg5Tf2ay61xZc1rzHpaQKSN9PZLap8hXcea8pA/UJP247KUuiEhPyCB2gpn647cOtLo6QMYQesAQOIEl7qHbNwP36qbHZ4G+1krn5yZ+AdyH5dBh6usN31hUNq3JdvdxWttR9i0z5aCPun9eLS8p/CRoXk2yhvSmjdM6cATsBtlrZpnI9i4B5ip9io8HSn0hHpCNsnEjp9OHSg0iS4nYkIcUoJBmNqpwZ4gSVc0ntuJU+OliZ7LQF+eT4w/iPf5o+LzV+nqC7VNZYNNVhtNMkqltstbQ6huCQUh7fK+qinPQcRX4dHjpvngB5hXvUthslFqX9vWlVudoq2pXTsLX5qHUuqKElR2luIEfPJONphvxILtYbr47ubj2l7Suw6fGqq5umoC15Pk7HNi4bHybnErWB7LBgdOB7ou4Nlrasp81RIAyCfv3j7frw5U9MySNsWNsD9+9R2Frcl5X6IfCX8bDkT4hecWm+XdHdbrZ9V6noaV2mNZSBNtNc6wHV28VO7/AK7aj5fqSAtcJSZ6k34mfxU9EfDF5d2+qvdC5qnWN9qPLtWl6Otbp6p1tMKdqX1KkssISRClJJWtSUpxuI/NmNHLqXP3KFOKCgpJA2hBBEEHqD0gzIjtE8auZD911VqC4Xa8XO5Xu+3FXnVdyrKtdVV1Ss+px5xSlqMT1VPb248PDMQlDgfL1CjOrZSw5591130//a2dLVmqaGnvnJG/W2zPJZTV1dJqJqqqKZReh5SGS2gOISz6kpC0qUsFJKRB4u54bPiveHzxbM0LWj+Z+nmrzcGHXk2S9OC03RkNgFwLaeITIGfStQKQopJCTH5ZbuwqjcAgpVlJSn+Ufb+uvCYsF2G1jzBJKW1epCfrnjlPYad5wzyqK25yR891+x1FxYulIzU0r9PUU9Q15rDzDqXmqhM4UhaSUqTI+YE9OuOIvzq566J8OujRqLXuqrFo6wKqEUSa67VaaZhx9YJSykn53CErO1MmEqPQE8fmV8G3xUeeHgQtH7H5d6xVT6cVVGtXYblRtXC2vOlO1R8twbkiACQ0tAJSCQTwn8bPxGua/wAQa/Wet5kXtmopbAwlmhtltZNHbWXCCF1Ip0qKA84CApfUgQIEjiHHwzKZMF3l+q9ff4wzIHm+i62/Fy+PBpXkjywb0pyM1VprW+t9T07iXr5bKpNfb9NUqkqQpSXG1QqsVMtJkhtMuGTsSeEtO3KQFbiBBE4/U5/onjy0n8Q93IKiBPY9Zx/lj24UtN5PUqM9RBOOG612yKkbhnXmUoXW5SVL8v8AkvoOxrAjuf8AtH+nbhw0Rox/mTrOgsVLU0NHU3NzyUvV1QllhBEn1LVhPTv3j34aax4JISCOwn3zwidcCT1A3kAiPlnrI7j/AF4nyvA2Q2GM5ytN4YNuq32lLacWy4psqbO5BIJSSkjqMcNbx8xRKgBGVQZjt+v368KahwrJkqjuOsDuQP6/lx7tlvNU6CsScHb3Of8AHiAfO7ZFWkMZkrZbaFSyoKG0rMwpM9/bv/X6yG0aeqHXElxkpE+neCAOm0QY7SZ6e46xIOVGlmHdUW6orqf8XQtPoW80MF9IIJQB23dP19uLu/Fn8TXIfxN1/Lxnkny4Xoqt0/avIvZ/DoYFY8dpbbCUE+aUlKyXDBV0zJHE9lE9pBI2KGyXCNxc0O3CpzpynW0gNs+altO3a55Y24g9T1E/pCZ69ZPfLodPaWdcdSpLtShYQhRy6SdsxEqk7xGJUkgZAHHuisTNtbK1gvuqQElSCBux2OZMmJyTgdYmDc0NS/i3kBDgbYUZZCnAUgJCSRk5EDbAIB3HJKhwUncaOAvPtEbIRTNFdOGD2RzKuFqrUFVcbuLpWVDj71a4S+oKHmLUVesqx1KlEYjJGO3DPfqZNqriVIaUhMCEZBnMBXQESB9gY7g7n3vO0o+0n5qR4lKSr1BJKkgSBnHYQP4jh6asX7b0nT1aUtqQCEJJ9Ux+WN3pjOOs4VmeKlrKEuInjX3vRVTSDTvCedDanqKa2teS84XWFKRhPpJjdEYI9p7wPYy702unLU+p1CFJadPmjbkBXZM5kdMkmJ7cDWw1ZtNyW0pQAIKcIwcgnABE9J6/6yVyrWGxUNoJplubnPUCXATEDB9R9Q7jr0BgI94tOpxk05yi8L2t8pRIe5mWrXFpSxWPhdQhJKlElQT2B3Gc4P8AH643s8r2LolC6Cvf3rAkT0USIJ64BAgGcHIxwKmaS2XF1S1OoYIKlqKnCOkE+pPWZAOcQSOoBcbJVV2mlLVR17yKYepxK3ISkE9ZzEj/AAOcjhPmtpjGIiR6HkpxcCNkRqnkvfklKmKlSzuSE7mNu5IBOfVOBBkgjPbB42s8l9bUoU429TVJQCkE75X6oChnImTHTPeSOGex+JC6WB9LDzm9xMGT68dp7T1zGZJx3ea/xeXQ04SLhSpQoFKMphQjrA6jAkECI6dwPfDXZ0hoKiSmRo2TVqjSGo9NNPLq6NohMqSralIGSYP6xH/tEdeBRr/WztK04PI2KEz6AQDBETkZ75Mex4lWv/FZX1TIDlSxUhM7vVu2Kic9BETjH0nPAD11zbrb3ULUtlCgRuSENiB2ASf4Y7k8Mdnt85IMg3SncppSTqCb7pqL9t39igbo3krqnkMpVkwVKAHpT1iRhOZ+vHUzwCVWstHFmmu6qhbKFbWnFkkiMDr+nXioXwjfDXWeJvxI097q7S0/pDRhNfWvOD0Gs2j8MymCASFkOdClIQJHqE9aqDl9SadCfJbSlKTkhIjHDnGPDZpAVU8ROL5w3sjny718LhRMtvyVEQcDOOHrUGkKa+MhbYAKhBHAd07dVW2oTtUZSSBnKuCjo7V/mNJQVgpH5T3HsONkGOyg+qOXKqZxY2SRgdo4g990SmfU3EHI29v0/wDGeLL1drYvtJvRG5QgE9fsf69+IRqnQ6WStQbQPLHvtAHWST0Ed+w4wrNXVU38UuoND8j+SV91Dr922N6WpmSH6OsbRUN3JwpV5dO20oEOuuEFKU7ZzOACeOBfPPUHKzntq6uuGn9I0XKmprFjyKWjq3qi2trn/wBQL3FAVJktQhO2QiMcXI/tBnjksXiB5uad5faGvVtv+k9EtLrq6ut9QH6WtujyShSEOJO1YYaARIJ9TjkEjjnGXCqpKpSIzPvwyWynMTNWSCVCqoWyjDhlXu8LH9nzufPjQ9l1ZX809Mq09e2vxNKvT1C/XqqW9ykkBbwZShQUkggpMEHi3nKX+zo8k9HJbXfU6r1i+k5/H3H8Iyuf/tU4Sof/AOw9OHD+zl+KrS3NPkvUcmWdNsad1No1td0NZTuKUzqVt1076he8yipSQlKkj0qQgKSBtUOOoNFy7Hl/9MJIjHc54iV1fUtkLNWAuNPboAMlu6qnyd+Htyu5JhP/AAxy90pZ3EAJL7NsaW+QOhLqwpaj9Zngx0fL6QkLSpxKR6d+dvt+n04LTPL8BCjskJTtPsTPQ/y4cKbQ4Tkp3d54Evkc45ccogI2NGGjCE9LofaTsR2gyIAHDnT6I2lKdgMeoADA9v6+/twUafR6Ns7NpAAInBHCyn0gltcEAZkyPrPGi2x6oaUmjdx3QoAZwP8AP/Thwp9Ipj5UwDn6D7fw4ITGl0oVgfNhX1HCpNgQiRgjAP2j/fjFnJQKn0eCSdgA6H2iev34WM6XLYHowBiBk/TibM2ZDQEgEZyf6zx7XbG0JCjCzmPqOMXhKirGmkbAoJC0+wEkcK2LEEJTAJ2HOM/pw/pZbQSARMY7ZH0/rvxji0MjcMdp/wA/48YsBXIn+04/D205dvD7/wDr/Y7caDV+mq6ktmoHKSnSlF1oahZaRUVBEfvGXPLSFxJS7CiQkRxB0c7trUTuUNwgDoPp/n+nH7CedHLSx89+Uup9Eakp01OntW2uptNwbUmf3LrZQVAf3kkhQ+qU8flT8SHgm174M+Z2pbHqy0umm0zqBzTi7uyPMt9fUpp26gJacgBSzTusuqTG5PmQQCI4cbBWa2eG7mPyQeujLX6uhTvpxlL9vbDYSErSVKUsjJAPSBOZPQjuRjht1NZ0O0y005bKDIRtSSrpPpAzkgZV7z78L9C07lRQmSvy0gjbvIKsJ7jIxt69zjpiSOULTZLSEhbiyP8Ap+ohI6EIOQRBIJ7oyMGXuOEuZqfsEIrrk4uEMIyUC9Q6HXSVDgUhRWCQUmQEH2+sfWftieI5V2w0Z2QOskx8w/y7ffPB61BpYISfM2B9UhaCoEmMEz1HykZHbHXiF1nLh15LjimzBMbts+YfpnOfb/bjm23ucctCXq64xw7SO3Q6qEtKccLDbjLMnZvVuUMDE9/fH09uPiWd8JJIAgerpnM/+OJXdtB1DB3EEdxHbt1/UfXhmetC6VSiQdoI6ffjY0zmHcIQK6OX2SkLI2JHQH36E/1PGxx7yhEiMD0yTPt/X8eMcSULJGBGMYP9f58LdO6qa0yzc0vWiiuxrqJdK2agE/gVk/8AWRH5hGJ9xnjnJIGjAUmKEvOSmKqqEtgbiAI7d/8AX2/qeENS7O6em7aCeg69Y+38+NlUVTEdOh9x7f5/x4yho11Dg24HWB3ziP67cQXOLtgiLYwwZK+UFrXXOpwoARgCI6fy4nektDrdI8xMpyTHUjb/AOP0+3HrQ+k/xDwJbhKQSSkSBPTHtP8AHi3PgW8AWtvG3rZWntDWxusq6OmNQ+48sJYQgKAKlEwAkqEde4AHB630DQ3XJsB3Sxd7o/V4UO5PZBmz2Cm03Z1vuJRtAPRUFWZOe/369B2y20FAwbi7V1SYacG5tBTt3JGCfuAJMxjPq4KPiu5MX3w8c0blo3VVGKG5WlzyKunUQApcjsZ9OZGSIjIyOBJUA1tRDQQl2pQts+qcCB6QcmCBG49hAyCDU8zGAMaNggVBTyOLnvO7uf8Aha9QXRdXSkqKypSStwpSZ9kyeoBBORJ7ZxA2vV182vWtT+Gx5oWBjZCVBUhQBTtJyB6sHaT0leurl/yiUpCUCoTkOLKkhKgVbVhI6EBPYA9SYBPEDcU3WVyQWvxAeT5QSp7KVeXCQVJBCinbuIxuHsI4TLvOXyaScqwLPTtjjy0K6ml7+liqcQ44ksVAU0szKlE9CTA9UZ6mARHE75Jupo1XCyuhMj10zjySSQmSEZ7GFAAnOfYcQ7lXybvHMnQmv9U2+ts1NR6AoW7tcGqytSw+6hbwaT5CFZdXKspTkDHdIKvSFX+JubdY2t5LzCIdWlsrSlQJBMgj6SAeqfpHAqztEkPhvX1Neasw1PjRHONjj3A/kU8a40240866lDiCAdiiQAnP8VTB6EdMyclkpdd1WmKxKiFLS8Av0wSEpkESQMyAkq9gAJngzt6aReqVtatnneUFkLSSESJBj2A+mZycRwNeaXK521KIaTuQpRWpM4Mdz1nv79+0DjW58O5aXjkplBxLDKQ0ndJHrzR6kCXaN5DS3CS60pfloZImQIOZ9j3CZM7uPIuryHgFKJH/AKe4BS/mIwPuE9ROfmEcQWkoKq11DikJdO31KS50cVgJ9IyICsx+hmTw+Iurjill9KFHfkbNwCumCDBAB7nEEDhDqbH3TXDVNIzlO1U+i5IUhTzp8s+lM+kHGSR7wZOQZiY6R6u0g9dcmrXChCo9H0knoSZnGJnI7LqZ9mpQA064N235U9ukBJ69R1kAH+G01gbUEkuqIQHNqwAqJxGQQCTIxn7SeBTrJURguaFK+8Ru2BUSf0aKR8qcDjrQMBSFA7T3GAMZBjBj34ZrjpVVW3LQfCIkbsIAjp16A4gfY9uCM66iCp5CkpUCVHYVKKZghREScH+sceHKpFMyspS00tCSkJMbgYz/AAMg/wAI7CO0Tx82qPUUjJBqWeDHm5d/B14kdP64oGqi4U9KtVPX0CXik3GidGx5nPpST6VJVBhSUzMcdw+Vuv7bzy5Saf1laGK1m2aloEV1O3Vo2OoSvG1Y6YIIkYIgjrxwoqdu9QCUoSkBAWsylJjp7erAiB0meo47W/Deor9VeBPlw7qR2oernLe4qm8+N4ovPcFIOkx5AbInO0pHE2OR7x5gq140tccLWVDeZOPf/lTVykUw6SoylWMCCOHiw3pVO4gAgqiCmMK9uN90s4DihAVH8R34ph8TP4ht58Gt509YNGUdruWo6+ncuFeitaU43RMFSU00gEZdUl0Rn0QZEg8d42lx0hItPSyVDxFEMkq/NdztsfKvST1+1PfbPp2xUZCamvudW3SUzJV0SpxwhO49hMnMA8cwPi//ANoI5X80fDvrvlHymYv2qa3V1tXaH9VNf8hbqRpak+cllLiS9UBaAWydrYIckKxxTv4rvjia8dnMLSj1ot1ztWnNN2pLJo618bV1zii4+6EAlHo/6SFkbilJmJCRTC729TaiJCSCCo5nMmM5n+X04J0tMzIceaKfwJ7GB8/Psosthurqdj76KRsJUQrYSEwDCQB0kwB7cNPmT7yQMDrw53ZoNLJJnHv1nrw0vvAHtPbGBwwM3QKpwHbLpr/ZYGV1vxH7i2thx+nb0XcnFKFKhxDK/NpwlRWQVNEytIKSCokpJiRx+iM21pLZIIxII44Jf2TC/u2vn9zmp0IpQ27pehfU4pP74FNbtCEnptO4lXeUo47l/wDFxqQF/wAZ6jhdun83C4xuySQpEaZrzUylBUFSk/3evT9D/Dj6otNpj0iI/SMf19+IurVZxC8ggifbhO7qgqUsEhIKvTnGR/X+3A7ZdFLjVNoAgpEGInAPTjwbmhsgn1T3HSOIkrUsOKMnasyRA9PtxrXqMkbh6k9579OnGYXoUuRc0gkCB6h1PHpy5pLiVBRSDII9vrwNNf8AOGxcotA3jVWprtT2TT9gpl1dwuFTu8qkZHVagATAkYAJ+/A3uPxO+QFt5W1Gs186OXK9O0ydzlQ3eG1vLyQEpph/zClkgwkNkn7CeN2QyPGWgn3LV0jG80R/F34vNNeCzw/3zmJq1bgtFicp2iwyQl2qdffbZQ2gHqr1lX2QrhJ4UPGdoXxp8lbRr7QN2auNhu29vZUFLFZQutkJdYeaJlDiCpO7qmFoIJCgT+fz433xi3/iGa+Z0foOru1Byf007vYbdKqdWpqwK/8Ar3mSApCEglDLapKRKyApRAotbNU3CzMPtUlbV07dQ0ph5LTym0uNqKFKSYMQotokd9iZ6DhggsgfCC84cfp6Iaa0+IdPsr9S2p/jV+GHSWtq/T9bzj0wu4W2uat1T+GQ/UU7T7igmEvNoLTjaSYW42pSEQdxHB90pzKsfMS2m4advdn1Db/MWwuqtle1WsJcQtSFJ8xtShuCkqSRPVCh1B4/HrVa7uVRpijsiXUJt1vfW+w15aQpKlYMqAkiB07SeDR8Pvx5aq8BniGsmtrI5UVtJTti33O1uVK0sXG2qJ305SPSFJ3FbaiPQ5CsyrjrJYWaMsO68bXPzuF+rI1gU7ukeqN0nEjHHNn+0p82eX1P4S7Nom9XY/8AHlTemL7YLfS+WtxgNbkPvVXqCmmXGnFJSoetxaYSFJSspG3Ov+04afuPhBau2g7C/aub90rlUP7IuSfxVJZWkAKVXF0AJdSvclLbfzbt+4QgFXIPm3zk1Hz25nXjWmsbzUXvUuoapVZX1zxBcqHD0iBCEJACUoSNqEgJSABHGlptcrZfFl2AW1VUtIwN0TtEXHzmfLQCPLbCAdgEkmJI6bYkx/GI4le8UtOunpQtTz6AXVD1BQgjqCSREgDIJgyDwGNCauLA8rdCGwsEkCIgSRjt26R/IGmx01Ezb6Sqcfpq124IU44ht4l+mg7Ql3cAEqMBQifT3PTizraxsxDT0SNea4UsRewblKrRpf8AbJDjwKwtJJjptAyCegSNvQGOo+vE95N6e0baeZForddWm43vSNFUJXdqO3ueTUVjUEFttwwEq6HqPliRjhttC37mhDTaUIkCEoaO0+oDM5EHABwD2mOJ5yn5iWzkVzBs+oK7Ttm1Y3aKs1K7PdEKXR10JILbkdpIPQ524IkcOsdPGyIgD9Cfiqjmnnqqge/3oG8ydN2pd5rl2+ndorUt9xVOy64HHadpSz5aVqj5gmCSBnJ6cDO96IXWKCtnlpMmVQAM9QPacR9IHBu1zczrDU9wuDFAzQoq3nKgU7DZ8imQVqUG0SZCE4AIyPSPpxELjZ/wrhO9tK4CCSASPzHrJj0RPfAxKeA9W0P2TTQ0/gDPVB6t0d5bSitojqEpGCIGZHvkA9e3DFX6cU88EpbSqVdQMKJjP+P8uCzdLQt7e4pBPkYG4AFcxCVQcQMkzI/WRGF2wVr0NNoUmZEAAqKfyzORGMzPbrwvVUIB0t5php5dtTioAnTa3FICE7wqQmFyFmYGe3frMj3zxIbFoRaUNfukqLyQoH3BJGD9ZH6QZI4nlj5fOUbfnKStVQ4CjdtymACECY98gduCdyF5aaZvnMC1o1ld6iwaaqalLNyudPTmpet7KgoechoH17fT6UnAPEqhsxzrk/JBrrxCxrdMZ9ENbBpj9mttpUElQxPXZHSYGD9O/frxazwAePHWXw+9U1mpdJCkqXrjSimqKB9nzG6hoGU+YAQEkKBjacZ6HHAMGmWbpfax5Ky5bad1TTZKQ2pxtMknr0IxkmFGM9eG2/3E17pU4FinQs7UIQf3YBjKQIEmAMEgdJggnJYG6Cxw27d0txyFzw9rvN37KVeMPxNai8XvPi7631O9SqvV2cLz/lbUJ2hP7tCBAIbCIgnsR9eBnarXUOUq3mE1inW9qEfuzucCUmD7hM9FT6YUdxMDiWaT5U12obTU1cOspYaipWUhJQFKgqV3MqgGPoOp4uP8GPwsaG8Vfi5tOk9XNpXZWWHq9VCl0oFctsAhuTmFdSnrASOo4gXKP7pSmplG3Qd+yK2yVlbVikidv1PbG5XMTXdoradSi4yWfwju5W/OSlSyVEYMwlMxtPqM+mRGax1TD8uLWpSW5G9RDkeakp3QohMwckR6e8if0K/2i74a3JXk14Q6TXWlrFa9G3uirUW9LVKShuubUlRMozuUnYk4GQADHH58r22ikujTRbbbceWg+pwgpIASU4OVBJSAQIBB6ySK8dK6Rxc/mrNbEyNgDOStk7pOot9/rrPW079LVKaJLTjam1gSSAoKKYE9AqIwMGOHXR7irXeGFOJUoPN+W7Hq2GIJAOAPYwIIgR14d+dHiJvviB551Wu9VKp6m/3mpLta8wyltp5cJgACIT2KQTgmPfh35ycxaXmnrk3hmx2rTbNQlDS6GhbhDRQhKSsTkFZyT23ZxI44wSCF+tp6r6fbGaqmEb24ON9+R7eqsF4ZtG/8aaauQ/aVBTfsSnRUIpnH9jlYS4EwykYUYztGOpnB4W849F0zlBTgDZ5o2kJAGxUgGTMxP2n9eBz4eb04zXU5Q6ht5kgt7HPSR3CcH056kmVdTwYNe3KlqqaHBKakBQXkhC+gBg94VjrgdAOLIonRzw5VN3KKooapxaUDrbyhTc72djJeUFiBClQD+YxkEQTu3YxmOJVe/CY3dbck06gl1KAlKkypCpwEp9oMffqYPSXaUQhirc8wJekENK3lKR8sCPeIzA6nGJ4l3/Ei2G2yydkrCTEqJMwD9TPbE/y40ksdPKCCFtBxlWRuBB2CqJzD8Ntz0sXalhsutA+ptRKSg5xumCf9sRgj20Wm/XW5tW+jp362tUS0ilbZLi3zMGEpzk/cycdJ46BMst6vpFoqUuB1aMqURCZPdWAn8xgkjPUCBxH+Xmjr34f+a9o1vpJqmpr9p+oFTTLfpEPIQSCn1oPUkE+0YiDwu1vDOn+WSQmqj44MkZJHmHLoFSKuuz1kuamK2hXR1TDhbUghaFoUkGUZMpIOCD0gn5shM5WsvtBRP7o/lKtoG4dOsQcgAD7zxZ7m7ykuvOjmJetT3ZlK77qKpcq615hISgPLJUopSkSPUOsCAIkzwNtY+G+rs9OVtUj7IUCjrt6iFfKBP8wTERJ4ES8LTvGprdkwUfGtN5WSOAceYyhPT0H7WcLKNzqnAWA0hsrW+tXpgYg5URAEnoOp471+B3Tev7L4VtI23mUyhGq6CmNO4ylfmPMUzYCadD5GC+lsDcU4+UEbgeOa/wAKXkZYH/EDdNba6RTUemeVtr/4gerKp0CmpqlLqU0/miCVlK1FSUpglSERu+Utvjx+KVrbxF62uNq0dfLvpfl4y4yaSlp1inrK0sqUU1DjrZ8wFaxuDYUANiJkgyqVVpkMpYBhc73M66ytpIAAG7lx9eg+HNWe+Lt44m+Ta9NaR0jqZyl1QxcBdrwLbVAqp6ZoQildUk+lTiiSpBztTkdOOYPOvmzeedWua7UeqLs9dbxXqBfcc3Qlv/02UemQ2iVAAACBJkk8MVXRrStSkoWSSqXPNlaiSSSSATvmCcwSep4a6yzvhaN+5ZTnegFM5yAY+43dBPuY41bbHjyhFLXb6aiiAG7up6qOagdRtUVKAgElIhJUJxBV6oyevv3xwPtULTBSnoSQdvUfr9B7x29uCRe7QhhtQUlPpG8pTChEEH0jAzg5kwfpwP8AVSUpAMCRCUkwe8kiBHt0+v0HE+ChcweZDrtOwtOEN76j1kwR3AjGc/58NCmt7hzw+35I3rhJBOMQIz/t/XZrZYU4/EZkp+hPEwNxsqxqW5k2XWH+y7cl9TUPMnmJzFbfoGdKizosDjSqgKq6qqU8h9tQaSfQ2lLa/WsQoqARJSuOyKLqSJBycj69JB4/OH8HDnvU+Gzx4aHublwrqHTl9rRZL61T1IbYqGalKmWi+D6XENPradjr6PSQTx+iV5a2S42sKStpRQtA6gg5/hnhbuzSJQ48isEJYMHqnRd1UFGTuSoEAzmPtx4VefMUUpWN6oOD/P8AThv84qWQTJ6gz04+CFLnuoZnr+vAtbBqcRcySMkdlDrxsTX7YAUrarBHb7/z4bWG3KhQSlLi1kQQASSOv+/EF8RHia0Z4XuU1+1hq29W2koLFROVRpE1rYq61YBCKZludynXFgIAjBJJgJPGzGF5AbutXEDmVRH+04+J6z6b8Pej+V1NcaZ/VF/vDeoKm3SSuloGGX2m3nABAC33VBAJBJZKoITPHEBxP7zfsQFp/NthX34JHih8R2p/F9z41JzE1Y6h696jqA6ttrFPQspSEM0zQJ9LTTYShPUmJJJUeB082tIAAGcDM/x/hw+0NJ93gEZ59UtzzeI8uXlf7sAH1CZ+o48IQpxQJTEdc8bmqIrKsn0CfofcwOFztjeoassPJ8tSYmFBQAIB6j6EYnvniYAtQeiTUzG3b2kyN2P4nrHCxhlLRCnCpSk9j16dPrP+XGM7W1Ao6yMz/wDt7fT+saHKoIAKd0kelSjJP6Dj1bZAStyrCI3ET7R8oye2ONKVqq3gEpSRPTt+o40NsqqCCsKCTlKT39v8e3DpRMBCJMJEQBjpx2jiLyh1ZWhgx1S+x1KretsoUvcj5FBW1QwYiPafr/hwUOWuqEtrSStIQgAq3kpEE+wPuev/AI4F9MhTywkJMq+k8SCw3E0DqYXtkpx7/wBR269ODtC7wSMpQryajIKuXzGXo7l27ZBpjVStTprLG1W3N1ujLX7OrXEr8ylUlWCUmAVjIkkAdOB6m9LrHj5soITMOboUdw++4YA7Z+3A2suoVLphC3VKCt5ROCRISE4nImOkR1PQF7kbyH1RzkrbizpbTl41NU2ikdudY3RMuKFJTsplx1wpMwJEq6FS8EzAOtr3OGMofR2+KJ2rC6DeDj4RulvEd4JbxzCuWp00NxpmnnGGUbFoY2IBlwnA/rpxzi1/aGLRX1DSw0tNIqAtDgAcKiNuxMndAiSACAqSenE5s3ii1do3l3Vadst3rKSxVwmop26hXluABIVuSD8smNqhJj3BPEFprVV6vr3amq3eUV9VLVK1E4CTJTAnp7HpMnjvBBK7LSckntyU6+Xei8Nojbp081EqyyP6jrFuwkUnslIdLqQVYEjpHeJGJPs/WPQSLczTrISl1YSQkqlaI7gzMxtHWSB7DibWjSDYfU4SXX0GQ4tuFJ+iQYgYjABEe2RJrNpFLSEuuypRKlrEgYAEkxnqR16wr2gMNDw6NQfJzVV3jjA6SyM7LX4ZvDdT86ebumdNXG9WrTLF8q00qq+uhFLR4MFZxJMiOmSn9N3N3kqxonmne9OUNxobxTWGufoW6+lWPw9YGlFPmtLPRJgkdR7ZyXJu2mueKVN7mxuKkbDtg9x39sT7E9+GTX2qGdJ06vJUPOIKEBIKcZzjpicgxAPuYYJbfHTxmR58uPr+aSIrtPVTiKP2j9AkHOnkjqrltoG03q4WC423Tl8So2+vU0UM3HaU7g0uJUd2T9B0iQA9SMftCrbUGg4htaVJQlsALMGClOITJEjrAOc4NvO7x98x/EFyM0ny11Fc2rno/RhAs1L+CQh0CFN7nVp9T0AqQOkD368Bu6PI0vZlmohwIO4Heny1L9XynsSMzB/mTwlySOlcXyDAGfl0VnU8AhjbHGck4+fVK9QcyHbFY2aBrym1g+gh0n09ZicwARiCPb5QYty28RWouS+u6W+aduj9ruiXErQ+w8pDiV7UwoqHdMjIiSRISCRxGLzcF1fnvrS6VPAqCAknYArAwD1KcCcRiAMsValVTVKX+/KFrUv5inaVCAo+lI2gqVIMSpYyY4VL1cXVX4Tt2jkE6WC1spPxWDDjzKM3it+I1za8ZlLbmNf6vu+oKa2ILlM2/Vq2toAhS0JQkx6gApUnEzAI4gPIbwxM8/dMcyLw9rPRmk6rl/ppV/pqO8rLb+oQlYCaWlQgq3vwR6fUEbfUOpEBulUuqpUIQ6vy3WULUkJIIMn8o2mAdwIjJSqCCUkxq4bnHWGtop23tym1yQVJkq3iBCv7uD2TJESFF7Gt9lOoe52NSti60uqs6XGSVPU5MJnDQATKIIEYIPScdo4eqe/oqLNQ1O0NGfKWsSopUI6f3ZHZMkQT78NFpS48hTPlmpgggmSJwIOR1J/XcehJ4+K3W+oXSOKJSXEraWsFPl5jp+VPpEyZxPXji+IBxaV9D2q5amCUFHHkPqiltOoVmsqhTIJ3JKAFyqYg9iTnPbcQB1JO2pta2HUWolm3IrWrW4pCtlQ4HakNSgORgCcqInoMSDJ4pnoy8uN1Bplr3ONyWvmAcA6TkmYk+8k9uCbpvWpNQyipdVJbhveohxaTPzAHpA6yMDIABgzaa7wsRu5LnxLZW1sf3iPnhWYTo2lpa41NsW+7Yn3lPUFXUsJS48yJSArJAUUwCkEpBA691DrbCCUggAOl0grHpGE5Bge3XOCPpwM9Mc3nWbSzSJr/ADKRClbGFrISwV+o7RgAHoSB2mOFl25kOPU61BAcWgSSr0oKhgxIGSRjPsIEwXWOraQCqentrmZaQjto7VmnrRoA21yzuft9Nal5Fx/EENt04RlotQQVlXfrtPcEcSGzU9Nd9pUykJScFO1cqMmQTgARP+WRxWGk5hor2FrDqnVjKSHAU4k+3pyCYHuBjiTaZ5+v2UJQXHGlpABWrcUAexHWMdQDiTExMiKSFx3KDVEE8LC5i6JeCrweaf5v11bVXhpTlDSJT5iEHYt1R6SeoBAnBmBHbiS+PD4fOj7Dyiqr7YqZdGq3gFxrzCUlJOYnoeKyeDj4lyOQ99dcqkC4UVdDdUwp1KVgjoQRgHJwe3fideMX4vFo5v8ALxdisducoaOr9dQ888lSlAdRAwB1mT+o4kupa91wY6F4+77Z3GMdcg9VWUkkzQ4uY81OrynfvtvyxjmufnNfSVwp7BdLBQrUxZ7pXM1ldTCA3VutBYZ3EkBYbLjigBPrIOYECh7lwmnKvU8W4WslLcl8g4UEomRkZPWMkEYJmteblJerk6ufKgglK1AKWnP5Vd+pEiZAwOvEI1RzIapVSlX40pCVp/eQmQQdxz0gRBxAkdchbhRUonc9gG5X0LYrtcTTta8b4UZq9AMstrGxKFIlKwSncAAdogHcmMCMYMe44jGqLZS2fDigysnosBSlEEdhmJJgR7Zxx41ZzLqnEoQ3UtwmFtJY+VB2hKSCD6gAYMenPUHPA3vd3qLsSEqU4D6tqDlafcGCCAfbpGTHC5UzUsR0tCb6c1kgy8rVrC9U6gpDSQuDgtqkdYBlRkqOBgjP8COrxNWsqbb80rUcqyTJ9x1wPc9veOJ0dD1t2dUU0617VbRCTtT8wMrIABmJyfyxnHGxXLIJcPnPNrWRuIEq2n82foc4Hb6cAZp3PPkC3npzp85QXuVgVUICgkyepEGfv279Ppwjt2ng5UhO3G4gbcnt0+33/T3LWo9OMUowkubfSDv+UDp1wP1iJGeGG22eKlISkpgSYMQY6T1H8uMjpnPO6VawRRyZB5J65d6I8xTaxClBSVDPQ9ck/YfwxMcdqPhG+Je+c6OUV405qm7Vd8vej321s3GufU9V1lE/u2eYpXqUWloWgE9UKR145H6Hs7qw2UI8wyFbidpRMRnpAkd+/wDC53wpufdn5B8+69eprixa7Fqu2G3uVzy4pad9LqXW1rVPpbkFG8zt3ZIkx3uVkL6Jxa3LgM+qF1V4pyQ3K6tJSU9R16KAyT0/x4a9f66tnK7QV51NfKlNHZdP0Ttxrn9m4ssto3LUEjJOIAHUqHA85reOTl1yr1HbbAq6r1HqS7+SaO22cIfU8HQC2VPKUhhtKkncNzmE56QeKceOn4vGj+bfhm1loqwaf1Zbr9eA1Qtv1IpnKbamoSt07krMyhsgQFAhXWRPCbR2Ctm0uMZDCcZ5e9RKi8UsZcwSAuA5Ku3j/wDi98w/EZWXqyaSq6nRXLSvpzSGgQhCLncmgQpbj9SkkoKikfu2VAITKSVlU8c+L5TNrfU88sqd3KUSVFTiyeuVGZOcnJ6kk8EnVlPdb+8pbqS2Y3BCpjaREkQTtwfYYJnIBg14tlPbaglxzzFqwkBXb2+v6Dh3jo4qduiMYS6KmSRxc481GXaZTii22ghAVjqCfof58fRZShne4JEjamMAdvr9uHI16W5DKEIn8wP17d+3CKuqgidyj9ACSY65Pt/pxjgOa7brQpCkLGyYSc4AGcRHfhMt0JG1IPpx6TP8+nvxrqazzMJISB7Hr/DhTZLu1Qpr/OoKevTVUymWlOqI/CqUcOpiPUAI9uOWcnAXrnhoyUicdU7BBTkYAEgzxtYodz0ncFHvPT6E8bqelhcJgrOJmRPsP5dOHO327ecgjd1AzxLgpi85KEV1wEY2K0U1HsJTECOh6DHX78KW2ZEyBAyomeNymwlWY9/m4Q1tb+72pkDqRuM/rj69+nBAhsbUFYXzuyVvfuQpyQgmDjr1/wAuv+PCm01ZXUNZ2K3RMYT2M/TIGfY+x4YWlFdTJVAhQJ7Z/o/5cPtlShtLfllCirJlM9ziD0EE/rP3EXxi4qcIGsaiFpZ8W5tDQTUbkr2kj0qwDIQesSRIj78GTkb4j9bcha25VmjNS3XT1fd7cu1V7tE5sU9TOJAcZXIKYJIIMggztjrwDrI8VF5ISIdcRvSckzJHt/eInoSB168ErRNrVd/KcWVOF3BBUpWf+4kkmZA7D6nHDFbmiTAS5dZvBaXZwpFpzT/7Rq4KSloOGG0EKAglIknCgJAwcT1HE+sNgLTrLbIQXm8l2ApM+wJ+u72Gf4J9L21TqErMJBSYCfT+VOfftEdY6RJHE4tVm8gNpPoLchRgREA56Z956Y4tG0WwNbqwqO4hv7nuLCUmt1gSy00qJVEJkSCZ3Awe0R1yY9uHFu1l59SQColUkYCcd56dCPoP8HJdEmip0egyEgbSCCnrjpiD7e/XHCC6XhFgoy+VtoViCTlJyAf1k9epmIyeGcxRxMy/kkHx5ah+iMZJSLWOqaXSenV+tKFuDzFrkAIGRHbB6yDn7cCnmXobUgoLHcLjS+TQaop1VdtW3UJWaxlK1NFxLaSSIWmBIhUpiRu4Rax1XV6xva2WFK2NEhSk5nMGAYwQDkwMEDhsrLoq1Ujeyrc8xoBTSt0hlICjuBAmRP8AAyADHCFebi6qk0tPkCtjhyyR0EWs7yuSMOJtdN5ywpLphotFpSgtAwoZkxIj2xkg8QbU9bVXmC3O1SJTuJUpCAqYEEKB6kqwkCehVl9TUN3q+0tOpYRSpcKEQoJSgAyT0Pqgz3gk9+p7+KRyq5D8oLNoRPJXWdfqw3SzIqL8h4pP4WrMIIV6QUrUSRsztgZPCrXSF0RIOB+qdrcAyoaxw8x+n+VULU9yDtQ4tpDiEOSUBKZiDsO2QSrCU4EEArBIHVpSUs1CXyny22kLWlTQDhlIOE9QrrkmUknMAceKx4PPvFLiHVuqELKvMWgLAUpJgQomTMDMQSkE8aq65KZp3yhXzrSspDigt3bJKgIOEJbBM+qTkkdEuZ+2VYkEeCAm281CWlklJQlhYkFQ3IlJDfqAHoBAkHHzn8ySWNttNS7Spa2pcQ6ASEYUsEJ3ExBImSeme/QrawrFclCylpDywle9ClKbT6UlYGFKnIxOERmJKFa1NtbgyyXKn0KSVFQb3DcCBJSYgFWARtIwZ4FOdlFGq19GULo2Slxhbji1llrchSpGxEQMqjpBJBxKogiRrsabpscT+K8kvKAUYbIhRmVZSOsyJA78RfRN8S7TN07j9RteLcjzAougH+4ImAFQYgncVEgRxbr4fXgvu3jI5u09gs79PSKqWhVuVji1OhoJHqUkmQYJEHJmc+8mVoJVgWG6lrC0nAAVYam1P2dpipdQpryklQIWQnacmRAO0ycf/jjiQ2y6JqqJSAotOpBS4JTCUkQQRIAEHOCOkzPFo/Hx4Eb34O9eKsl8eRcRXNmpYrky4p9BMq64CiQQVKySoRxTK/tvaIvK0ZU15hTvTkoMA9YgRuKgR6v048FMfaarCor7EIQHnIPI9FPLPfaj5FvFK0YeAWFKViRnuYJxHeM9OJBqbVzabxVN2x2tqKMHaz+KWhNTG0TIk5yQB/dH3HA5pb05W+elRAcpWEnadxGAIiSQPcDp8sYE8SO06iU4wW6gJS6WikPEFSXRuSNpzhJV2Ez7HEFoZZGswhFcaSV+chOy9bCpeC1oJUVSkKA2ORtOFJzCZwMJO7oMDh6ptbI1A3RttNUDQp0JbCkKBVVrKioLWZIKlExJ27toG0HPCO3N2q6PIKvLaUuf3bioCjlRIJTIwR0gyTHsH1nQ9luiN4rmm3FwVNpJdKIk/LM7oExOBu7jiSyQu5HdBJqBrTvuEhGrTS0u1K3m/SpCkpTvUUyCkKVHcpmCYkdY4Zr7zGfdfV1DhUCtRQDKwQBlWAUzkgEkDpIw/wBfoS0UTCUftTzNxPpCxKQRlQITtAmE/afoSxXfT9pokEIeYc2wSlgnd0lQBGEgyrEY3cd9VQ4YD1xaKCJwcWclELxq2pUpNOULKEynYYlK49SiCDnb7SQSc5A4jdfeaivIHmKLa1nYtCtqVKBMlSlYBHbEdepPE9XpW1vEJWhLbY9SUxtwFRCVLHuDJ2/MP1H2k0vahUbkOU7jajv85DYEkFMbSogQZEEAmQQMTEF9vmfsXIoziajhadLAhqxZKm7PrKW1FKlpVJ3KkHr82SEzGI+4jh9t+i6ilaUpSEKTO1tRUlKSoJMiCAlUSoz6sj3PE2rnKG0UYCEMhUeW+FhLhBhM5nvlUTiehHEO1DWVl5q3kp3SswtBhITBIiJBIBSSQrHUn341/hMMQydyoc3GT5Dpi2HuX2kt6bxeqa3suGrq33UstNhyBuMJQkrUAkCTGfb9eGXXdur9MXeutdYymnrLW84y8224HEU7qFbVgbRCvUCJJORjoOFFBp119bbaVqcSte5DSBswBknqCSJPUwSJMRMhsPJ1qutyqmsXVoDbIcZKGVKC1ASEzAgkkgkYkdenHpgYwbDCiurKipOS5BW60jr7xSrcAlJCQMewkYg9T/Gfrw46OsbCKtKqtZCSsDYAqSR1iSAIPbrkZHXjZrK21FFWODyQwEuQlsNgLT0wB1GB/nPbhrttmr6twlhsulBASveClAORk4GJzMe3vxzZOyN+ojOFBq7W9zCC7BVibzetKU2j6NVA35dWw0ZbCwJPYZODj37HoCeIxcubv4BhSqVxlhMy0WhCdu5MnfAMSfy9MicZjunOVd1uOz8R5gZZKWlhCQNqiFdFrhKCMdZEicKE8Smm5SN6baTUVrlMisUNohAUpJyZEkDp0k5HU4I4OV96mrJQ6JgYAANvRI1Dw1T2+ItnlLySTlx33OfkOiZGK+46gU6KJh8oX6gYLaEpMYUTlSfUkQY94JEhlvDVPaadT1Q+XXylKUICpeWraJCQMAHtEjBg44ll4ti31hlpC6dBVKVrWlbpXgewAkAdNvzdcwY9dNKsUdMXVu+W7ghRQoOQTkEnpEScgAjv2FzSycnlSYqWJ7vwW7d0M9X3x6qQUNJdQyYKEGEmI646kD7x2zkDm/ocaCifVkhQ/vmepPTJ9z7cE3WtdT2111LJRugFaSfUknOQSBJye+MnJngVX1T1xdUPLIJUQoBIJBHacdgR0/04DSPLnbKbI2OJuCd0z1dyS0kgKBOQVbsf0ONf7Hqaiwm5hKPwfn/hiouDcF7d0FPzEQOsRx9fsrzpUkqWokwYPy/fHGpNuU04ScAyoR269D7Z+/Ggie7oh0tW1oSRthSwYT37YHf+v04X0lt2rJGVHIkEdv44+n+/CiioiYSlIBBgR2/zP68PlBYlKbCyN3ZIGZPtjp1j+PE+npN0Fq7h2SCjoNpSYmehERw6Jb/DsmY3EEgEwTH+PC6jsYp/UspKUpJkyEgAT/D/AB4QXVRW4pKQkkTAIyn+GOh+vQcEyxsbMlA8unkGybLjVlaylJICvQAcDp0PX/zHvwo0fqShsH7T/HWC2ag/aVseoKYVq3EmgeVGyqb2qEutjoFSn3B4TOUS33QI2lQI6fMABIHvj+OI+u2ntvp3JCvKWSn0rk4kn9I9v4Znga8l5RuLTG3CR0dKWfKnaAFASVxI6wYyT+nY8PFrpv3KJOwuKCJ29dwEGfuekZk+/C2zaddqllhhtBUoRAiSoZgHp16nv/LiX0nLattT6nHmKimG/elC1SsddqnI/n0kjHfjpFSvfyC4z10TdnFadH2NTq1FaSGwtKCmSAuUqT3ndEdSoRJAHUA1aD0+XgQsrBGXdxG87v730JIAyTHuOInpDRooyhIbSFBfqGwkkxOCInGOvX9eDVpLUdksXLq4W+pspqLvW1DJpbgH1b6RCCsLR5eQvzCQN3UbcdTw+8PUAZ55Oiq7iy8Of+FD1Sq33NFqAS0hK1IykhQMQmSe5IjvPtjsH2wXRx2uZStzaCrcgIXhIMAkQTBEASf7o6xxBrG887ucdSFvvrG1sNJKnSo9NoJKs/TqQMDiY21//hukNW+pZqCZSVLAQADJJwdyuhj1RuyDMh9irQ0ZJ2VWVVqfISGjfuuufw6/hqcsecHhNpNQ6iYRcbxekuFxwOwbeQSAI7KHUz78ckfGzY2NKc8L7pm31jL1vs9a7T+aj1JWhCoUspAgfKPlESesHiT2rxicwuVOn6+0WjUd4sNK+Ciopm6ksI3QYBBPQpjvMH3jiunMPmE5U3RdSuoZffW6XFKdTA3T7L6gkT9CcjB4WKqaoYZnzzFzXnIB/pHYfkm210EB8GOCANdGCHOByXHueyn+n/CfzB1jyNvet9OaTu1forTJT+1rq2gq8lfpEJGZ2pJMAqCRBJMjiuV81S9cXnEIKoVO5sBKipUmYSM5VECIwT8pjiyXJr4wXNnw7+FPUfJnTjtic0tqtp9p9yspCqvpi6CHShcpELbHqJSqAcGTitVgs6E0z1S+gimQ6FMoIiRvA6j0jPyqAGcwCY4UZK2WokMYGGg7e71Vjw0ENLGJHHLj+foldO83p+nFc6Q+4lKtiJSCmTJndEmZkgZnCRIJgOobs5WPONrceDgBVsSolSlFKkHaMiUz3lXYdZDtrTUQr9zXmJbC0ltkAypcHaencj8wGCe8nhitWlrpq4VdbbqKtrRRMqfrk0TKyzRU4MFwqR8qQoknaNo7kTt4DXOqyfDYdgmC0URaPGkG5SOhWmoDqxKfJRCkyVNrCikGDJAAyrPcTkiOGW73NDtI6gKeU6rYlajJwEK9QJAzComJAChkcLq2qabom0s7CsKGwlG/zITnB9MJO07QDBM4IHEcK2qirbLSQd6wFKcCnCkmcDochPuJPWIgA55PLgJhhZg5K8tVaXqcFwsBxQ29JPpklUGSpJJSQfoYzwnfrWm22UOJUolIbUlJTlJgwVZn0xmB1VkRn29dSEpWhb6kpSGwCqCiSYAPQJnIgY68JH0ht55AdZbCHTsG4EJO4BRmDMdZnHUexglS1Z+mqjaUbypwIUSpKg8tsKCZAAyqPyhGcFKgojcscWQ8Gni71H4U9f2++6dudLRXWlchJC/NSsKSApsj/wBRJkSB7/ljgGNWlTrLW1twJeUtLZSolYUAgByff0wFA5XJMJ+ZvYZrLLtSlDrexRJW2YjcVESYJgbgokzO36JPDHPQOaMrLbeA04CuR4uvG7qzxW67/wCIdZ1TQfQwGGSxLDTTeSNqSNwAJTMiYVPcEA/U1pZ1FTrlTcQpLpMrS0okgEmAVbSCqeg6SIjiMcwOfd65x3Rusv1ciormKVm2pU2whr8O1TJS0hASjb6wlABVgkkrBKscI9L8wm7bdAKh0KQstqcQFD1gKBKFJTvMmQAJOdvckiEZPD2TnS3ESRhp2C9nQ7lFVoVSPq2bFJR6wlZ7AQpXy/XEwDtOYIGgNLPXhva+CpYQnGw7QkErKvUJjI7EAA4kmZHy+uemNbWKrqf20TfKYNCmpfJSoVExvBUkwjan5ZGZEwRPE65Va3tvLXU7NzrrHZdS25lCgbY8tSaWqCmylKiU7T6FKQcADAmQSOJ8Dm+1nK0mkcMlqgOoOXbdjcDDTyUIW4krCVlxJkwEgwkKIJA3AxITt68Ru43Fy2sqVDzLAKgUqUN6IkkSDtKon05mJxOSNcL7RVBIUo7Q6skEgt7lEJ3ZgDdBECZVtOR8zRVaFotRMJS04PLUpxsLYAVtBBMyBkJG7MyANuYHEvwwd2ryOrkaMZQxuWs2kLKU/im2JhYUQorG4xO1RmUyJmevSeGw6wZYfSlSqplQTI3oKNhEnA6kSIJM/oDHE/u/htuaWEPp3uMLJCiCdySZxBlMyoCZG0Z9iYpceSz9AFKWxWMqeKzJUSEAABAidylYJED7Y6do9bNsIHXvmc4kOSKj1tQobOXXlLCVje3lJk42kSfdIAxCczkSHTd/OoKlKGXXnVFKVJKQlPlyc7TKQcqiUxIIz1HDVYOV5VUswyEOpKXN+5aRuJABURtEkJ2pVOd4wOoNHJ/l5bdMUrYddUmoUvchS0klO1JI27uqc9ComBJPXgtbKZtRMIn+UHqlG9V1RSQOmblxHQcymuzcp6nUVG665TVSA8lKqby1EpJk/M4pUpAkxsGTOEkjhwo/C9XlgrqEIQwAhCQ4VJTsG4wCpPp6YA29THQ8GG2a7sui6XznXGFLSkOEIdBKslJgAhWPf2IBgQDDNd+KJNwqagISy8XdywhsJHf1EkZMKBx1iMQDBG822mpX+Gx2pReF75VVkHjTxlh7FR5PI2ktFKX6hbGzzIO9SR2AUIGRCdpI7AHHqE9Ofh6aa8OT/g+vCdX/APCa7sF1Bua7mUoqUtx+72boUAE9CMzMknjlLdOZlfe6tTg8wqUN6FJfK1OKiPV6QQN0GMbQFCPZwtd3fTSjzzVKQYUoEk+mQFbcEyDPQDM/qoVUHiYA2Vhw3EOj0OJHLcc1Fue+i9Pr5lXYWdDtVa231uUyluAJCPMMEgpJJg/ZJH68Ri32UUDi/LRTJcA8kKKQ44PUUlCt2E9VKiM5T2AE3vd7tqHVOvVFKhCnNihuLu0kJ+nqVgSZIjGBwq0NetI3S2X5Vw1NTWJ222pyqt5cpis3qpb2g0yFI+UqCyRuMYIPXjyOiYN3lZdb65zNMXNP3hY5d6Z1Pzf0+xrnUruk9PtvFqovD1MKgUaUI3pV5WT6lBAP5fUJEQVQbmFR2+l1Rdm6K5Lr6ZmuWG6hLSg4tALgQsFIlCSACU7knIGSOIDqXm4urU47vrHN6hMCQ4QjJlUx17FPzI6CIi72ta68PJRSvVIp24Lfl7lLcgRMEjatRxmCIVlREK8lqo2jRFuUuQxSF3i1B2Uu1BrCksTADSUpWlG1IQkFwjduMJBGyTJyT0x2BgGp77W6geS0t3y1Pq2BlB3rRBIOI3bZCsjJj8wnhyt2lap5JePniUpWVEJAgKBz9jsMDbE4gSOHeh5fihUsLStbjhLZO1JXGATBwe0iDBg9N3HkVpqandwwFyq+KqalboYc+5C3/hAV9G2t5wKDilLiNziQZUJBEwSSMxGZj8zbU6SQpRS2hKhugQqEnbkyCM+0GB/Lg1q0Gq4+S08ryt6kjy5UPKBOwKCeqiOwPsY6ji2nh9+B1zB8TfIir19ZmEN28oWumYeqEocrggGQgH5ozHST+o4kzWuOlALygAvctYSGBcy7np4qDiS2dpJCEnsJ+0DH8YM+3EarbEfMWopED5s4n+vpweuZ2hHNDXmpoHabynqVam6gLSIQpKoOf0VPTp2zxDtM6Ho9Q6ytlLeK5Vpt9XUIZqazyS+aJoqhbvlJ9S9o9UdTHvjjnLTBq5MncUPrbZ8oSlKyVAwQY3exMxg8GjkHz61D4ebNrKmsdHYqlGuLG9YLgbjQN1SqdhwgqUwpf/ScBAheD34ilTbbbZ7nUIo3/wAcy08tLNSoBsPpCvSoggwNoBg9N4HGmofcq1qaBUlskBIIMKkCAR1VP0/w47RRNaMqJM9z3YTNdkOOqUC6hKimTKjIIEgGYjPvgduGty1CQlQDhWnfsg5T1BGe8AicfwB4kX7Dfq9oSHBuSSEmVAnInGTkHqInEcOlLoV1sphEKW6pCTnKZ7jGP1zH6caGmfIchdBWxwjBKgqLPtBU4kGSApK0EoSJggRjt2z/AIcOFHYgppBQkqIkSBAGYAgTkwc9ZBAAjiY0OhgKlpKw4qFbwABvT6pKiMSraMDvuxmOHG1acYolFSkLSUgKQoHaE7e5mB7dTH0njvDbTncKNPdgRscqQ+GZdu5ac0rFeb1bG7tbrZXNVVbQOK2IqmkupK0KMwQtMJmeihxcv4uXi/5YeNzmfpS6ctNH/wDC1NaLb+Gq3HKRqlXULUQoBSWzG1tIKQr/AE4plpehbu1XTNLUmmYWr1vJRv8AJABlQSYO0TkSCMntl5FWGGypltG0jY1CQpSjIIMdTJBzkkkxjbwepaOEFr3ZyP3ySxWVU79QaR5v0/fxUw5SafsdVrixUF3vKLHb6+sZpqq4lpTrdqZcdCVulIyoNo9UgmUgjHaUc6OWFg0Vzz1JYtHaiRrXTVtqVtW29tNbGri3tBKwAduCFCRg7ZE44GdDSppW3Ha0wlKUjy43z/8AEyJ9HzHtgAdONrWp6ispSzQn8FQEhK1KIH4lUwIT3j5gEiQYMgxwabKWbM7cv1KXXUeo6nHOTz/QKcV17t2hmkJQ4hypebP735incBCW0pCcT2AG4pncEkAsdw1Mupfcq64OM7UoV5O2EgfMoiIOTJATgE4joYpUXxizredSFv1LeHXFwAVBJAnOMmf06cNPMjmDZXdF0VLRpqG9RNVjrdXUIfSqlcaVCadDKQNylT5ilqIgEpiAJ4jTVohGuU7qZTW505EcIwDzPVbde8zn7hV1C0uBThlW70kJjad2O8GDGY+44gztS7dSqpdD61OKAS0JK90DaBnHUTImDPc8Rw3tC3TUrUhakIWEhbwKwrrIgdBn8xAmJnPF3vge+GnRXio8ZFksvMCqpkWQtVB/CuOhCnnEJBbZCiBlZmTGYMRPCvU3UzAud0TjS2ZtOGtZzKqTZtGOVDwqKoFTSE7w33fCgr1pBPqSSAJwOwzE/NVPvGiQltva03IASIUCqMjdJk9AfqImAB1w+Pf4GOU/hWpNJ1Whaduy193LgqremoKw2hMBLmTuAVJEf9v345Ya2oWG6FLrSw3uQdqkp8vYAARmTA749jHaJFFC2WlMse2QodXVmKtbDJvgoTVlCp1Zd3tsMuOOJSVStG6VAkiUgSMgq29Mg9OJ54dPHHr/AMIlt17TaDuVtoUcyNPuaXvyn6NNQ5UUbiVFwMqXJS4QpSt6SQNwiYBA+1RWtU9xdbQ+2ryX/MDiD5beEkbkKE+mCSInuT3mKVVxcS3uKVkJSlpxJTLaiVZIPQgbgAlXcjuSAlVbhqIVi0bSWgrTVVSVtJSGkObdq0ArKwpCUhQEDpuEGTklMCIJGtuxu1d3dt1CuouhL5ZZ/DMEOVCSIKdplad2PSATgkgTBQ1z6U7vKBKyPJTKxuV+UAQSZOSe24xPfh/5M84NQeHrmhY9aaYqzaNRafqzW2+saOKR4IKUrAUkjaRIgzuJ7dOBxKIAKPsuOW+6OI2JbcCvKIU0FhCyTIUnJ+UEJ25Bjpw3u1RaaR6i2W0oASViU5MSB1gT1yMdwTwoutTUXq6OVDzrztU+tTjriQVKU6pQK5KcZKiSR2P24SfiF1oZ85SkturK1An05hMwcAnAn2A6wZ0Kz1XT3x9+B+7/AA/ufC9B3i+W+9LRQs1Ka+lSoSl1EbFGNycJI9WAAekiAdXpbr6Mo8unUtpIHlpaKEtBRkJSMnpJABgwDKtxHD4ddVevWU1V3qX6p14FtbtQ+HnSoN7Q2A5J2gHA3TCSokbEgtVfTpTWhLTjJp9wH4cLAQ6CsBaACTIkpJ69IAJE8WDE7U3D90rVFOY362bBRS5MOUoJQ8EUyFbVIUJlM4BHqgQUwYGR3JMIKy4ipS8yVKS/6/JKCoBBUZIM4iSQCCFSJkpB4k9PTKeo0pab85lCSpaXQkkwJCcYACRMAYBUenDfUaVUXgFNoUVuKaKUO5CVeiEk+lUkYxB2me5Aiut59toRi3XQtIa8poqb7UtqU6HFOFsJcQqJgJQAApBkbSUCEkhJ3AJTEw40XNG50byqeqXvSkHakqV5iwYSoJXtMEwCZ9MA9zw1CzPOpaQW1IUhzalQktnamSd0GVbjG4iMx1HpS/sU+WoBThp2l7mkqUk7TJKk5BAXG0bT1gkiIhf8KRvspmiuDe6k55x1LzjPlvVCCpSQk7pKFFJhZJkiAE9twkQZHD3p/n+7RvB92oWghRCm0oLi0K/LshQCSAIzBhGI6AXP2MIBCP3LmUBQchvOIJEYMmEiTkGIBlN+x1lzY4hKVESjzDKNpAEzMT9gAM5kYz71LHyUqORsmDlWd0V4ylaatpaefBYXgjaE7IHUyYCZ2gT7kdJPDoeeln1PUOLcqXApa9iwl1ptCyBISteYylcqnG38wA4CV48NGttHcsqHWNVa7hbtOXFXlIqmWFFDK0kHMqyodj0//KeID+9aSChbYcbIhPmrSUjACdxBkxnEYSTI2xxJjus4ADlu9oI55CvTyncteu01q3rmxTtopws7lqQREH0oT+ZJJJhMgnrIkxrUevk2CqXT0qm0KYX5aVIaUreJQBB6boiZmeh6EGplr1VXCrCqd8tIcT/1QA2kpKoRtCSJEpUTnqZJ6wutdZXXBKDv/dqhspMrUuT8oKiQkkqJntkDrgrFepS3DBuhVXDTkZcOSOt45oU1Y8lTtZU1K94UttpQndIEiJO6SSQBMkwesxuu5p01O2tLBR5iUStJCYQdwweyDJSZwqScSeItTaWdvDLanAypLY3O1BJcSNqOoBJAGDnr6QCT1Hi4cunUKaR5LLK0FSAkErcaPTZuIBWqYmIOFEyAeMlqqlx1HdQ2eByZhOV45xVyKZxJcNOlO8k7FHCSd0JMCTJM5BLsCD0YrxzZr6h5QVU53BtsqJI3xhKBPqTAgbfSO27B4dnfC1qWm5dta1ftdYxpsXAWxNzCUltqt27koJPyqEqJgCQCR8oAhzWji04lNHTEu+UtBUoEqTtmVKxIAAVtEYKR6sZjGeY7LdzGtGcpS3rKqutXDZeUtwBPr9RUqD3TgmffqQYMHj5+KqrygtNuKdStahtQk5B9iqRHo/L7xuE8OVl5cVVwec8xUlW9S2RkdYTu2EFQgHCj+WYJIkh6K5NUvmoNQ6K1xJCyhO4oKQBAEAST1BIBH96eJ9FZqqreNkGuN+paNmXHdQrS2jm7ptqAyp9zeHSt9BAAEQcSYEgQFQSTAjgjWTQYpaBK1JbJWhMQkBpRIG70gZJBgACDAIz0L/KDkhp/Utm1M9XaksWnntN2Z2606LhUFtV2W0oJFKykAfvlQQBP5cTmB9cOY9FTIBp0NvLd3ALWRCxnAUDJ9KQCEnqSI4daGz0dICZCNQ/VVpcOIq2uOmAYaUpoNLCjdDiQ2Gqckh95SUhCRISRgkExEEjAAjvw7aeotFq05flXTVSbLW2m3Kq7ewihW4LvUhaQKZKgSEFQUsgr6BJmOnA3vGsay8kuuv7Q2gLQpRAbWvegle0TIkKzIUQmOs8MNUlKyhRadeJAUHcpDaYPqQkndGTJPXAIzB3nrmnaMYXCmtrtnVLs+ieKzVKqi4FSEtBQUtKoSPUYn0jBkwnJyYVIwAb6eFT432rPDP4alaBatdnrkU9O4igfckLogo+ojbAWmVAgDIM9ccc5F3sUrCg2xSMlTkKCmgVISEqKjBVA6xCifpA6OXKHSi+b/M6wWCqvFLZaW81rFE5X1rxbapm3nAkKWs/KhJUZk9jkxPASsLZR+MM4R+mzDnwjhe+b/MVjUeqrlXuFl6prqlx5SgkDbkkuSrO0kkwBIkZE4YdN8jtVc1ba9XWyzVVwpGE+a8aZpag2M/N3iBGR24JfxHvB/Q+BbxIVmhrdq21a1pKWmZqkVlEUjaVoK/JUEqgKEdAYIInPF6Pgj/Fb5R+Dbkhq7T+v6ByjuNwqE1TD7FKahVchLYQWFEiYkSJ/vHvwNqpHOGYm6s4RODS04kOFycv+n16ZuQp3kKdqW0pGxKSBuIg7Se5xJEAmdoAgcJKO2ru1UhsNEpcWCEIAUpeBEbesx0j6nJwUvFPru2c8ef2oNR2W3G02+63B55mjaAKadtTpKUo2kDoY2nAM+5ATaT0ApTqXHVpNQ6opASkLIgABUYgkd5z+s8EKC3vmeGgIJc7tHTtLiQpd4ZvDVWc5dW26z2+mFXX3F9FPTtoTAqFLMgfQDPTp17Rxfnm38AXWfKjlK7qPfQXD8HT+fVUtLhTQgFRAjMZmOvfgOfD95jU3h2506Y1JXMB6ms9Wl19JysJMhX2I3GO478deufHxVuU6eRFyftt7buVyudCtqnoW0HzCpSYhQ7DPDNWUdwo3wR0dPra/2jjPX6bdVWtPdbfX/en1tWY3MHkAOM7fXfovzx8wtIv2S6+TsKvKWUKlMBJzmT0Oc5MfXA4i9DpOpu1YEpaWjyyQtcFKUAfLIPc9P4ZweD/zBoWL9qh6qO1G98vKKsATJEDsAJz37e3EbvVytWnrK4EIUrYFbpwlZKSDuI75Eye0z6eC9RZACXyENaotu4kcWNijaXOKg9LokUTLUK8pttsqQt0wlSgQSe5JEiCDORJmeEVXdqS1vw1tdcKlNFwNlSAgETtAJUT2+p9xI4W6juFw1Ql5Sj+FpEggNghGZCJBGQIWemOmZnaz195pbKk+Ult1xKlOIVI3LEiJVHSEnA/7pJJkhZjE04ZyHVNVO2Vzcy7nsP1WxNmNaymruZUimIKW0AfMTMIxgYMnMxuJ6EjRW3ldzqHktKShptMvHyyskZ9IPUAZBznvPThhcvbt7fW4VhIKStb64KWkwJSPr1BzjiL655koprYKWl3oo2kFRQVDJ7E46n0iDI6AlJE8RJayGFhceX1KmxW6eolDBufoP9pXrLWdNbWjToCkra2oXvPrWSCBtI9x9gBkkdRBa6sqa4oeqAGE7fM8wKiIAgj2EKmB6vVkAAwjrKx78Q8p9SkrQVl0mQSZBKZx1UQJPUkSIAPG6lt4tri3Nhc/DuqbKXUh3zFgHaYxE5IBONoKpHCDX3B9TMTyCsy22qKkhxjLuqdqOiCHVObS0mnDTTqAlXpAIykHqkekCQJMDHeXaa5qVvL64U1Ta3naaqQUvMqaUptwGEhOxSTKiAMiYwcxgD126ob8kqhKVPqc9RCggSFKVuIABKSTMdgCO/H2kqg3TF55xlRXueLJUd7YETAIiNpAA7dZSN3G9POGclzqKXxOYRh1hz71DzJqUVupL3X3N0qS0l+sqlvLkpxC1yI6GBGCDgSOBrrrmA/W1DiGtwLSf3ZCCQIPuAYE5yIxI3do9e9Rfgmi048ErUCVo3StQBhRKpPZSSI/VOI4i1Zc6httaAlbKlpkpCQ2CoiJIkZxIPaE53CeNq2+O8IwxnmuVDYImzCd45Jdcn3rhUPpLakBLa1OiFGUneohSVKjckAHMRBMZVLReFr/ABLjKXijzVo8zcd/ykwdwJOSATEjIHYHgweA/Tug9eeLHQdn5o3den+XlyurVNeaxDiYZYKVbVKgEJBVsTChCdyjETNqv7RP4a/Dj4bOcWkKPkBe6CtZu1p82/UduuouVHblBSUtOBxKjtccQDuQqRASY+UBaOT5imgYBAC5x1G5x5RcVKTlaZICTgEkKEKyDj6gdIHHSDwh+BDwwc2PhNc0+Y+seZJsPNSw1FS1arZ+PaS6HGm9zLYpzKny+Soen5ds8c5326r8SmrSy8lCyIWBtTuT6iEGACQDgAYCUnthE/dnkUzO6EhnaiVD1FMkgAg5BxMZxOO+rTg5Wzhnkpbr7mBbdcaO0nbqfS1utN4sjD7Vxu1E8pTt9StaVM+Y38jamW5SCkkqCgSeEnPTS1i0DzJuVq03q2k1xp+kDCqa90lvco26vzGmnFANPBLiFtrUplU9S0YJTxEFn8QraooCm4bG31JjPTMDoAD0j9Txo9T0ubdnZXqyepjrnH+A9+MJysyrsczfEvaOcHMav1DatHaf0Ja7jSMUyrPaajyaVry2EoW4CtSzvcHqcELkqA2y4Bw4ou72pqNbVS4relDrqA4yE7nV7IKVhJVuCl/KpKgMbVEE7a6aW1Ahyrp3SpQRUtlCoQUNGBBQCDjC/USTPpySsgE+vtSNGaT0veXLtpm4q1JRPOop6GtS9V24NOqaWKxmUeS47s3hoqUFhZODkscFW7n2UJ7G8ip7bLax+3aQVLiGwry1Ph353uikrKcbpTtMHMe3Bg1Fyz0zedLKdRUodUtMHykIHRCQobzhQAMEkHrMQSDXzT9+WpgslldQlL/lOB5twqB2j0yCCdwXjPzKRABnbOKjm4bly8t2naez0VA7YXqmoVXMU5FbWodDY2PSY2t7DAMBJdUCZJh0snEVPA18U0evUPkkfiPhuoqpI5aaYx6TnA6+iIvOnwCcxOSfLDT2t9WaVq7dpjWIQmjuKsl1WxSkJVA3pCh6hv2lXq9QngTVegEVIUp5a3nIUttJQklwHqorBBiSB+bpH5uCx4h/iR82PEp4fdJ8uNS3+gqdL6R2LoVt04Q44lCNjbjqmgoubWVKCYSkggmM8Aml1BcLCkofSpKvMPqQ5LYViTMztGBuB2wrHvwPgNO/aZmN+i71JqogDE75pxqOUzyWQGfMZ2JTiVgGSem7cP8AtJAA7yTB400XK1yjeSUISg7/ADAjeQCBE+n2lAKsCRuwDHE95Aau0TqHU71HzCut80/Zk0FSW6m3UYq3V1nlqNO2tE/9NTiUoUfyzMpT6uG2jvpqnEOPKbYcSsKKkJMqJI/MkCYIjqJ6j34kNs1HLks9FBHEddD7aInMrxe61154TLPylrW0N6YstZ+JZWlKXDG6QnHSDt6K7bsDHFdajlgKdCXVIDOQWlOjZ5UZ2gbTkkyOvQAk9OCFW1i0KS22ttalpOTASkSCoqgSAEiCOvXAxw3UGoaCxakp66so2qyjYeQ5UMqUsofQIUpO/sCAUkyFRmSRPEaexQMy7CIU/FVTJ5cqMUmjNjKFlaghKgpSA4dqVknaQFerqBMAnIEz0lul9Aodd2rbSmG0ndMg9BEd8AnBBkBXRIgicw9RNeLfxDX1/lLyzqrLT1AcraTTdtLlW7RsMU8vrxBUJDqzEABXzTgOPh61JYk3FX49sJWhxKFbVbiIGAmIBSEwAPvOOk6yWqCombANie6HcR8RVFFTvqXAuDeg5plodIooqR5RKw6zMJU4CVbk9ciCogpSDuT1SACY4ba6wuVbnUFS1QFMoCd53JgknI/LBRlI6nqeC1zJvFjav77lndFRSqCgkqZ2K2QEpXGIVJmZMT78Rmr1S21TLh9xaCVJVtAAdG443dDI6iTIjIM8H6nhtjHlpPJLdBxxJJE1wB3781B1aVvFz08baaypRbVLFT+DL6gwl4D/AKgRO3zClJzM/N177aDkq0w6pS3VKbSopU4kJkCIVG8JEqAmIIjcO44fKzWdCl4JWoguQ4oU5IU3KfmGJPXoTjaDBOSyXbUNRdg6UMU6AILinlFtpsp9KgqfUScnsRuVEZ4iG2U0Z5ZKmniWslHtYCckUdnsjSE/MtKgQ0D5ra0gnBmMED7iSmTPDbeuZaUUv4WnpVAiQVuulanTtMkZGCE9OhIEqGOGt2203nbqm4gtok7W/SYxClKx0ke84k9uNC9Z2jTrS26SnogYkIEOEHAAkjpP92AYBnIjsC5o2OgIbNUCR4JaXu+iT3mvvWsStbqXVNbhmPQ0DnuAJCY/KN0caG9KKSy47VVakectSpEEOgHHX1KIG3MSYUJMbuGi981Kisd3NNhtIJ3VDxJISZEJQI9IAkTkQfbhLqyx6itOkbPqCrdZVb9Qh80ZYqWlv7miG3C41uLjY3KTAWhO5MlMgSB01RTt5kuKIwwVbwAAGD6p5uNVTUdSVIaUQJhdSvZB3dSSSRInIJMHBE4YL3qAKWtJdCHQsNKQlA2LV/dM4lWMEHOfdPDPT0z90c8xwvKQg+YlO+VqH1OYMdDAJOI6J4dKbRRQlC1+YzTwdqXG43JEhR6QTux9SJI7cRw58v8ALauznxQfzXZKa13AMuIUhourKlQVLEuJCjncRtAOe3QTk8am6qrr/MCFKebUoqcUBO8kzJHyicH69568ShvS9tpadYJW69O1aHQAhgHaUGAI/vEg4iPm4V1KGrXTpKklTu1C/LCTAB+UlShJmTnJg/fjYUJzl68fdWYwwKM1Gj6hawalCgNkpLiyklEY6diM47YzMl0suiXq4LlBQ0mCpayUBXQ4yEnJGZzEAGOJHY7Ebw75r6EBCyklJgKcEzBIMzgZxOegJ4mFo0+0H3F1I2q3iCvcFSTAMYkmT0jr24K0llD3ZIwPVL9z4j8JuAd/T9FHNJ8tUpZBUgBxKtylLJBQTM4jHvhMwD0kcEiz0NDpy3pS6yhbjigpZBA2pzJ+oyM46TnoH/RnJ+5aq5aas1LbBQVFs0Wyy7cQ7UtNPoQ6vYkpbJCljcIJSDtj7HgXXjUKHEEVD4bpwobokFZIzmDIPt1jMcNFP4NID4YyRsfRItQ+ouTsOJ0/mpQNXVFSSGUFKSsALSoIC4zHvP8A3fSCOnE35Cc+9OcteZ7FdqjRrevbKhh5ldncqlU6X1qbKUuFaQSCgkKg9oMjB4DaKev1KyNjAo6FtJ8x4pgARClDsrp3Kj6pMdvtRqe36etqkMrWt11vzHSuApQCyEpVgERBJAI3bs9xxCqb2NDmHJz6kfUbhFaPhUPcHEAY9M/PKfrle6u8PbWB5QSvcYWkpBzAKp9xBI9u8kcRO+XujsxK11FHV1TcBkBY8hoA/UErJHsOgSSep4jGseb9RVOJpKVQNIQAIbW0l5sFSZ2rKlAjPXcTOZMAQ+vrCismoKnKhTaZUtsDaT0ABEgQSQJzEnsCs1l4mndlyd7fw/T0rQ1v/tPV/wBf1lY44lsuLWAStSVGDEAkGCTnbg4x2nEZW+usfacfLjhXKdrhBSoASIyBsAnqQOo77uNdK6h5QK0hpICU7lrCUpQSYPtjBJIkwB7wjuFf57bhDaPIQrzl78hcLgFQ+aDunqRJABgmA8tTpGuQpjp6MuIZGEmump3a8ww4UsFQKwkkAp3EeoiNxicDunIAJPEVHlqblAWlZaClbUNkNgpPcknJwcA/NOYHC66ViHLgoy3UshxI3rbWTUtJ/IITuSFETJ27SoCAUcM1UVuLaZQSp9W2nQC0loOLBUFLBCgEeooBJEQTKsAlQr690rtinW325kDduaVOsihaCnUrZQEqWUrBKS3sSUqAUTBVuc9RTHyx88hHX3NxRcpw2ZbXKEJRKUOGUyNpAwUExEQ2MQI4SVr6nN623AG3CopLbYKXCAolyPYSlIPcJkiAY1vLUKJ5e90tpIQBEpG5BAWkKTJK5UoElJwkEdJGiXCI6M7JYh5TCg88pxPkblpb8zzFJVu3+YU4KmiASNpyUJz04SXS6O07boSahJaP76HE7Wdx2qSpSTCjCkgHPUpPRQKC5VqXANxSkJCH/wDo4dkJ3Ix1CSRAxuCZx3bbm+l6pUt9aklSNiFJWHFEDYCd0dQEnviAYHfwzu6LBCMp0oLZW3x5VPQMVVatphbpaoWFvOIS0hS1OKCc7EphaiISkbj2KeGRTrTTbLhbS4pSDKSUFtQBKj83ypmT0g7o6Agyrlhzu1RyNvNzuej75WadrLza6vT9Y5TqG6ooKtHl1DCpGQ62SVGJGIAUBxDK15KEtqa/d7AXQVJgxO1PXB2AACIwD7ccMldgMJVR3B2ibDoUlgtKMlELQDGSqAPTBSIknr7gceaqpaq1uLDyVBLhg7yd0jr2OcSoAElPQmeEVe8tK1rG6Hwpzes/NBIEKAz0juCQT7cfKja1WLCiFlFQoJJAEwCJKuv6TGCep48WKa3fxAapv/Ja08tnryv/AIO07d6q/wBBb1tso/DV1Uhtp51LoHmHclCRtUspGyREq4YtIctr1r+k1BUWWyV17p9OW1d2ujlHSKWi10aVoaVUPFKf3bYU42lSyAAVYIJkMjZkZS2D+Yx8qRGenYDMZMx7jj7b71WW2iqksVD9MiuQaZ4NOKbTUtSFKaUUwCkHaSk+nCcSAeMXmegW+0WOr1FWoYpqZysqXHEBLbTe8vAdCAnqPYAdBGSDCW+WOssFc5S1yV0dUw8ULQ6Cktmckp/rp9eJFR1N95K6xttexUs092pm6e5UjrLiXsOJStsSCU7gDlJEgjj7q6/ap8S3My9XutS9fNS3RL10r1ssoQVJbRvdXAACdqETjsDkk8YsIX//2Q==" />
These are GGUF quantized versions of Envoid/Fish-8x7B.
The importance matrix was trained for 100K tokens (200 batches of 512 tokens) using 'URL'.
The IQ2_XXS and IQ2_XS versions are compatible with URL, version '147b17a' or later. The IQ3_XXS requires version 'f4d7e54' or later.
Some model files above 50GB are split into smaller files. To concatenate them, use the 'cat' command (on Windows, use PowerShell): 'cat foo-Q6_K.gguf.* > foo-Q6_K.gguf' | [] | [
"TAGS\n#gguf #not-for-all-audiences #en #license-cc-by-nc-4.0 #region-us \n"
] | [
31
] | [
"passage: TAGS\n#gguf #not-for-all-audiences #en #license-cc-by-nc-4.0 #region-us \n"
] | [
0.016246788203716278,
0.13549204170703888,
-0.009164837189018726,
0.05377594009041786,
-0.03637194260954857,
0.08497514575719833,
0.18622490763664246,
0.044540584087371826,
0.16163523495197296,
-0.032311975955963135,
0.11922280490398407,
0.009323467500507832,
-0.01138903945684433,
-0.05263134837150574,
-0.03180311992764473,
-0.05605296418070793,
0.00994391180574894,
0.04647710174322128,
0.09504728764295578,
0.02576516754925251,
0.03023965284228325,
-0.006605185568332672,
-0.03456639498472214,
-0.04243160039186478,
-0.11300323158502579,
-0.05400640144944191,
0.06143096461892128,
-0.015387783758342266,
0.04679226130247116,
0.030199656262993813,
0.06497565656900406,
0.11497696489095688,
-0.053746242076158524,
-0.1139727383852005,
0.021186845377087593,
-0.03742412105202675,
-0.18268807232379913,
0.02607908472418785,
0.021201537922024727,
0.06602587550878525,
0.11941754817962646,
0.15599796175956726,
-0.1400631070137024,
0.08214541524648666,
-0.22659002244472504,
-0.14799228310585022,
-0.07358530908823013,
0.03393694385886192,
-0.03134128451347351,
0.05797988176345825,
0.0009508179500699043,
0.06654758751392365,
-0.1856636255979538,
-0.042769093066453934,
0.04724618420004845,
-0.2923770546913147,
0.03485812991857529,
0.23824477195739746,
-0.016943074762821198,
0.06249874085187912,
-0.1455625742673874,
0.12928083539009094,
0.03909803926944733,
-0.016656288877129555,
-0.08309011161327362,
-0.06938938796520233,
0.01711396686732769,
0.057257331907749176,
-0.016498547047376633,
-0.04735462740063667,
0.25525572896003723,
0.07431049644947052,
-0.03641221672296524,
0.08863496780395508,
0.00213811662979424,
0.00041381249320693314,
-0.03121008165180683,
0.07475648820400238,
0.05940784886479378,
0.16938844323158264,
0.006971420254558325,
-0.06280998140573502,
-0.1553136259317398,
-0.0736115351319313,
-0.14339229464530945,
0.0808836817741394,
-0.04768504574894905,
0.07186280936002731,
-0.08141466975212097,
0.021306268870830536,
-0.24453949928283691,
-0.026306482031941414,
-0.08998031914234161,
-0.07236810028553009,
0.11108189076185226,
-0.00032716052373871207,
-0.03204228729009628,
0.21723346412181854,
0.1478722095489502,
0.11298615485429764,
-0.09980206191539764,
0.0035793620627373457,
-0.07274463027715683,
0.1581660658121109,
-0.007643151096999645,
-0.10923783481121063,
0.03069605492055416,
0.24982669949531555,
0.0952674150466919,
-0.08970227092504501,
0.058138228952884674,
0.027749797329306602,
-0.09969884902238846,
-0.004594655241817236,
-0.22818700969219208,
0.1401631087064743,
-0.004845944233238697,
-0.044903598725795746,
-0.08131953328847885,
0.08502565324306488,
0.16445766389369965,
0.08015657216310501,
-0.04488004371523857,
0.06130296736955643,
0.047928545624017715,
-0.0919923409819603,
-0.07931339740753174,
0.07604841142892838,
0.13343407213687897,
0.0696403905749321,
-0.13782548904418945,
-0.0035952278412878513,
0.06920387595891953,
0.10967285186052322,
0.14040957391262054,
-0.029765333980321884,
0.06031670421361923,
-0.09933698177337646,
-0.04116975888609886,
0.0736328735947609,
-0.06370171904563904,
0.0037987318355590105,
0.023848246783018112,
0.09331679344177246,
0.0650588646531105,
-0.04908900335431099,
-0.056994445621967316,
-0.0874059796333313,
-0.11422902345657349,
0.1421206146478653,
0.007626840844750404,
0.010849334299564362,
-0.15800385177135468,
-0.03575902059674263,
-0.07049747556447983,
0.03755199536681175,
0.010717585682868958,
-0.04162868112325668,
-0.13676100969314575,
0.07594430446624756,
0.0276754442602396,
-0.012725825421512127,
-0.08830959349870682,
0.014231030829250813,
-0.09163746982812881,
0.19017943739891052,
-0.045793239027261734,
-0.01993473619222641,
0.18340903520584106,
-0.11423604190349579,
-0.10721955448389053,
0.10027667135000229,
0.06687156856060028,
-0.08973851799964905,
0.036001767963171005,
0.28189152479171753,
-0.03781391680240631,
-0.14512883126735687,
-0.01139070838689804,
0.18916940689086914,
-0.10116668790578842,
-0.06621717661619186,
0.13964024186134338,
-0.1289820820093155,
-0.1830635964870453,
0.0016489095287397504,
-0.09755441546440125,
0.062219973653554916,
-0.023281503468751907,
-0.06814593076705933,
0.006571050733327866,
-0.023772697895765305,
0.09340347349643707,
0.024370111525058746,
0.017243534326553345,
-0.04068658873438835,
-0.0160082895308733,
-0.19863398373126984,
0.05843454599380493,
0.08893995732069016,
0.01853802427649498,
-0.10121898353099823,
0.07470303773880005,
-0.019544491544365883,
-0.0024703906383365393,
0.030391458421945572,
-0.05319644510746002,
0.00041352552943862975,
0.022338002920150757,
0.14115135371685028,
0.19082500040531158,
0.025533005595207214,
-0.06594184041023254,
-0.08580475300550461,
0.027244502678513527,
-0.022618617862462997,
0.0015015325043350458,
0.05851148068904877,
-0.08693584054708481,
0.12749680876731873,
0.014389325864613056,
0.05298228934407234,
-0.10495330393314362,
-0.055581774562597275,
0.26399219036102295,
-0.09151407331228256,
-0.09851115196943283,
-0.02003241516649723,
0.08520470559597015,
-0.032979920506477356,
0.07950515300035477,
0.012439305894076824,
0.12917983531951904,
0.05569044128060341,
-0.11323901265859604,
0.1196119412779808,
-0.03529460355639458,
0.17901384830474854,
0.14250867068767548,
-0.0750579982995987,
0.006286097224801779,
-0.028502395376563072,
-0.00880376622080803,
0.028940094634890556,
0.039538897573947906,
0.06763143837451935,
0.10191711783409119,
-0.0891733318567276,
-0.0037187200505286455,
-0.029521683230996132,
-0.005022362340241671,
0.04386882483959198,
-0.05210266262292862,
-0.050492897629737854,
0.02120044082403183,
0.15992696583271027,
-0.12047377973794937,
0.1697545200586319,
0.30280745029449463,
0.03194693848490715,
0.11720865964889526,
-0.04780742898583412,
0.03306391462683678,
-0.10833577811717987,
0.04919173941016197,
0.0023382927756756544,
0.1738935112953186,
-0.16574493050575256,
0.06203823164105415,
0.04560529440641403,
0.0041847312822937965,
0.058929409831762314,
-0.1703638732433319,
-0.18167713284492493,
-0.05561596155166626,
-0.08289340138435364,
-0.21012789011001587,
0.04422181099653244,
-0.13241349160671234,
0.03329070657491684,
-0.002155158668756485,
-0.031730491667985916,
0.18551620841026306,
0.005064783152192831,
-0.1050942987203598,
0.06296943873167038,
-0.14219005405902863,
-0.11822299659252167,
-0.05616883933544159,
-0.036955587565898895,
-0.051034487783908844,
0.07793280482292175,
0.08645354211330414,
-0.09253835678100586,
-0.05170764774084091,
0.043087586760520935,
-0.14871414005756378,
-0.12305939197540283,
-0.02195020392537117,
0.01910811848938465,
0.031706683337688446,
0.029127212241292,
-0.07543016970157623,
-0.0221877321600914,
-0.06297050416469574,
-0.16148817539215088,
0.09317095577716827,
-0.01568920537829399,
0.11479800194501877,
0.09923379123210907,
0.0987614095211029,
0.028238367289304733,
-0.07073622196912766,
0.13325132429599762,
-0.12630626559257507,
-0.12406283617019653,
0.058522019535303116,
0.009258103556931019,
0.002418736694380641,
0.09834107011556625,
0.14023801684379578,
-0.10309092700481415,
-0.053174592554569244,
-0.019868455827236176,
-0.11499903351068497,
-0.15000657737255096,
-0.0212460495531559,
-0.08621197938919067,
0.10452929884195328,
-0.014626028016209602,
0.108127661049366,
0.1680736094713211,
0.06510138511657715,
-0.019542943686246872,
-0.03505618870258331,
0.008783702738583088,
-0.001114679267629981,
0.2193504273891449,
-0.01513329241424799,
-0.053028404712677,
-0.08484329283237457,
0.07733716815710068,
0.19634085893630981,
0.07831422984600067,
0.06792277097702026,
0.1576288491487503,
0.1805621087551117,
0.1782720535993576,
0.11119034886360168,
0.15759962797164917,
-0.055215295404195786,
0.027075516059994698,
-0.045365381985902786,
-0.022560350596904755,
-0.03978462889790535,
0.02753949910402298,
-0.020112482830882072,
0.0688886046409607,
-0.24547073245048523,
0.03931097686290741,
-0.29494938254356384,
0.06829935312271118,
-0.03225485980510712,
0.06517491489648819,
0.07353466004133224,
0.060000281780958176,
0.048781413584947586,
0.14970535039901733,
0.011914154514670372,
0.1213596910238266,
0.031140895560383797,
-0.03015762008726597,
0.06119689345359802,
0.0114122424274683,
0.025069942697882652,
-0.02323002554476261,
0.03203140199184418,
-0.07633650302886963,
-0.12435972690582275,
-0.0004895993624813855,
0.10940637439489365,
-0.27121293544769287,
0.15562930703163147,
0.06381604820489883,
-0.0061836810782551765,
0.0020366546232253313,
-0.09376704692840576,
0.06044842675328255,
0.1654088944196701,
0.2204250544309616,
0.06734366714954376,
-0.07736003398895264,
-0.09365478157997131,
-0.04361995309591293,
0.04092930629849434,
0.07046680897474289,
-0.04278676584362984,
-0.15852299332618713,
0.009498324245214462,
0.07419823855161667,
0.057118721306324005,
0.012277870438992977,
-0.14515864849090576,
-0.04723236337304115,
0.1158524677157402,
0.15874594449996948,
0.015249882824718952,
-0.040901150554418564,
0.0518496073782444,
-0.11712323874235153,
0.10835839062929153,
-0.23813332617282867,
-0.002813353668898344,
-0.05730518698692322,
-0.10295043885707855,
0.007205932401120663,
0.004740116652101278,
0.00041848537512123585,
-0.06293895840644836,
-0.08881822228431702,
-0.11646213382482529,
-0.1407783031463623,
0.10875598341226578,
-0.06962963938713074,
-0.08320632576942444,
-0.004013653378933668,
0.1325872391462326,
-0.030690578743815422,
0.07257160544395447,
0.0025348144117742777,
0.022466307505965233,
0.03971865028142929,
-0.19437162578105927,
0.13660740852355957,
-0.17536218464374542,
0.005961649119853973,
0.0067777070216834545,
0.04739956185221672,
0.02838609740138054,
0.04867669939994812,
-0.09052669256925583,
0.12496798485517502,
0.4782823920249939,
-0.03939371556043625,
0.16605965793132782,
0.3588009476661682,
-0.07261940836906433,
-0.1884508728981018,
-0.10127781331539154,
-0.2578360140323639,
-0.19015011191368103,
0.030621996149420738,
-0.19061000645160675,
-0.0443388856947422,
0.19855503737926483,
-0.12013384699821472,
0.3134317696094513,
-0.11768538504838943,
-0.02550463192164898,
0.09504447877407074,
-0.004909860901534557,
0.4131777882575989,
-0.14645053446292877,
-0.11298084259033203,
0.06810154765844345,
-0.12753398716449738,
0.12883973121643066,
-0.04517942667007446,
0.11323771625757217,
0.008069314062595367,
-0.0640907734632492,
-0.013905475847423077,
-0.02292492985725403,
0.18920579552650452,
-0.01315402239561081,
0.0687018558382988,
-0.05283815413713455,
-0.1155165284872055,
0.2522318661212921,
0.048500172793865204,
-0.08934351801872253,
-0.10982055217027664,
-0.03512589633464813,
-0.008965296670794487,
-0.008567769080400467,
-0.0792737677693367,
0.11968431621789932,
-0.017398132011294365,
-0.11342033743858337,
-0.16073933243751526,
0.0026327341329306364,
-0.11191035807132721,
0.0018831404158845544,
0.15913255512714386,
-0.11465153098106384,
0.06202888488769531,
0.048004623502492905,
0.04195107892155647,
-0.09633322060108185,
-0.08101285994052887,
-0.07072003930807114,
-0.08107288181781769,
0.0697515681385994,
-0.12376680225133896,
-0.003888534614816308,
0.04011455178260803,
0.030045993626117706,
0.06056973710656166,
0.08774790912866592,
-0.033042021095752716,
0.11187239736318588,
0.13590162992477417,
-0.14554579555988312,
-0.09456014633178711,
-0.012857847847044468,
-0.04548652097582817,
0.1261676549911499,
0.013028264977037907,
0.018929490819573402,
0.05200450122356415,
0.032316114753484726,
0.012876993976533413,
0.019070299342274666,
-0.10879684239625931,
-0.08426334708929062,
0.021621573716402054,
-0.028752408921718597,
-0.12577183544635773,
0.09731847047805786,
0.042221736162900925,
-0.028270432725548744,
-0.06544362753629684,
0.011456262320280075,
-0.06898531317710876,
-0.10381888598203659,
-0.2355221062898636,
-0.06528016179800034,
-0.13940685987472534,
-0.0364765003323555,
0.018614118918776512,
-0.06563809514045715,
-0.046335119754076004,
0.046631474047899246,
0.0034382652956992388,
0.1479358971118927,
0.059017814695835114,
0.06101129576563835,
0.05769525095820427,
-0.07452522218227386,
-0.24193888902664185,
0.006374972406774759,
-0.07618175446987152,
-0.08254856616258621,
0.048755016177892685,
0.05963541939854622,
-0.07254138588905334,
-0.06688497215509415,
-0.12017348408699036,
0.04092836007475853,
-0.009355228394269943,
0.03239612281322479,
-0.09170528501272202,
0.024414269253611565,
0.009728961624205112,
0.0014745095977559686,
-0.01695597730576992,
-0.012431890703737736,
-0.07604123651981354,
0.03290138021111488,
0.08195780217647552,
0.05278319865465164,
-0.06024280935525894,
-0.02379356510937214,
0.04613485187292099,
0.06472742557525635,
0.17070315778255463,
0.08891843259334564,
0.022910894826054573,
0.08074129372835159,
-0.2884860932826996,
0.007439220789819956,
0.07515506446361542,
-0.04662906378507614,
-0.08988194167613983,
0.08446824550628662,
0.009833188727498055,
0.009455026127398014,
-0.05585147812962532,
0.06998903304338455,
-0.03857304900884628,
-0.12644438445568085,
-0.08132705092430115,
0.04857433959841728,
-0.06526488810777664,
-0.011815403588116169,
-0.0938950851559639,
0.1698630303144455,
0.03688252344727516,
0.07725980132818222,
0.02276080660521984,
-0.038043513894081116,
0.007442230824381113,
-0.02938603051006794,
-0.010271962732076645,
-0.09282509237527847,
-0.07528038322925568,
-0.00003826682223007083,
-0.07752823829650879,
-0.015726037323474884,
0.30453842878341675,
-0.05798837170004845,
-0.2054697871208191,
0.040027059614658356,
0.10825991630554199,
0.06543620675802231,
-0.02851417288184166,
0.29256513714790344,
0.01714840903878212,
-0.04387294873595238,
-0.14764253795146942,
0.06101759523153305,
-0.04886339604854584,
-0.29180604219436646,
0.05447065457701683,
-0.051135219633579254,
-0.0774797797203064,
-0.06593026965856552,
0.0389639176428318,
-0.11032279580831528,
-0.015670083463191986,
0.0898561030626297,
0.034418392926454544,
-0.014318641275167465,
-0.022647347301244736,
-0.07653820514678955,
0.17619255185127258,
0.016608063131570816,
0.020682865753769875,
-0.05053301528096199,
-0.021955322474241257,
-0.1225261390209198,
-0.12362121790647507,
0.034194573760032654,
-0.10952982306480408,
0.1014748141169548,
-0.03924209251999855,
0.05565211549401283,
0.15039785206317902,
0.015197248198091984,
-0.10205058008432388,
0.018404951319098473,
-0.07041021436452866,
-0.04687085002660751,
0.03406192734837532,
-0.026154568418860435,
-0.03523268178105354,
-0.1312113106250763,
-0.07922740280628204,
0.041400205343961716,
-0.14027126133441925,
0.029425567016005516,
-0.0041165477596223354,
0.054974522441625595,
-0.05434655770659447,
-0.13539637625217438,
-0.04604210704565048,
-0.04977991059422493,
0.10907735675573349,
0.030881909653544426,
0.20561857521533966,
-0.024406809359788895,
0.005916055757552385,
0.09200219810009003,
0.07417040318250656,
0.013892737217247486,
0.006672530435025692,
-0.008875206112861633,
0.07872273772954941,
-0.1406538039445877,
0.07900360971689224,
-0.03908838704228401,
-0.01774238608777523,
0.12678281962871552,
0.18000665307044983,
0.15530450642108917,
-0.1072651669383049,
0.03553967922925949,
-0.0608183853328228,
0.01951795630156994,
0.1060711070895195,
0.1332281082868576,
-0.03577513247728348,
0.22374117374420166,
-0.10499130934476852,
-0.026278220117092133,
0.026347056031227112,
0.0974307730793953,
-0.045841850340366364,
0.053453464061021805,
0.038206685334444046,
0.006401240359991789,
-0.10315156728029251,
0.09193481504917145,
-0.12522770464420319,
0.14718683063983917,
0.10668721795082092,
-0.003033002372831106,
0.1052246242761612,
-0.02886725775897503,
-0.06340668350458145,
0.03156779706478119,
0.05253202095627785,
-0.0989774540066719,
-0.09064667671918869,
-0.14604823291301727,
0.02761601097881794,
-0.2690942585468292,
-0.05994480848312378,
0.08353559672832489,
0.1702696830034256,
0.24261093139648438,
-0.021057721227407455,
0.12495321780443192,
0.006303875707089901,
0.006804750766605139,
-0.08928600698709488,
0.18369457125663757,
0.027517031878232956,
-0.1264003962278366,
-0.12677280604839325,
-0.08121442794799805,
-0.004653862677514553,
-0.008225004188716412,
0.03993155434727669,
0.08930313587188721,
0.0573541633784771,
0.18842309713363647,
-0.05781339854001999,
0.008884484879672527,
0.02177247405052185,
-0.1126050129532814,
0.033789582550525665,
-0.1409948170185089,
0.0010304995812475681,
-0.12986010313034058,
-0.03807957097887993,
0.00403286749497056,
0.062945157289505,
-0.10370957106351852,
-0.018254276365041733,
0.0824250653386116,
0.05342443287372589,
0.2544509172439575,
-0.009574304334819317,
-0.08437587320804596,
0.001239005126990378,
-0.011702688410878181,
0.15217290818691254,
-0.0732964426279068,
0.005498813930898905,
0.08726103603839874,
-0.048726364970207214,
-0.01198987290263176,
-0.2607085406780243,
0.01859160140156746,
-0.07946765422821045,
-0.02340335212647915,
-0.06048467382788658
] |
null | null | stable-baselines3 |
# **PPO** Agent playing **LunarLander-v2**
This is a trained model of a **PPO** agent playing **LunarLander-v2**
using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3).
## Usage (with Stable-baselines3)
TODO: Add your code
```python
from stable_baselines3 import ...
from huggingface_sb3 import load_from_hub
...
```
| {"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "225.40 +/- 73.06", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | smcaleese/ppo-LunarLander-v2 | [
"stable-baselines3",
"LunarLander-v2",
"deep-reinforcement-learning",
"reinforcement-learning",
"model-index",
"region:us"
] | 2024-02-08T18:24:13+00:00 | [] | [] | TAGS
#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
|
# PPO Agent playing LunarLander-v2
This is a trained model of a PPO agent playing LunarLander-v2
using the stable-baselines3 library.
## Usage (with Stable-baselines3)
TODO: Add your code
| [
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
"TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n",
"# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.",
"## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
39,
41,
17
] | [
"passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code"
] | [
0.03942384943366051,
0.04900386184453964,
-0.005304091144353151,
0.026427261531352997,
0.107408307492733,
-0.026511888951063156,
0.11188238859176636,
0.0814051404595375,
0.10722193866968155,
0.04762078449130058,
0.08338645845651627,
0.06030960753560066,
0.05080918222665787,
0.2571701407432556,
0.04754156619310379,
-0.22987541556358337,
0.036159250885248184,
-0.04869936779141426,
0.12395193427801132,
0.07178173214197159,
-0.0038484656251966953,
-0.06485428661108017,
0.020415637642145157,
-0.013290755450725555,
0.05367108806967735,
0.04282612353563309,
-0.01716216839849949,
-0.08207534998655319,
0.07169748842716217,
-0.06345846503973007,
0.06986866891384125,
0.07677983492612839,
0.13218913972377777,
-0.17832116782665253,
0.029566360637545586,
0.02571309357881546,
-0.07189024239778519,
0.01342033501714468,
0.008019951172173023,
0.05120139941573143,
0.17303818464279175,
0.019879888743162155,
0.07844575494527817,
-0.0025605305563658476,
-0.15412317216396332,
-0.018950799480080605,
0.0436202734708786,
0.12546207010746002,
0.08808347582817078,
0.04605821147561073,
0.01970590092241764,
0.17503218352794647,
-0.054352790117263794,
-0.028833400458097458,
0.21759237349033356,
-0.2881564497947693,
-0.031460098922252655,
0.321048766374588,
0.06997483223676682,
0.09725230932235718,
-0.07540661096572876,
-0.03619609400629997,
0.007783263456076384,
-0.013137873262166977,
-0.028666524216532707,
-0.07447073608636856,
0.17313385009765625,
0.05152064561843872,
-0.05057951435446739,
-0.09541505575180054,
0.16948209702968597,
0.006921638268977404,
0.0018855923553928733,
-0.019282981753349304,
0.009060598909854889,
0.07402525842189789,
-0.016097044572234154,
-0.07255112379789352,
0.057438433170318604,
0.05330665782094002,
0.019649166613817215,
-0.1435653269290924,
-0.10762494057416916,
-0.022740179672837257,
-0.008012006990611553,
0.17786912620067596,
-0.009255532175302505,
0.042902372777462006,
0.003065188182517886,
0.10384012013673782,
-0.12480384111404419,
-0.03354184702038765,
-0.0454259067773819,
-0.07565800100564957,
-0.0223417766392231,
-0.02058211714029312,
-0.03580251708626747,
0.07184842973947525,
0.11971849203109741,
0.027368178591132164,
0.09350208193063736,
0.047715865075588226,
-0.03206788748502731,
0.06343851238489151,
0.05555703118443489,
0.14222665131092072,
0.05807621404528618,
0.012854371219873428,
0.13179877400398254,
0.055213116109371185,
0.033023182302713394,
-0.0613492950797081,
-0.18252409994602203,
0.07489913702011108,
-0.07031869143247604,
0.007941240444779396,
0.12051256000995636,
-0.04480670019984245,
-0.1183447614312172,
-0.037500523030757904,
-0.017392054200172424,
-0.06224250793457031,
-0.025395862758159637,
0.0547584593296051,
-0.02883218228816986,
-0.03973718360066414,
0.0011496668448671699,
0.09384800493717194,
0.00953749567270279,
-0.1752052903175354,
0.03303423151373863,
-0.025042934343218803,
-0.10782608389854431,
0.009975161403417587,
0.0022444494534283876,
0.03394931182265282,
0.04408763721585274,
-0.11822668462991714,
-0.30899152159690857,
-0.07652641832828522,
0.05490870401263237,
-0.06516939401626587,
-0.18425025045871735,
-0.13193942606449127,
0.02454492449760437,
-0.09037084132432938,
-0.044885024428367615,
-0.12759265303611755,
-0.028549788519740105,
0.01743689924478531,
0.011519349180161953,
0.10758619755506516,
-0.0106219332665205,
-0.012188062071800232,
-0.1571401208639145,
0.008273907005786896,
-0.20951123535633087,
0.0890483483672142,
-0.019150104373693466,
0.037884220480918884,
-0.032381169497966766,
-0.07404014468193054,
0.030707746744155884,
0.052499737590551376,
-0.01474119070917368,
0.13510210812091827,
-0.15592676401138306,
-0.03691192343831062,
-0.007996266707777977,
-0.13611900806427002,
-0.04786273464560509,
-0.10358831286430359,
-0.04357128217816353,
0.13354332745075226,
0.018664736300706863,
0.15356586873531342,
-0.08709818124771118,
-0.0722038671374321,
0.20489206910133362,
-0.010411538183689117,
-0.12820468842983246,
-0.076752208173275,
0.10165707021951675,
0.021510310471057892,
-0.056606587022542953,
-0.02523270808160305,
-0.1839766949415207,
-0.0152357779443264,
-0.04550420492887497,
-0.047039128839969635,
0.01796751655638218,
-0.010888241231441498,
0.13837894797325134,
0.08494598418474197,
0.05018039792776108,
-0.06086122244596481,
-0.006730288732796907,
0.10779471695423126,
0.08823856711387634,
0.008680110797286034,
0.023406028747558594,
-0.05774238705635071,
0.09552932530641556,
-0.04003755748271942,
-0.0142367510125041,
-0.08283266425132751,
-0.036246106028556824,
-0.026256313547492027,
0.17507147789001465,
0.09440762549638748,
0.2257927656173706,
0.09567736834287643,
0.039160262793302536,
0.031270865350961685,
-0.13181598484516144,
-0.1425403207540512,
-0.0017254541162401438,
0.09020978957414627,
-0.14270411431789398,
-0.04119925573468208,
-0.08974775671958923,
-0.17768175899982452,
-0.12202505767345428,
0.0006432619411498308,
-0.17960017919540405,
0.06390921026468277,
0.05408334732055664,
-0.035177867859601974,
0.03272094577550888,
0.13032332062721252,
-0.011533179320394993,
-0.03967514634132385,
0.0831870287656784,
0.0379033200442791,
-0.041234664618968964,
-0.021742934361100197,
0.11885567009449005,
0.15673065185546875,
0.13124459981918335,
-0.03511447086930275,
0.004914294462651014,
0.07076404243707657,
-0.02309088408946991,
0.06539414077997208,
0.0558244064450264,
0.20973342657089233,
0.188301220536232,
0.038996949791908264,
0.008822928182780743,
-0.07048165798187256,
0.0855446457862854,
-0.0742373839020729,
-0.14302679896354675,
-0.05579735338687897,
0.08729292452335358,
0.016605578362941742,
0.023469142615795135,
0.08711627870798111,
0.024545932188630104,
0.09132762253284454,
0.15968108177185059,
0.01990218088030815,
-0.09659269452095032,
-0.050218869000673294,
0.01175848301500082,
0.027713103219866753,
0.04794301092624664,
-0.04514073207974434,
-0.00937939714640379,
0.017020760104060173,
-0.10303554683923721,
0.031789086759090424,
-0.1413339376449585,
-0.1358717679977417,
0.044326696544885635,
0.003906996920704842,
0.010907664895057678,
0.02786896750330925,
-0.0038291432429105043,
0.019039705395698547,
0.04351753741502762,
-0.06975466758012772,
0.047416772693395615,
-0.024745507165789604,
-0.020031947642564774,
0.03340689837932587,
-0.057257164269685745,
-0.205775648355484,
-0.17696654796600342,
0.00013708483311347663,
-0.09910997003316879,
0.10194740444421768,
0.018308809027075768,
-0.12373185902833939,
0.047737859189510345,
-0.05822649225592613,
0.027574289590120316,
-0.01875593699514866,
-0.049130141735076904,
0.10507171601057053,
0.1525275856256485,
-0.016146350651979446,
0.018018173053860664,
-0.04865182936191559,
-0.10157987475395203,
-0.19632206857204437,
0.0691583976149559,
0.04680244252085686,
0.014610917307436466,
0.10669491440057755,
0.018072687089443207,
0.02367905154824257,
-0.007674071006476879,
-0.016521066427230835,
-0.011659215204417706,
-0.08781040459871292,
0.31909599900245667,
0.04510033503174782,
-0.025173069909214973,
0.02041010931134224,
-0.0043001663871109486,
-0.028083480894565582,
0.03263787180185318,
-0.0985708013176918,
-0.07548979669809341,
-0.08774089068174362,
-0.04367410019040108,
-0.09784720093011856,
0.053299110382795334,
0.05916472524404526,
0.003188040340319276,
-0.07727594673633575,
0.04221395403146744,
0.11369874328374863,
-0.0923808291554451,
-0.07137343287467957,
0.07477962225675583,
0.0972946360707283,
-0.07331304252147675,
0.00012658814375754446,
0.00874367356300354,
0.023951783776283264,
0.037102166563272476,
0.06778035312891006,
-0.03966575115919113,
0.08589404821395874,
-0.19917890429496765,
0.0372927263379097,
0.106058269739151,
0.023754918947815895,
0.0638108178973198,
0.07643651217222214,
-0.1058402881026268,
-0.008500572293996811,
-0.032518330961465836,
-0.21341575682163239,
0.1668180525302887,
0.1355515867471695,
0.06788124144077301,
-0.025637222453951836,
-0.00461410591378808,
-0.0649740919470787,
0.05773647129535675,
0.02723747305572033,
-0.14758841693401337,
0.004883295856416225,
0.06064270809292793,
0.026899009943008423,
0.01614922471344471,
0.07971042394638062,
0.014697225764393806,
-0.1801026314496994,
-0.014406266622245312,
0.10730406641960144,
0.002390873385593295,
0.0053148469887673855,
-0.03175045922398567,
-0.1755964607000351,
0.0751047357916832,
0.004285442177206278,
0.07233936339616776,
-0.1676585078239441,
0.14297930896282196,
-0.10089799761772156,
0.07726949453353882,
-0.004285062663257122,
-0.021311495453119278,
0.02507244050502777,
-0.0541163794696331,
0.15163759887218475,
0.01058570109307766,
-0.021810131147503853,
-0.1200498715043068,
-0.1717042326927185,
-0.019227758049964905,
-0.11788936704397202,
-0.11679866164922714,
0.050424277782440186,
0.062185097485780716,
0.04923136904835701,
-0.061147067695856094,
0.1518532931804657,
-0.047422297298908234,
0.060713399201631546,
-0.06893875449895859,
-0.06755045056343079,
0.03764858841896057,
-0.12588608264923096,
-0.08176055550575256,
0.05573027580976486,
0.19166934490203857,
0.15833087265491486,
-0.02816431224346161,
-0.03472423925995827,
-0.047419581562280655,
-0.006212298292666674,
-0.007802055217325687,
0.0275666993111372,
0.023223137483000755,
0.07315318286418915,
-0.07681374251842499,
-0.11649256944656372,
0.033787861466407776,
-0.06713802367448807,
-0.055589709430933,
-0.015439179725944996,
0.1513158082962036,
0.04671623185276985,
0.07720734924077988,
-0.018946662545204163,
0.03887668624520302,
-0.001724981120787561,
-0.056474871933460236,
0.16197094321250916,
0.03885216265916824,
-0.05193585529923439,
0.06837689876556396,
0.053174007683992386,
0.043745119124650955,
0.03011113777756691,
-0.026783017441630363,
0.206032395362854,
0.1980147808790207,
0.014206883497536182,
0.2175983190536499,
0.03177616000175476,
-0.03772832080721855,
-0.1300560086965561,
-0.065880686044693,
-0.006372632458806038,
0.03559038043022156,
0.08070417493581772,
-0.18207235634326935,
-0.015011128038167953,
-0.05689644813537598,
-0.034518610686063766,
-0.15059494972229004,
-0.28553900122642517,
-0.05957856774330139,
0.20075850188732147,
0.14706264436244965,
0.27519428730010986,
-0.10432573407888412,
0.035197313874959946,
0.02663275972008705,
-0.04912831634283066,
-0.006501141935586929,
0.00018665487004909664,
0.10268618166446686,
-0.15421873331069946,
0.1176437959074974,
0.08486983180046082,
-0.019002694636583328,
0.01058861706405878,
-0.1619086116552353,
0.00936629343777895,
-0.12191236019134521,
0.05354422330856323,
0.1400289237499237,
-0.048128653317689896,
-0.054873593151569366,
0.14033560454845428,
-0.024562934413552284,
-0.22685599327087402,
-0.04648222774267197,
-0.043600670993328094,
-0.010640020482242107,
0.026607351377606392,
-0.1013401448726654,
0.04101909324526787,
0.1330099105834961,
0.009380043484270573,
0.1147187277674675,
0.11749245226383209,
-0.052566803991794586,
0.10792597383260727,
0.2257719188928604,
-0.018785694614052773,
0.04689010605216026,
-0.12743118405342102,
-0.0012336712097749114,
-0.028270328417420387,
0.013657891191542149,
-0.09504974633455276,
-0.09938385337591171,
0.02366873063147068,
0.02872389927506447,
0.009118586778640747,
0.0921793207526207,
-0.029922157526016235,
0.0759170651435852,
0.06817561388015747,
-0.13014446198940277,
-0.16288450360298157,
0.015828335657715797,
-0.007344507612287998,
0.08354310691356659,
0.00027861111448146403,
0.08878035843372345,
-0.11932205408811569,
-0.018093237653374672,
-0.03153328225016594,
-0.03319635987281799,
-0.130486860871315,
-0.07138993591070175,
0.06156524643301964,
0.028095467016100883,
-0.06602972000837326,
0.1398407518863678,
0.026440169662237167,
0.15942534804344177,
0.049197953194379807,
0.012499804608523846,
0.07227300107479095,
-0.05345509201288223,
0.1283530443906784,
0.13818155229091644,
-0.00868943240493536,
-0.05460423603653908,
-0.1013643890619278,
-0.10236792266368866,
0.08925779908895493,
-0.05773641914129257,
0.07476430386304855,
-0.14885357022285461,
-0.06675903499126434,
0.015772046521306038,
0.016141414642333984,
-0.09562095999717712,
0.02571965754032135,
-0.01625603251159191,
-0.18119946122169495,
0.056570518761873245,
-0.048285093158483505,
0.0440407395362854,
-0.06347788125276566,
-0.1110161691904068,
-0.17226378619670868,
0.06091433763504028,
0.08593481779098511,
-0.053876690566539764,
-0.12229149043560028,
0.011023230850696564,
-0.00012518465518951416,
-0.06341652572154999,
-0.05023367330431938,
0.09722746908664703,
-0.11020902544260025,
0.031452205032110214,
-0.012567701749503613,
0.08853451162576675,
-0.03510405123233795,
-0.011538895778357983,
0.044220831245183945,
-0.08039166033267975,
-0.009481523185968399,
0.03534642979502678,
-0.026372017338871956,
-0.04127239063382149,
-0.2689029574394226,
0.0036654395516961813,
0.0341104120016098,
0.02497158572077751,
0.07856601476669312,
0.011906822212040424,
0.021174922585487366,
0.03993808850646019,
-0.15396519005298615,
-0.013395369984209538,
0.14574195444583893,
-0.07689505815505981,
-0.022186370566487312,
0.05703273415565491,
-0.09054436534643173,
0.013882770203053951,
-0.030287226662039757,
0.1345842480659485,
0.023923413828015327,
0.06404478847980499,
-0.0851147472858429,
0.10106813907623291,
-0.1451139897108078,
-0.04998219385743141,
-0.01244612317532301,
0.09761348366737366,
0.07019034773111343,
-0.10272270441055298,
0.014697125181555748,
0.04210108891129494,
0.19416837394237518,
0.016384804621338844,
-0.0356343574821949,
-0.03396720811724663,
0.004015897400677204,
0.22076453268527985,
0.03044266067445278,
0.10457023978233337,
0.07281364500522614,
-0.026583973318338394,
0.12624378502368927,
0.09929762035608292,
0.11280370503664017,
-0.055645186454057693,
0.13904185593128204,
0.04667386785149574,
0.038641396909952164,
0.0614289753139019,
0.06836545467376709,
0.09098632633686066,
-0.0008288522367365658,
0.1138714924454689,
0.013811973854899406,
-0.02422109805047512,
-0.021335409954190254,
0.17759373784065247,
0.10501719266176224,
-0.14769648015499115,
0.029047364369034767,
-0.01258957851678133,
0.039933037012815475,
-0.014194529503583908,
-0.15634691715240479,
-0.07240267097949982,
-0.3315149247646332,
0.1226184144616127,
-0.07119352370500565,
0.019930170848965645,
0.007913772016763687,
-0.037425633519887924,
-0.03296699747443199,
-0.04477746784687042,
0.13151589035987854,
-0.013641550205647945,
-0.006079165264964104,
-0.04815853759646416,
-0.015360191464424133,
-0.11607866734266281,
-0.11200575530529022,
-0.013207737356424332,
-0.13671602308750153,
-0.010119039565324783,
0.05595948174595833,
0.003977729007601738,
0.01821410097181797,
-0.03142618387937546,
0.0024383175186812878,
0.06541839241981506,
-0.05751744285225868,
0.056182678788900375,
0.12097269296646118,
0.08766137808561325,
-0.1058853268623352,
0.031048951670527458,
0.2011747509241104,
0.04359564557671547,
-0.12483977526426315,
0.01449228823184967,
0.1819491684436798,
0.004885740112513304,
0.017068125307559967,
-0.006097703706473112,
-0.0540788508951664,
-0.07554277032613754,
0.1251034289598465,
0.08296554535627365,
-0.09985227137804031,
0.015833314508199692,
-0.0726347416639328,
-0.01594804972410202,
-0.06374675035476685,
0.10130585730075836,
0.09538925439119339,
0.04440245032310486,
-0.10621760785579681,
-0.08487539738416672,
-0.10891728103160858,
0.040588874369859695,
-0.08629853278398514,
-0.07311757653951645,
0.09629398584365845,
-0.07057105004787445,
-0.07029950618743896,
0.025521177798509598,
-0.17978744208812714,
-0.009467960335314274,
0.1711762249469757,
-0.24654000997543335,
-0.0916430801153183,
-0.10857923328876495,
0.14477859437465668,
0.016497576609253883,
0.1013975441455841,
-0.006207061931490898,
-0.007889035157859325,
-0.20577777922153473,
0.024890204891562462,
-0.05293011665344238,
-0.02073732763528824,
0.07814782857894897,
-0.09476397186517715,
0.22629831731319427,
-0.08276885002851486,
0.020940175279974937,
0.012659613974392414,
0.0870661810040474,
-0.030675338581204414,
0.09283176809549332,
-0.03660329803824425,
-0.12576518952846527,
-0.03620953485369682,
0.03001813031733036,
0.013904244638979435,
0.10071761906147003,
0.09772487729787827,
-0.03414725139737129,
0.03389119729399681,
0.09747414290904999,
0.04172342270612717,
-0.023843804374337196,
0.0360250361263752,
-0.17077107727527618,
0.02182629331946373,
-0.018498148769140244,
-0.06935930997133255,
0.03687669709324837,
-0.06603235751390457,
0.1639697551727295,
0.04022442549467087,
0.0670473501086235,
-0.036152735352516174,
0.0073931049555540085,
-0.014454689808189869,
-0.013775371946394444,
-0.026180334389209747,
-0.17259705066680908,
-0.10422050207853317,
-0.1347656100988388,
-0.012701659463346004,
-0.034971047192811966,
0.04591470584273338,
0.023234914988279343,
-0.0003200018545612693,
-0.014577031135559082,
-0.12090865522623062,
0.04360328987240791,
0.11146783083677292,
-0.04631396010518074,
-0.026193076744675636
] |
null | null | transformers |
# Model Trained Using AutoTrain
- Problem type: Text Classification
## Validation Metrics
loss: 0.4244471490383148
f1: 0.0
precision: 0.0
recall: 0.0
auc: 0.7067708333333333
accuracy: 0.8275862068965517
| {"tags": ["autotrain", "text-classification"], "datasets": ["autotrain-wewh0-x81n8/autotrain-data"], "widget": [{"text": "I love AutoTrain"}]} | text-classification | fecia/admissionBERT | [
"transformers",
"safetensors",
"roberta",
"text-classification",
"autotrain",
"dataset:autotrain-wewh0-x81n8/autotrain-data",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] | 2024-02-08T18:28:04+00:00 | [] | [] | TAGS
#transformers #safetensors #roberta #text-classification #autotrain #dataset-autotrain-wewh0-x81n8/autotrain-data #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Model Trained Using AutoTrain
- Problem type: Text Classification
## Validation Metrics
loss: 0.4244471490383148
f1: 0.0
precision: 0.0
recall: 0.0
auc: 0.7067708333333333
accuracy: 0.8275862068965517
| [
"# Model Trained Using AutoTrain\n\n- Problem type: Text Classification",
"## Validation Metrics\nloss: 0.4244471490383148\n\nf1: 0.0\n\nprecision: 0.0\n\nrecall: 0.0\n\nauc: 0.7067708333333333\n\naccuracy: 0.8275862068965517"
] | [
"TAGS\n#transformers #safetensors #roberta #text-classification #autotrain #dataset-autotrain-wewh0-x81n8/autotrain-data #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Model Trained Using AutoTrain\n\n- Problem type: Text Classification",
"## Validation Metrics\nloss: 0.4244471490383148\n\nf1: 0.0\n\nprecision: 0.0\n\nrecall: 0.0\n\nauc: 0.7067708333333333\n\naccuracy: 0.8275862068965517"
] | [
68,
16,
50
] | [
"passage: TAGS\n#transformers #safetensors #roberta #text-classification #autotrain #dataset-autotrain-wewh0-x81n8/autotrain-data #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Model Trained Using AutoTrain\n\n- Problem type: Text Classification## Validation Metrics\nloss: 0.4244471490383148\n\nf1: 0.0\n\nprecision: 0.0\n\nrecall: 0.0\n\nauc: 0.7067708333333333\n\naccuracy: 0.8275862068965517"
] | [
-0.11592216044664383,
0.09330125153064728,
0.0011471929028630257,
0.08775173127651215,
0.13442178070545197,
0.004703921265900135,
0.07163295149803162,
0.10951437801122665,
-0.05404298007488251,
0.06440943479537964,
0.19271282851696014,
0.0916927307844162,
0.020448241382837296,
0.1793353259563446,
-0.16781127452850342,
-0.17790716886520386,
0.05271759256720543,
-0.009525765664875507,
0.10500781983137131,
0.11728972941637039,
0.09369052201509476,
-0.10540691018104553,
0.10533355921506882,
-0.007435874082148075,
-0.23297475278377533,
0.020976059138774872,
0.11773740500211716,
-0.1406164914369583,
0.1086789071559906,
0.06829383969306946,
0.13207240402698517,
0.026968181133270264,
0.10514272749423981,
-0.07699859887361526,
-0.012034042738378048,
0.020905476063489914,
-0.010662543587386608,
0.07336726039648056,
0.0647086575627327,
-0.05282662436366081,
-0.034881412982940674,
0.0020369214471429586,
0.06326121836900711,
0.04426942393183708,
-0.12164893746376038,
-0.02474004589021206,
-0.07328201830387115,
0.058490052819252014,
0.09034614264965057,
0.10988080501556396,
-0.028220556676387787,
0.21305613219738007,
-0.08770738542079926,
0.09566647559404373,
0.031655505299568176,
-0.2928547263145447,
-0.035423003137111664,
0.1347864270210266,
0.029751859605312347,
-0.037309467792510986,
-0.10288581252098083,
0.08012840896844864,
0.14794723689556122,
0.00478071253746748,
0.03294037654995918,
-0.01848207786679268,
-0.11399898678064346,
0.04145170748233795,
-0.09390603005886078,
-0.0024065186735242605,
0.25316324830055237,
-0.012416353449225426,
-0.04977601766586304,
-0.06904719769954681,
-0.06325772404670715,
-0.13064990937709808,
-0.019353553652763367,
-0.0699576810002327,
-0.04715151712298393,
-0.020486980676651,
-0.050102364271879196,
0.08312331140041351,
-0.10187720507383347,
-0.062628373503685,
-0.1721242070198059,
0.10282373428344727,
-0.01950713060796261,
0.042270079255104065,
-0.0584387481212616,
0.06569942831993103,
-0.05841606482863426,
-0.10534393787384033,
0.03928713500499725,
-0.07000728696584702,
-0.06519833952188492,
-0.07128055393695831,
-0.03721078857779503,
-0.059857044368982315,
0.0209905244410038,
0.14682069420814514,
0.05582364276051521,
0.03719474747776985,
-0.04200117662549019,
0.007984702475368977,
-0.012724536471068859,
0.21878662705421448,
-0.08652486652135849,
-0.07484076172113419,
0.031252309679985046,
-0.0033660922199487686,
0.03399277105927467,
-0.05192197486758232,
-0.10554184764623642,
-0.08983081579208374,
0.18677304685115814,
0.04374435916543007,
-0.0046240463852882385,
0.06604865193367004,
-0.061786629259586334,
0.030369484797120094,
0.0018751376774162054,
-0.056903913617134094,
0.04051322862505913,
-0.04787468910217285,
-0.07454626262187958,
-0.059677109122276306,
0.05191930755972862,
0.06628450751304626,
-0.01326794270426035,
0.08606042712926865,
-0.1047176644206047,
-0.018231898546218872,
-0.04343864321708679,
-0.13093332946300507,
0.011610911227762699,
-0.04063675180077553,
0.005709691438823938,
-0.1846354454755783,
-0.2054208219051361,
-0.0381900928914547,
-0.03687773644924164,
-0.0942080169916153,
-0.024952132254838943,
-0.049228884279727936,
-0.05976269394159317,
0.1095316931605339,
-0.017335793003439903,
0.03594543784856796,
-0.0289973933249712,
0.028727509081363678,
0.028726186603307724,
0.08889970183372498,
-0.09557688981294632,
0.023702943697571754,
-0.09015830606222153,
-0.007479938678443432,
-0.13650064170360565,
0.029129235073924065,
-0.04066511616110802,
0.09283772855997086,
-0.08184796571731567,
0.015754716470837593,
0.05444580316543579,
0.01896657608449459,
0.059234555810689926,
0.18227621912956238,
-0.13864079117774963,
-0.007744001690298319,
0.07514667510986328,
-0.06354741007089615,
-0.10904831439256668,
0.12725406885147095,
-0.04462074115872383,
0.057082004845142365,
0.06650649011135101,
0.1348658800125122,
0.03424401581287384,
-0.06090747192502022,
-0.046246130019426346,
-0.005154004320502281,
-0.06994875520467758,
-0.07555725425481796,
0.009415935724973679,
0.016381699591875076,
-0.18171001970767975,
0.03702395409345627,
0.12775255739688873,
0.013341784477233887,
-0.07917804270982742,
-0.06885793060064316,
-0.054242365062236786,
-0.06936129927635193,
0.08806905895471573,
0.01170909870415926,
0.1144399344921112,
-0.11099855601787567,
-0.07721566408872604,
-0.03439619392156601,
0.05615885555744171,
0.020747456699609756,
-0.02893153578042984,
-0.09378119558095932,
0.08372519165277481,
-0.1290663778781891,
-0.02729232981801033,
-0.15016408264636993,
-0.06101524457335472,
-0.014757007360458374,
0.15140032768249512,
-0.04718249663710594,
0.000025944314984371886,
0.06744065880775452,
0.01943366974592209,
-0.05036463215947151,
-0.031060993671417236,
0.16836197674274445,
0.046971630305051804,
-0.1613246202468872,
-0.09943263232707977,
0.07111972570419312,
-0.043542828410863876,
0.16780632734298706,
-0.21565714478492737,
-0.009510954841971397,
-0.00017416177433915436,
0.10320806503295898,
0.01213905494660139,
0.050559207797050476,
-0.02263629250228405,
0.028323357924818993,
-0.07841749489307404,
-0.03316245228052139,
0.028683748096227646,
-0.007051798049360514,
-0.09578320384025574,
0.11513442546129227,
-0.23013782501220703,
0.21951133012771606,
0.16139118373394012,
-0.06819363683462143,
-0.11658389866352081,
0.024869561195373535,
0.022325849160552025,
-0.025098400190472603,
-0.06650467962026596,
-0.01677779108285904,
-0.0052334824576973915,
-0.03749576583504677,
0.16485466063022614,
-0.03453110530972481,
-0.026844071224331856,
0.0375058650970459,
-0.05413775146007538,
-0.08422806859016418,
0.12077637761831284,
-0.056305985897779465,
-0.19231267273426056,
0.14249053597450256,
0.12970413267612457,
-0.06929721683263779,
0.1137324646115303,
0.041290659457445145,
-0.011502108536660671,
-0.02559695392847061,
-0.08584193140268326,
0.006644001696258783,
0.06113557517528534,
0.002228635363280773,
-0.02895924635231495,
0.0478794202208519,
0.018054235726594925,
0.001330960774794221,
-0.10576096177101135,
-0.02397259883582592,
0.007502860389649868,
0.03974229842424393,
-0.015324815176427364,
0.02839057333767414,
0.02670014277100563,
0.14385251700878143,
0.007694899570196867,
-0.1256784349679947,
0.09806862473487854,
0.01855909451842308,
-0.1046738550066948,
0.19537438452243805,
-0.10834464430809021,
-0.2436276376247406,
-0.19060637056827545,
-0.0912269800901413,
-0.0793665274977684,
0.04502337425947189,
0.03428899124264717,
-0.09842853248119354,
-0.08284134417772293,
-0.07335300743579865,
-0.05176805332303047,
-0.04119238629937172,
0.027577901259064674,
-0.036802370101213455,
-0.01888134889304638,
0.041122738271951675,
-0.07812738418579102,
-0.056232038885354996,
-0.02462089993059635,
-0.014536147937178612,
0.15829071402549744,
-0.04911820963025093,
0.11942296475172043,
0.16573594510555267,
-0.05939663574099541,
0.028561610728502274,
-0.03823399543762207,
0.15314531326293945,
-0.04805467277765274,
0.004854699596762657,
0.13779577612876892,
0.00045894074719399214,
0.02250807359814644,
0.1952807605266571,
-0.028071800246834755,
-0.0739540383219719,
0.048415686935186386,
-0.010036901570856571,
-0.029593080282211304,
-0.19210831820964813,
-0.12503951787948608,
-0.004353462718427181,
0.04324052482843399,
0.0761064738035202,
0.024968238547444344,
0.1541845053434372,
0.10824672132730484,
0.03969120979309082,
0.07332516461610794,
-0.06867048889398575,
0.07654862105846405,
0.1415763944387436,
0.005636734887957573,
0.2194640040397644,
-0.061915479600429535,
-0.08187936991453171,
0.06357906013727188,
-0.067563995718956,
0.08428104221820831,
0.06420546025037766,
-0.0784435048699379,
-0.03187889978289604,
0.11112146079540253,
0.10771191120147705,
0.0965256318449974,
0.10106203705072403,
-0.07979406416416168,
0.007317596115171909,
-0.032638080418109894,
-0.057528603821992874,
0.07175973057746887,
0.006392584647983313,
0.021061744540929794,
-0.13620641827583313,
0.042582593858242035,
0.04780129715800285,
0.07538376748561859,
0.16360153257846832,
-0.40689125657081604,
-0.037922658026218414,
0.06107153370976448,
-0.01969759911298752,
-0.14388398826122284,
0.04164886102080345,
0.010316848754882812,
-0.12537603080272675,
0.03541340306401253,
-0.016773082315921783,
0.09075938165187836,
0.00707295211032033,
0.03689175471663475,
-0.08124265819787979,
-0.007806870620697737,
-0.0790376365184784,
0.08358149975538254,
-0.26836204528808594,
0.23330757021903992,
0.047456372529268265,
0.03363436833024025,
-0.08964953571557999,
-0.010703742504119873,
0.05635485798120499,
0.12653504312038422,
0.13505108654499054,
-0.024286266416311264,
-0.13862866163253784,
-0.21607716381549835,
-0.07909374684095383,
-0.015642357990145683,
0.04256643354892731,
0.044824738055467606,
0.06908809393644333,
-0.0711604431271553,
-0.0009622310171835124,
0.054285965859889984,
0.0015834325458854437,
-0.14259213209152222,
-0.06159241124987602,
0.005164507310837507,
0.13588100671768188,
0.010968348942697048,
-0.03178912028670311,
-0.08356253057718277,
-0.11858367919921875,
0.12036242336034775,
0.01466595008969307,
-0.02113918401300907,
-0.1559228003025055,
0.06301502883434296,
0.06732238829135895,
-0.0635867640376091,
0.04603922367095947,
-0.026852848008275032,
0.07883134484291077,
-0.002597623271867633,
-0.09334322810173035,
0.14216388761997223,
-0.10625922679901123,
-0.0806141048669815,
-0.05252121761441231,
0.05013673007488251,
-0.032952748239040375,
0.013601329177618027,
0.05118119344115257,
0.08690549433231354,
-0.06227809563279152,
-0.06878916174173355,
0.07929038256406784,
-0.011833605356514454,
0.14186930656433105,
0.08581971377134323,
-0.0002053191274171695,
-0.21009691059589386,
-0.03866664692759514,
-0.022556088864803314,
0.18595345318317413,
0.2139957845211029,
-0.1114957258105278,
0.021073980256915092,
0.08243662118911743,
-0.031378649175167084,
-0.3265971839427948,
0.04315907880663872,
-0.05913129821419716,
0.06729374080896378,
-0.038869064301252365,
0.045010756701231,
0.10320916026830673,
0.06292994320392609,
-0.05382867157459259,
0.004282016772776842,
-0.13100139796733856,
-0.13189005851745605,
0.2656359076499939,
-0.0019940733909606934,
0.2662869691848755,
-0.11180239915847778,
-0.04877118021249771,
-0.15368543565273285,
-0.10407855361700058,
0.038437239825725555,
-0.06319081038236618,
0.07796896249055862,
-0.0576011948287487,
0.06422404944896698,
0.058593977242708206,
-0.06891461461782455,
0.12153635174036026,
-0.04207872226834297,
0.08997338265180588,
-0.1201585978269577,
-0.0959610641002655,
0.0638599842786789,
-0.08080267161130905,
0.10415344685316086,
-0.010962100699543953,
0.07119370251893997,
-0.1278640478849411,
-0.018578974530100822,
-0.019641540944576263,
0.07267682999372482,
-0.024248138070106506,
-0.05244400352239609,
-0.0368226133286953,
0.016630511730909348,
-0.03297145664691925,
-0.09220633655786514,
0.03972427174448967,
-0.029619315639138222,
0.1158686950802803,
0.20293466746807098,
0.10553443431854248,
-0.05240875855088234,
0.10210778564214706,
0.051328860223293304,
-0.101889967918396,
0.09449011832475662,
-0.11863160878419876,
0.07837846130132675,
0.10463131219148636,
-0.04463525488972664,
0.1505025178194046,
0.0745944157242775,
-0.011816377751529217,
-0.028180241584777832,
0.111924909055233,
-0.17324945330619812,
0.023072347044944763,
-0.058920495212078094,
-0.009991055354475975,
-0.029450058937072754,
0.04115636274218559,
0.1174120157957077,
-0.06924775242805481,
0.002875402569770813,
-0.016741488128900528,
0.029397251084446907,
0.00022139158681966364,
0.18784166872501373,
0.07035602629184723,
0.04521794244647026,
-0.09883106499910355,
0.08586262911558151,
0.08059936761856079,
-0.025149507448077202,
0.07678521424531937,
-0.010775404050946236,
-0.0937720388174057,
-0.06898867338895798,
0.03135945275425911,
0.21998676657676697,
-0.16603592038154602,
-0.07990618795156479,
-0.06724350154399872,
-0.14378729462623596,
0.049875080585479736,
0.21642643213272095,
0.07579676061868668,
0.03990079462528229,
-0.00879586860537529,
-0.09090970456600189,
-0.11204182356595993,
0.09308163076639175,
0.1197546124458313,
0.06755737215280533,
-0.193644180893898,
0.11959737539291382,
-0.07744719833135605,
0.034906744956970215,
-0.0736222043633461,
0.012466921471059322,
-0.18076695501804352,
0.0065626525320112705,
-0.18667080998420715,
0.007393297739326954,
-0.006832887418568134,
-0.015038706362247467,
0.013548613525927067,
-0.034112732857465744,
-0.06629504263401031,
-0.007038666866719723,
-0.05470353364944458,
0.008242407813668251,
0.022436287254095078,
0.036672063171863556,
-0.09136586636304855,
-0.08499261736869812,
0.040113694965839386,
-0.006531159859150648,
0.061000630259513855,
0.08310315012931824,
0.027413291856646538,
0.024070607498288155,
-0.09600801020860672,
-0.032779064029455185,
0.06823179870843887,
-0.0340617299079895,
0.07687785476446152,
-0.09719009697437286,
0.012548205442726612,
0.058772843331098557,
0.03136979416012764,
0.07851789146661758,
0.08431244641542435,
-0.08382699638605118,
0.0051159667782485485,
-0.0590810552239418,
-0.09359566122293472,
-0.0971946269273758,
0.052492622286081314,
0.11317302286624908,
-0.0036691685672849417,
0.14970505237579346,
-0.10452187806367874,
-0.003525486681610346,
-0.10853247344493866,
0.022630637511610985,
-0.057632364332675934,
-0.13452287018299103,
-0.07479961961507797,
-0.02369842864573002,
0.069973424077034,
-0.01710914447903633,
0.12048577517271042,
0.06700962036848068,
0.014409136027097702,
0.038041677325963974,
0.08015542477369308,
-0.0065958937630057335,
0.01948452740907669,
0.17642885446548462,
0.08689416199922562,
-0.04248460382223129,
-0.014367743395268917,
0.08328834921121597,
0.13113006949424744,
0.045402202755212784,
0.058560263365507126,
0.045713357627391815,
-0.07281939685344696,
0.1578092724084854,
-0.013541062362492085,
-0.0915730819106102,
-0.0551028847694397,
-0.031636595726013184,
-0.17278847098350525,
0.04940308630466461,
-0.02358822710812092,
-0.05303853005170822,
0.16086135804653168,
-0.06597238779067993,
-0.03047276847064495,
-0.14793860912322998,
-0.09851459413766861,
-0.18360371887683868,
-0.064018614590168,
-0.15035995841026306,
-0.0740986242890358,
0.005627274978905916,
-0.07594625651836395,
0.04138340428471565,
0.09881705790758133,
0.04550014063715935,
-0.03477035090327263,
0.10754971206188202,
-0.11919169872999191,
0.010980728082358837,
0.06821098923683167,
-0.02910366654396057,
0.0007084790267981589,
-0.053564395755529404,
-0.01687110774219036,
-0.02179638110101223,
0.0253762174397707,
0.007474392652511597,
0.0021732028108090162,
0.011180865578353405,
0.06653992831707001,
-0.02506432496011257,
-0.10506047308444977,
-0.020625699311494827,
0.05135166272521019,
0.05981244891881943,
0.0376640185713768,
0.03337449952960014,
0.045169807970523834,
0.005379633978009224,
0.2875962257385254,
-0.11991073936223984,
-0.0502462163567543,
-0.12452080100774765,
0.2453794926404953,
0.022067047655582428,
0.10232636332511902,
-0.03732249513268471,
-0.10467018932104111,
0.04366275668144226,
0.19854196906089783,
0.20981484651565552,
-0.06052018702030182,
-0.006444203667342663,
-0.040186598896980286,
-0.013428281992673874,
-0.0032756966538727283,
0.08191371709108353,
0.04714633524417877,
0.04791043698787689,
-0.06489028036594391,
0.050841934978961945,
-0.008950942195951939,
-0.07309887558221817,
-0.023239556699991226,
0.14282065629959106,
0.0016820486634969711,
-0.009166429750621319,
-0.05393240228295326,
0.06920421123504639,
-0.03635181114077568,
0.07932595163583755,
0.013591310009360313,
-0.13638482987880707,
-0.12765493988990784,
0.052777718752622604,
0.06962455064058304,
-0.035015229135751724,
0.08489429205656052,
-0.0650186762213707,
0.0080787418410182,
-0.04138227924704552,
-0.009963255375623703,
-0.10441918671131134,
-0.055481839925050735,
0.02243414893746376,
0.0759381651878357,
0.20640544593334198,
0.005126791074872017,
0.09133369475603104,
0.13894245028495789,
-0.006212497595697641,
-0.12342414259910583,
0.10455835610628128,
-0.016559943556785583,
-0.04406824707984924,
0.12235962599515915,
0.07083509117364883,
0.012948769144713879,
0.058238908648490906,
0.030188528820872307,
-0.16919729113578796,
0.035589877516031265,
-0.08676992356777191,
-0.03404771164059639,
-0.07694994658231735,
0.03130864351987839,
-0.03170868009328842,
0.15303748846054077,
0.09874198585748672,
-0.05948003754019737,
0.0340394526720047,
-0.03523007780313492,
0.0035161336418241262,
0.0009485479677096009,
-0.0301506370306015,
0.0026941506657749414,
-0.14460442960262299,
0.061546340584754944,
0.07085223495960236,
0.01356885489076376,
-0.31592440605163574,
-0.027901094406843185,
-0.08171091973781586,
-0.053216129541397095,
-0.07270894199609756,
0.11695773899555206,
0.13381221890449524,
0.021218836307525635,
-0.03987812250852585,
-0.08443514257669449,
0.021485527977347374,
0.11426527798175812,
-0.039736006408929825,
-0.11949695646762848
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text2text-generation | language-plus-molecules/molt5-large-caption2smiles-LPM24 | [
"transformers",
"safetensors",
"t5",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T18:32:55+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #t5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #t5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
58,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #t5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.053328532725572586,
0.16120538115501404,
-0.005120371468365192,
0.022602224722504616,
0.09686747193336487,
0.013199392706155777,
0.07261143624782562,
0.11177206039428711,
-0.020693831145763397,
0.1128523200750351,
0.0323781855404377,
0.09778297692537308,
0.11381756514310837,
0.15530984103679657,
-0.0018252237932756543,
-0.23414164781570435,
0.051169246435165405,
-0.12603329122066498,
-0.039110470563173294,
0.11734651774168015,
0.14655858278274536,
-0.10434788465499878,
0.07780920714139938,
-0.029932111501693726,
-0.010786613449454308,
-0.030950399115681648,
-0.06109464541077614,
-0.04963193088769913,
0.05158040300011635,
0.07096312940120697,
0.06875279545783997,
0.009741154499351978,
0.09293358027935028,
-0.2676756680011749,
0.021060682833194733,
0.07436702400445938,
-0.0019205488497391343,
0.07644513249397278,
0.05394738167524338,
-0.07786445319652557,
0.08801496773958206,
-0.053122974932193756,
0.14802159368991852,
0.08166222274303436,
-0.09144649654626846,
-0.19256246089935303,
-0.08630277216434479,
0.10201671719551086,
0.17971307039260864,
0.050409309566020966,
-0.02338344417512417,
0.10295069962739944,
-0.08843041211366653,
0.012706292793154716,
0.059160783886909485,
-0.06515879184007645,
-0.05482804775238037,
0.0630323737859726,
0.08173035830259323,
0.0787791833281517,
-0.12468571215867996,
-0.018215585500001907,
0.011311499401926994,
0.00691694812849164,
0.08102929592132568,
0.022060219198465347,
0.14176861941814423,
0.03922285884618759,
-0.1292058527469635,
-0.047744158655405045,
0.10315844416618347,
0.04381343349814415,
-0.04969092458486557,
-0.24839195609092712,
-0.028692634776234627,
-0.03409173712134361,
-0.029329892247915268,
-0.041139665991067886,
0.04428756237030029,
-0.010770969092845917,
0.08322557806968689,
-0.008045176975429058,
-0.07979845255613327,
-0.03690612316131592,
0.06324487924575806,
0.05645342543721199,
0.024454401805996895,
-0.008984005078673363,
0.006743076257407665,
0.1175178587436676,
0.10636600106954575,
-0.12631633877754211,
-0.05289403349161148,
-0.06528059393167496,
-0.0853288322687149,
-0.04429693520069122,
0.03338160738348961,
0.04351643845438957,
0.04334709793329239,
0.24920088052749634,
0.011966975405812263,
0.05556565150618553,
0.03878911957144737,
0.011687099933624268,
0.06360286474227905,
0.11270952969789505,
-0.05845928564667702,
-0.09383665025234222,
-0.033332064747810364,
0.09301437437534332,
0.008503437042236328,
-0.0402098223567009,
-0.06047673895955086,
0.06078295037150383,
0.015703821554780006,
0.12211526930332184,
0.087046779692173,
0.002870776690542698,
-0.07195370644330978,
-0.06478150933980942,
0.19285908341407776,
-0.15949691832065582,
0.047871991991996765,
0.03357849270105362,
-0.040312062948942184,
-0.0005020854296162724,
0.01165273692458868,
0.023987481370568275,
-0.021567439660429955,
0.0924374982714653,
-0.05500924214720726,
-0.03761355206370354,
-0.10879732668399811,
-0.03591866046190262,
0.03197222575545311,
0.0022585385013371706,
-0.02967100404202938,
-0.033424828201532364,
-0.08920473605394363,
-0.0635172426700592,
0.09580977261066437,
-0.07413128018379211,
-0.05156254023313522,
-0.016345804557204247,
-0.0761859342455864,
0.026101797819137573,
0.01702207140624523,
0.08535456657409668,
-0.0213642455637455,
0.037230201065540314,
-0.05421315133571625,
0.06241346150636673,
0.10910454392433167,
0.0320611298084259,
-0.053984515368938446,
0.06094928830862045,
-0.2412392497062683,
0.10316064208745956,
-0.07156267017126083,
0.05108866095542908,
-0.15137021243572235,
-0.025331947952508926,
0.04665522649884224,
0.009590202011168003,
-0.011478574015200138,
0.14007656276226044,
-0.2198302298784256,
-0.029333066195249557,
0.1640782356262207,
-0.09730498492717743,
-0.08055570721626282,
0.059064920991659164,
-0.054139286279678345,
0.10999192297458649,
0.04003598168492317,
-0.023768696933984756,
0.06297750771045685,
-0.14250542223453522,
-0.0039275879971683025,
-0.041889119893312454,
-0.01720282807946205,
0.16010744869709015,
0.07506491243839264,
-0.06698185205459595,
0.077672079205513,
0.022212913259863853,
-0.023321649059653282,
-0.04393244534730911,
-0.022494852542877197,
-0.10826845467090607,
0.009565223939716816,
-0.06269361078739166,
0.02424052357673645,
-0.023944495245814323,
-0.0903024971485138,
-0.029575346037745476,
-0.1770460456609726,
-0.013402442447841167,
0.08679109811782837,
-0.010982494801282883,
-0.019886262714862823,
-0.11693590134382248,
0.012033592909574509,
0.032231178134679794,
0.0004325093177612871,
-0.13445010781288147,
-0.05658498778939247,
0.0273329745978117,
-0.16240260004997253,
0.031236927956342697,
-0.05114622414112091,
0.04928715154528618,
0.03406677767634392,
-0.03175085783004761,
-0.031348153948783875,
0.01572313904762268,
0.006510823033750057,
-0.013680041767656803,
-0.24737438559532166,
-0.02852414920926094,
-0.022412575781345367,
0.16979394853115082,
-0.2190135270357132,
0.04012007266283035,
0.07135825604200363,
0.15074580907821655,
0.006911954842507839,
-0.03669405356049538,
0.005606858059763908,
-0.0768459290266037,
-0.03284264728426933,
-0.0623927041888237,
-0.008401541970670223,
-0.03721899166703224,
-0.054593876004219055,
0.051287684589624405,
-0.16718235611915588,
-0.031153932213783264,
0.1028679683804512,
0.06780845671892166,
-0.13963541388511658,
-0.01705223321914673,
-0.04106766730546951,
-0.043112557381391525,
-0.05709490180015564,
-0.05539087578654289,
0.11148729920387268,
0.05757083371281624,
0.04828811436891556,
-0.06848311424255371,
-0.0756818875670433,
0.006132613401859999,
-0.0179264098405838,
-0.021222935989499092,
0.0928845927119255,
0.07583390921354294,
-0.12310270220041275,
0.09178637713193893,
0.10549022257328033,
0.0892157256603241,
0.10119049996137619,
-0.02137933485209942,
-0.08691582083702087,
-0.04892461374402046,
0.0229446180164814,
0.016364475712180138,
0.13983985781669617,
-0.016759416088461876,
0.05310053750872612,
0.04020100086927414,
-0.012910815887153149,
0.011883769184350967,
-0.09328193217515945,
0.02934250421822071,
0.03636814281344414,
-0.019501443952322006,
0.040251899510622025,
-0.03908125311136246,
0.020790016278624535,
0.08787564933300018,
0.04434992000460625,
0.03818633407354355,
0.013980780728161335,
-0.04370194673538208,
-0.11091572046279907,
0.17051653563976288,
-0.12536633014678955,
-0.239797443151474,
-0.14147889614105225,
0.001731917611323297,
0.041165996342897415,
-0.01159723661839962,
0.0031763319857418537,
-0.06770002096891403,
-0.11874829977750778,
-0.09346967190504074,
0.015001182444393635,
0.04228860139846802,
-0.080612413585186,
-0.05524664744734764,
0.05777253210544586,
0.040611669421195984,
-0.143319234251976,
0.020423002541065216,
0.04869217798113823,
-0.08989228308200836,
-0.00900039542466402,
0.08071441948413849,
0.06998268514871597,
0.17929090559482574,
0.009512054733932018,
-0.020932139828801155,
0.03292093798518181,
0.2157505750656128,
-0.13771237432956696,
0.11451084166765213,
0.14277678728103638,
-0.0911637470126152,
0.08293474465608597,
0.1991184800863266,
0.03884927183389664,
-0.10264625400304794,
0.03326369449496269,
0.022328944876790047,
-0.028676386922597885,
-0.2503291964530945,
-0.06918580830097198,
0.0007976540364325047,
-0.05238448083400726,
0.07527847588062286,
0.08888168632984161,
0.09494108706712723,
0.01729334332048893,
-0.09416709095239639,
-0.08025584369897842,
0.04901478812098503,
0.10409125685691833,
0.010409193113446236,
-0.01156378723680973,
0.09060908854007721,
-0.03323452174663544,
0.01843860000371933,
0.09313460439443588,
0.004041523206979036,
0.17060963809490204,
0.05550962686538696,
0.18336638808250427,
0.07643263041973114,
0.0721396952867508,
0.015671607106924057,
0.013079277239739895,
0.02304760180413723,
0.021578695625066757,
-0.0033059304114431143,
-0.0851421132683754,
-0.009511260315775871,
0.11862117052078247,
0.06801546365022659,
0.020754681900143623,
0.009507957845926285,
-0.033934496343135834,
0.08064714074134827,
0.17465052008628845,
-0.0009437129483558238,
-0.1870066076517105,
-0.06896740943193436,
0.08026526123285294,
-0.08972865343093872,
-0.10345284640789032,
-0.02900044620037079,
0.0354950949549675,
-0.17372116446495056,
0.02448408491909504,
-0.018045885488390923,
0.11108683049678802,
-0.1356782615184784,
-0.01890929788351059,
0.06319493800401688,
0.07008420675992966,
-0.0016097982879728079,
0.06208989396691322,
-0.16155508160591125,
0.10791012644767761,
0.01390943955630064,
0.06503470987081528,
-0.09786296635866165,
0.10111832618713379,
-0.006267238408327103,
-0.007413685787469149,
0.14043578505516052,
0.009255880489945412,
-0.07051325589418411,
-0.08343593031167984,
-0.0979004055261612,
-0.010649190284311771,
0.12877127528190613,
-0.14879846572875977,
0.08456916362047195,
-0.0322830006480217,
-0.04405250772833824,
0.005208021495491266,
-0.10768675804138184,
-0.12857580184936523,
-0.18887875974178314,
0.05537694692611694,
-0.13356289267539978,
0.033175256103277206,
-0.1055491715669632,
-0.0408647358417511,
-0.02885887771844864,
0.19630752503871918,
-0.22321896255016327,
-0.0670507624745369,
-0.15318840742111206,
-0.09096445143222809,
0.14798617362976074,
-0.049908362329006195,
0.08374498039484024,
-0.005065108183771372,
0.18742504715919495,
0.01894373446702957,
-0.024415504187345505,
0.1011786088347435,
-0.09638315439224243,
-0.19627197086811066,
-0.08534666895866394,
0.15457913279533386,
0.13537167012691498,
0.0351712740957737,
-0.004617651924490929,
0.03167666867375374,
-0.0189940445125103,
-0.12101218104362488,
0.022920187562704086,
0.17696480453014374,
0.07036592066287994,
0.024736741557717323,
-0.02639835514128208,
-0.11453131586313248,
-0.06600044667720795,
-0.032452553510665894,
0.02982977218925953,
0.18294402956962585,
-0.07586611062288284,
0.18679921329021454,
0.13732017576694489,
-0.05770440772175789,
-0.1956426501274109,
0.01923983357846737,
0.04058924317359924,
0.00837375782430172,
0.032165057957172394,
-0.20239581167697906,
0.08806682378053665,
0.0007347199134528637,
-0.05074144899845123,
0.13624143600463867,
-0.17552010715007782,
-0.15046143531799316,
0.06929060816764832,
0.03642011433839798,
-0.19279520213603973,
-0.12030941992998123,
-0.08865538984537125,
-0.05107492581009865,
-0.17776648700237274,
0.10758756101131439,
0.02193085290491581,
0.00676411809399724,
0.033654287457466125,
0.026140762493014336,
0.014790141955018044,
-0.0396585576236248,
0.19431912899017334,
-0.02348872646689415,
0.030807901173830032,
-0.08293910324573517,
-0.07001609355211258,
0.05941145867109299,
-0.05705835670232773,
0.0775861069560051,
-0.022215960547327995,
0.013414059765636921,
-0.10643109679222107,
-0.04425564035773277,
-0.03175993636250496,
0.015691282227635384,
-0.09722420573234558,
-0.08909335732460022,
-0.050057362765073776,
0.09262266010046005,
0.0974174216389656,
-0.035089656710624695,
-0.03564268350601196,
-0.07118509709835052,
0.039714183658361435,
0.18831974267959595,
0.17605267465114594,
0.046182651072740555,
-0.08030564337968826,
-0.004098092205822468,
-0.011694483458995819,
0.042484745383262634,
-0.21906526386737823,
0.062426332384347916,
0.05058585852384567,
0.014059843495488167,
0.1173645630478859,
-0.01779606007039547,
-0.15810294449329376,
-0.06761486083269119,
0.05993710458278656,
-0.06326820701360703,
-0.19225671887397766,
0.0032602818682789803,
0.055388111621141434,
-0.16711848974227905,
-0.04538320377469063,
0.0430813767015934,
-0.005750913172960281,
-0.039257556200027466,
0.01613711006939411,
0.08359149098396301,
0.0031580389477312565,
0.07040093839168549,
0.05520293489098549,
0.086640864610672,
-0.10250966250896454,
0.07937785238027573,
0.08386688679456711,
-0.08347215503454208,
0.028158824890851974,
0.09330378472805023,
-0.06144890934228897,
-0.029910072684288025,
0.032212331891059875,
0.08255140483379364,
0.012964491732418537,
-0.04401125758886337,
0.008184057660400867,
-0.10146338492631912,
0.0627170279622078,
0.09755739569664001,
0.03206513822078705,
0.011901181191205978,
0.03383762761950493,
0.04645882546901703,
-0.07481352984905243,
0.11842621862888336,
0.025973208248615265,
0.01822328381240368,
-0.04273592680692673,
-0.04516541585326195,
0.027133917436003685,
-0.02340707741677761,
-0.007566304877400398,
-0.03583317995071411,
-0.06988023966550827,
-0.01722576655447483,
-0.16493180394172668,
-0.01076561864465475,
-0.044063083827495575,
0.008020744659006596,
0.026847293600440025,
-0.0369400717318058,
0.008594665676355362,
0.009077225811779499,
-0.07577309012413025,
-0.06240518018603325,
-0.02245018258690834,
0.0914878100156784,
-0.16343435645103455,
0.023352261632680893,
0.08310231566429138,
-0.12098916620016098,
0.09322582185268402,
0.018653366714715958,
-0.0019369579385966063,
0.02680385299026966,
-0.15561461448669434,
0.0368269607424736,
-0.027320701628923416,
0.014671673998236656,
0.045705173164606094,
-0.21818207204341888,
-0.0014451020397245884,
-0.03558654710650444,
-0.059982262551784515,
-0.010693925432860851,
-0.037350837141275406,
-0.11245633661746979,
0.10088492184877396,
0.012412267737090588,
-0.08672942966222763,
-0.03157110512256622,
0.03652326017618179,
0.08053763210773468,
-0.02631879225373268,
0.15205731987953186,
-0.0010786735219880939,
0.07447176426649094,
-0.1738860309123993,
-0.0210786834359169,
-0.0090115275233984,
0.02177848480641842,
-0.016872623935341835,
-0.01564885675907135,
0.042430613189935684,
-0.026671668514609337,
0.18584245443344116,
-0.027355844154953957,
0.03733034059405327,
0.06316441297531128,
0.01770097203552723,
-0.021354418247938156,
0.10755398869514465,
0.06012963131070137,
0.02173144742846489,
0.019801700487732887,
0.0075409491546452045,
-0.041807159781455994,
-0.018543899059295654,
-0.19347810745239258,
0.07164526730775833,
0.14044208824634552,
0.08769161999225616,
-0.012164209969341755,
0.08067302405834198,
-0.10084949433803558,
-0.11743459850549698,
0.11121641099452972,
-0.059808436781167984,
-0.0022669173777103424,
-0.06652101874351501,
0.13155525922775269,
0.14582572877407074,
-0.19254228472709656,
0.07050827890634537,
-0.06511960923671722,
-0.05269601568579674,
-0.11906112730503082,
-0.1953776627779007,
-0.05703132599592209,
-0.054343048483133316,
-0.015079263597726822,
-0.05059242993593216,
0.07498416304588318,
0.05622640252113342,
0.010858895257115364,
0.0015552249969914556,
0.06971994787454605,
-0.019759170711040497,
0.001521410304121673,
0.032095473259687424,
0.06417544931173325,
0.014362066984176636,
-0.03133942559361458,
0.018592869862914085,
-0.008470231667160988,
0.03991629183292389,
0.0633486732840538,
0.04155107960104942,
-0.028110865503549576,
0.01659207232296467,
-0.0337030366063118,
-0.10854189842939377,
0.04278707876801491,
-0.028698457404971123,
-0.08063279837369919,
0.13984808325767517,
0.025403661653399467,
0.009562181308865547,
-0.022226108238101006,
0.241981640458107,
-0.07480388879776001,
-0.09265431761741638,
-0.14692139625549316,
0.1055137887597084,
-0.04348868504166603,
0.06415078788995743,
0.045384783297777176,
-0.10421041399240494,
0.012057800777256489,
0.12658540904521942,
0.1625804305076599,
-0.0438871793448925,
0.019560009241104126,
0.03037482313811779,
0.00398933095857501,
-0.03853052854537964,
0.05252939090132713,
0.06827457249164581,
0.14848913252353668,
-0.050116557627916336,
0.09223522990942001,
0.0050886585377156734,
-0.09908851981163025,
-0.034064266830682755,
0.11810369789600372,
-0.019035303965210915,
0.019260596483945847,
-0.05601469427347183,
0.11788773536682129,
-0.06368034332990646,
-0.233087420463562,
0.06406685709953308,
-0.07426205277442932,
-0.14131881296634674,
-0.024826664477586746,
0.07676053047180176,
-0.014309047721326351,
0.027850469574332237,
0.0722186341881752,
-0.07654546946287155,
0.19937579333782196,
0.03671684116125107,
-0.058611851185560226,
-0.05623113736510277,
0.07896319031715393,
-0.11419995129108429,
0.27488458156585693,
0.015893742442131042,
0.045155949890613556,
0.1038452610373497,
-0.013412448577582836,
-0.13435201346874237,
0.01833420805633068,
0.09638454020023346,
-0.08846497535705566,
0.04018587991595268,
0.20595665276050568,
-0.0028567397966980934,
0.11962885409593582,
0.07707620412111282,
-0.08087631314992905,
0.049051105976104736,
-0.09828304499387741,
-0.07230360060930252,
-0.08931835740804672,
0.09120666980743408,
-0.07232820242643356,
0.14308606088161469,
0.1311190128326416,
-0.05265164002776146,
0.00968363881111145,
-0.029376711696386337,
0.045510269701480865,
0.004632700700312853,
0.10403459519147873,
0.008749093860387802,
-0.1797543615102768,
0.02403045818209648,
0.01841445453464985,
0.10992073267698288,
-0.1701374351978302,
-0.09734909981489182,
0.043629229068756104,
-0.0012522460892796516,
-0.06121290475130081,
0.1290796846151352,
0.05957380682229996,
0.05011506378650665,
-0.043520737439394,
-0.0211784765124321,
-0.008504665456712246,
0.14072857797145844,
-0.10404830425977707,
-0.00016830587992444634
] |
null | null | null | Distilbert trained on deepfaketextdetect | {"metrics": ["accuracy"]} | null | lyon0210/distilbert-deepfake-text | [
"region:us"
] | 2024-02-08T18:33:39+00:00 | [] | [] | TAGS
#region-us
| Distilbert trained on deepfaketextdetect | [] | [
"TAGS\n#region-us \n"
] | [
6
] | [
"passage: TAGS\n#region-us \n"
] | [
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | peft | ## Training procedure
The following `bitsandbytes` quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: False
- bnb_4bit_compute_dtype: float16
### Framework versions
- PEFT 0.4.0
| {"library_name": "peft"} | null | Sakil/llama2-chat-hub-new-model-finetuned | [
"peft",
"region:us"
] | 2024-02-08T18:34:27+00:00 | [] | [] | TAGS
#peft #region-us
| ## Training procedure
The following 'bitsandbytes' quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: False
- bnb_4bit_compute_dtype: float16
### Framework versions
- PEFT 0.4.0
| [
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16",
"### Framework versions\n\n\n- PEFT 0.4.0"
] | [
"TAGS\n#peft #region-us \n",
"## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16",
"### Framework versions\n\n\n- PEFT 0.4.0"
] | [
9,
154,
11
] | [
"passage: TAGS\n#peft #region-us \n## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16### Framework versions\n\n\n- PEFT 0.4.0"
] | [
-0.06979092955589294,
0.021877916529774666,
-0.0024098597932606936,
0.1378258466720581,
0.10862318426370621,
0.07474296540021896,
0.10942144691944122,
0.12339992076158524,
0.0572756864130497,
0.09055759757757187,
0.09443594515323639,
0.05080708488821983,
0.07617583870887756,
0.15425844490528107,
-0.02472618594765663,
-0.030937515199184418,
0.05461447685956955,
0.0024960634764283895,
0.003683591727167368,
0.07837489992380142,
0.05202331766486168,
-0.036243125796318054,
0.037176959216594696,
-0.09029994904994965,
-0.1683429628610611,
-0.0007763148751109838,
0.010803398676216602,
0.03168312832713127,
0.04643981158733368,
0.03667012229561806,
0.061239324510097504,
-0.01880047656595707,
-0.024256091564893723,
-0.2065766304731369,
-0.008815872482955456,
0.12412996590137482,
-0.022375915199518204,
0.0724588930606842,
-0.08079487085342407,
0.13203540444374084,
-0.039120372384786606,
-0.02855866029858589,
0.005213533993810415,
0.031375303864479065,
-0.08930839598178864,
-0.12567098438739777,
-0.06602675467729568,
0.061225954443216324,
0.022498471662402153,
0.0596780888736248,
0.0024417112581431866,
0.19193901121616364,
-0.1333344727754593,
0.08866282552480698,
0.10110592097043991,
-0.24191753566265106,
-0.03025546856224537,
0.14384306967258453,
-0.024540316313505173,
0.15543513000011444,
-0.07628224045038223,
-0.10719222575426102,
0.08085530251264572,
0.0496523417532444,
-0.056598056107759476,
-0.0009009919012896717,
-0.09088310599327087,
0.006839253939688206,
-0.13980844616889954,
-0.05230165645480156,
0.15110687911510468,
0.026294633746147156,
-0.037592023611068726,
-0.031246718019247055,
-0.0950726643204689,
-0.35348814725875854,
0.024415571242570877,
-0.002826953772455454,
-0.06987019628286362,
0.048015810549259186,
0.035156507045030594,
-0.008995379321277142,
0.004349078983068466,
-0.09262323379516602,
-0.04143654555082321,
0.10029355436563492,
0.04808283597230911,
0.04402211308479309,
0.019573308527469635,
0.1058531403541565,
-0.12883538007736206,
-0.021110432222485542,
-0.041084397584199905,
-0.02869381383061409,
-0.050102680921554565,
-0.014695894904434681,
-0.07354352623224258,
0.17733292281627655,
0.06763829290866852,
0.09666602313518524,
-0.1813267171382904,
0.1257096230983734,
-0.02475627325475216,
0.05742272362112999,
-0.035531818866729736,
0.014435730874538422,
-0.11885007470846176,
0.12319590896368027,
0.0007600552635267377,
0.13899193704128265,
0.02280823327600956,
-0.04473873972892761,
-0.0730285570025444,
-0.007154371589422226,
0.1233254000544548,
0.004622994922101498,
-0.11530132591724396,
0.007921767421066761,
-0.1411384791135788,
-0.04554228112101555,
0.0911121591925621,
-0.07906953990459442,
0.015561599284410477,
0.041121818125247955,
-0.04623573645949364,
0.004239376168698072,
0.09626287966966629,
-0.05426957085728645,
-0.04919392615556717,
-0.03332843258976936,
-0.10088252276182175,
0.00408816896378994,
-0.1039951965212822,
-0.13126124441623688,
0.051543787121772766,
-0.1563965529203415,
-0.00575046194717288,
-0.04191170632839203,
-0.06613527983427048,
0.02011139690876007,
0.019768383353948593,
-0.08437694609165192,
0.05859275907278061,
-0.08150110393762589,
-0.1557578295469284,
-0.03541033715009689,
0.018620938062667847,
0.007860139012336731,
-0.02999168075621128,
0.09800084680318832,
0.032462093979120255,
0.10065783560276031,
-0.1774347871541977,
-0.0032585663720965385,
0.013272988609969616,
0.06698489189147949,
0.016430236399173737,
0.11852112412452698,
-0.0984061062335968,
-0.029854416847229004,
-0.05219494178891182,
-0.06776446104049683,
-0.10671226680278778,
-0.009978118352591991,
0.12339237332344055,
0.081203892827034,
-0.15919211506843567,
-0.010960342362523079,
0.07894296199083328,
-0.034127745777368546,
-0.08150023221969604,
0.1462218016386032,
-0.043517038226127625,
0.10362964868545532,
-0.03142445161938667,
0.07441626489162445,
0.2393224835395813,
-0.13294187188148499,
0.0026927399449050426,
0.11512340605258942,
0.04870401695370674,
-0.024512527510523796,
0.001410747179761529,
0.06937900930643082,
-0.142485573887825,
0.03395999222993851,
0.067036472260952,
0.03290358558297157,
-0.059946075081825256,
-0.05717241391539574,
-0.03879120573401451,
-0.054277822375297546,
0.11310963332653046,
0.028706887736916542,
0.012219887226819992,
-0.07421029359102249,
-0.07091692090034485,
0.14475229382514954,
0.13386821746826172,
-0.01724834553897381,
-0.005964277777820826,
-0.1111484244465828,
0.002316186437383294,
-0.04553208500146866,
0.027614165097475052,
-0.12323106080293655,
0.024642376229166985,
0.08063367009162903,
0.01609816960990429,
-0.005867816042155027,
0.048622094094753265,
0.06404208391904831,
0.02775740437209606,
-0.054996591061353683,
0.023086974397301674,
-0.05230040103197098,
-0.002152500906959176,
-0.10709255933761597,
-0.07857770472764969,
-0.013062533922493458,
-0.010289276950061321,
0.18266798555850983,
-0.14896303415298462,
0.03612537682056427,
0.10309912264347076,
0.004995338153094053,
-0.010612043552100658,
-0.036244068294763565,
-0.07481951266527176,
0.11431097984313965,
-0.012485682964324951,
-0.03552611544728279,
0.04104119539260864,
0.034952957183122635,
-0.03309863805770874,
-0.1580384373664856,
-0.09931837022304535,
0.05646728724241257,
0.13403663039207458,
0.07577018439769745,
-0.06703833490610123,
-0.047553446143865585,
-0.020618250593543053,
-0.03879788890480995,
0.056729573756456375,
-0.05360506474971771,
0.023469123989343643,
0.012125842273235321,
0.06819912791252136,
-0.10050064325332642,
-0.03629688918590546,
0.06696230918169022,
-0.011795254424214363,
-0.04454666003584862,
0.11060863733291626,
0.01751997321844101,
-0.07169803977012634,
0.07787928730249405,
0.05941096320748329,
-0.14069285988807678,
0.08998789638280869,
-0.004688691347837448,
-0.023223139345645905,
-0.09080317616462708,
0.17682680487632751,
0.024636752903461456,
0.12305090576410294,
-0.1265403926372528,
0.10609687864780426,
-0.010145329870283604,
0.014946194365620613,
0.07392288744449615,
-0.20129480957984924,
-0.004805627278983593,
-0.0471431165933609,
-0.09572996944189072,
-0.05821099132299423,
-0.015557871200144291,
0.0170791894197464,
0.043467629700899124,
-0.005354706663638353,
0.06044595688581467,
0.142664834856987,
-0.018936002627015114,
-0.08370210230350494,
0.17457811534404755,
-0.21770337224006653,
-0.20481646060943604,
-0.23560066521167755,
0.0029743367340415716,
-0.11851438134908676,
-0.037289947271347046,
-0.05196189507842064,
-0.07753796130418777,
0.03229401260614395,
-0.08353889733552933,
-0.040616557002067566,
-0.019290516152977943,
0.003061749739572406,
0.05178794264793396,
0.011813749559223652,
0.1695105880498886,
-0.0735611766576767,
0.023279182612895966,
0.046429067850112915,
-0.02350066788494587,
0.10787586867809296,
-0.08872182667255402,
-0.03777230158448219,
0.12821447849273682,
-0.009709981270134449,
0.02801087126135826,
0.01423876266926527,
0.3048967123031616,
0.010030297562479973,
0.02763240598142147,
0.09710080921649933,
0.009545043110847473,
0.06174734979867935,
0.07832355052232742,
0.015707092359662056,
-0.10942601412534714,
0.06433121860027313,
0.04781096801161766,
-0.09661031514406204,
-0.13850480318069458,
-0.047762222588062286,
-0.07341255247592926,
0.006593871861696243,
0.08610724657773972,
0.07147743552923203,
0.07943940162658691,
0.07025924324989319,
0.023519491776823997,
0.11204782873392105,
-0.004567211028188467,
-0.005818400532007217,
0.1201663389801979,
-0.020266585052013397,
0.04491112753748894,
-0.030482962727546692,
0.03723101690411568,
0.07022300362586975,
0.14435730874538422,
0.08234275877475739,
-0.0826135203242302,
-0.0026839443016797304,
0.04897516965866089,
0.2717514932155609,
-0.01714148186147213,
0.10665575414896011,
-0.06918495893478394,
-0.013546466827392578,
0.0014700466999784112,
-0.03347883000969887,
-0.07884032279253006,
0.04058695584535599,
-0.02208472229540348,
0.07936602830886841,
-0.017719898372888565,
-0.027150988578796387,
0.07915937900543213,
0.12511037290096283,
0.15299907326698303,
-0.28453946113586426,
-0.11869988590478897,
-0.013159751892089844,
0.11041011661291122,
-0.10107292234897614,
0.0223299078643322,
0.21577490866184235,
0.008470140397548676,
-0.08149175345897675,
-0.03681793436408043,
0.03811902552843094,
-0.01879933662712574,
0.016045289114117622,
0.12573954463005066,
0.14544515311717987,
0.005644232500344515,
0.07689831405878067,
-0.2924128472805023,
0.023194577544927597,
0.05471178889274597,
0.046213310211896896,
-0.04598768800497055,
0.008340165950357914,
-0.04420474171638489,
-0.06625235080718994,
0.029148653149604797,
0.01076830830425024,
0.1823798418045044,
-0.28289252519607544,
-0.06890449672937393,
-0.02097264863550663,
0.1087518259882927,
0.08112131059169769,
0.04162963479757309,
0.01758452132344246,
0.056389037519693375,
0.06615620851516724,
0.03817380219697952,
-0.0343405120074749,
-0.09718122333288193,
0.0010788479121401906,
0.15619537234306335,
-0.14491763710975647,
-0.05582432448863983,
-0.06452815234661102,
-0.024293042719364166,
0.05778651684522629,
-0.1548195481300354,
-0.05276089906692505,
-0.06254495680332184,
0.015085814520716667,
0.1334981471300125,
-0.023490145802497864,
-0.01085786335170269,
-0.0201820470392704,
0.024849023669958115,
-0.04709408059716225,
-0.09744754433631897,
0.11661811918020248,
-0.0471058264374733,
-0.1306038200855255,
-0.019828403368592262,
0.1378549486398697,
0.08238520473241806,
-0.015439215116202831,
-0.0857001468539238,
-0.04882924631237984,
0.027058644220232964,
-0.15197721123695374,
0.005223190877586603,
0.09445921331644058,
-0.0594605952501297,
0.09092224389314651,
-0.11116138100624084,
0.22619131207466125,
-0.04533985257148743,
0.09459085017442703,
0.061151355504989624,
0.2978513538837433,
-0.08713971078395844,
0.018928516656160355,
0.04804879426956177,
-0.018178772181272507,
-0.2438703179359436,
0.04353063553571701,
0.05640896409749985,
0.040471117943525314,
-0.03873369097709656,
-0.1723128706216812,
0.035184476524591446,
0.09765327721834183,
0.009074017405509949,
0.20277278125286102,
-0.32115042209625244,
-0.05401403456926346,
0.04298613592982292,
0.06709850579500198,
0.15412142872810364,
-0.05032355710864067,
0.005308091174811125,
0.00299404701218009,
-0.025177884846925735,
0.14879994094371796,
-0.10914967209100723,
0.11319706588983536,
-0.028690336272120476,
0.024565158411860466,
0.009507894515991211,
-0.03613625466823578,
0.14931310713291168,
0.013264713808894157,
0.089763343334198,
0.019067833200097084,
-0.04837853088974953,
0.05236007273197174,
-0.08060209453105927,
0.048760365694761276,
-0.055018678307533264,
0.08426039665937424,
-0.0716785341501236,
0.014209864661097527,
-0.0624924935400486,
-0.02230324223637581,
-0.07208869606256485,
-0.03789118677377701,
-0.098033107817173,
0.07829388231039047,
-0.014118213206529617,
-0.0252530500292778,
-0.030577998608350754,
0.06026613339781761,
0.06182622164487839,
0.4425763189792633,
-0.06215111166238785,
-0.059386324137449265,
0.0778389647603035,
0.09745035320520401,
-0.018147382885217667,
0.10684601962566376,
-0.12679900228977203,
0.04500621557235718,
0.12372613698244095,
-0.0008733426802791655,
0.12444817274808884,
0.0908789336681366,
-0.11470408737659454,
-0.01735471934080124,
0.03941637650132179,
-0.13241484761238098,
-0.072132408618927,
-0.02437761053442955,
-0.0033349243458360434,
-0.10859358310699463,
0.02123371511697769,
0.10324949026107788,
-0.028791457414627075,
0.054798197001218796,
0.03556765988469124,
0.044085126370191574,
-0.13439145684242249,
0.13734261691570282,
0.03844649717211723,
0.0715968981385231,
-0.09430023282766342,
0.08732116222381592,
0.023660900071263313,
0.012141572311520576,
0.04477553442120552,
-0.0023568833712488413,
-0.10149598866701126,
0.010059066116809845,
-0.038228947669267654,
-0.09548848867416382,
0.09702998399734497,
-0.04586396366357803,
-0.0412009172141552,
-0.11459630727767944,
0.016487689688801765,
0.07441749423742294,
0.06058561056852341,
0.10161381214857101,
-0.0266474150121212,
0.01351380068808794,
-0.1312156319618225,
0.0554811991751194,
-0.034641385078430176,
0.028874071314930916,
-0.15108218789100647,
0.060451265424489975,
-0.01823197305202484,
0.07453461736440659,
-0.019939351826906204,
-0.013813626021146774,
-0.22570104897022247,
0.021174952387809753,
-0.033057328313589096,
0.003854290582239628,
0.029487796127796173,
0.03671148419380188,
0.02527959644794464,
0.05630463361740112,
-0.031085221096873283,
0.04344148188829422,
-0.03895607963204384,
-0.051280442625284195,
0.04188956692814827,
0.0007596524665132165,
-0.035418447107076645,
-0.046238139271736145,
0.04512936249375343,
-0.11377303302288055,
0.037379465997219086,
0.037038300186395645,
-0.05313664302229881,
0.07457074522972107,
0.0551333986222744,
0.020679296925663948,
0.09771078824996948,
0.05872461944818497,
0.05539431795477867,
-0.05644817277789116,
0.038207992911338806,
-0.013830473646521568,
-0.0014212807873263955,
0.05458271503448486,
0.13561320304870605,
-0.05696135014295578,
-0.05802462622523308,
-0.1356644332408905,
-0.022950179874897003,
-0.04388374090194702,
0.04029588773846626,
0.15611368417739868,
0.10005352646112442,
0.08508609980344772,
-0.08406435698270798,
-0.023177621886134148,
-0.1434895098209381,
-0.0838560089468956,
0.057367514818906784,
-0.04613112285733223,
-0.013442962430417538,
-0.03551625832915306,
0.07219497859477997,
0.007350239437073469,
0.15389199554920197,
-0.09119492769241333,
-0.09884331375360489,
-0.050555817782878876,
-0.17701447010040283,
-0.1346471607685089,
-0.005777652841061354,
0.2509252429008484,
0.031908661127090454,
-0.030132247135043144,
-0.06317965686321259,
0.00790130253881216,
0.0689205676317215,
0.13760150969028473,
0.04632747173309326,
0.07801228016614914,
-0.12434402108192444,
0.10745592415332794,
0.04290035367012024,
-0.05061715841293335,
0.12014038860797882,
0.30707380175590515,
-0.08199357986450195,
0.011254170909523964,
-0.08713367581367493,
0.08879105001688004,
0.02187724970281124,
-0.1383187174797058,
0.013609658926725388,
-0.03021937422454357,
-0.15772560238838196,
-0.1141202449798584,
0.03438706323504448,
-0.06414132565259933,
-0.20812147855758667,
-0.02861359901726246,
-0.10091375559568405,
-0.07548198103904724,
0.1385963410139084,
0.036400459706783295,
-0.01720549911260605,
0.1921585202217102,
-0.07791507244110107,
0.047225624322891235,
-0.010112340562045574,
-0.01755616068840027,
-0.004292663186788559,
-0.03010612167418003,
-0.10022464394569397,
0.14229239523410797,
0.0026088182348757982,
0.11017946153879166,
0.0076710316352546215,
0.08859197050333023,
0.05187401920557022,
-0.024778150022029877,
-0.040657348930835724,
-0.0020166346803307533,
0.015454727225005627,
-0.06949343532323837,
0.1287740021944046,
0.06015017256140709,
-0.08151866495609283,
-0.08153397589921951,
-0.020808378234505653,
-0.08484478294849396,
-0.021831484511494637,
-0.1572553515434265,
0.28017404675483704,
-0.03308238461613655,
0.10164105892181396,
-0.009170068427920341,
-0.05692457780241966,
-0.10435646772384644,
0.14374536275863647,
0.120839424431324,
-0.13673727214336395,
-0.012003384530544281,
0.08866086602210999,
-0.007145259529352188,
-0.07515248656272888,
0.155784472823143,
0.09963296353816986,
-0.014980965293943882,
0.02876802533864975,
-0.023070234805345535,
-0.03320426493883133,
0.0033824562560766935,
-0.005501914769411087,
-0.034212253987789154,
0.005797996185719967,
0.03244750574231148,
-0.13621188700199127,
-0.017982242628932,
-0.06249391660094261,
-0.06658949702978134,
0.1939048320055008,
-0.1203932911157608,
-0.08973969519138336,
-0.0308162122964859,
-0.09773550927639008,
-0.10429038852453232,
0.02302357368171215,
-0.11399734765291214,
0.059150729328393936,
0.08660008758306503,
-0.05918910726904869,
-0.015915945172309875,
-0.07759834825992584,
-0.004432562738656998,
0.01995953358709812,
0.06680567562580109,
-0.012953863479197025,
0.07061396539211273,
0.10941770672798157,
-0.02109832689166069,
-0.05627770721912384,
0.09945710748434067,
0.005617896094918251,
-0.058368053287267685,
-0.14522160589694977,
0.017044585198163986,
-0.029553061351180077,
0.10308431833982468,
0.024984009563922882,
-0.06643787026405334,
-0.03581478074193001,
-0.17625966668128967,
-0.011637690477073193,
-0.1381247639656067,
-0.08038759231567383,
-0.07788003236055374,
0.10739508271217346,
0.19482040405273438,
-0.05468401312828064,
0.025051061064004898,
-0.03843652456998825,
0.03321219980716705,
-0.04963286593556404,
0.04273268207907677,
-0.005887014791369438,
-0.14593636989593506,
0.05899544060230255,
-0.05112109333276749,
0.004382130224257708,
-0.28370237350463867,
0.001215865253470838,
0.005580263677984476,
-0.033779751509428024,
-0.05040258169174194,
0.1377897560596466,
0.00042426568688824773,
0.07757376879453659,
-0.062357865273952484,
-0.24849353730678558,
-0.0690150111913681,
0.14145798981189728,
0.0006318931700661778,
-0.07864443957805634
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# perioli_vgm_v8
This model is a fine-tuned version of [microsoft/layoutlmv3-base](https://huggingface.co/microsoft/layoutlmv3-base) on the sroie dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0120
- Precision: 0.9340
- Recall: 0.9274
- F1: 0.9307
- Accuracy: 0.9981
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 2
- eval_batch_size: 2
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 3000
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| No log | 0.32 | 100 | 0.0817 | 0.4667 | 0.2787 | 0.3490 | 0.9800 |
| No log | 0.64 | 200 | 0.0494 | 0.5981 | 0.5925 | 0.5953 | 0.9864 |
| No log | 0.96 | 300 | 0.0351 | 0.7578 | 0.7400 | 0.7488 | 0.9909 |
| No log | 1.29 | 400 | 0.0296 | 0.8333 | 0.7260 | 0.7760 | 0.9924 |
| 0.0728 | 1.61 | 500 | 0.0242 | 0.8480 | 0.8103 | 0.8287 | 0.9940 |
| 0.0728 | 1.93 | 600 | 0.0237 | 0.7564 | 0.7564 | 0.7564 | 0.9929 |
| 0.0728 | 2.25 | 700 | 0.0196 | 0.8067 | 0.8407 | 0.8234 | 0.9949 |
| 0.0728 | 2.57 | 800 | 0.0192 | 0.7982 | 0.8337 | 0.8156 | 0.9946 |
| 0.0728 | 2.89 | 900 | 0.0149 | 0.8642 | 0.8642 | 0.8642 | 0.9961 |
| 0.015 | 3.22 | 1000 | 0.0146 | 0.8695 | 0.8735 | 0.8715 | 0.9964 |
| 0.015 | 3.54 | 1100 | 0.0140 | 0.9346 | 0.9040 | 0.9190 | 0.9973 |
| 0.015 | 3.86 | 1200 | 0.0149 | 0.8886 | 0.9157 | 0.9020 | 0.9969 |
| 0.015 | 4.18 | 1300 | 0.0147 | 0.8924 | 0.9133 | 0.9028 | 0.9971 |
| 0.015 | 4.5 | 1400 | 0.0153 | 0.8628 | 0.9133 | 0.8874 | 0.9964 |
| 0.0075 | 4.82 | 1500 | 0.0155 | 0.9007 | 0.9133 | 0.9070 | 0.9971 |
| 0.0075 | 5.14 | 1600 | 0.0135 | 0.9190 | 0.9297 | 0.9243 | 0.9977 |
| 0.0075 | 5.47 | 1700 | 0.0153 | 0.8970 | 0.8970 | 0.8970 | 0.9970 |
| 0.0075 | 5.79 | 1800 | 0.0148 | 0.9348 | 0.9063 | 0.9203 | 0.9975 |
| 0.0075 | 6.11 | 1900 | 0.0158 | 0.8857 | 0.9251 | 0.9049 | 0.9970 |
| 0.0034 | 6.43 | 2000 | 0.0145 | 0.9189 | 0.9016 | 0.9102 | 0.9972 |
| 0.0034 | 6.75 | 2100 | 0.0140 | 0.9303 | 0.9063 | 0.9181 | 0.9975 |
| 0.0034 | 7.07 | 2200 | 0.0137 | 0.9062 | 0.8829 | 0.8944 | 0.9971 |
| 0.0034 | 7.4 | 2300 | 0.0150 | 0.9099 | 0.9227 | 0.9163 | 0.9971 |
| 0.0034 | 7.72 | 2400 | 0.0124 | 0.9195 | 0.9368 | 0.9281 | 0.9979 |
| 0.0022 | 8.04 | 2500 | 0.0119 | 0.9329 | 0.9438 | 0.9383 | 0.9980 |
| 0.0022 | 8.36 | 2600 | 0.0124 | 0.9370 | 0.9063 | 0.9214 | 0.9976 |
| 0.0022 | 8.68 | 2700 | 0.0124 | 0.9261 | 0.9391 | 0.9326 | 0.9979 |
| 0.0022 | 9.0 | 2800 | 0.0120 | 0.9409 | 0.9321 | 0.9365 | 0.9981 |
| 0.0022 | 9.32 | 2900 | 0.0122 | 0.9340 | 0.9274 | 0.9307 | 0.9981 |
| 0.0017 | 9.65 | 3000 | 0.0120 | 0.9340 | 0.9274 | 0.9307 | 0.9981 |
### Framework versions
- Transformers 4.28.0
- Pytorch 2.1.0+cu121
- Datasets 2.2.2
- Tokenizers 0.13.3
| {"license": "cc-by-nc-sa-4.0", "tags": ["generated_from_trainer"], "datasets": ["sroie"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "perioli_vgm_v8", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "sroie", "type": "sroie", "config": "discharge", "split": "test", "args": "discharge"}, "metrics": [{"type": "precision", "value": 0.9339622641509434, "name": "Precision"}, {"type": "recall", "value": 0.927400468384075, "name": "Recall"}, {"type": "f1", "value": 0.9306698002350177, "name": "F1"}, {"type": "accuracy", "value": 0.9981452418319304, "name": "Accuracy"}]}]}]} | token-classification | atatavana/vgm_v8 | [
"transformers",
"pytorch",
"tensorboard",
"layoutlmv3",
"token-classification",
"generated_from_trainer",
"dataset:sroie",
"license:cc-by-nc-sa-4.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | 2024-02-08T18:40:05+00:00 | [] | [] | TAGS
#transformers #pytorch #tensorboard #layoutlmv3 #token-classification #generated_from_trainer #dataset-sroie #license-cc-by-nc-sa-4.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
| perioli\_vgm\_v8
================
This model is a fine-tuned version of microsoft/layoutlmv3-base on the sroie dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0120
* Precision: 0.9340
* Recall: 0.9274
* F1: 0.9307
* Accuracy: 0.9981
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 2
* eval\_batch\_size: 2
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 3000
### Training results
### Framework versions
* Transformers 4.28.0
* Pytorch 2.1.0+cu121
* Datasets 2.2.2
* Tokenizers 0.13.3
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 3000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.1.0+cu121\n* Datasets 2.2.2\n* Tokenizers 0.13.3"
] | [
"TAGS\n#transformers #pytorch #tensorboard #layoutlmv3 #token-classification #generated_from_trainer #dataset-sroie #license-cc-by-nc-sa-4.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 3000",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.1.0+cu121\n* Datasets 2.2.2\n* Tokenizers 0.13.3"
] | [
76,
97,
4,
35
] | [
"passage: TAGS\n#transformers #pytorch #tensorboard #layoutlmv3 #token-classification #generated_from_trainer #dataset-sroie #license-cc-by-nc-sa-4.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 2\n* eval\\_batch\\_size: 2\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 3000### Training results### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.1.0+cu121\n* Datasets 2.2.2\n* Tokenizers 0.13.3"
] | [
-0.11467330902814865,
0.1045001819729805,
-0.0019301355350762606,
0.12248644977807999,
0.1583670675754547,
0.022767098620533943,
0.13266205787658691,
0.12350881844758987,
-0.060378119349479675,
0.027160849422216415,
0.1335083693265915,
0.13550476729869843,
0.026432795450091362,
0.15382246673107147,
-0.046036817133426666,
-0.24979080259799957,
-0.012830570340156555,
0.04436175897717476,
-0.05808768793940544,
0.13654614984989166,
0.09798716008663177,
-0.12582649290561676,
0.09427221864461899,
0.008224229328334332,
-0.20600447058677673,
-0.017963457852602005,
0.026740504428744316,
-0.0476793497800827,
0.14987710118293762,
0.02907484956085682,
0.12945456802845,
0.0225067101418972,
0.1030711829662323,
-0.15788589417934418,
0.01241059135645628,
0.048107489943504333,
0.006241658702492714,
0.10428646206855774,
0.03885545954108238,
0.016075683757662773,
0.04815295338630676,
-0.07401233166456223,
0.057206764817237854,
0.012602753937244415,
-0.1330523043870926,
-0.209299236536026,
-0.09290273487567902,
0.05093935504555702,
0.09093198925256729,
0.08236317336559296,
0.0009662179509177804,
0.15255875885486603,
-0.06275396049022675,
0.07691864669322968,
0.16943535208702087,
-0.28935331106185913,
-0.06974988430738449,
0.06979169696569443,
0.023015087470412254,
0.05846939980983734,
-0.10303880274295807,
-0.020609289407730103,
0.03680786117911339,
0.03686673566699028,
0.14477328956127167,
-0.027230991050601006,
-0.024204423651099205,
0.013701101765036583,
-0.136224165558815,
-0.03739393502473831,
0.1535636931657791,
0.049880657345056534,
-0.039198294281959534,
-0.054015666246414185,
-0.04525590315461159,
-0.1259753555059433,
-0.03488096222281456,
-0.0019990215077996254,
0.03840269520878792,
-0.02734639123082161,
-0.10838425904512405,
-0.029369989410042763,
-0.10745371878147125,
-0.0683988407254219,
-0.05855065956711769,
0.11728410422801971,
0.009235787205398083,
0.010690326802432537,
-0.013907566666603088,
0.11703846603631973,
-0.0020214112009853125,
-0.1292656660079956,
0.03472856059670448,
0.020963724702596664,
-0.03266650065779686,
-0.0656518042087555,
-0.0426168255507946,
-0.048057518899440765,
-0.014125501736998558,
0.11832387745380402,
-0.009651564061641693,
0.02332969196140766,
0.02858077920973301,
0.05153549090027809,
-0.09816049784421921,
0.1944194734096527,
-0.05184009671211243,
-0.034972548484802246,
-0.0003848263877443969,
0.09046602994203568,
0.02190125547349453,
-0.016842765733599663,
-0.14929071068763733,
0.007021480705589056,
0.08366751670837402,
0.010327079333364964,
-0.04399559274315834,
0.055830396711826324,
-0.06479106098413467,
-0.04135294631123543,
0.058712251484394073,
-0.07375072687864304,
0.026829972863197327,
-0.015317332930862904,
-0.07583200931549072,
-0.04992710053920746,
0.002797864843159914,
0.03074885904788971,
0.015638289973139763,
0.11895502358675003,
-0.10830038785934448,
0.022106464952230453,
-0.08952971547842026,
-0.10459408164024353,
0.016583001241087914,
-0.09419792145490646,
0.01674767956137657,
-0.09680194407701492,
-0.18520675599575043,
-0.016080383211374283,
0.06070813536643982,
-0.03653766214847565,
-0.07379553467035294,
-0.04178498685359955,
-0.061803899705410004,
0.00983976200222969,
-0.014926780015230179,
0.12727488577365875,
-0.061423443257808685,
0.10366902500391006,
0.012766553089022636,
0.05483433976769447,
-0.052199140191078186,
0.04393034428358078,
-0.09516048431396484,
0.03304052725434303,
-0.1467573195695877,
0.03210924193263054,
-0.033813074231147766,
0.06780073046684265,
-0.11027887463569641,
-0.0866309404373169,
0.019435115158557892,
-0.013530418276786804,
0.06342773884534836,
0.0865996852517128,
-0.18262527883052826,
-0.07005850970745087,
0.14243444800376892,
-0.05490151420235634,
-0.12791132926940918,
0.12834471464157104,
-0.0662209615111351,
0.06085043027997017,
0.05607107654213905,
0.17339545488357544,
0.08240167051553726,
-0.08980151265859604,
0.022850828245282173,
0.010659785009920597,
0.06167276203632355,
-0.08573763072490692,
0.10157979279756546,
-0.003072981256991625,
0.03219573572278023,
0.005352990701794624,
-0.06674963235855103,
0.06213633716106415,
-0.08243446797132492,
-0.09152733534574509,
-0.010440479032695293,
-0.09342999011278152,
0.05919385328888893,
0.06361786276102066,
0.06452750414609909,
-0.08319608867168427,
-0.08699297159910202,
0.06599768251180649,
0.08694709092378616,
-0.04389767721295357,
0.021139975637197495,
-0.07992484420537949,
0.07783471047878265,
-0.07904328405857086,
-0.031241806223988533,
-0.15584705770015717,
-0.05027537792921066,
0.007892311550676823,
0.033366817981004715,
0.01911040022969246,
0.02233336865901947,
0.06180812418460846,
0.05810266733169556,
-0.06370805948972702,
-0.0182064026594162,
-0.029058706015348434,
-0.00020378758199512959,
-0.12929630279541016,
-0.19148559868335724,
-0.05672509968280792,
-0.029582975432276726,
0.17460457980632782,
-0.21614107489585876,
0.03439762070775032,
0.000013209391909185797,
0.09514900296926498,
0.0384637787938118,
-0.022949982434511185,
-0.03681236505508423,
0.07170247286558151,
-0.034978024661540985,
-0.058426372706890106,
0.07902128994464874,
0.020563047379255295,
-0.11482222378253937,
-0.016005465760827065,
-0.12301257997751236,
0.15975899994373322,
0.1216045394539833,
-0.0784076675772667,
-0.07393798977136612,
-0.03389131650328636,
-0.04415028542280197,
-0.02868913672864437,
-0.05146118998527527,
0.007848815061151981,
0.14193464815616608,
0.015539668500423431,
0.16351836919784546,
-0.06886567920446396,
-0.04784790053963661,
0.025089066475629807,
-0.030469851568341255,
0.007806910667568445,
0.10957568138837814,
0.10393565148115158,
-0.11286148428916931,
0.15490077435970306,
0.16462258994579315,
-0.06435291469097137,
0.13311690092086792,
-0.03159294277429581,
-0.06676989048719406,
-0.0473029725253582,
-0.023726075887680054,
0.013840904459357262,
0.13741204142570496,
-0.08235730230808258,
-0.009860438294708729,
0.025019925087690353,
0.017841847613453865,
0.0007899576448835433,
-0.22703664004802704,
-0.04862797260284424,
0.041823290288448334,
-0.036211833357810974,
-0.028991922736167908,
-0.014397719874978065,
-0.011174047365784645,
0.09614573419094086,
0.032029327005147934,
-0.09086672961711884,
0.050822433084249496,
0.00039390934398397803,
-0.07801583409309387,
0.1937953382730484,
-0.0642637386918068,
-0.15049265325069427,
-0.1530614048242569,
-0.08760346472263336,
-0.037353504449129105,
0.020305363461375237,
0.027173176407814026,
-0.06142913177609444,
-0.019972940906882286,
-0.07565271109342575,
-0.018878860399127007,
-0.013584987260401249,
0.016792716458439827,
0.009224548935890198,
-0.001057144720107317,
0.066916823387146,
-0.076927050948143,
-0.005631026346236467,
-0.038103051483631134,
-0.026752552017569542,
0.035679981112480164,
0.015182062052190304,
0.114218570291996,
0.15584227442741394,
-0.01164241787046194,
0.010400974191725254,
-0.044687289744615555,
0.21864335238933563,
-0.08890388906002045,
-0.021478213369846344,
0.14661268889904022,
-0.031104881316423416,
0.056087832897901535,
0.13733157515525818,
0.07350340485572815,
-0.08068802952766418,
0.002571525750681758,
0.016236741095781326,
-0.047190625220537186,
-0.18732309341430664,
-0.04191315174102783,
-0.05873526260256767,
-0.004243781790137291,
0.10139921307563782,
0.015855083242058754,
0.019942354410886765,
0.06928200274705887,
0.035773858428001404,
0.08056981861591339,
-0.04085077717900276,
0.07615335285663605,
0.1040266677737236,
0.044378861784935,
0.1364961713552475,
-0.03526437655091286,
-0.05262461304664612,
0.03760777413845062,
0.036597661674022675,
0.2016662359237671,
0.01568562537431717,
0.1641910970211029,
0.03837493807077408,
0.15951445698738098,
0.009580880403518677,
0.04650605842471123,
0.00894853100180626,
-0.03671557828783989,
-0.020729515701532364,
-0.03059365600347519,
-0.03103710152208805,
0.033223677426576614,
-0.01197786908596754,
0.04062287509441376,
-0.1021047905087471,
0.006875202059745789,
0.04423249885439873,
0.23720955848693848,
0.0620327889919281,
-0.3476886451244354,
-0.09992433339357376,
0.009903430938720703,
-0.020975429564714432,
-0.02014482580125332,
0.00442655012011528,
0.1121731847524643,
-0.0994284600019455,
0.01953524723649025,
-0.08524417877197266,
0.09015914797782898,
-0.06676755100488663,
0.03638334572315216,
0.08371564745903015,
0.076542429625988,
-0.004988760221749544,
0.07300553470849991,
-0.2563894987106323,
0.2969544529914856,
0.0172538161277771,
0.04993468150496483,
-0.06386158615350723,
-0.010156984440982342,
0.028968660160899162,
0.07584232836961746,
0.08936048299074173,
-0.007279790472239256,
-0.030984073877334595,
-0.22060920298099518,
-0.06449519097805023,
0.007772837765514851,
0.07295435667037964,
-0.0628521665930748,
0.09464161098003387,
-0.03938893601298332,
0.0037660961970686913,
0.06372606009244919,
0.015473535284399986,
-0.014936272986233234,
-0.09434086084365845,
0.012187490239739418,
0.02448434755206108,
-0.04310521483421326,
-0.0675046443939209,
-0.11139409244060516,
-0.09936255216598511,
0.14240454137325287,
-0.030404765158891678,
-0.030300915241241455,
-0.11691928654909134,
0.08393853902816772,
0.0709340050816536,
-0.08698856830596924,
0.027821118012070656,
0.0006560888141393661,
0.10246230661869049,
0.016114400699734688,
-0.037219688296318054,
0.11126760393381119,
-0.06700579822063446,
-0.1628357321023941,
-0.07537949085235596,
0.1186608225107193,
0.011404537595808506,
0.07357680797576904,
0.0017759841866791248,
0.02781275101006031,
-0.03186025470495224,
-0.06474635004997253,
0.04896042123436928,
-0.030404608696699142,
0.06648533046245575,
-0.0018853915389627218,
-0.025228533893823624,
0.038743600249290466,
-0.05787276476621628,
-0.04162409156560898,
0.17688313126564026,
0.27752766013145447,
-0.10378319025039673,
0.02363092452287674,
0.029910212382674217,
-0.05740077421069145,
-0.19379696249961853,
0.0534406416118145,
0.0472220703959465,
0.023525020107626915,
0.05479646846652031,
-0.16362982988357544,
0.07595274597406387,
0.09122121334075928,
-0.030704999342560768,
0.0850071981549263,
-0.29586261510849,
-0.12537819147109985,
0.08760416507720947,
0.12276840955018997,
0.09560784697532654,
-0.1235618069767952,
-0.03481872379779816,
-0.01942422427237034,
-0.11949031800031662,
0.1118106096982956,
-0.06302856653928757,
0.11346128582954407,
-0.0107572665438056,
0.08794380724430084,
0.010469531640410423,
-0.055965449661016464,
0.1317312866449356,
0.008784990757703781,
0.0870901420712471,
-0.051759183406829834,
-0.050035733729600906,
0.0600472129881382,
-0.0497751459479332,
-0.0032850599382072687,
-0.06451523303985596,
0.01967802084982395,
-0.11539410054683685,
-0.02152220346033573,
-0.0739189013838768,
0.017381850630044937,
-0.02945602871477604,
-0.07107751071453094,
-0.02936251275241375,
0.06110810860991478,
0.04244178533554077,
-0.01658104546368122,
0.1483290195465088,
0.010072045959532261,
0.136368989944458,
0.11419662833213806,
0.0882418230175972,
-0.053321562707424164,
-0.06437960267066956,
-0.020459365099668503,
-0.033633969724178314,
0.05427759513258934,
-0.1457405686378479,
0.026834998279809952,
0.13146038353443146,
0.026205873116850853,
0.14485600590705872,
0.07232699543237686,
-0.028295814990997314,
0.0192742720246315,
0.06684461981058121,
-0.1450251042842865,
-0.0911698117852211,
-0.009741758927702904,
-0.029592128470540047,
-0.13838064670562744,
0.025739211589097977,
0.12127205729484558,
-0.061764005571603775,
-0.009263704530894756,
0.0066359033808112144,
-0.0015974328853189945,
-0.04841427132487297,
0.18001247942447662,
0.06728239357471466,
0.05673143267631531,
-0.08684858679771423,
0.056509826332330704,
0.06707274913787842,
-0.06370893120765686,
-0.008158361539244652,
0.036020249128341675,
-0.09925765544176102,
-0.04033735394477844,
0.01277156826108694,
0.1406225562095642,
-0.086281418800354,
-0.02935723215341568,
-0.14725135266780853,
-0.09962354600429535,
0.05990995466709137,
0.1437893956899643,
0.10471054911613464,
0.005620874464511871,
-0.04910239204764366,
0.0016285431338474154,
-0.1154409870505333,
0.09806947410106659,
0.03930594027042389,
0.07748138904571533,
-0.15216021239757538,
0.1606949269771576,
-0.01371016912162304,
0.051266852766275406,
-0.020028933882713318,
0.029541825875639915,
-0.10178820788860321,
0.01594052091240883,
-0.10509432852268219,
-0.027144409716129303,
-0.03412671014666557,
-0.0020047754514962435,
-0.0044886586256325245,
-0.060047697275877,
-0.04636096581816673,
0.0029743569903075695,
-0.11455775052309036,
-0.02208290435373783,
0.03855257108807564,
0.052210088819265366,
-0.09915801137685776,
-0.040000393986701965,
0.026155469939112663,
-0.058312688022851944,
0.07461000233888626,
0.003170153358951211,
0.034991055727005005,
0.03279949352145195,
-0.09191545844078064,
0.015160548500716686,
0.03522947058081627,
0.017185179516673088,
0.07137299329042435,
-0.0937284380197525,
-0.00883392058312893,
-0.02029620297253132,
0.038133688271045685,
0.03075164370238781,
0.08935289084911346,
-0.12614065408706665,
0.0033869787584990263,
-0.0065309409983456135,
-0.06520219147205353,
-0.06294918060302734,
0.049494896084070206,
0.06792623549699783,
0.04718019813299179,
0.1997133195400238,
-0.06780093908309937,
0.039024535566568375,
-0.2012561708688736,
-0.003065363736823201,
-0.014319614507257938,
-0.10523534566164017,
-0.10612428933382034,
-0.07208500802516937,
0.05954402685165405,
-0.059471044689416885,
0.11698155850172043,
0.03482184186577797,
0.06412822753190994,
0.04002264887094498,
-0.003003936493769288,
0.041533421725034714,
0.016770372167229652,
0.1772177815437317,
0.03528248146176338,
-0.03522773087024689,
0.07081503421068192,
0.042623136192560196,
0.08413641154766083,
0.11668219417333603,
0.1722528338432312,
0.13654941320419312,
0.0160619355738163,
0.07512573152780533,
0.04765112325549126,
-0.04619385302066803,
-0.18680442869663239,
0.017661619931459427,
-0.041243989020586014,
0.09927477687597275,
-0.02652086317539215,
0.2000088095664978,
0.07425189018249512,
-0.1793426126241684,
0.021655626595020294,
-0.06079810857772827,
-0.08192403614521027,
-0.09603558480739594,
-0.08675557374954224,
-0.07974429428577423,
-0.11633538454771042,
0.0005412364844232798,
-0.09698773175477982,
0.005126882344484329,
0.1518334299325943,
-0.0036486603785306215,
-0.01509329117834568,
0.12722137570381165,
-0.007089408114552498,
0.025071153417229652,
0.054443418979644775,
0.011224404908716679,
-0.012878520414233208,
-0.10411011427640915,
-0.06480304896831512,
-0.0137572530657053,
-0.03101225569844246,
0.03445672243833542,
-0.0749705508351326,
-0.022684277966618538,
0.022444305941462517,
-0.00843113660812378,
-0.11149152368307114,
0.006428553722798824,
0.02213628962635994,
0.06348845362663269,
0.04309231787919998,
0.006950520910322666,
0.03101981431245804,
-0.015395795926451683,
0.2262720912694931,
-0.07684368640184402,
-0.048580095171928406,
-0.117264524102211,
0.2462080717086792,
0.0012849133927375078,
-0.024811850860714912,
0.0237041637301445,
-0.0734269991517067,
0.02774062752723694,
0.23002976179122925,
0.19710016250610352,
-0.12472434341907501,
-0.007343771867454052,
0.014918111264705658,
-0.009332379326224327,
-0.027532657608389854,
0.1150049939751625,
0.08441095799207687,
0.012392732314765453,
-0.0961885079741478,
-0.04785433039069176,
-0.06607883423566818,
-0.01813466101884842,
-0.00978049635887146,
0.05596357583999634,
0.03541038930416107,
0.016286788508296013,
-0.056825827807188034,
0.06580521166324615,
-0.046207524836063385,
-0.0995304137468338,
0.06834880262613297,
-0.21616992354393005,
-0.1672075241804123,
-0.014242188073694706,
0.07635574787855148,
-0.004489521961659193,
0.06224461644887924,
-0.0367046482861042,
0.02068568579852581,
0.05938142538070679,
-0.020255835726857185,
-0.06706834584474564,
-0.07556548714637756,
0.10765599459409714,
-0.0800890251994133,
0.21344195306301117,
-0.05892101675271988,
0.06503789126873016,
0.12425526976585388,
0.0584675632417202,
-0.08146416395902634,
0.046453189104795456,
0.05928538367152214,
-0.04036543890833855,
0.031702250242233276,
0.0939360111951828,
-0.036108627915382385,
0.11628895252943039,
0.052300501614809036,
-0.1347203105688095,
0.019327621906995773,
-0.08120965212583542,
-0.054878681898117065,
-0.04825417324900627,
-0.038662347942590714,
-0.04920746758580208,
0.1530056893825531,
0.2009054273366928,
-0.036761436611413956,
-0.01900325156748295,
-0.05691525340080261,
0.0014848012942820787,
0.07911408692598343,
0.03182461857795715,
-0.07479224354028702,
-0.20469102263450623,
0.00011701721814461052,
0.03859871253371239,
-0.01821938157081604,
-0.2457868605852127,
-0.09431388229131699,
0.002298662206158042,
-0.06616362184286118,
-0.06992589682340622,
0.101884625852108,
0.07873667031526566,
0.04867977648973465,
-0.06493356823921204,
-0.03929635509848595,
-0.06422235816717148,
0.12800395488739014,
-0.14527547359466553,
-0.08859479427337646
] |
null | null | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="Jarles/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Jarles/q-FrozenLake-v1-4x4-noSlippery | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-08T18:44:44+00:00 | [] | [] | TAGS
#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 FrozenLake-v1
This is a trained model of a Q-Learning agent playing FrozenLake-v1 .
## Usage
| [
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
"TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
40,
39
] | [
"passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage"
] | [
0.04578453302383423,
-0.08074592798948288,
-0.00430759321898222,
0.10720831900835037,
0.05034215748310089,
-0.040469273924827576,
0.11997015029191971,
0.018999949097633362,
0.20601962506771088,
-0.010012076236307621,
0.1455274522304535,
0.007022971753031015,
-0.006192410364747047,
0.1867983490228653,
0.04572829231619835,
-0.26324528455734253,
0.01831899583339691,
-0.09495259821414948,
-0.07281816750764847,
0.11870454251766205,
0.05470194295048714,
-0.01901467889547348,
-0.0007633853238075972,
0.056141503155231476,
-0.0673527717590332,
0.0007737681735306978,
0.031996939331293106,
-0.012976245954632759,
0.19804789125919342,
-0.02254498563706875,
0.06641989201307297,
0.054705578833818436,
0.0758768692612648,
-0.1998077929019928,
0.0358855277299881,
-0.04215473681688309,
-0.09439758956432343,
-0.03934839740395546,
-0.018780618906021118,
0.05878105387091637,
0.053356342017650604,
0.03858819976449013,
0.058354366570711136,
0.09384993463754654,
-0.0773480236530304,
0.04328357055783272,
0.04280758649110794,
0.024811049923300743,
0.04589218273758888,
-0.0237203948199749,
-0.027002155780792236,
0.08246652781963348,
-0.22182892262935638,
0.10318073630332947,
-0.010159241035580635,
-0.5270710587501526,
-0.00633762264624238,
0.24088262021541595,
0.11517096310853958,
0.05707438662648201,
-0.06903956830501556,
0.10566288232803345,
0.03913382440805435,
-0.007209456991404295,
0.03210983797907829,
0.02150118350982666,
0.12817370891571045,
0.06009242683649063,
-0.09581366181373596,
0.040699947625398636,
0.13722525537014008,
0.012822695076465607,
0.020306183025240898,
-0.08888901025056839,
0.0410032719373703,
-0.03461858257651329,
-0.007679527159780264,
-0.09758518636226654,
0.05478060990571976,
0.012466507963836193,
-0.0934976264834404,
-0.09247440844774246,
-0.04236573353409767,
-0.06708304584026337,
0.11252415925264359,
0.046419668942689896,
-0.0874939113855362,
0.03884070739150047,
-0.06760413944721222,
0.05918780341744423,
-0.16863860189914703,
0.02074250765144825,
-0.06627868115901947,
-0.09376336634159088,
-0.11799788475036621,
-0.01683047041296959,
-0.07946427166461945,
0.009092256426811218,
0.056664444506168365,
0.1447116881608963,
0.22076484560966492,
0.06690320372581482,
0.09728849679231644,
0.07456006109714508,
0.06531001627445221,
0.1538129299879074,
0.10918238013982773,
0.019075315445661545,
-0.015266558155417442,
0.0948706716299057,
-0.06445580720901489,
-0.1351388692855835,
-0.15579092502593994,
0.005488025024533272,
0.0983937531709671,
0.08871900290250778,
-0.044080477207899094,
-0.006702381651848555,
-0.024641724303364754,
0.08566431701183319,
-0.11314457654953003,
-0.024612564593553543,
-0.002267979085445404,
0.06882024556398392,
-0.024801667779684067,
0.020378148183226585,
-0.06242705136537552,
0.12715265154838562,
0.04222423583269119,
-0.059924717992544174,
-0.055308472365140915,
-0.03053177334368229,
-0.014276440255343914,
-0.027539284899830818,
0.02446848154067993,
-0.07659092545509338,
0.04767750948667526,
-0.16766095161437988,
-0.042871296405792236,
-0.04784649610519409,
0.025697942823171616,
-0.03907240927219391,
-0.13557587563991547,
-0.17699143290519714,
-0.048906855285167694,
-0.022438718006014824,
0.03549358621239662,
-0.038111843168735504,
0.006551501806825399,
-0.006318534724414349,
-0.1583600640296936,
0.09783563017845154,
0.09784027189016342,
-0.03643378987908363,
-0.02749447710812092,
0.056263517588377,
-0.07194498926401138,
0.1561182290315628,
-0.21054518222808838,
-0.054014235734939575,
-0.044764336198568344,
-0.06595750898122787,
0.19673264026641846,
0.012690845876932144,
-0.01202624011784792,
0.19873127341270447,
-0.29073721170425415,
-0.06078760325908661,
0.12533614039421082,
-0.07834373414516449,
-0.0936407670378685,
0.06941844522953033,
-0.04206686094403267,
0.023345354944467545,
0.046047765761613846,
0.36345911026000977,
-0.02069227211177349,
-0.16197136044502258,
-0.021782705560326576,
0.13971707224845886,
-0.1184760183095932,
0.059895481914281845,
0.04240793362259865,
0.12543781101703644,
-0.04250509291887283,
-0.018672896549105644,
-0.09023164212703705,
0.05999075248837471,
-0.05241934582591057,
-0.09016361832618713,
-0.03393383324146271,
-0.07645075023174286,
0.13294468820095062,
-0.0629684180021286,
0.05601520463824272,
-0.03255095332860947,
-0.07133250683546066,
-0.050324998795986176,
-0.016492370516061783,
0.04460815340280533,
0.05951254442334175,
-0.12794871628284454,
0.11029167473316193,
0.13025271892547607,
-0.0006193425506353378,
-0.07498852163553238,
-0.17872096598148346,
0.003240168560296297,
0.009576505981385708,
0.039837226271629333,
0.17141658067703247,
0.12209978699684143,
0.033295199275016785,
0.008770671673119068,
-0.06389404833316803,
-0.18276847898960114,
0.058129217475652695,
-0.056212130934000015,
-0.14230976998806,
-0.052409034222364426,
-0.0728459507226944,
0.017381802201271057,
-0.0859743058681488,
-0.017379917204380035,
0.021926190704107285,
0.006908397190272808,
0.02990424446761608,
-0.026645656675100327,
-0.049561817198991776,
0.021254703402519226,
0.06490101665258408,
-0.0037617047782987356,
0.12023693323135376,
0.008277264423668385,
-0.18308481574058533,
0.07930773496627808,
0.08478537946939468,
0.09196605533361435,
0.013250201940536499,
0.02685922384262085,
-0.021522263064980507,
-0.08061408251523972,
-0.054420311003923416,
0.02957955375313759,
0.11417073011398315,
0.1317172348499298,
0.2361993044614792,
0.08753683418035507,
0.04697408527135849,
-0.02164587564766407,
-0.016415923833847046,
0.002810494042932987,
-0.06318057328462601,
-0.029935607686638832,
0.10614971816539764,
0.05865858122706413,
-0.067733034491539,
-0.04576427489519119,
0.09590928256511688,
0.02732124738395214,
0.21205885708332062,
-0.03342745825648308,
0.01286078616976738,
-0.10957037657499313,
-0.06550975888967514,
-0.031982194632291794,
0.09201868623495102,
0.09498392790555954,
0.009755023755133152,
-0.022056059911847115,
-0.04259001836180687,
0.0012916827108711004,
-0.1334889680147171,
-0.10375088453292847,
0.026475343853235245,
0.013400445692241192,
-0.11206940561532974,
0.11674030870199203,
-0.11352457851171494,
0.039504457265138626,
0.06024791672825813,
-0.13837239146232605,
0.04428480193018913,
-0.029713207855820656,
-0.07886212319135666,
0.16866780817508698,
-0.11075661331415176,
-0.094340018928051,
-0.08831550180912018,
0.004082420375198126,
0.0075836325995624065,
-0.03922267258167267,
-0.009283260442316532,
-0.19952571392059326,
-0.005375816952437162,
-0.03544965013861656,
0.013616434298455715,
-0.06988783925771713,
-0.11287739872932434,
-0.010957922786474228,
0.07084179669618607,
-0.043388739228248596,
-0.07803605496883392,
0.007967432029545307,
-0.08923084288835526,
-0.10623309016227722,
0.028189711272716522,
0.019765101373195648,
-0.022883659228682518,
0.16152891516685486,
0.01816628873348236,
0.05626589432358742,
-0.03298520669341087,
0.30665266513824463,
-0.038163769990205765,
0.08371731638908386,
-0.02993497997522354,
-0.07433546334505081,
0.06130730360746384,
-0.022327827289700508,
0.06086638569831848,
-0.020221687853336334,
-0.02362890914082527,
0.0077952733263373375,
-0.08579335361719131,
-0.18365982174873352,
-0.05417544022202492,
0.03724347800016403,
0.195254847407341,
0.031118987128138542,
0.01910330168902874,
-0.0488768145442009,
-0.010547760874032974,
0.1665220558643341,
-0.10005921125411987,
0.04030545800924301,
-0.05366240441799164,
0.11506262421607971,
-0.08640182018280029,
0.06195629760622978,
0.020486772060394287,
0.04266135022044182,
-0.04877188801765442,
0.09486009180545807,
0.0826394334435463,
0.1121082529425621,
-0.02206910029053688,
0.046257395297288895,
0.019012698903679848,
0.07383184134960175,
0.11073657125234604,
0.0368414968252182,
-0.0729052945971489,
0.001982470043003559,
-0.006313489284366369,
-0.039427030831575394,
0.11933320760726929,
0.17963355779647827,
-0.11991413682699203,
-0.05106910318136215,
0.27167606353759766,
0.0031242913100868464,
0.19481229782104492,
-0.01315275114029646,
0.043591804802417755,
-0.04484925419092178,
0.04572054371237755,
-0.05338600277900696,
-0.04086209088563919,
0.2094656229019165,
0.08045925945043564,
-0.17165091633796692,
-0.08549032360315323,
-0.05912299454212189,
0.07081323862075806,
0.10728751868009567,
0.0013539529172703624,
-0.04156802222132683,
0.0004610282776411623,
0.0014198932331055403,
0.08339415490627289,
-0.14520122110843658,
0.11816094070672989,
-0.03172019124031067,
0.05612684786319733,
0.017555562779307365,
-0.045326150953769684,
0.04264266416430473,
0.07474290579557419,
0.26618310809135437,
0.0904107540845871,
-0.040318213403224945,
-0.0892091691493988,
-0.12260187417268753,
0.010461576282978058,
0.029102616012096405,
-0.03534553572535515,
0.0037547778338193893,
-0.020087555050849915,
0.0318896509706974,
0.008264793083071709,
0.016230624169111252,
-0.08987458795309067,
-0.03175399824976921,
-0.027736429125070572,
-0.023839212954044342,
0.10733365267515182,
-0.09495144337415695,
-0.1444292515516281,
-0.15713949501514435,
0.04191131144762039,
-0.0766405463218689,
-0.056593164801597595,
-0.054507751017808914,
-0.05239389091730118,
-0.0311186034232378,
-0.03773957118391991,
0.09099467098712921,
-0.0021037792321294546,
0.14807306230068207,
-0.1920108050107956,
-0.04220759496092796,
0.051812779158353806,
-0.07607918977737427,
-0.08729588985443115,
0.03410962224006653,
0.12136995792388916,
0.05116051807999611,
0.11504370719194412,
0.013609255664050579,
0.09567681699991226,
0.0045484392903745174,
-0.06713183224201202,
0.15302421152591705,
-0.14069625735282898,
-0.27875974774360657,
-0.03836318850517273,
0.016946332529187202,
0.1615200787782669,
-0.05613167956471443,
0.031766023486852646,
0.3335736393928528,
0.27782970666885376,
-0.1428707242012024,
0.25916144251823425,
0.019178593531250954,
0.004398873541504145,
-0.19130495190620422,
-0.10125631093978882,
0.025324683636426926,
0.04740457236766815,
0.12032642960548401,
-0.14564448595046997,
-0.010732659138739109,
-0.04543145373463631,
-0.025908485054969788,
0.10386138409376144,
-0.12300799041986465,
-0.07263197749853134,
0.07765276730060577,
0.039809420704841614,
0.1808302253484726,
0.03932500258088112,
0.0014799144119024277,
0.13626977801322937,
0.06612244248390198,
0.019124457612633705,
0.05216038227081299,
0.08028066903352737,
-0.018944554030895233,
0.14207926392555237,
0.05448179319500923,
-0.02551644667983055,
0.052681710571050644,
-0.0054580713622272015,
-0.03219012916088104,
0.015605825930833817,
-0.183198019862175,
-0.10147556662559509,
-0.0561356320977211,
-0.10798973590135574,
-0.04978342354297638,
0.056853994727134705,
-0.12395523488521576,
-0.007896827533841133,
-0.03841273859143257,
0.03718273714184761,
-0.07831971347332001,
-0.09360362589359283,
-0.036494381725788116,
0.1351792961359024,
0.07210618257522583,
0.04471297934651375,
0.035655103623867035,
-0.07390819489955902,
0.07097936421632767,
0.21671734750270844,
0.08159157633781433,
0.028919655829668045,
-0.19545674324035645,
-0.024042490869760513,
-0.0803457647562027,
0.06306298077106476,
-0.08856996893882751,
-0.016788700595498085,
0.11923003196716309,
0.08616556972265244,
0.05413002520799637,
0.09640096127986908,
-0.045083072036504745,
0.021686913445591927,
0.02684609219431877,
-0.15131035447120667,
-0.18501274287700653,
-0.08534606546163559,
-0.03519878163933754,
0.11561143398284912,
-0.06398691236972809,
0.10897188633680344,
-0.13615410029888153,
0.010051886551082134,
-0.006060056854039431,
0.02693452313542366,
-0.03596206381917,
-0.11251141875982285,
0.15348562598228455,
0.11999429017305374,
-0.06767056882381439,
0.03127254918217659,
-0.09527092427015305,
-0.04423454403877258,
0.12686803936958313,
-0.013623855076730251,
-0.0371493324637413,
-0.054547641426324844,
-0.03628576174378395,
0.15247689187526703,
-0.03436964750289917,
0.008244883269071579,
-0.041229065507650375,
-0.18217355012893677,
0.0798322781920433,
0.09045056998729706,
0.019827889278531075,
-0.031874191015958786,
-0.09797266125679016,
-0.010231015272438526,
-0.0011165260802954435,
0.11730700731277466,
-0.10696814209222794,
-0.10933240503072739,
-0.15144047141075134,
0.06713984161615372,
-0.0007159380475059152,
0.18502596020698547,
-0.06394898891448975,
-0.08904669433832169,
-0.12429379671812057,
0.02344517596065998,
-0.0027384376153349876,
-0.042264558374881744,
0.01618490368127823,
0.07992301136255264,
-0.04095321521162987,
0.02075677551329136,
-0.06651144474744797,
0.06372585147619247,
-0.11786920577287674,
0.09625071287155151,
0.01063506118953228,
0.016993753612041473,
-0.0417880080640316,
-0.01618220843374729,
0.039470795542001724,
-0.057925306260585785,
0.07921463251113892,
0.011758086271584034,
0.0010938759660348296,
0.10196787863969803,
-0.0034960443153977394,
0.06409632414579391,
-0.05372481048107147,
-0.023290161043405533,
0.06578411161899567,
-0.05874887853860855,
-0.03370826691389084,
-0.1573946475982666,
-0.0709633082151413,
0.020051732659339905,
-0.04775108024477959,
0.002077929675579071,
0.03673801198601723,
0.062159497290849686,
-0.06937079131603241,
-0.12125655263662338,
-0.043812792748212814,
-0.028638383373618126,
0.021301284432411194,
0.10829301923513412,
-0.07526551932096481,
0.1547859013080597,
-0.052787959575653076,
-0.00020603960729204118,
0.07437096536159515,
0.04048224538564682,
0.01393822580575943,
-0.10422444343566895,
-0.04698587954044342,
-0.11035211384296417,
0.1502903699874878,
-0.007902312092483044,
-0.03533121198415756,
0.03719403222203255,
-0.11946307867765427,
-0.1572723090648651,
0.03418220207095146,
0.10199101269245148,
0.0448341928422451,
0.025807438418269157,
0.027079269289970398,
-0.04042419046163559,
-0.021270349621772766,
-0.07034418731927872,
0.0882953479886055,
-0.12085357308387756,
-0.09669415652751923,
0.09555385261774063,
0.12178351730108261,
-0.0036850625183433294,
-0.07441367954015732,
0.11554073542356491,
-0.021787192672491074,
0.05525410920381546,
-0.02971339225769043,
0.10308072715997696,
0.0796005055308342,
-0.12273547053337097,
0.005693064536899328,
-0.036891788244247437,
-0.0741485133767128,
-0.12975730001926422,
0.019545545801520348,
-0.061916105449199677,
-0.13383042812347412,
0.12179028987884521,
-0.09376577287912369,
0.030037038028240204,
-0.10506992787122726,
0.021338803693652153,
0.01864001713693142,
0.061665527522563934,
-0.10988292098045349,
0.08575301617383957,
0.13424484431743622,
-0.043199893087148666,
-0.07184189558029175,
-0.12455986440181732,
-0.05022053420543671,
-0.04231856390833855,
-0.13957437872886658,
-0.11600435525178909,
0.0100301094353199,
-0.023418782278895378,
-0.05818291753530502,
0.0015462689334526658,
-0.03659068048000336,
0.008594646118581295,
0.021907730028033257,
0.04032021388411522,
-0.02693161368370056,
0.05134565755724907,
-0.057569269090890884,
-0.052510857582092285,
0.11489357799291611,
0.04113486409187317,
-0.03561042994260788,
-0.052359987050294876,
0.12997733056545258,
-0.11959461867809296,
0.07662346214056015,
-0.020313527435064316,
0.017129231244325638,
-0.06435854732990265,
0.17131924629211426,
0.11673715710639954,
-0.1367570012807846,
-0.005008010193705559,
-0.08210669457912445,
0.020409544929862022,
0.023555370047688484,
0.13693512976169586,
-0.03411718085408211,
-0.0012358218664303422,
-0.1580323874950409,
0.018575575202703476,
-0.18557456135749817,
-0.03716109320521355,
0.04671547934412956,
0.09917585551738739,
0.15293832123279572,
-0.0034432117827236652,
-0.1263325810432434,
0.10424192249774933,
-0.2118520885705948,
0.0907607227563858,
0.05121984705328941,
-0.11874113976955414,
-0.06765396893024445,
-0.06795281916856766,
0.1198519766330719,
0.009196433238685131,
0.2040700763463974,
-0.013615905307233334,
-0.09132910519838333,
-0.07060808688402176,
-0.01980910450220108,
-0.030524181202054024,
0.09714830666780472,
0.041414931416511536,
0.04653804749250412,
0.12821412086486816,
0.00368314771912992,
0.07533777505159378,
0.060310911387205124,
0.02759413793683052,
-0.012300663627684116,
0.04076618701219559,
0.08261215686798096,
-0.14588621258735657,
-0.1659701019525528,
0.1326720416545868,
0.025149408727884293,
0.11792458593845367,
0.03658788278698921,
-0.1549617499113083,
0.06687124073505402,
0.2523096203804016,
-0.11147607117891312,
0.02505038119852543,
0.12737524509429932,
-0.0366884209215641,
0.0672016367316246,
0.1144871786236763,
-0.02633814327418804,
-0.05217865854501724,
-0.011363590136170387,
0.10233135521411896,
0.028660254552960396,
-0.04646271467208862,
-0.02340836264193058,
-0.03373933956027031,
-0.019070526584982872,
-0.011738128960132599,
-0.0909019410610199,
-0.1543993502855301,
-0.10471053421497345,
-0.16619662940502167,
0.04399140924215317,
-0.04626438021659851,
0.13418889045715332,
0.09469578415155411,
-0.012723101302981377,
0.04568437114357948,
0.028575526550412178,
0.07275456190109253,
0.07916246354579926,
-0.02939477376639843,
-0.036159269511699677
] |
null | null | null |
# **Q-Learning** Agent playing1 **Taxi-v3**
This is a trained model of a **Q-Learning** agent playing **Taxi-v3** .
## Usage
```python
model = load_from_hub(repo_id="Jarles/Taxi-v3", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.48 +/- 2.73", "name": "mean_reward", "verified": false}]}]}]} | reinforcement-learning | Jarles/Taxi-v3 | [
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | 2024-02-08T18:47:20+00:00 | [] | [] | TAGS
#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
|
# Q-Learning Agent playing1 Taxi-v3
This is a trained model of a Q-Learning agent playing Taxi-v3 .
## Usage
| [
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
"TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n",
"# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
32,
33
] | [
"passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage"
] | [
0.048862796276807785,
-0.16549694538116455,
-0.005485367961227894,
0.02960980497300625,
0.1345081776380539,
-0.01784728653728962,
0.11895976960659027,
0.07759871333837509,
-0.07461097836494446,
-0.055395450443029404,
0.1418241262435913,
0.09088201075792313,
0.055222880095243454,
0.05699880048632622,
0.09511256217956543,
-0.27440664172172546,
0.048217080533504486,
-0.02918700873851776,
0.05621987581253052,
0.11878681182861328,
0.0670095682144165,
-0.040441032499074936,
0.061956584453582764,
0.11818158626556396,
-0.1018151044845581,
-0.007344264071434736,
0.035402704030275345,
-0.09440053254365921,
0.17413531243801117,
0.07204403728246689,
0.12337774783372879,
0.05132639780640602,
0.179361954331398,
-0.12762396037578583,
0.024310702458024025,
-0.0010275895474478602,
-0.10138072073459625,
-0.03909514099359512,
-0.012415820732712746,
-0.08349097520112991,
0.03230205550789833,
0.23522862792015076,
0.07199250161647797,
0.06632792949676514,
-0.17707863450050354,
-0.06584878265857697,
-0.04375573247671127,
0.069611094892025,
0.14951466023921967,
0.03758616745471954,
-0.033800311386585236,
0.1684885323047638,
-0.2564343810081482,
0.05066783353686333,
0.037275806069374084,
-0.42313119769096375,
0.017119819298386574,
0.1507398933172226,
0.15090937912464142,
0.06909667700529099,
-0.10573802888393402,
0.013512322679162025,
0.051325585693120956,
-0.0005318621988408267,
0.024325110018253326,
0.006554204970598221,
0.15601307153701782,
0.08537693321704865,
-0.1487821787595749,
-0.058576688170433044,
0.17441977560520172,
-0.03788546845316887,
-0.02613203600049019,
-0.039745692163705826,
0.0067160045728087425,
-0.06427708268165588,
-0.004067842848598957,
-0.1777995079755783,
0.00734262028709054,
0.06666424125432968,
-0.014348524622619152,
0.014901017770171165,
-0.035522811114788055,
-0.0966939702630043,
-0.023098144680261612,
-0.08592145889997482,
0.01677769608795643,
-0.006319406442344189,
-0.10187895596027374,
0.05002119392156601,
-0.061138734221458435,
0.0014382408699020743,
-0.05123179033398628,
-0.15047866106033325,
-0.049055423587560654,
-0.03481535613536835,
0.1474713832139969,
-0.0044205985032022,
-0.01873963139951229,
-0.03164304047822952,
0.15474793314933777,
0.049551334232091904,
-0.05370146036148071,
0.05625450983643532,
0.07605006545782089,
0.23867930471897125,
0.10401605814695358,
0.10196955502033234,
-0.06798075139522552,
0.10180158913135529,
-0.12330973148345947,
-0.08915644884109497,
-0.17508824169635773,
0.11820860952138901,
0.00015364694991149008,
0.1317785084247589,
-0.12023144960403442,
0.07898581773042679,
-0.067511186003685,
0.013453764840960503,
0.01636839471757412,
0.0820009782910347,
-0.012399360537528992,
0.10676060616970062,
-0.005061192903667688,
-0.06941985338926315,
0.014177112840116024,
0.05935845896601677,
0.03754841163754463,
-0.038601722568273544,
-0.03192409873008728,
-0.05762290954589844,
-0.05065649375319481,
-0.10128600150346756,
-0.06447898596525192,
0.018573462963104248,
-0.007677143905311823,
-0.1833900660276413,
-0.06407523155212402,
0.00897200871258974,
0.015712225809693336,
-0.03988850116729736,
-0.05148044601082802,
-0.15265507996082306,
-0.042461175471544266,
-0.015450406819581985,
-0.03500641882419586,
-0.06214277446269989,
-0.0383245050907135,
0.046435944736003876,
-0.07560601085424423,
0.013364278711378574,
0.023342855274677277,
0.05405820533633232,
-0.025881100445985794,
0.06068144738674164,
-0.08357544988393784,
0.09493788331747055,
-0.1540430635213852,
-0.03271956741809845,
-0.025445878505706787,
-0.041183918714523315,
0.1752462536096573,
0.06099751964211464,
-0.015994304791092873,
0.15260063111782074,
-0.17141541838645935,
-0.058121129870414734,
0.15596486628055573,
0.008629098534584045,
-0.09967197477817535,
-0.003560945624485612,
-0.09397093951702118,
0.1428760588169098,
0.08571921288967133,
0.2478504776954651,
0.12005335837602615,
-0.22748184204101562,
0.055358242243528366,
0.12515293061733246,
-0.14365963637828827,
0.10365243256092072,
0.07344598323106766,
0.005470725707709789,
-0.18886831402778625,
-0.06843198090791702,
-0.06121627986431122,
0.1053021252155304,
-0.08522345870733261,
-0.0776243582367897,
0.09323626756668091,
-0.05086790770292282,
0.24641476571559906,
-0.028281206265091896,
0.06174173951148987,
-0.026681531220674515,
-0.1389324963092804,
-0.01723906397819519,
0.060955192893743515,
0.05258452147245407,
-0.024835573509335518,
-0.25895482301712036,
0.13646544516086578,
0.048650871962308884,
0.025074828416109085,
0.004106190986931324,
-0.05691491439938545,
0.016934165731072426,
0.1511998474597931,
0.020012924447655678,
0.13717477023601532,
0.027723990380764008,
0.0706823319196701,
-0.006239562761038542,
-0.10560829937458038,
-0.04169593006372452,
0.061916545033454895,
-0.08518962562084198,
-0.06641357392072678,
0.011197872459888458,
-0.06935211271047592,
-0.11783787608146667,
-0.12166737765073776,
-0.026334572583436966,
-0.02980303019285202,
-0.07444227486848831,
0.02368103712797165,
0.06536602973937988,
-0.06702698022127151,
-0.0023908785078674555,
0.007125476840883493,
-0.011537045240402222,
0.16434046626091003,
0.011393417604267597,
-0.007796820718795061,
0.1328643560409546,
-0.11533161997795105,
0.12461213022470474,
0.049438029527664185,
-0.024806302040815353,
-0.04662557691335678,
0.0014137453399598598,
-0.057529181241989136,
0.029044216498732567,
-0.04390640929341316,
0.02774495631456375,
0.20111067593097687,
0.02772962674498558,
0.11389166116714478,
-0.0656520202755928,
0.04385066404938698,
-0.007961965166032314,
-0.009693224914371967,
0.018563594669103622,
0.07608018070459366,
0.07813210040330887,
-0.1324140727519989,
0.02262016013264656,
0.22455167770385742,
0.1385764330625534,
0.18313980102539062,
-0.010877152904868126,
0.06325667351484299,
-0.04875868931412697,
0.027505528181791306,
0.024100203067064285,
0.10314226150512695,
-0.10732068121433258,
-0.0322517491877079,
-0.025407759472727776,
0.023599207401275635,
-0.08197105675935745,
-0.1055799350142479,
-0.090115025639534,
0.01222382951527834,
-0.03125503659248352,
-0.15570329129695892,
0.13300658762454987,
-0.10451057553291321,
0.01802753657102585,
0.04692702740430832,
-0.22163605690002441,
0.11530312895774841,
0.014291439205408096,
-0.10303618758916855,
0.11281087249517441,
-0.12051989883184433,
-0.08699832111597061,
-0.05777236074209213,
-0.18658851087093353,
0.05280197039246559,
0.04673841595649719,
0.05166793242096901,
-0.18521739542484283,
0.024835903197526932,
0.05545609071850777,
0.13426995277404785,
-0.09743253141641617,
-0.07142634689807892,
-0.15038461983203888,
0.016068490222096443,
-0.033661190420389175,
-0.16029728949069977,
-0.005609163548797369,
-0.032781440764665604,
-0.18849676847457886,
-0.04539939761161804,
-0.15086813271045685,
-0.034627582877874374,
0.20464378595352173,
0.026907702907919884,
0.09480511397123337,
-0.07926445454359055,
0.3802889585494995,
-0.042039383202791214,
-0.06146497279405594,
-0.01321389526128769,
-0.07072482258081436,
0.02512686513364315,
0.13271741569042206,
0.0036099457647651434,
-0.017886579036712646,
-0.0037857077550143003,
0.0024592927657067776,
-0.06234965845942497,
-0.13400450348854065,
0.0028710351325571537,
0.03905198723077774,
0.1874423623085022,
0.004639793653041124,
0.06659388542175293,
0.03133883699774742,
0.057546284049749374,
0.07748064398765564,
0.030926106497645378,
0.0011591583024710417,
-0.01591806672513485,
0.06604493409395218,
-0.11684755235910416,
0.042466625571250916,
-0.030429253354668617,
-0.10143838077783585,
-0.013183288276195526,
0.07950251549482346,
0.12755028903484344,
0.17849206924438477,
-0.04790908098220825,
0.17489230632781982,
0.13580141961574554,
0.16576050221920013,
0.049315933138132095,
-0.020801831036806107,
-0.08773037046194077,
-0.06118565797805786,
0.004774159751832485,
-0.031952597200870514,
0.04869702458381653,
0.3231290578842163,
0.037619613111019135,
-0.09036035090684891,
0.11149907857179642,
0.009480619803071022,
0.05359881371259689,
0.022797370329499245,
-0.11162138730287552,
0.11170321702957153,
0.07968773692846298,
-0.06341761350631714,
-0.07602835446596146,
0.16758501529693604,
-0.1109386757016182,
-0.26646625995635986,
-0.11410990357398987,
-0.012305386364459991,
0.07903840392827988,
0.005651174578815699,
0.05498376116156578,
-0.11829282343387604,
-0.16034497320652008,
-0.034191906452178955,
0.1335442066192627,
-0.3077351450920105,
0.2065143585205078,
-0.0198091771453619,
0.06707923114299774,
-0.039657969027757645,
-0.07026876509189606,
0.09694647043943405,
0.13174086809158325,
0.29124146699905396,
0.01396956667304039,
0.04841272905468941,
-0.15176129341125488,
-0.0976925864815712,
0.0018439020495861769,
0.015482662245631218,
-0.02563396655023098,
0.028520405292510986,
-0.0540912002325058,
0.008404579944908619,
-0.018086453899741173,
0.2102297693490982,
-0.11316607892513275,
0.004344627261161804,
-0.06968966871500015,
-0.11707738786935806,
0.19409789144992828,
-0.07178345322608948,
-0.04543264955282211,
-0.14959357678890228,
-0.15512511134147644,
-0.004174166824668646,
-0.02413962036371231,
-0.019664527848362923,
-0.17603960633277893,
-0.18804074823856354,
-0.05204557999968529,
-0.005645004566758871,
-0.003464865731075406,
0.05867868289351463,
-0.07517234236001968,
-0.04805335775017738,
0.1009904220700264,
-0.07743175327777863,
-0.056063808500766754,
-0.1103200614452362,
0.1391381323337555,
0.06248528137803078,
0.16743235290050507,
0.05907081440091133,
0.0006117874872870743,
0.11471151560544968,
-0.02913086675107479,
0.11103474348783493,
-0.11291708797216415,
-0.17145049571990967,
-0.08334989100694656,
-0.018775060772895813,
0.09519003331661224,
-0.04789286106824875,
0.0028788831550627947,
0.2550160884857178,
0.14880181849002838,
-0.0897710770368576,
0.27680760622024536,
0.04414956644177437,
-0.09375058114528656,
-0.18432219326496124,
-0.15961645543575287,
0.03759992495179176,
0.060025621205568314,
0.13095876574516296,
-0.057205069810152054,
-0.08483537286520004,
-0.08492398262023926,
-0.07478608191013336,
-0.13140805065631866,
-0.24232175946235657,
-0.030598774552345276,
0.22874866425991058,
0.08656918257474899,
0.08219650387763977,
-0.012482990510761738,
-0.01186054851859808,
0.00526038184762001,
0.02680150233209133,
0.12018456310033798,
-0.13341329991817474,
0.11107480525970459,
0.022198403254151344,
0.044267985969781876,
0.009712530300021172,
0.07929777354001999,
0.03375575691461563,
-0.003218587953597307,
-0.0006439819699153304,
-0.0988350659608841,
-0.2596651017665863,
0.0816885456442833,
-0.01623627357184887,
-0.09960969537496567,
0.014988959766924381,
0.02061903104186058,
-0.2089255303144455,
0.011128270998597145,
-0.019883770495653152,
-0.03150356933474541,
-0.06483490765094757,
-0.10664787143468857,
-0.056551624089479446,
0.04928823933005333,
0.10853826254606247,
0.011660109274089336,
0.05354316532611847,
-0.0404130220413208,
0.07917837053537369,
0.0826287642121315,
0.15132710337638855,
0.06795957684516907,
-0.190711110830307,
-0.10953907668590546,
-0.0414445661008358,
0.12121522426605225,
-0.12505418062210083,
0.036917757242918015,
0.053161121904850006,
-0.016534561291337013,
0.14621229469776154,
0.1070784479379654,
-0.07452095299959183,
0.11915595084428787,
0.08904775977134705,
-0.04094788804650307,
-0.23367151618003845,
-0.07120766490697861,
0.11133213341236115,
0.07195597887039185,
-0.03961895406246185,
0.018120890483260155,
-0.04960581287741661,
-0.013980977237224579,
0.048759616911411285,
-0.0538676381111145,
-0.07230538129806519,
0.004421027842909098,
0.1247575581073761,
0.1029362753033638,
-0.04655474051833153,
0.01296416949480772,
0.037371400743722916,
0.003788623260334134,
0.04730486497282982,
0.0407949760556221,
-0.08269952982664108,
-0.04124005511403084,
0.02782733179628849,
0.37552911043167114,
-0.010165480896830559,
-0.020456433296203613,
0.018555615097284317,
-0.19949445128440857,
0.09135842323303223,
0.13205479085445404,
0.04697350412607193,
0.004247748292982578,
-0.08139242231845856,
0.026877427473664284,
-0.010625290684401989,
0.09936143457889557,
-0.07806670665740967,
-0.05493134260177612,
-0.21631066501140594,
-0.025010565295815468,
0.017490221187472343,
0.24077683687210083,
-0.08458559215068817,
-0.12801732122898102,
-0.20628872513771057,
0.13128381967544556,
-0.11333390325307846,
-0.03695881739258766,
-0.024473199620842934,
0.03926658630371094,
-0.01989821158349514,
0.06291737407445908,
-0.0710630789399147,
0.006373001262545586,
-0.11024709790945053,
0.055267609655857086,
0.04204455390572548,
0.1229788213968277,
0.014207782223820686,
0.02016810141503811,
0.05822525918483734,
-0.01837925612926483,
0.07173580676317215,
-0.06203491613268852,
-0.04550490900874138,
0.14224006235599518,
-0.020255116745829582,
-0.04152837023139,
-0.0483345128595829,
-0.036874305456876755,
0.11981741338968277,
-0.05059147998690605,
-0.007141099311411381,
-0.054929375648498535,
-0.06906463205814362,
0.03462086617946625,
-0.009175732731819153,
-0.008798843249678612,
0.06801853328943253,
0.04024988040328026,
-0.026994358748197556,
0.005263668950647116,
0.03447828069329262,
-0.10330043733119965,
-0.04955084249377251,
0.16955432295799255,
-0.0749620869755745,
0.10274054110050201,
-0.031069839373230934,
0.018015999346971512,
0.005847334861755371,
-0.022399673238396645,
-0.015360680408775806,
-0.1457086056470871,
-0.06137600541114807,
-0.09489979594945908,
0.11565322428941727,
0.08146517723798752,
0.03358805552124977,
0.04274565726518631,
0.019532648846507072,
-0.04414922371506691,
-0.038583990186452866,
0.12961317598819733,
0.08133101463317871,
0.012996876612305641,
0.01137041300535202,
0.01941833831369877,
-0.020302120596170425,
0.0028480992186814547,
-0.01250747125595808,
-0.07239153981208801,
-0.05874783173203468,
0.09400010108947754,
0.1600283533334732,
-0.06127211079001427,
-0.13325586915016174,
-0.020593497902154922,
0.04988488554954529,
0.0014717020094394684,
-0.08777432143688202,
0.04833676666021347,
0.15805292129516602,
-0.05623878911137581,
0.03216489031910896,
-0.09984751045703888,
-0.07263360917568207,
-0.16060975193977356,
-0.10029061883687973,
-0.06092562898993492,
-0.28350353240966797,
0.09752398729324341,
0.006392303854227066,
-0.014731393195688725,
0.059529416263103485,
0.051305368542671204,
-0.052508849650621414,
0.07068239152431488,
-0.18146829307079315,
-0.007054794579744339,
0.03497592359781265,
-0.13212306797504425,
0.02475893869996071,
-0.2378365397453308,
0.10198072344064713,
-0.04623803123831749,
-0.1519704908132553,
-0.04004510119557381,
0.0641569048166275,
-0.09540136158466339,
-0.01822364516556263,
-0.0475153923034668,
-0.01922670193016529,
0.01624443754553795,
-0.009348669089376926,
-0.031147832050919533,
0.13716529309749603,
0.02827494591474533,
-0.03268734738230705,
0.005254602525383234,
0.0223685409873724,
0.03955082967877388,
-0.0969657450914383,
-0.05986930429935455,
0.08311155438423157,
-0.031056145206093788,
0.14728976786136627,
0.000341245875461027,
0.04181376099586487,
-0.06758682429790497,
0.2593761384487152,
0.2023983597755432,
-0.12479214370250702,
0.008118697442114353,
-0.021801479160785675,
0.012670028023421764,
-0.041751839220523834,
0.13110700249671936,
0.013386172242462635,
0.12186761200428009,
-0.17513342201709747,
-0.01036517322063446,
-0.0818324014544487,
-0.04501292482018471,
0.06702108681201935,
0.14714950323104858,
0.15742522478103638,
0.03436789661645889,
-0.07328428328037262,
0.06722653657197952,
-0.30119743943214417,
0.20540550351142883,
-0.1346001923084259,
-0.01498429011553526,
-0.040251150727272034,
-0.058389630168676376,
0.061147745698690414,
0.11309876292943954,
0.10832664370536804,
-0.021150551736354828,
-0.0905047357082367,
-0.04486766457557678,
-0.039378076791763306,
-0.13019338250160217,
-0.02718670479953289,
0.1654091775417328,
0.06799814850091934,
0.31520840525627136,
-0.017577875405550003,
0.07702425122261047,
0.034410297870635986,
0.06451138854026794,
0.004519328009337187,
0.09537279605865479,
0.07960964739322662,
-0.06345855444669724,
-0.07373003661632538,
-0.001637450186535716,
0.05033271387219429,
0.14567798376083374,
-0.03826142102479935,
-0.18691548705101013,
0.15858715772628784,
0.07192251086235046,
-0.13762691617012024,
-0.05777517706155777,
0.08409425616264343,
-0.0739973932504654,
0.0550808347761631,
0.08115427941083908,
0.015876613557338715,
-0.017793258652091026,
-0.004664506763219833,
0.06074233725667,
0.024694660678505898,
-0.02343848906457424,
0.003570882137864828,
-0.08337053656578064,
-0.04151543974876404,
0.07267895340919495,
-0.0844460055232048,
-0.20546193420886993,
-0.0957019031047821,
-0.07551700621843338,
0.030557552352547646,
-0.0649830624461174,
0.12575586140155792,
0.1717868149280548,
0.0593598335981369,
-0.03307248651981354,
-0.10721943527460098,
-0.035562749952077866,
0.07602505385875702,
-0.044773899018764496,
-0.09409699589014053
] |
null | null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.8.2 | {"library_name": "peft", "base_model": "mistralai/Mistral-7B-Instruct-v0.2"} | null | pikaduck/Mistral-7B-Instruct-v0.2-cv-editor | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"region:us"
] | 2024-02-08T18:48:42+00:00 | [
"1910.09700"
] | [] | TAGS
#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-Instruct-v0.2 #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
### Framework versions
- PEFT 0.8.2 | [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
"TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-Instruct-v0.2 #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact",
"### Framework versions\n\n- PEFT 0.8.2"
] | [
42,
6,
3,
54,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4,
11
] | [
"passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-Instruct-v0.2 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2"
] | [
-0.12410319596529007,
0.2010166496038437,
-0.0028893498238176107,
0.02710896171629429,
0.08296072483062744,
0.01904967427253723,
0.052783653140068054,
0.12995868921279907,
0.03010331466794014,
0.11474749445915222,
0.07281026989221573,
0.1246524378657341,
0.11124330013990402,
0.20343951880931854,
0.005492560565471649,
-0.16418711841106415,
0.028107086196541786,
-0.08119428902864456,
0.016911618411540985,
0.12729473412036896,
0.14302164316177368,
-0.10105469077825546,
0.07827022671699524,
-0.019680403172969818,
0.003734445432201028,
-0.03514719754457474,
-0.0676296204328537,
-0.017137311398983,
0.05334393307566643,
0.0260662529617548,
0.06058273836970329,
-0.016222983598709106,
0.09149807691574097,
-0.2670385241508484,
0.018769847229123116,
0.04243418201804161,
0.008953949436545372,
0.08478713780641556,
0.1020808219909668,
-0.039719514548778534,
0.11745807528495789,
-0.020938202738761902,
0.13943594694137573,
0.09182620793581009,
-0.09294260293245316,
-0.2322562038898468,
-0.06288547813892365,
0.07646185159683228,
0.18977797031402588,
0.08811664581298828,
-0.04585372284054756,
0.13324467837810516,
-0.06533069908618927,
0.023692261427640915,
0.05868341028690338,
-0.10365474224090576,
-0.06595955044031143,
0.06907257437705994,
0.13893724977970123,
0.06974613666534424,
-0.11929887533187866,
-0.03888959065079689,
0.03501056507229805,
0.04709905385971069,
0.05147114023566246,
0.008162870071828365,
0.16042667627334595,
0.03068653680384159,
-0.14914821088314056,
-0.05700765177607536,
0.14465714991092682,
0.006209301296621561,
-0.04471925273537636,
-0.21986714005470276,
-0.0017105763545259833,
-0.09096468240022659,
-0.029785465449094772,
-0.057076796889305115,
0.033241648226976395,
0.012998118996620178,
0.13023634254932404,
-0.041996389627456665,
-0.09516381472349167,
-0.02276061847805977,
0.10196305811405182,
0.0537891685962677,
0.02434694766998291,
-0.023223552852869034,
0.014226345345377922,
0.12359470129013062,
0.07607105374336243,
-0.1321575939655304,
-0.06573349237442017,
-0.07548417896032333,
-0.039750099182128906,
-0.032794807106256485,
0.03757920861244202,
0.016408022493124008,
0.06335418671369553,
0.27120932936668396,
-0.01988113671541214,
0.06158364564180374,
0.05218353867530823,
0.019749661907553673,
0.03842940181493759,
0.10705602914094925,
-0.03324436768889427,
-0.1591789275407791,
-0.013680425472557545,
0.1027100533246994,
-0.0029301634058356285,
-0.031166184693574905,
-0.046749986708164215,
0.029272882267832756,
0.045267410576343536,
0.11936883628368378,
0.1145344078540802,
-0.021738169714808464,
-0.08144217729568481,
-0.05754102021455765,
0.19460640847682953,
-0.1561046689748764,
0.042514022439718246,
0.02497181110084057,
-0.0033909329213202,
-0.06405963003635406,
0.005479773972183466,
0.02067570947110653,
-0.030656415969133377,
0.0668933168053627,
-0.06567777693271637,
-0.042296040803194046,
-0.1267937272787094,
-0.029879439622163773,
0.03426501527428627,
0.019413180649280548,
-0.04652857407927513,
-0.04621002823114395,
-0.08210966736078262,
-0.11106777936220169,
0.10873367637395859,
-0.053001683205366135,
-0.05658300966024399,
-0.027196334674954414,
-0.08428024500608444,
0.020672641694545746,
0.030444534495472908,
0.07233098149299622,
-0.02653544396162033,
0.048880793154239655,
-0.0018393327482044697,
0.06260377168655396,
0.08586197346448898,
0.031089920550584793,
-0.08586600422859192,
0.06668511033058167,
-0.1940956711769104,
0.07390007376670837,
-0.07962337881326675,
0.03753335773944855,
-0.1659746617078781,
-0.009758904576301575,
0.00730910012498498,
0.03118695318698883,
0.042738255113363266,
0.1615987867116928,
-0.2184610217809677,
-0.023008311167359352,
0.15677569806575775,
-0.11259480565786362,
-0.13492459058761597,
0.04524976387619972,
-0.03756643459200859,
0.18281100690364838,
0.02820627950131893,
0.01388672087341547,
0.08356999605894089,
-0.1607387214899063,
-0.024809645488858223,
-0.028796304017305374,
0.006961580831557512,
0.07446914911270142,
0.08818746358156204,
-0.09269585460424423,
-0.004778424743562937,
0.009624771773815155,
-0.06440450251102448,
-0.01457137055695057,
-0.042774610221385956,
-0.10505311191082001,
0.003389041405171156,
-0.08510596305131912,
0.021626079455018044,
0.006051310338079929,
-0.10088779777288437,
-0.009337589144706726,
-0.15419022738933563,
-0.06117285043001175,
0.08837836980819702,
0.003436786588281393,
-0.020544283092021942,
-0.10206453502178192,
0.0655185803771019,
-0.039029963314533234,
-0.02375018782913685,
-0.14741362631320953,
-0.020607884973287582,
0.017986992374062538,
-0.14173021912574768,
-0.00777188828215003,
-0.11856295168399811,
0.0669332817196846,
0.004967313259840012,
-0.046384334564208984,
-0.045790329575538635,
-0.002593749901279807,
-0.0021016946993768215,
-0.05897143855690956,
-0.22953595221042633,
-0.026958411559462547,
-0.05431513488292694,
0.15697221457958221,
-0.23199377954006195,
0.04287084937095642,
-0.0006575930747203529,
0.1138000339269638,
0.002258235588669777,
-0.06576001644134521,
0.021034864708781242,
-0.06367801874876022,
-0.027234621345996857,
-0.0694202333688736,
-0.0029602739959955215,
-0.0005214597331359982,
-0.022617410868406296,
0.016286877915263176,
-0.11958853155374527,
-0.06237909197807312,
0.09583486616611481,
0.056932609528303146,
-0.15001294016838074,
0.007309744134545326,
-0.03755912929773331,
-0.05939666926860809,
-0.07143530994653702,
-0.07032694667577744,
0.08730877190828323,
0.057858362793922424,
0.03947750851511955,
-0.07754365354776382,
-0.06501077115535736,
0.0033470625057816505,
-0.02844887226819992,
-0.014653410762548447,
0.12209482491016388,
0.0772131085395813,
-0.09792868047952652,
0.09060956537723541,
0.07679019123315811,
0.018177559599280357,
0.06825857609510422,
-0.02470150962471962,
-0.1070401519536972,
-0.02786204218864441,
0.05909346789121628,
0.012765456922352314,
0.16712158918380737,
-0.07411017268896103,
0.05555039271712303,
0.04832817614078522,
-0.04894062504172325,
0.04458760470151901,
-0.08713903278112411,
0.005571435205638409,
0.002418435411527753,
-0.012463239021599293,
0.03790893405675888,
-0.020551640540361404,
0.00434570387005806,
0.08263204991817474,
0.06021817401051521,
0.022444281727075577,
0.016874724999070168,
-0.03545399755239487,
-0.14367610216140747,
0.1808919757604599,
-0.09178680181503296,
-0.23821739852428436,
-0.15497145056724548,
0.05720202997326851,
0.05665547400712967,
-0.011232015676796436,
0.031683921813964844,
-0.05561789125204086,
-0.09925535321235657,
-0.08572676032781601,
0.008156844414770603,
0.036012426018714905,
-0.06058713048696518,
-0.06670568138360977,
0.03893832117319107,
0.039515554904937744,
-0.11778721213340759,
0.027364082634449005,
0.06470848619937897,
-0.010850037448108196,
-0.0012279885122552514,
0.049670420587062836,
0.09720010310411453,
0.19374659657478333,
0.0005513874348253012,
0.0006568313110619783,
0.06210177391767502,
0.27663084864616394,
-0.15596731007099152,
0.1064300537109375,
0.15220075845718384,
-0.06765604764223099,
0.07136226445436478,
0.18558739125728607,
0.023910965770483017,
-0.09549623727798462,
0.024791844189167023,
0.02822188287973404,
-0.018860895186662674,
-0.2736659049987793,
-0.053117986768484116,
-0.017228785902261734,
-0.08058833330869675,
0.08031898736953735,
0.08753632754087448,
0.08681897819042206,
0.038510728627443314,
-0.059503063559532166,
-0.10761744529008865,
0.02825724519789219,
0.11377029865980148,
-0.011964253149926662,
0.0015402027638629079,
0.08238743990659714,
-0.046604812145233154,
0.005163230467587709,
0.08555135875940323,
-0.019335702061653137,
0.1341121792793274,
0.05706215649843216,
0.09903840720653534,
0.08108504116535187,
0.09050989896059036,
-0.009613562375307083,
0.03516557812690735,
0.009375021792948246,
0.0219536405056715,
0.020888591185212135,
-0.08731455355882645,
0.006910230498760939,
0.1127997413277626,
0.02171499654650688,
0.022362161427736282,
0.01768806017935276,
-0.05411510169506073,
0.03689267858862877,
0.1980794221162796,
0.029457712545990944,
-0.2176842987537384,
-0.08729077875614166,
0.05134613811969757,
-0.07673999667167664,
-0.16056466102600098,
-0.007958224974572659,
0.018778232857584953,
-0.1603054255247116,
0.01210764516144991,
-0.04447836056351662,
0.10479702800512314,
-0.07418752461671829,
-0.038989678025245667,
0.1098288968205452,
0.04774298518896103,
-0.02324816770851612,
0.05178719758987427,
-0.1959376484155655,
0.10592316091060638,
0.026451386511325836,
0.07147755473852158,
-0.08764348924160004,
0.09223330765962601,
-0.0020246999338269234,
-0.01933877356350422,
0.16840746998786926,
-0.0034848875366151333,
-0.06223803386092186,
-0.08466429263353348,
-0.08692798763513565,
-0.0043763634748756886,
0.07759454101324081,
-0.12676900625228882,
0.07821814715862274,
-0.03891296684741974,
-0.02440192736685276,
-0.006242895498871803,
-0.08496065437793732,
-0.12896519899368286,
-0.15303200483322144,
0.05879655480384827,
-0.09301026165485382,
0.030709851533174515,
-0.08292851597070694,
-0.05555327609181404,
0.027738114818930626,
0.18124821782112122,
-0.20799577236175537,
-0.1057131215929985,
-0.1418704092502594,
-0.11038675904273987,
0.16187523305416107,
-0.0420379601418972,
0.08532092720270157,
-0.005378061905503273,
0.16069838404655457,
0.005667721852660179,
-0.010719613172113895,
0.08596894890069962,
-0.09526142477989197,
-0.1912127435207367,
-0.04924310743808746,
0.1647324413061142,
0.13891617953777313,
0.02700139582157135,
-0.008506519719958305,
0.032515011727809906,
-0.06922445446252823,
-0.1091255396604538,
0.028861921280622482,
0.15426728129386902,
0.07169631123542786,
-0.019406909123063087,
-0.037406403571367264,
-0.10578691959381104,
-0.0649622455239296,
-0.04465539753437042,
-0.004816913977265358,
0.20358677208423615,
-0.06573144346475601,
0.14491750299930573,
0.11280136555433273,
-0.059244461357593536,
-0.21086691319942474,
0.03252866119146347,
0.04201432317495346,
0.019656194373965263,
0.04145071655511856,
-0.18712441623210907,
0.09135694801807404,
-0.01298889983445406,
-0.08375605195760727,
0.1736239343881607,
-0.18779730796813965,
-0.1318461000919342,
0.11317607760429382,
0.025007523596286774,
-0.20633426308631897,
-0.14734181761741638,
-0.10163825750350952,
-0.017777442932128906,
-0.12265390902757645,
0.03728831931948662,
0.020455244928598404,
0.008876300416886806,
0.01366375107318163,
0.020413819700479507,
0.04369619861245155,
-0.0538116991519928,
0.21189740300178528,
-0.04310271516442299,
-0.006460869684815407,
-0.05677029862999916,
-0.07586601376533508,
0.015245414339005947,
-0.053524162620306015,
0.12220840156078339,
-0.010396245867013931,
0.03562389686703682,
-0.16481198370456696,
-0.0445837676525116,
-0.05417162925004959,
0.03582264110445976,
-0.09256159514188766,
-0.080452561378479,
-0.040812499821186066,
0.0905209332704544,
0.08815445750951767,
-0.016387157142162323,
0.01054933201521635,
-0.09582687169313431,
0.08853982388973236,
0.19723457098007202,
0.19489292800426483,
0.0788634791970253,
-0.053322628140449524,
0.030755653977394104,
-0.03706871345639229,
0.04420938715338707,
-0.2084701955318451,
0.040779855102300644,
0.0648471787571907,
0.018882639706134796,
0.06577204912900925,
-0.007609706372022629,
-0.15836678445339203,
-0.08151598274707794,
0.08579933643341064,
-0.058091774582862854,
-0.16868966817855835,
-0.026639780029654503,
0.03904619440436363,
-0.2138693928718567,
-0.047230105847120285,
0.036008212715387344,
-0.01598323881626129,
-0.04123258218169212,
0.024398166686296463,
0.08290894329547882,
-0.023660460487008095,
0.09724966436624527,
0.0880955383181572,
0.09174808859825134,
-0.09182809293270111,
0.05387898534536362,
0.07673577964305878,
-0.03097637929022312,
0.027789825573563576,
0.1268376260995865,
-0.03726270794868469,
-0.049149561673402786,
0.08226066827774048,
0.11605004966259003,
-0.0031336680985987186,
-0.056250110268592834,
0.00587568711489439,
-0.044257938861846924,
0.058898504823446274,
0.10247617214918137,
0.02873624674975872,
-0.004942453000694513,
0.07650740444660187,
0.02757342718541622,
-0.09146638214588165,
0.12312479317188263,
0.05862504616379738,
0.025292271748185158,
-0.04693218320608139,
-0.03961494565010071,
-0.016947977244853973,
-0.00788995809853077,
-0.016896843910217285,
-0.0004530868027359247,
-0.08577961474657059,
0.005245373118668795,
-0.12706540524959564,
0.017588218674063683,
-0.07660529017448425,
0.0019533121958374977,
0.033341679722070694,
-0.0503024123609066,
0.002360518556088209,
-0.005558492615818977,
-0.07387258857488632,
-0.05547299236059189,
-0.02121906541287899,
0.07999098300933838,
-0.14151383936405182,
0.03503906726837158,
0.07881216704845428,
-0.1045379489660263,
0.06721489131450653,
-0.008728310465812683,
0.009360571391880512,
0.00017638974532019347,
-0.14702174067497253,
0.058483466506004333,
-0.025967372581362724,
-0.0068121906369924545,
-0.0020625328179448843,
-0.1904023438692093,
-0.007335529197007418,
-0.03533678874373436,
-0.0687810629606247,
0.016880115494132042,
0.0009893294190987945,
-0.12199243158102036,
0.11150610446929932,
0.006448720116168261,
-0.06400388479232788,
-0.01921859197318554,
0.039566073566675186,
0.08283372223377228,
-0.007640472147613764,
0.1212635189294815,
-0.029633942991495132,
0.08298641443252563,
-0.17615458369255066,
-0.006950009148567915,
-0.016769153997302055,
0.057643767446279526,
-0.016023054718971252,
-0.04023636505007744,
0.05745045468211174,
-0.02252732589840889,
0.16358813643455505,
-0.006107073277235031,
0.07168685644865036,
0.05159496143460274,
0.01301510538905859,
0.04658519849181175,
0.07414140552282333,
0.06052171438932419,
-0.011848571710288525,
-0.002738693729043007,
0.034350812435150146,
-0.000028201460736454464,
-0.05256810039281845,
-0.14900416135787964,
0.06299556791782379,
0.18326662480831146,
0.06357476860284805,
0.02572336606681347,
0.006351623684167862,
-0.12749691307544708,
-0.07391927391290665,
0.10654070973396301,
-0.020253000780940056,
-0.03248732164502144,
-0.06469862908124924,
0.22561006247997284,
0.1440858691930771,
-0.19123812019824982,
0.07266645133495331,
-0.06280723959207535,
-0.03856344893574715,
-0.1423911601305008,
-0.1734597086906433,
-0.05848124250769615,
-0.0470263697206974,
-0.02805337868630886,
-0.056551121175289154,
0.047676339745521545,
0.04214883968234062,
-0.0049680061638355255,
-0.02410033717751503,
0.10285825282335281,
0.029881322756409645,
-0.03433713689446449,
0.042509518563747406,
0.05520700290799141,
0.04173598438501358,
-0.0948147252202034,
0.008655120618641376,
0.002851796569302678,
0.01364042703062296,
0.06763856112957001,
0.023438580334186554,
-0.07359297573566437,
0.03003004938364029,
-0.021295083686709404,
-0.12327025085687637,
0.04447251185774803,
-0.004893502686172724,
-0.027765894308686256,
0.149532288312912,
0.03866618499159813,
0.00960999820381403,
-0.011475211940705776,
0.23604647815227509,
-0.07244839519262314,
-0.08253564685583115,
-0.13467007875442505,
0.0780881717801094,
-0.07230719178915024,
0.024576395750045776,
0.01996026746928692,
-0.12221863120794296,
0.014945339411497116,
0.16973578929901123,
0.11451959609985352,
-0.01655866578221321,
0.006180240772664547,
0.04829459264874458,
0.005904082674533129,
-0.04154635965824127,
0.01428845152258873,
0.05701562017202377,
0.2032882273197174,
-0.07915740460157394,
0.06002949923276901,
-0.019239120185375214,
-0.07284606248140335,
-0.02558174915611744,
0.09680742770433426,
-0.01388328243046999,
-0.006623120047152042,
-0.061960138380527496,
0.14927618205547333,
-0.0754377469420433,
-0.20746295154094696,
0.057789355516433716,
-0.06455498933792114,
-0.13971345126628876,
-0.044710252434015274,
0.039163388311862946,
-0.026234451681375504,
0.006238372530788183,
0.062338687479496,
-0.04757266119122505,
0.1757270097732544,
0.02755853533744812,
-0.040321994572877884,
-0.09431149810552597,
0.05743449926376343,
-0.14257776737213135,
0.2850838005542755,
0.024792293086647987,
0.05443957820534706,
0.1118428111076355,
-0.022589324042201042,
-0.14643193781375885,
0.015809999778866768,
0.11255357414484024,
-0.06439529359340668,
0.07097096741199493,
0.15858381986618042,
0.003936937544494867,
0.12721814215183258,
0.06751634180545807,
-0.0518413707613945,
0.03665494918823242,
-0.07913793623447418,
-0.04930124804377556,
-0.12390561401844025,
0.07825948297977448,
-0.09394824504852295,
0.15530800819396973,
0.11564930528402328,
-0.0716809332370758,
0.002328893169760704,
-0.0226058978587389,
0.08586639165878296,
0.016556650400161743,
0.1145932599902153,
0.007989857345819473,
-0.19067500531673431,
0.04224180802702904,
0.008731488138437271,
0.09664896875619888,
-0.20374362170696259,
-0.05386274680495262,
0.0430329255759716,
-0.02220054902136326,
-0.07850324362516403,
0.11384803056716919,
0.03693259134888649,
0.02049495466053486,
-0.03668023273348808,
-0.040239542722702026,
0.01582854613661766,
0.15718567371368408,
-0.10779841244220734,
-0.01871471293270588
] |
null | null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | text-generation | aidonuts/corgy-001 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | 2024-02-08T18:52:42+00:00 | [
"1910.09700"
] | [] | TAGS
#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by:
- Funded by [optional]:
- Shared by [optional]:
- Model type:
- Language(s) (NLP):
- License:
- Finetuned from model [optional]:
### Model Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Hours used:
- Cloud Provider:
- Compute Region:
- Carbon Emitted:
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
| [
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
"TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Model Card for Model ID",
"## Model Details",
"### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
"### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Downstream Use [optional]",
"### Out-of-Scope Use",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"## How to Get Started with the Model\n\nUse the code below to get started with the model.",
"## Training Details",
"### Training Data",
"### Training Procedure",
"#### Preprocessing [optional]",
"#### Training Hyperparameters\n\n- Training regime:",
"#### Speeds, Sizes, Times [optional]",
"## Evaluation",
"### Testing Data, Factors & Metrics",
"#### Testing Data",
"#### Factors",
"#### Metrics",
"### Results",
"#### Summary",
"## Model Examination [optional]",
"## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
"## Technical Specifications [optional]",
"### Model Architecture and Objective",
"### Compute Infrastructure",
"#### Hardware",
"#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Model Card Authors [optional]",
"## Model Card Contact"
] | [
60,
6,
3,
82,
28,
3,
4,
9,
9,
10,
42,
20,
3,
4,
5,
9,
11,
13,
3,
12,
5,
4,
5,
3,
4,
9,
53,
9,
8,
6,
3,
14,
8,
7,
9,
4
] | [
"passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | [
-0.04654794931411743,
0.16618601977825165,
-0.005445904564112425,
0.01853804849088192,
0.0981811136007309,
0.011998992413282394,
0.06433123350143433,
0.11398410052061081,
-0.0230073444545269,
0.11406639218330383,
0.03047988750040531,
0.10172267258167267,
0.11317981779575348,
0.14841650426387787,
-0.002152352826669812,
-0.22403094172477722,
0.050844956189394,
-0.12105348706245422,
-0.033293843269348145,
0.11749980598688126,
0.1483822613954544,
-0.09928343445062637,
0.07274559140205383,
-0.029687678441405296,
-0.012143402360379696,
-0.030057786032557487,
-0.05890674889087677,
-0.046214159578084946,
0.04651786759495735,
0.06640566885471344,
0.06770290434360504,
0.0071083661168813705,
0.09012923389673233,
-0.2696533799171448,
0.018959321081638336,
0.07145345956087112,
-0.002759667346253991,
0.06957992166280746,
0.06404146552085876,
-0.07107418030500412,
0.10337356477975845,
-0.05106033384799957,
0.14650006592273712,
0.08365883678197861,
-0.09081148356199265,
-0.1895141303539276,
-0.08866965025663376,
0.09882009029388428,
0.17572562396526337,
0.04925641790032387,
-0.02320658043026924,
0.09761467576026917,
-0.08769196271896362,
0.015438909642398357,
0.04981724172830582,
-0.07620415836572647,
-0.05378096550703049,
0.05986575037240982,
0.07907199114561081,
0.06627275794744492,
-0.12434766441583633,
-0.02885502204298973,
0.005009706597775221,
0.010980482213199139,
0.0769270583987236,
0.01728810742497444,
0.146672785282135,
0.0338633768260479,
-0.12615777552127838,
-0.04880760237574577,
0.09869225323200226,
0.03395522013306618,
-0.04422314465045929,
-0.24749068915843964,
-0.03152675926685333,
-0.030810698866844177,
-0.029386121779680252,
-0.03716538846492767,
0.04340358078479767,
-0.007673026993870735,
0.08638741075992584,
-0.0060646249912679195,
-0.07403432577848434,
-0.03937075287103653,
0.06169692054390907,
0.0672287791967392,
0.02999979443848133,
-0.013745363801717758,
0.010938193649053574,
0.11620724946260452,
0.1095694974064827,
-0.12054188549518585,
-0.05555335059762001,
-0.06393084675073624,
-0.08656639605760574,
-0.040790557861328125,
0.034162238240242004,
0.03456587344408035,
0.05349370837211609,
0.25305667519569397,
0.015654386952519417,
0.059652652591466904,
0.034477248787879944,
0.007892133668065071,
0.05848940089344978,
0.11044429242610931,
-0.06018859148025513,
-0.10444226115942001,
-0.02648012898862362,
0.08843598514795303,
0.008199662901461124,
-0.03287925571203232,
-0.05088530853390694,
0.06019928678870201,
0.01946467161178589,
0.11926145106554031,
0.09061790257692337,
0.010536285117268562,
-0.07121123373508453,
-0.061038948595523834,
0.1891259253025055,
-0.16544590890407562,
0.04322727024555206,
0.035097137093544006,
-0.03903156518936157,
0.00019933005387429148,
0.013914269395172596,
0.016625655815005302,
-0.025983380153775215,
0.09017423540353775,
-0.054113563150167465,
-0.04145489260554314,
-0.11186197400093079,
-0.03383193537592888,
0.033762916922569275,
0.008953776210546494,
-0.035059962421655655,
-0.033713940531015396,
-0.08351044356822968,
-0.07577689737081528,
0.09320491552352905,
-0.07346344739198685,
-0.04878907650709152,
-0.01804324984550476,
-0.07530532777309418,
0.022395428270101547,
0.019394835457205772,
0.07707412540912628,
-0.02362251654267311,
0.04399976506829262,
-0.05189276114106178,
0.05863580107688904,
0.11207318305969238,
0.03570080175995827,
-0.05736649036407471,
0.06062258034944534,
-0.23834340274333954,
0.09552820026874542,
-0.07409077137708664,
0.05591456592082977,
-0.153293639421463,
-0.024439791217446327,
0.04788333550095558,
0.008784620091319084,
-0.009650949388742447,
0.13416339457035065,
-0.21702027320861816,
-0.02536402828991413,
0.1717337965965271,
-0.10057014971971512,
-0.07069246470928192,
0.05619903281331062,
-0.04835370555520058,
0.10988964140415192,
0.03825836628675461,
-0.025690359994769096,
0.06171267107129097,
-0.1267417073249817,
0.003717758459970355,
-0.05005312338471413,
-0.017048977315425873,
0.1548657864332199,
0.07182947546243668,
-0.07217690348625183,
0.07399354875087738,
0.025708531960844994,
-0.0246540866792202,
-0.04625825211405754,
-0.015164627693593502,
-0.10536660254001617,
0.014689887873828411,
-0.06369215250015259,
0.014470234513282776,
-0.020807426422834396,
-0.09071163833141327,
-0.027962757274508476,
-0.17504668235778809,
-0.03014434315264225,
0.08651752024888992,
-0.008693269453942776,
-0.01803150773048401,
-0.1178668737411499,
0.009341353550553322,
0.04177580401301384,
0.0061247628182172775,
-0.13462838530540466,
-0.04812471568584442,
0.02780051715672016,
-0.1600649207830429,
0.034652888774871826,
-0.05392369255423546,
0.04932025074958801,
0.025790516287088394,
-0.028889117762446404,
-0.026493212208151817,
0.021633783355355263,
0.005992184858769178,
-0.011999987065792084,
-0.24343903362751007,
-0.028118690475821495,
-0.024888472631573677,
0.1682123839855194,
-0.20917098224163055,
0.03546025976538658,
0.07867541164159775,
0.15366052091121674,
0.011240328662097454,
-0.04177491366863251,
0.005974748637527227,
-0.06935794651508331,
-0.02736494317650795,
-0.05875484645366669,
-0.0047869328409433365,
-0.03310677409172058,
-0.04545191675424576,
0.04568447172641754,
-0.16510973870754242,
-0.032636504620313644,
0.09776268899440765,
0.06289951503276825,
-0.13922683894634247,
-0.020621931180357933,
-0.03630133345723152,
-0.049253206700086594,
-0.04911839962005615,
-0.0605199858546257,
0.10893940925598145,
0.05891856551170349,
0.04574795812368393,
-0.05928509309887886,
-0.07568105310201645,
-0.001827909960411489,
-0.013898161239922047,
-0.017864689230918884,
0.09759635478258133,
0.0751434788107872,
-0.13251115381717682,
0.09224759042263031,
0.09603385627269745,
0.07919023185968399,
0.09113933145999908,
-0.02355697751045227,
-0.08261934667825699,
-0.045987509191036224,
0.031442027539014816,
0.020124373957514763,
0.13039541244506836,
-0.024294709786772728,
0.04352088272571564,
0.042134687304496765,
-0.019369594752788544,
0.014752166345715523,
-0.08687400817871094,
0.033972494304180145,
0.028472330421209335,
-0.016721390187740326,
0.050190530717372894,
-0.03876714035868645,
0.02440318465232849,
0.08830609917640686,
0.045322712510824203,
0.03507532551884651,
0.015493292361497879,
-0.05206458270549774,
-0.1083620935678482,
0.16405931115150452,
-0.12714070081710815,
-0.22483378648757935,
-0.13936103880405426,
0.0037376401014626026,
0.035628627985715866,
-0.015835661441087723,
0.002417160663753748,
-0.059374887496232986,
-0.12220635265111923,
-0.08858037739992142,
0.015140829607844353,
0.04942670464515686,
-0.09028962254524231,
-0.06437795609235764,
0.058117836713790894,
0.03889724239706993,
-0.14560972154140472,
0.017612040042877197,
0.04854894429445267,
-0.09789852797985077,
-0.006774199660867453,
0.08094939589500427,
0.0698540136218071,
0.1770169734954834,
0.017703235149383545,
-0.021850809454917908,
0.032354529947042465,
0.20614571869373322,
-0.13538233935832977,
0.11083246022462845,
0.13607586920261383,
-0.09041404724121094,
0.08072979003190994,
0.19951270520687103,
0.03932560607790947,
-0.10153959691524506,
0.031980328261852264,
0.02283124253153801,
-0.0284719280898571,
-0.24526868760585785,
-0.07212468236684799,
-0.004402178805321455,
-0.058010730892419815,
0.07660572230815887,
0.09286724030971527,
0.08215958625078201,
0.012304253876209259,
-0.09310996532440186,
-0.08154371380805969,
0.05942574888467789,
0.10367169976234436,
0.024584239348769188,
-0.010839897207915783,
0.08998730033636093,
-0.034100502729415894,
0.019626356661319733,
0.0853661298751831,
0.005239574704319239,
0.17840281128883362,
0.05159219726920128,
0.18830420076847076,
0.07925192266702652,
0.07219027727842331,
0.009912233799695969,
0.013080619275569916,
0.018877580761909485,
0.03300119563937187,
-0.002769160782918334,
-0.08440786600112915,
-0.02248465269804001,
0.11566436290740967,
0.06668911874294281,
0.010815348476171494,
0.015172341838479042,
-0.04104290530085564,
0.07965951412916183,
0.1831512451171875,
-0.007656289264559746,
-0.1783534437417984,
-0.057547420263290405,
0.07553383708000183,
-0.09879875183105469,
-0.09854305535554886,
-0.013454320840537548,
0.03072015568614006,
-0.17046253383159637,
0.023390959948301315,
-0.02239842526614666,
0.1106182336807251,
-0.14194999635219574,
-0.020490378141403198,
0.07218493521213531,
0.07199500501155853,
0.004729843698441982,
0.05758659541606903,
-0.16417601704597473,
0.10671813786029816,
0.008950476534664631,
0.06779605895280838,
-0.09610627591609955,
0.1008887067437172,
-0.004196076653897762,
-0.02063460275530815,
0.1393408179283142,
0.002700034761801362,
-0.06884108483791351,
-0.0763031542301178,
-0.08754398673772812,
-0.009632662869989872,
0.12754282355308533,
-0.1419651061296463,
0.08767123520374298,
-0.037212442606687546,
-0.0424150750041008,
-0.0017086371080949903,
-0.10206665843725204,
-0.11638247221708298,
-0.18888559937477112,
0.06001543253660202,
-0.13492922484874725,
0.03152317553758621,
-0.10799519717693329,
-0.032371897250413895,
-0.030304040759801865,
0.19337286055088043,
-0.23447458446025848,
-0.07199826091527939,
-0.1475764364004135,
-0.10233612358570099,
0.1443224400281906,
-0.0501345656812191,
0.08485390990972519,
-0.007241467013955116,
0.16846685111522675,
0.019060896709561348,
-0.02531743235886097,
0.0971490666270256,
-0.09173708409070969,
-0.19302815198898315,
-0.07869284600019455,
0.15662524104118347,
0.13260218501091003,
0.031680017709732056,
-0.002461588243022561,
0.036563750356435776,
-0.015421539545059204,
-0.11935004591941833,
0.015969349071383476,
0.1787186712026596,
0.06237189099192619,
0.02331034652888775,
-0.027346095070242882,
-0.11273157596588135,
-0.06900003552436829,
-0.028530338779091835,
0.03054865077137947,
0.17762407660484314,
-0.07057618349790573,
0.18207968771457672,
0.14163152873516083,
-0.05922834202647209,
-0.20400173962116241,
0.010538800619542599,
0.03055560030043125,
0.0009220078936778009,
0.02591954916715622,
-0.20123432576656342,
0.08688826113939285,
0.004683020059019327,
-0.05110127478837967,
0.13194532692432404,
-0.17217805981636047,
-0.14451217651367188,
0.0765485092997551,
0.038384392857551575,
-0.19559739530086517,
-0.12913893163204193,
-0.09174312651157379,
-0.045869920402765274,
-0.18591414391994476,
0.09569250047206879,
0.0305706188082695,
0.010893458500504494,
0.03030681423842907,
0.029179483652114868,
0.019487828016281128,
-0.0418255440890789,
0.18391458690166473,
-0.024792250245809555,
0.026594700291752815,
-0.08539514988660812,
-0.06927408277988434,
0.03743394836783409,
-0.052842434495687485,
0.07349982857704163,
-0.023486759513616562,
0.007861839607357979,
-0.10348054021596909,
-0.042148489505052567,
-0.03735732287168503,
0.015448716469109058,
-0.09657872468233109,
-0.08514349907636642,
-0.045032672584056854,
0.09675803780555725,
0.09690850973129272,
-0.033646680414676666,
-0.028050623834133148,
-0.07533035427331924,
0.04412057250738144,
0.19926515221595764,
0.1785389482975006,
0.042153384536504745,
-0.08034496754407883,
-0.004150947090238333,
-0.010121207684278488,
0.04310847446322441,
-0.20463712513446808,
0.06283636391162872,
0.05450061708688736,
0.01973269321024418,
0.11436162889003754,
-0.019565396010875702,
-0.15359151363372803,
-0.07263088971376419,
0.06303015351295471,
-0.060181066393852234,
-0.19620554149150848,
0.00867035984992981,
0.060603946447372437,
-0.16371412575244904,
-0.04535605385899544,
0.04643881320953369,
-0.005620351992547512,
-0.038163937628269196,
0.021896906197071075,
0.09194854646921158,
0.0026654244866222143,
0.07427921891212463,
0.05387866869568825,
0.0827430784702301,
-0.10537070035934448,
0.08090532571077347,
0.08839722722768784,
-0.08452684432268143,
0.023530138656497,
0.10478579998016357,
-0.059433579444885254,
-0.03440561518073082,
0.020135708153247833,
0.08153781294822693,
0.01775863952934742,
-0.040019966661930084,
0.013229827396571636,
-0.10452935844659805,
0.05954122915863991,
0.08839859813451767,
0.032507482916116714,
0.016702456399798393,
0.03425082191824913,
0.04607953503727913,
-0.07238735258579254,
0.12142276018857956,
0.031868141144514084,
0.017129309475421906,
-0.036505792289972305,
-0.040896978229284286,
0.019542274996638298,
-0.03214648738503456,
-0.005015232600271702,
-0.03023446537554264,
-0.07695909589529037,
-0.014793801121413708,
-0.1626158058643341,
-0.011131818406283855,
-0.05648450180888176,
0.010329355485737324,
0.03204665705561638,
-0.032609567046165466,
0.008124498650431633,
0.009250079281628132,
-0.07695289701223373,
-0.0663459524512291,
-0.020460480824112892,
0.09540658444166183,
-0.16213038563728333,
0.022481130436062813,
0.08244425803422928,
-0.12187694013118744,
0.09281346201896667,
0.016204802319407463,
-0.006236857734620571,
0.025038830935955048,
-0.1475188434123993,
0.034843120723962784,
-0.03386561945080757,
0.010836300440132618,
0.04373383894562721,
-0.21569781005382538,
-0.00004886732858722098,
-0.033673107624053955,
-0.06639216095209122,
-0.009451326914131641,
-0.03672455996274948,
-0.11508306115865707,
0.1058407872915268,
0.007236586883664131,
-0.08753558248281479,
-0.03186136856675148,
0.029325377196073532,
0.0838974118232727,
-0.021959776058793068,
0.15145497024059296,
-0.008370938710868359,
0.07429654151201248,
-0.16209737956523895,
-0.018623165786266327,
-0.006028574425727129,
0.022658247500658035,
-0.01664556935429573,
-0.01111356820911169,
0.044031109660863876,
-0.022746501490473747,
0.17925859987735748,
-0.030318550765514374,
0.02272745408117771,
0.06815794110298157,
0.019072026014328003,
-0.030184008181095123,
0.10406795144081116,
0.04094860330224037,
0.02014910988509655,
0.018591465428471565,
0.003289656015112996,
-0.04647882282733917,
-0.03173251822590828,
-0.19407226145267487,
0.07288651913404465,
0.15608493983745575,
0.09729263186454773,
-0.016707008704543114,
0.07954329252243042,
-0.10199416428804398,
-0.1109243705868721,
0.12477338314056396,
-0.04797708988189697,
-0.002418199321255088,
-0.07150927931070328,
0.13247236609458923,
0.1437523066997528,
-0.1859612911939621,
0.07269313186407089,
-0.0699717253446579,
-0.04708027467131615,
-0.10980689525604248,
-0.19441905617713928,
-0.05561789125204086,
-0.049456022679805756,
-0.016053348779678345,
-0.04698808491230011,
0.07504211366176605,
0.054538097232580185,
0.006766852922737598,
-0.0023397188633680344,
0.06506035476922989,
-0.031050674617290497,
-0.0037882844917476177,
0.032597362995147705,
0.06591679900884628,
0.012734474614262581,
-0.030802709981799126,
0.016619903966784477,
-0.013545602560043335,
0.045626189559698105,
0.06578011065721512,
0.04976864159107208,
-0.02938537672162056,
0.014603170566260815,
-0.038539156317710876,
-0.10249634087085724,
0.043612558394670486,
-0.024421939626336098,
-0.0789753645658493,
0.15477414429187775,
0.023680059239268303,
0.007779473438858986,
-0.020137663930654526,
0.23901568353176117,
-0.0738423764705658,
-0.0964353010058403,
-0.14737580716609955,
0.10557299107313156,
-0.038081806153059006,
0.05800395458936691,
0.04625935107469559,
-0.10226529091596603,
0.018044332042336464,
0.1338089406490326,
0.16182038187980652,
-0.039008259773254395,
0.020095856860280037,
0.031135575845837593,
0.00566398398950696,
-0.03622615709900856,
0.04847532883286476,
0.06906453520059586,
0.16569648683071136,
-0.04632584750652313,
0.09100406616926193,
0.0019041687482967973,
-0.09579581767320633,
-0.038361791521310806,
0.11069868505001068,
-0.016052277758717537,
0.019335128366947174,
-0.05818064883351326,
0.11742528527975082,
-0.06386786699295044,
-0.23783175647258759,
0.06453443318605423,
-0.0684293657541275,
-0.13765870034694672,
-0.02378307841718197,
0.08207765966653824,
-0.012955902144312859,
0.027587108314037323,
0.0730307325720787,
-0.07240920513868332,
0.201939657330513,
0.03798431158065796,
-0.05499868467450142,
-0.055047210305929184,
0.0805421993136406,
-0.10008571296930313,
0.2739645540714264,
0.01557221356779337,
0.04601577669382095,
0.10384146869182587,
-0.009341772645711899,
-0.13838784396648407,
0.019836371764540672,
0.09581108391284943,
-0.10502193123102188,
0.04196618124842644,
0.19815568625926971,
-0.0014755994779989123,
0.12389086186885834,
0.07657600939273834,
-0.07551808655261993,
0.0478031262755394,
-0.08054235577583313,
-0.06760486960411072,
-0.09260394424200058,
0.09703279286623001,
-0.07772123068571091,
0.14251399040222168,
0.13876807689666748,
-0.05074559152126312,
0.012724342755973339,
-0.031311117112636566,
0.044293127954006195,
-0.00010600237874314189,
0.10321761667728424,
0.004272161517292261,
-0.1832672357559204,
0.024692710489034653,
0.005650998093187809,
0.10749758034944534,
-0.16033467650413513,
-0.09566054493188858,
0.042343202978372574,
0.003505636239424348,
-0.0672195628285408,
0.1290110945701599,
0.05665452033281326,
0.04342988133430481,
-0.03997718170285225,
-0.03521440550684929,
-0.0060732318088412285,
0.13561366498470306,
-0.10713256150484085,
0.0009933578548952937
] |
null | null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# BOLETIN_4bit_13
This model is a fine-tuned version of [bertin-project/BOLETIN](https://huggingface.co/bertin-project/BOLETIN) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1.41e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.7.1
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.1 | {"license": "openrail", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "bertin-project/BOLETIN", "model-index": [{"name": "BOLETIN_4bit_13", "results": []}]} | null | versae/BOLETIN_4bit_13 | [
"peft",
"tensorboard",
"safetensors",
"generated_from_trainer",
"base_model:bertin-project/BOLETIN",
"license:openrail",
"region:us"
] | 2024-02-08T18:53:32+00:00 | [] | [] | TAGS
#peft #tensorboard #safetensors #generated_from_trainer #base_model-bertin-project/BOLETIN #license-openrail #region-us
|
# BOLETIN_4bit_13
This model is a fine-tuned version of bertin-project/BOLETIN on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1.41e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
- mixed_precision_training: Native AMP
### Training results
### Framework versions
- PEFT 0.7.1
- Transformers 4.37.2
- Pytorch 2.2.0+cu121
- Datasets 2.14.6
- Tokenizers 0.15.1 | [
"# BOLETIN_4bit_13\n\nThis model is a fine-tuned version of bertin-project/BOLETIN on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1.41e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- PEFT 0.7.1\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.14.6\n- Tokenizers 0.15.1"
] | [
"TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-bertin-project/BOLETIN #license-openrail #region-us \n",
"# BOLETIN_4bit_13\n\nThis model is a fine-tuned version of bertin-project/BOLETIN on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1.41e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3\n- mixed_precision_training: Native AMP",
"### Training results",
"### Framework versions\n\n- PEFT 0.7.1\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.14.6\n- Tokenizers 0.15.1"
] | [
44,
35,
6,
12,
8,
3,
104,
4,
39
] | [
"passage: TAGS\n#peft #tensorboard #safetensors #generated_from_trainer #base_model-bertin-project/BOLETIN #license-openrail #region-us \n# BOLETIN_4bit_13\n\nThis model is a fine-tuned version of bertin-project/BOLETIN on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 1.41e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.7.1\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.14.6\n- Tokenizers 0.15.1"
] | [
-0.1255156546831131,
0.10732937604188919,
-0.0018351870821788907,
0.06742889434099197,
0.13059139251708984,
-0.011080823838710785,
0.11219768226146698,
0.09278897196054459,
-0.08765602111816406,
0.09975109994411469,
0.09259475022554398,
-0.003950811922550201,
0.05341366305947304,
0.15785059332847595,
-0.010295131243765354,
-0.2811092138290405,
0.01830008625984192,
0.01916777342557907,
-0.06739117205142975,
0.09097842872142792,
0.11433453112840652,
-0.08528942614793777,
0.04811011627316475,
0.04461422190070152,
-0.15275876224040985,
0.03701477497816086,
-0.030724937096238136,
-0.0679372176527977,
0.07796746492385864,
0.03962085023522377,
0.1260586827993393,
-0.0030480525456368923,
0.09163504838943481,
-0.18303203582763672,
0.00999118946492672,
0.07656360417604446,
0.05349556356668472,
0.10303228348493576,
0.09551950544118881,
0.0453813336789608,
0.1026729941368103,
-0.09924053400754929,
0.09437979012727737,
0.03715737909078598,
-0.069048210978508,
-0.19270433485507965,
-0.09680823236703873,
0.1436537206172943,
0.06991060078144073,
0.0712747797369957,
0.004534195177257061,
0.1377352923154831,
-0.06615722924470901,
0.05774744227528572,
0.21412788331508636,
-0.2715483009815216,
-0.102420873939991,
0.024710794910788536,
0.08029805123806,
0.06376952677965164,
-0.11278852820396423,
-0.043705325573682785,
0.08638988435268402,
0.03195445239543915,
0.06495727598667145,
0.014975085854530334,
-0.014592735096812248,
-0.026602866128087044,
-0.1609697937965393,
-0.012365786358714104,
0.179563507437706,
0.07417372614145279,
-0.07639028877019882,
-0.08583340048789978,
-0.029866356402635574,
-0.1278601735830307,
-0.027163513004779816,
-0.025040164589881897,
0.013431034050881863,
-0.040455423295497894,
-0.038507651537656784,
-0.06292150169610977,
-0.08684089034795761,
-0.07175388187170029,
0.02296328730881214,
0.14714348316192627,
0.06444933265447617,
0.00797970499843359,
-0.03165614604949951,
0.09310509264469147,
-0.009996474720537663,
-0.09089457988739014,
-0.003678566310554743,
-0.03275779262185097,
-0.022527635097503662,
-0.06077426299452782,
-0.031756527721881866,
-0.02135453186929226,
0.007909346371889114,
0.14458979666233063,
-0.12025593221187592,
0.06926661729812622,
-0.0006105806678533554,
0.03663421794772148,
-0.03627393767237663,
0.11810174584388733,
-0.058989185839891434,
0.062326256185770035,
0.005935652181506157,
0.10517515242099762,
0.0006127501837909222,
0.01013200543820858,
-0.07158071547746658,
-0.011703279800713062,
0.08385162800550461,
0.04779311269521713,
-0.04818187281489372,
-0.006382702849805355,
-0.04099283739924431,
-0.03061904013156891,
0.05637663975358009,
-0.11067534983158112,
0.03672021999955177,
-0.002803554991260171,
-0.0722050815820694,
0.03502793610095978,
0.02687794528901577,
0.008402991108596325,
-0.022888077422976494,
0.08434560894966125,
-0.10810016095638275,
-0.020109234377741814,
-0.10208006203174591,
-0.0687989667057991,
0.03361479192972183,
-0.020208250731229782,
-0.004697564523667097,
-0.10697347670793533,
-0.14619088172912598,
-0.023521840572357178,
0.028987348079681396,
-0.05466166511178017,
-0.08007031679153442,
-0.005480632185935974,
-0.06673192232847214,
0.02033618837594986,
-0.025541264563798904,
0.1437087059020996,
-0.021809730678796768,
0.08992140740156174,
0.0007511846488341689,
-0.0013133661122992635,
-0.02249962091445923,
0.028180034831166267,
-0.05993964523077011,
0.05432213470339775,
-0.10486732423305511,
0.025063475593924522,
-0.12218798696994781,
0.04886893928050995,
-0.1246299147605896,
-0.12108048051595688,
-0.03983562812209129,
-0.010139106772840023,
0.10852429270744324,
0.10460635274648666,
-0.10226745158433914,
-0.03453566133975983,
0.16826343536376953,
-0.1009598970413208,
-0.10066498070955276,
0.10946982353925705,
-0.022380532696843147,
0.022987157106399536,
0.04281345009803772,
0.14815695583820343,
0.10269059240818024,
-0.1541232466697693,
-0.016518179327249527,
0.013334772549569607,
0.10168899595737457,
-0.0298161618411541,
0.08491069078445435,
-0.02125919982790947,
-0.06897043436765671,
0.015702510252594948,
-0.018095575273036957,
0.02675781212747097,
-0.10551959276199341,
-0.06462951004505157,
-0.04300270974636078,
-0.09371072053909302,
0.04940140247344971,
0.020687973126769066,
0.046183910220861435,
-0.07193814963102341,
-0.11889537423849106,
0.13346458971500397,
0.14220011234283447,
-0.034140948206186295,
0.013125733472406864,
-0.0702115148305893,
0.07515795528888702,
-0.06015996262431145,
-0.0450005941092968,
-0.1662239283323288,
-0.11168272793292999,
0.04697501286864281,
-0.08278761804103851,
0.011988358572125435,
-0.008379099890589714,
0.055018164217472076,
0.07957031577825546,
-0.028724675998091698,
-0.04078642651438713,
-0.12348336726427078,
-0.016088930889964104,
-0.1134108230471611,
-0.1748754382133484,
-0.07135701924562454,
-0.03227679803967476,
0.1354910433292389,
-0.2564660310745239,
0.02370402030646801,
-0.025096086785197258,
0.15301775932312012,
0.035104621201753616,
-0.05823059752583504,
-0.012648340314626694,
0.06905662268400192,
-0.003284232458099723,
-0.06867850571870804,
0.03647756949067116,
0.027381666004657745,
-0.07623100280761719,
-0.049540676176548004,
-0.09276840090751648,
0.07114652544260025,
0.06675520539283752,
0.07561594992876053,
-0.09161181002855301,
-0.08808480948209763,
-0.10286413133144379,
-0.037651628255844116,
-0.05346217751502991,
-0.01590418815612793,
0.13005836308002472,
0.01583968661725521,
0.14800399541854858,
-0.09976162016391754,
-0.05832315981388092,
0.02487853914499283,
-0.02752738818526268,
-0.031445566564798355,
0.09806420654058456,
0.0916074812412262,
-0.0421861931681633,
0.09700632840394974,
0.06023326888680458,
-0.0885535180568695,
0.1576170176267624,
-0.06319788843393326,
-0.12328711152076721,
-0.00649529043585062,
0.05652191862463951,
0.011643982492387295,
0.14284640550613403,
-0.07372429221868515,
0.02295053005218506,
0.024521976709365845,
0.014965099282562733,
0.06463028490543365,
-0.17149651050567627,
-0.013594462536275387,
0.0005488356691785157,
-0.013092245906591415,
-0.012076683342456818,
0.005058457143604755,
0.01472078263759613,
0.0902780294418335,
0.03159976378083229,
-0.010178010910749435,
0.0066510410979390144,
0.010613302700221539,
-0.0822402611374855,
0.18826532363891602,
-0.10660979896783829,
-0.13019205629825592,
-0.16623060405254364,
0.04678114131093025,
-0.04083292931318283,
-0.043642397969961166,
0.016644537448883057,
-0.08034076541662216,
-0.04355567321181297,
-0.08547274768352509,
-0.021106066182255745,
-0.08569524437189102,
0.0041999113745987415,
0.050018370151519775,
0.009071198292076588,
0.1305955946445465,
-0.12356099486351013,
0.024006707593798637,
-0.009780789725482464,
-0.06499875336885452,
-0.018859725445508957,
0.0395088717341423,
0.09218832105398178,
0.10984667390584946,
-0.0025415662676095963,
0.017971016466617584,
-0.042204294353723526,
0.22964079678058624,
-0.04731786251068115,
-0.031540174037218094,
0.12455035001039505,
0.013704792596399784,
0.06880811601877213,
0.07427424192428589,
0.06077098101377487,
-0.07779726386070251,
0.04934479668736458,
0.07642018049955368,
-0.018774718046188354,
-0.25803789496421814,
-0.04259645938873291,
-0.03137337416410446,
-0.03581773489713669,
0.11393380910158157,
0.06963405758142471,
-0.002577340928837657,
0.06858138740062714,
-0.029033256694674492,
0.05772869288921356,
-0.054586268961429596,
0.11093302816152573,
0.07411284744739532,
0.02114134654402733,
0.0852951928973198,
-0.03159468621015549,
-0.027554886415600777,
0.06208023056387901,
0.003726950380951166,
0.2492937594652176,
-0.00901741348206997,
0.06713097542524338,
0.06760426610708237,
0.1684977412223816,
0.008415263146162033,
0.04065545275807381,
-0.007696415763348341,
-0.017619986087083817,
0.02730153314769268,
-0.07059445977210999,
-0.005786007270216942,
0.03237498924136162,
-0.04264397546648979,
0.09012381732463837,
-0.11932066082954407,
-0.01252875104546547,
-0.019668452441692352,
0.26969829201698303,
0.02749735862016678,
-0.2938276529312134,
-0.09866927564144135,
0.000027742251404561102,
-0.052498213946819305,
-0.10288489609956741,
0.03820633143186569,
0.10760655254125595,
-0.11477527022361755,
0.027626654133200645,
-0.06695505231618881,
0.11608842760324478,
0.007486123126000166,
-0.026969071477651596,
0.04454139247536659,
0.149954155087471,
-0.019266402348876,
0.08938728272914886,
-0.1931573450565338,
0.2225373089313507,
0.004019944928586483,
0.10006839036941528,
-0.009627485647797585,
0.041325025260448456,
0.005858305841684341,
0.05190775543451309,
0.08108130842447281,
-0.00444452278316021,
0.00977750588208437,
-0.19999587535858154,
-0.10293269902467728,
0.02897484041750431,
0.09837603569030762,
-0.023125579580664635,
0.0759328082203865,
-0.033367007970809937,
0.029121102765202522,
0.03166752681136131,
-0.056333303451538086,
-0.2205408364534378,
-0.10798301547765732,
0.014995450153946877,
0.031858451664447784,
0.003287144238129258,
-0.12737460434436798,
-0.11659417301416397,
-0.04256961867213249,
0.12073040753602982,
-0.03525484725832939,
-0.05089368671178818,
-0.13217230141162872,
0.10138162225484848,
0.13472776114940643,
-0.04788528382778168,
0.030606066808104515,
-0.010295629501342773,
0.1545175462961197,
0.007115511689335108,
-0.05442889779806137,
0.052550725638866425,
-0.08581000566482544,
-0.16713489592075348,
-0.059260446578264236,
0.15776962041854858,
0.06678067147731781,
0.03495228290557861,
-0.014310728758573532,
-0.008302809670567513,
0.0002488506434019655,
-0.09265726804733276,
-0.0011225346243008971,
0.07033748924732208,
0.031063618138432503,
0.02441342920064926,
-0.09940396994352341,
0.07413540780544281,
-0.02940279245376587,
-0.039797164499759674,
0.1007785052061081,
0.22367297112941742,
-0.07964777946472168,
0.03869553655385971,
0.08806568384170532,
-0.0710466206073761,
-0.14034758508205414,
0.06643027067184448,
0.16827119886875153,
0.020954933017492294,
0.01235933043062687,
-0.20506708323955536,
0.10339904576539993,
0.13560202717781067,
-0.03129309043288231,
0.09248471260070801,
-0.299375057220459,
-0.11178755760192871,
0.0835481733083725,
0.11326135694980621,
0.017187604680657387,
-0.15403711795806885,
-0.06516816467046738,
0.003081821370869875,
-0.10543251782655716,
0.08751959353685379,
-0.07649213820695877,
0.07532230019569397,
-0.009858046658337116,
0.09605097770690918,
0.02334316447377205,
-0.03701886534690857,
0.15891554951667786,
-0.03232422471046448,
0.05988893285393715,
-0.030137022957205772,
0.054881833493709564,
0.047524742782115936,
-0.06184596195816994,
0.058004606515169144,
-0.0040908497758209705,
0.06488107889890671,
-0.16072024405002594,
-0.029464738443493843,
-0.07537510246038437,
0.07855939120054245,
-0.0699969232082367,
-0.05164260044693947,
-0.04118367284536362,
0.05760855972766876,
0.011063138954341412,
-0.015920694917440414,
0.07425209134817123,
0.013993281871080399,
0.15366515517234802,
0.10421591997146606,
0.06421715766191483,
-0.02055095136165619,
-0.1090480387210846,
0.0030910286586731672,
-0.03787723183631897,
0.09178005903959274,
-0.11933191865682602,
-0.006498029455542564,
0.12043977528810501,
0.03409443795681,
0.12313509732484818,
0.03442360833287239,
-0.09549704194068909,
-0.0037366519682109356,
0.05273168534040451,
-0.09781208634376526,
-0.1293790340423584,
-0.044244978576898575,
0.14484281837940216,
-0.1639590710401535,
0.026478545740246773,
0.09654741734266281,
-0.10355023294687271,
-0.023817935958504677,
-0.006785728968679905,
-0.006801936775445938,
-0.026059690862894058,
0.15167391300201416,
0.07042260468006134,
0.06846900284290314,
-0.04900723323225975,
0.08015568554401398,
0.0782080590724945,
-0.10145608335733414,
0.036404795944690704,
0.06929466128349304,
-0.05875825136899948,
-0.004938593599945307,
0.03719271719455719,
0.12701529264450073,
-0.04176619276404381,
-0.06654047966003418,
-0.06631816923618317,
-0.09589662402868271,
0.054382629692554474,
0.09268069267272949,
0.02922859601676464,
-0.017846254631876945,
-0.015020369552075863,
0.05580771714448929,
-0.1417359858751297,
0.09725456684827805,
0.051517922431230545,
0.09159177541732788,
-0.17276078462600708,
0.10535196959972382,
0.008146277628839016,
0.020140578970313072,
-0.0025319589767605066,
0.010835787281394005,
-0.10126972198486328,
-0.019552985206246376,
-0.13207000494003296,
-0.0008124058367684484,
-0.029620029032230377,
0.008925783447921276,
-0.015511063858866692,
-0.052450839430093765,
-0.05323730781674385,
0.04760555550456047,
-0.06710127741098404,
-0.04231280833482742,
0.029058534651994705,
0.07586261630058289,
-0.10792004317045212,
0.01716478541493416,
0.031337346881628036,
-0.08511631935834885,
0.059321943670511246,
0.029561959207057953,
0.043509479612112045,
0.029199087992310524,
-0.0918014645576477,
0.04306052252650261,
0.013894896022975445,
0.013552727177739143,
0.040400430560112,
-0.0785653218626976,
-0.002719024196267128,
-0.03333231434226036,
0.049453895539045334,
-0.0020456372294574976,
0.009466390125453472,
-0.1498788297176361,
-0.0811905711889267,
-0.01132263895124197,
-0.03978559747338295,
-0.0629948079586029,
0.03970133513212204,
0.05955345928668976,
0.061039406806230545,
0.1224956288933754,
-0.08982067555189133,
0.022126713767647743,
-0.17959821224212646,
-0.01503453217446804,
-0.018236879259347916,
0.009954060427844524,
-0.025020578876137733,
-0.0220579132437706,
0.053864698857069016,
-0.04921261593699455,
0.09474607557058334,
-0.028572576120495796,
0.0612168051302433,
0.042985089123249054,
-0.09805404394865036,
-0.0046576885506510735,
0.014720918610692024,
0.24071983993053436,
0.05026311054825783,
0.013820081949234009,
0.052524153143167496,
-0.019298898056149483,
0.05121785029768944,
0.07727015018463135,
0.14528562128543854,
0.15295977890491486,
-0.05855491757392883,
0.05741085112094879,
0.0653456375002861,
-0.07815784960985184,
-0.10655678063631058,
0.06987491250038147,
0.009648077189922333,
0.09576494991779327,
-0.04970527067780495,
0.18860332667827606,
0.14037540555000305,
-0.14540386199951172,
0.022501066327095032,
-0.014118382707238197,
-0.09992730617523193,
-0.11212621629238129,
-0.05318595841526985,
-0.06220975145697594,
-0.1308804303407669,
0.018578553572297096,
-0.13047783076763153,
0.008930183947086334,
0.0586894229054451,
0.021380528807640076,
0.014453893527388573,
0.1596902459859848,
0.02522040531039238,
-0.001289943465963006,
0.0771515741944313,
0.007718491367995739,
0.008992713876068592,
-0.08560747653245926,
-0.08460590243339539,
0.05855602025985718,
-0.01904822699725628,
0.05229286476969719,
-0.04486951604485512,
-0.025539934635162354,
0.03136397898197174,
0.030837377533316612,
-0.06080515682697296,
0.029028698801994324,
0.007313680835068226,
0.03417707234621048,
0.05947684869170189,
0.027584247291088104,
-0.008646801114082336,
-0.06665123254060745,
0.2702465057373047,
-0.0717846006155014,
-0.02189251407980919,
-0.13939785957336426,
0.19612522423267365,
0.005857166834175587,
-0.020795095711946487,
0.04372021555900574,
-0.10436251014471054,
-0.02433924563229084,
0.16825208067893982,
0.09265994280576706,
-0.016117844730615616,
-0.022848889231681824,
0.0025446692015975714,
-0.02358883060514927,
-0.0864768922328949,
0.13710245490074158,
0.10491838306188583,
0.06451688706874847,
-0.04904182255268097,
-0.005273512098938227,
-0.04481637477874756,
0.0018391306512057781,
-0.09637045860290527,
0.06092794984579086,
0.026981595903635025,
-0.014425869099795818,
-0.05617382749915123,
0.08733970671892166,
-0.007947747595608234,
-0.1346425712108612,
0.007851806469261646,
-0.10385160148143768,
-0.17987965047359467,
-0.03816518187522888,
0.05059647932648659,
-0.01226649060845375,
0.04600769653916359,
-0.03330915793776512,
-0.0024727073032408953,
0.1267724335193634,
-0.02277993969619274,
-0.021303236484527588,
-0.1333666890859604,
0.09183488041162491,
-0.024242039769887924,
0.2442518174648285,
0.0033977897837758064,
0.06750480830669403,
0.10696590691804886,
0.010943696834146976,
-0.13880187273025513,
0.006068462505936623,
0.0766693726181984,
-0.08779507130384445,
0.011688278056681156,
0.15309830009937286,
-0.02869957871735096,
0.11268268525600433,
0.013624248094856739,
-0.1519322395324707,
-0.04367383196949959,
-0.060161180794239044,
-0.010785862803459167,
-0.09335596859455109,
0.04354335740208626,
-0.07929602265357971,
0.1341773122549057,
0.180014967918396,
-0.06958728283643723,
-0.04651665687561035,
-0.07256163656711578,
0.05099425092339516,
0.04824759066104889,
0.08197140693664551,
0.01010394748300314,
-0.22330106794834137,
0.006627009715884924,
-0.010518450289964676,
0.019687868654727936,
-0.29770514369010925,
-0.051055874675512314,
0.020195316523313522,
-0.03461706265807152,
-0.06526168435811996,
0.10472872108221054,
0.06219742074608803,
0.03464822843670845,
-0.03872738033533096,
-0.10935445874929428,
-0.06623483449220657,
0.14547741413116455,
-0.1226072683930397,
-0.03629358485341072
] |
null | null | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# my_awesome_qa_model
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.5819
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 250 | 2.1515 |
| 2.6574 | 2.0 | 500 | 1.6337 |
| 2.6574 | 3.0 | 750 | 1.5819 |
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.1
- Tokenizers 0.15.1
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "my_awesome_qa_model", "results": []}]} | question-answering | DanielAvelar09/my_awesome_qa_model | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"question-answering",
"generated_from_trainer",
"base_model:distilbert-base-uncased",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | 2024-02-08T18:53:46+00:00 | [] | [] | TAGS
#transformers #tensorboard #safetensors #distilbert #question-answering #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #endpoints_compatible #region-us
| my\_awesome\_qa\_model
======================
This model is a fine-tuned version of distilbert-base-uncased on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 1.5819
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.35.2
* Pytorch 2.1.0+cu121
* Datasets 2.16.1
* Tokenizers 0.15.1
| [
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
"TAGS\n#transformers #tensorboard #safetensors #distilbert #question-answering #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
65,
98,
4,
33
] | [
"passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #question-answering #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1"
] | [
-0.10708221793174744,
0.08787639439105988,
-0.0020266782958060503,
0.10252698510885239,
0.1274273842573166,
0.02112007699906826,
0.141371950507164,
0.11134063452482224,
-0.08131598681211472,
0.05829233676195145,
0.13410121202468872,
0.11583946645259857,
-0.0018265031976625323,
0.09120819717645645,
-0.0611470602452755,
-0.2005513310432434,
0.004526883829385042,
0.03799702972173691,
-0.09097661823034286,
0.11530310660600662,
0.08769755810499191,
-0.13717275857925415,
0.07678336650133133,
-0.0056874570436775684,
-0.17407989501953125,
0.026953130960464478,
0.006698115728795528,
-0.037090420722961426,
0.1213178038597107,
0.022234441712498665,
0.13187210261821747,
0.023392172530293465,
0.07922547310590744,
-0.19724278151988983,
0.016064628958702087,
0.06023797392845154,
-0.002444722456857562,
0.07741731405258179,
0.025374948978424072,
0.00522031681612134,
0.08328425884246826,
-0.08128440380096436,
0.06258498877286911,
0.029924798756837845,
-0.12490696460008621,
-0.2442893236875534,
-0.09979312866926193,
0.05106590315699577,
0.09691397845745087,
0.09645190834999084,
-0.012973576784133911,
0.13962721824645996,
-0.07699720561504364,
0.08457877486944199,
0.23785600066184998,
-0.3144238293170929,
-0.07457572966814041,
0.04030787572264671,
0.03424936532974243,
0.07267975062131882,
-0.10186595469713211,
-0.031949128955602646,
0.07634760439395905,
0.02774656005203724,
0.10123031586408615,
-0.03811340406537056,
-0.08034272491931915,
0.02562544122338295,
-0.14547140896320343,
-0.012132234871387482,
0.16400417685508728,
0.06514542549848557,
-0.04620467126369476,
-0.03429178148508072,
-0.06315664947032928,
-0.11850249767303467,
-0.029209623113274574,
-0.038746949285268784,
0.04297671839594841,
-0.036058444529771805,
-0.07818788290023804,
-0.016910377889871597,
-0.1026930958032608,
-0.09884173423051834,
-0.050167523324489594,
0.15796343982219696,
0.04260306805372238,
0.016289860010147095,
-0.024042820557951927,
0.10198324918746948,
-0.033326566219329834,
-0.14423881471157074,
0.002374351257458329,
0.020160887390375137,
-0.005382745061069727,
-0.05033143609762192,
-0.04691127687692642,
-0.05261760577559471,
0.03646983578801155,
0.1889425367116928,
-0.07942666113376617,
0.04144255816936493,
0.027733104303479195,
0.03756984323263168,
-0.09601793438196182,
0.14689302444458008,
-0.0625351294875145,
-0.016724655404686928,
-0.002582760527729988,
0.0813518762588501,
0.03259194642305374,
0.0043426938354969025,
-0.08961918205022812,
0.021114153787493706,
0.09003882110118866,
0.02003106102347374,
-0.04685822129249573,
0.05495336651802063,
-0.05019587278366089,
-0.0007271741633303463,
0.012994267977774143,
-0.08166541904211044,
0.026867516338825226,
0.009509325958788395,
-0.060018569231033325,
-0.040473684668540955,
0.02484549954533577,
0.021615348756313324,
0.020202718675136566,
0.09433086216449738,
-0.0980396494269371,
0.008519732393324375,
-0.08998142182826996,
-0.10547184944152832,
0.022492701187729836,
-0.06569727510213852,
0.033701442182064056,
-0.0928310751914978,
-0.16869968175888062,
-0.013243017718195915,
0.05997863784432411,
-0.03089512512087822,
-0.015934670343995094,
-0.04027499631047249,
-0.09035107493400574,
-0.00003611849751905538,
-0.013057121075689793,
0.09050161391496658,
-0.05944909527897835,
0.10105447471141815,
0.061352185904979706,
0.0693640485405922,
-0.03430663421750069,
0.03159473091363907,
-0.10945126414299011,
0.041360966861248016,
-0.18735328316688538,
0.003874248592182994,
-0.08583753556013107,
0.07179579883813858,
-0.09306249767541885,
-0.08517639338970184,
-0.015431683510541916,
0.0062622493132948875,
0.08747606724500656,
0.10318896174430847,
-0.14944259822368622,
-0.0581885389983654,
0.14982865750789642,
-0.09124893695116043,
-0.17624926567077637,
0.12427528202533722,
-0.047914113849401474,
0.061393290758132935,
0.04755978658795357,
0.16868218779563904,
0.0616682767868042,
-0.11862847954034805,
-0.018172746524214745,
0.003125043585896492,
0.059556711465120316,
-0.027621155604720116,
0.069667287170887,
-0.016716107726097107,
0.003911957144737244,
0.008871469646692276,
-0.06546089053153992,
0.034424856305122375,
-0.09911514818668365,
-0.08977389335632324,
-0.05385984107851982,
-0.10926553606987,
0.04224531352519989,
0.06640573590993881,
0.056745659559965134,
-0.11437124013900757,
-0.10020264238119125,
0.07629279792308807,
0.08758904784917831,
-0.07177864015102386,
0.01953268051147461,
-0.07853355258703232,
0.08141487091779709,
-0.07461990416049957,
-0.0283375047147274,
-0.1568889319896698,
-0.05427051708102226,
0.0030947246123105288,
-0.004757899325340986,
0.006649648770689964,
0.015061174519360065,
0.08364995568990707,
0.06679408252239227,
-0.07277575880289078,
-0.04576537013053894,
-0.05219518020749092,
0.014072518795728683,
-0.1144556924700737,
-0.2103353887796402,
-0.026713289320468903,
-0.03361331671476364,
0.10900475829839706,
-0.21285325288772583,
0.043190985918045044,
-0.0020271677058190107,
0.09231361746788025,
0.036184754222631454,
-0.009037263691425323,
-0.04559910297393799,
0.06368440389633179,
-0.03133980184793472,
-0.06026409938931465,
0.03968898206949234,
0.002780989510938525,
-0.09370192140340805,
-0.07623559981584549,
-0.1157163456082344,
0.1735924631357193,
0.12736397981643677,
-0.09160782396793365,
-0.079501211643219,
-0.0017041253158822656,
-0.06279008835554123,
-0.040141116827726364,
-0.03961801528930664,
0.005880035925656557,
0.11513597518205643,
-0.014275002293288708,
0.11783244460821152,
-0.08347894996404648,
-0.038106340914964676,
0.010217049159109592,
-0.05128463730216026,
0.018388114869594574,
0.1080067977309227,
0.11438195407390594,
-0.07793420553207397,
0.14664483070373535,
0.18690940737724304,
-0.11109071969985962,
0.11495424807071686,
-0.06606268882751465,
-0.08766110986471176,
-0.03343747556209564,
0.023377012461423874,
0.00860913097858429,
0.14631205797195435,
-0.13784223794937134,
0.018673941493034363,
0.015117805451154709,
0.011043009348213673,
0.02062302455306053,
-0.21574048697948456,
-0.04429280757904053,
0.027669047936797142,
-0.04530027136206627,
-0.014141418971121311,
-0.010482629761099815,
-0.008280049078166485,
0.08751733601093292,
-0.012050123885273933,
-0.06216656044125557,
0.039291467517614365,
-0.0028817809652537107,
-0.07470632344484329,
0.2108505219221115,
-0.07383661717176437,
-0.08231393247842789,
-0.11021465808153152,
-0.027316400781273842,
-0.03676888346672058,
0.01191480178385973,
0.06919683516025543,
-0.07928702235221863,
-0.035149816423654556,
-0.09274188429117203,
0.014241885393857956,
0.035347312688827515,
0.028293823823332787,
0.03302577883005142,
0.00197785347700119,
0.09445279836654663,
-0.11119365692138672,
0.007646216079592705,
-0.0434451699256897,
-0.06777635216712952,
0.03244328871369362,
0.04332584887742996,
0.12954339385032654,
0.13176003098487854,
-0.006293661426752806,
-0.003174192737787962,
-0.018926145508885384,
0.24626010656356812,
-0.06957513093948364,
-0.026417139917612076,
0.13200309872627258,
-0.009646299295127392,
0.04485779628157616,
0.1337425708770752,
0.07189981639385223,
-0.0981660783290863,
0.026833930984139442,
0.05833779275417328,
-0.018213531002402306,
-0.22848562896251678,
-0.015088150277733803,
-0.03629239648580551,
0.000045420805690810084,
0.07777763903141022,
0.030681850388646126,
0.040295474231243134,
0.07582857459783554,
0.021042656153440475,
0.045759521424770355,
-0.03376749902963638,
0.0639004036784172,
0.09280970692634583,
0.037512339651584625,
0.11627176403999329,
-0.049132365733385086,
-0.05257520079612732,
0.02913925051689148,
0.003403926733881235,
0.22885066270828247,
0.011365432292222977,
0.13645634055137634,
0.07914537191390991,
0.17988373339176178,
-0.019697509706020355,
0.06641747057437897,
-0.013529207557439804,
-0.064922995865345,
0.0009128026431426406,
-0.05508677661418915,
0.0013120006769895554,
0.04334177449345589,
-0.08823199570178986,
0.08494602143764496,
-0.0997040644288063,
0.025331957265734673,
0.07190584391355515,
0.24650606513023376,
0.05759034305810928,
-0.2938763499259949,
-0.09897506982088089,
0.01800764538347721,
-0.025016997009515762,
-0.01612672209739685,
0.030009277164936066,
0.13152241706848145,
-0.038725271821022034,
0.01120639406144619,
-0.0626465231180191,
0.08149950206279755,
0.011924086138606071,
0.03988318517804146,
0.06834918260574341,
0.08969014137983322,
-0.01107326801866293,
0.07226663827896118,
-0.2856568396091461,
0.271731436252594,
0.01860191859304905,
0.09337102621793747,
-0.04448726028203964,
-0.004649949260056019,
0.018219614401459694,
0.06607591360807419,
0.09256407618522644,
-0.027485761791467667,
-0.04307778924703598,
-0.17113527655601501,
-0.04461004212498665,
0.038364943116903305,
0.09722242504358292,
-0.029817651957273483,
0.10828591138124466,
-0.01972132734954357,
0.01194797083735466,
0.0911933183670044,
-0.012685544788837433,
-0.10899017751216888,
-0.07839783281087875,
-0.016333330422639847,
0.005076844710856676,
-0.04149710386991501,
-0.09716060012578964,
-0.09863841533660889,
-0.1091403141617775,
0.13526444137096405,
-0.04622833430767059,
-0.02194986678659916,
-0.10105779021978378,
0.06971652805805206,
0.09055378288030624,
-0.07615187764167786,
0.03999576345086098,
0.02138366736471653,
0.05533910542726517,
0.03994203358888626,
-0.04782804101705551,
0.11958404630422592,
-0.0817321166396141,
-0.17507250607013702,
-0.061528678983449936,
0.10911136865615845,
0.037768565118312836,
0.044988106936216354,
-0.012711377814412117,
0.004085663706064224,
-0.02373502217233181,
-0.09175744652748108,
0.026362668722867966,
-0.02753734402358532,
0.07150380313396454,
0.0161500945687294,
-0.04852674528956413,
0.04140724986791611,
-0.049215421080589294,
-0.020558204501867294,
0.13631883263587952,
0.29612913727760315,
-0.08491875231266022,
-0.015827137976884842,
0.062439993023872375,
-0.04892180114984512,
-0.19544535875320435,
0.06185584515333176,
0.03257249295711517,
-0.004850971046835184,
0.057877007871866226,
-0.1424984186887741,
0.14537838101387024,
0.10979026556015015,
-0.026831332594156265,
0.11165576428174973,
-0.30763980746269226,
-0.12517096102237701,
0.13092049956321716,
0.15578052401542664,
0.10271019488573074,
-0.17396891117095947,
-0.03480206057429314,
-0.012560687027871609,
-0.1473204791545868,
0.08108969032764435,
-0.1549307405948639,
0.0945407971739769,
-0.013403966091573238,
0.04998783394694328,
0.0037797512486577034,
-0.07027675956487656,
0.14607980847358704,
-0.000046318287786561996,
0.11850225180387497,
-0.04621842876076698,
-0.013441015034914017,
0.07169019430875778,
-0.043754030019044876,
0.03303072974085808,
-0.08953805267810822,
0.05375760793685913,
-0.05097579583525658,
-0.026208076626062393,
-0.06335466355085373,
0.0432647280395031,
-0.05150436982512474,
-0.06759865581989288,
-0.045696575194597244,
0.031161170452833176,
0.02781839855015278,
-0.011523814871907234,
0.14316794276237488,
0.02016613259911537,
0.15032581984996796,
0.11551767587661743,
0.07723074406385422,
-0.06723835319280624,
-0.05477220565080643,
0.011233268305659294,
-0.0318131148815155,
0.07324054837226868,
-0.15704317390918732,
0.041798677295446396,
0.1301533579826355,
0.031892213970422745,
0.13937824964523315,
0.07037341594696045,
-0.04625731706619263,
0.011523851193487644,
0.056804731488227844,
-0.16416457295417786,
-0.13765478134155273,
0.0004902769578620791,
-0.04848967865109444,
-0.14053121209144592,
0.08377081155776978,
0.09766854345798492,
-0.058544453233480453,
0.007872414775192738,
-0.0021101331803947687,
0.004452725872397423,
-0.060293179005384445,
0.1910732388496399,
0.0856165662407875,
0.050420865416526794,
-0.07316560298204422,
0.08543732762336731,
0.025754539296030998,
-0.08191429078578949,
0.0036028979811817408,
0.02972598373889923,
-0.06128942221403122,
-0.04768751561641693,
0.057786114513874054,
0.17645438015460968,
-0.017480891197919846,
-0.05138344317674637,
-0.15388798713684082,
-0.10391058027744293,
0.05379781499505043,
0.17472943663597107,
0.09876082092523575,
0.011043163016438484,
-0.016174687072634697,
0.03864007815718651,
-0.12078005075454712,
0.11567529290914536,
0.039784107357263565,
0.0825420618057251,
-0.15192361176013947,
0.10602233558893204,
-0.005278966389596462,
0.015046751126646996,
-0.021611301228404045,
0.05356408283114433,
-0.12369267642498016,
0.004818740300834179,
-0.16167838871479034,
-0.031912337988615036,
-0.04243628308176994,
-0.00007314696267712861,
0.016739998012781143,
-0.07804747670888901,
-0.0766838788986206,
0.024620801210403442,
-0.10337430983781815,
-0.01428994257003069,
0.06022421643137932,
0.052621982991695404,
-0.14696744084358215,
-0.03189612552523613,
0.041056301444768906,
-0.07057381421327591,
0.0632329210639,
0.0331825315952301,
0.029022322967648506,
0.05176157504320145,
-0.18470416963100433,
0.023166287690401077,
0.05453268811106682,
0.008255518041551113,
0.05301650986075401,
-0.09316183626651764,
-0.029816320165991783,
-0.00498243561014533,
0.055776119232177734,
0.00805442314594984,
0.048523422330617905,
-0.12184276431798935,
-0.006599885877221823,
-0.041501663625240326,
-0.06636733561754227,
-0.06084393337368965,
0.011216291226446629,
0.10473647713661194,
0.0061342609114944935,
0.20113077759742737,
-0.08018026500940323,
0.020927581936120987,
-0.21721096336841583,
0.006357924081385136,
0.0005479813553392887,
-0.07817442715167999,
-0.09612289071083069,
-0.040327563881874084,
0.0530112087726593,
-0.06777244061231613,
0.1314149647951126,
-0.03942599520087242,
0.03305871784687042,
0.036887556314468384,
-0.05675159767270088,
0.06102527305483818,
0.02875526063144207,
0.2587580978870392,
0.017544526606798172,
-0.027191393077373505,
0.012560263276100159,
0.03838387876749039,
0.09434543550014496,
0.07679316401481628,
0.17991912364959717,
0.1845041662454605,
-0.050945721566677094,
0.08982273936271667,
0.054168540984392166,
-0.06128912791609764,
-0.11643441021442413,
0.06968209892511368,
-0.024204589426517487,
0.07028414309024811,
-0.014917990192770958,
0.21106743812561035,
0.11576888710260391,
-0.1583249568939209,
0.013773822225630283,
-0.05545544996857643,
-0.08638123422861099,
-0.10067977756261826,
-0.030203357338905334,
-0.08510877937078476,
-0.17570649087429047,
0.016593754291534424,
-0.1270909309387207,
0.0009135941509157419,
0.09953097254037857,
0.010672074742615223,
-0.014867765828967094,
0.2160998284816742,
0.031276196241378784,
0.053685739636421204,
0.03416353464126587,
-0.0031266885343939066,
-0.03791229426860809,
-0.07161640375852585,
-0.0726158618927002,
0.015160031616687775,
-0.034109894186258316,
0.020006928592920303,
-0.05643845722079277,
-0.05833569914102554,
0.04029514640569687,
-0.005345369689166546,
-0.09742140769958496,
0.0028150181751698256,
0.02954794466495514,
0.04776023328304291,
0.05714746564626694,
0.015563864260911942,
0.023946579545736313,
-0.01110083982348442,
0.21950305998325348,
-0.08350735157728195,
-0.070400170981884,
-0.12050648778676987,
0.1965944468975067,
0.025354061275720596,
0.01319186482578516,
0.017012549564242363,
-0.10124767571687698,
0.030138062313199043,
0.207011878490448,
0.16067950427532196,
-0.08113617449998856,
-0.0012133243726566434,
0.005652322433888912,
-0.009823101572692394,
-0.07486828416585922,
0.05726049095392227,
0.12612278759479523,
0.02443198673427105,
-0.08212122321128845,
-0.06962821632623672,
-0.0440346822142601,
-0.0178349781781435,
-0.041994258761405945,
0.04103654995560646,
0.044749293476343155,
0.007497087586671114,
-0.04636310413479805,
0.06092584878206253,
-0.03342123329639435,
-0.12664338946342468,
0.0767127200961113,
-0.1735985279083252,
-0.14392557740211487,
-0.0183744877576828,
0.12503382563591003,
-0.01228486467152834,
0.053399212658405304,
-0.04180862754583359,
-0.012640140019357204,
0.08096010982990265,
-0.019735140725970268,
-0.06453187018632889,
-0.08844248205423355,
0.08284731209278107,
-0.09417755901813507,
0.23162773251533508,
-0.031115693971514702,
0.0654124766588211,
0.13982054591178894,
0.0345631018280983,
-0.084713414311409,
0.08391895890235901,
0.06642034649848938,
-0.10540223121643066,
0.011755467392504215,
0.07862038165330887,
-0.03128759190440178,
0.120674267411232,
0.05946990102529526,
-0.15106318891048431,
0.0053291283547878265,
-0.03741956129670143,
-0.0729675367474556,
-0.07949789613485336,
-0.031817272305488586,
-0.06557213515043259,
0.13159750401973724,
0.18833664059638977,
-0.04668503627181053,
0.016340022906661034,
-0.04423931986093521,
0.04671303182840347,
0.07153702527284622,
0.053498659282922745,
-0.030362121760845184,
-0.22725465893745422,
0.05558815971016884,
0.05898962914943695,
-0.022746507078409195,
-0.24096186459064484,
-0.09315942227840424,
0.021570716053247452,
-0.05836266651749611,
-0.07721637189388275,
0.06337779760360718,
0.1329357922077179,
0.06411603093147278,
-0.05933794751763344,
-0.11032743752002716,
-0.08574248850345612,
0.15853965282440186,
-0.1339440792798996,
-0.09521511942148209
] |
null | null | null |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-cased-bn-adapter-895K-squad-model1
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 4
- seed: 76
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-base-cased-bn-adapter-895K-squad-model1", "results": []}]} | null | varun-v-rao/bert-base-cased-bn-adapter-895K-squad-model1 | [
"tensorboard",
"generated_from_trainer",
"dataset:varun-v-rao/squad",
"base_model:bert-base-cased",
"license:apache-2.0",
"region:us"
] | 2024-02-08T18:54:27+00:00 | [] | [] | TAGS
#tensorboard #generated_from_trainer #dataset-varun-v-rao/squad #base_model-bert-base-cased #license-apache-2.0 #region-us
|
# bert-base-cased-bn-adapter-895K-squad-model1
This model is a fine-tuned version of bert-base-cased on the squad dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 4
- seed: 76
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
### Framework versions
- Transformers 4.35.2
- Pytorch 2.1.1+cu121
- Datasets 2.15.0
- Tokenizers 0.15.0
| [
"# bert-base-cased-bn-adapter-895K-squad-model1\n\nThis model is a fine-tuned version of bert-base-cased on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 4\n- seed: 76\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
"TAGS\n#tensorboard #generated_from_trainer #dataset-varun-v-rao/squad #base_model-bert-base-cased #license-apache-2.0 #region-us \n",
"# bert-base-cased-bn-adapter-895K-squad-model1\n\nThis model is a fine-tuned version of bert-base-cased on the squad dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 4\n- seed: 76\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3",
"### Training results",
"### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
49,
45,
6,
12,
8,
3,
90,
4,
33
] | [
"passage: TAGS\n#tensorboard #generated_from_trainer #dataset-varun-v-rao/squad #base_model-bert-base-cased #license-apache-2.0 #region-us \n# bert-base-cased-bn-adapter-895K-squad-model1\n\nThis model is a fine-tuned version of bert-base-cased on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 4\n- seed: 76\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0"
] | [
-0.0894874855875969,
0.03411871939897537,
-0.0005927752354182303,
0.07275237888097763,
0.17212897539138794,
0.012622317299246788,
0.12911853194236755,
0.09410377591848373,
-0.11854589730501175,
0.03419187664985657,
0.05006055906414986,
0.04325467348098755,
0.03712443262338638,
0.09137928485870361,
-0.017690148204565048,
-0.2365936040878296,
-0.0012330833123996854,
0.03492705523967743,
-0.11791003495454788,
0.08915336430072784,
0.11089775711297989,
-0.12128310650587082,
0.06529467552900314,
0.025153471156954765,
-0.2215052843093872,
0.05520622432231903,
-0.010973661206662655,
-0.03718232363462448,
0.10494109988212585,
0.023250669240951538,
0.13859322667121887,
-0.008468594402074814,
0.14591363072395325,
-0.16384561359882355,
0.007799128070473671,
0.09262869507074356,
0.035755131393671036,
0.08967388421297073,
0.019267743453383446,
0.04433157667517662,
0.08020452409982681,
-0.08393730223178864,
0.08739744871854782,
0.03502456843852997,
-0.075468048453331,
-0.20389100909233093,
-0.07572910934686661,
0.038987260311841965,
0.10280486941337585,
0.07216808944940567,
-0.005637888330966234,
0.1325550228357315,
-0.05930975452065468,
0.061474572867155075,
0.19888953864574432,
-0.2908451557159424,
-0.09066556394100189,
0.09269938617944717,
0.05713158845901489,
0.07761667668819427,
-0.09879308938980103,
-0.012628952041268349,
0.06408122181892395,
0.048029132187366486,
0.1049959734082222,
-0.014116380363702774,
-0.1126946285367012,
0.009392967447638512,
-0.14412686228752136,
0.031609516590833664,
0.1478167027235031,
0.038643572479486465,
-0.061440397053956985,
-0.01993054896593094,
-0.05618898570537567,
-0.0703851506114006,
-0.054967377334833145,
-0.04850879684090614,
0.06625118106603622,
-0.03522516414523125,
-0.09012258797883987,
-0.08125653862953186,
-0.08411361277103424,
-0.09699376672506332,
-0.02134568989276886,
0.13033518195152283,
0.0368560291826725,
0.02309439331293106,
-0.06301677972078323,
0.10314981639385223,
-0.07403990626335144,
-0.1122455894947052,
0.01003890298306942,
0.01358009409159422,
-0.050949711352586746,
-0.06601804494857788,
-0.06855551898479462,
-0.05728256702423096,
0.0072582680732011795,
0.1334250420331955,
-0.049217261373996735,
0.0649813637137413,
-0.02211337722837925,
0.020753460004925728,
-0.05010131001472473,
0.15675537288188934,
-0.06272446364164352,
-0.022758662700653076,
0.02124810591340065,
0.07968751341104507,
0.012343522161245346,
0.006080899853259325,
-0.11531973630189896,
0.017346525564789772,
0.06412991881370544,
0.011822891421616077,
-0.08148845285177231,
0.03070100024342537,
-0.023479225113987923,
-0.027241311967372894,
-0.0030484613962471485,
-0.09860505163669586,
0.029265567660331726,
-0.018033528700470924,
-0.0905800387263298,
-0.03352869302034378,
0.023009173572063446,
0.014436095952987671,
0.00585779407992959,
0.08990482985973358,
-0.1066366508603096,
0.011886530555784702,
-0.128017395734787,
-0.10926085710525513,
0.0064332447946071625,
-0.08891330659389496,
0.0021843702998012304,
-0.06784044206142426,
-0.17507800459861755,
-0.007535062730312347,
0.0529060959815979,
-0.057703811675310135,
-0.030895261093974113,
-0.028887102380394936,
-0.0757640078663826,
0.003140918677672744,
-0.010585064068436623,
0.17574356496334076,
-0.04138559475541115,
0.059383295476436615,
0.03955215588212013,
0.03164388984441757,
-0.07422447949647903,
0.043886054307222366,
-0.06900088489055634,
0.025035571306943893,
-0.2103659063577652,
0.0412929467856884,
-0.10563358664512634,
0.04375024139881134,
-0.11126066744327545,
-0.07999447733163834,
0.009380683302879333,
0.0027722662780433893,
0.10531549900770187,
0.07646551728248596,
-0.19010134041309357,
-0.043263860046863556,
0.10745719075202942,
-0.07299956679344177,
-0.09819704294204712,
0.08215456455945969,
-0.055312514305114746,
0.07329285144805908,
0.06595523655414581,
0.1488543450832367,
0.04403937980532646,
-0.1346016228199005,
-0.002681573387235403,
0.02227836661040783,
0.10897818207740784,
-0.03302868455648422,
0.04700559377670288,
-0.007684778422117233,
-0.006212822627276182,
0.023284928873181343,
-0.07007969170808792,
-0.006706451065838337,
-0.09373927116394043,
-0.06694665551185608,
-0.051806870847940445,
-0.10117567330598831,
0.009002876468002796,
0.03687736764550209,
0.05446910113096237,
-0.07716456800699234,
-0.10370860993862152,
0.2124837040901184,
0.1183871403336525,
-0.050913989543914795,
0.025067733600735664,
-0.09331252425909042,
0.07237383723258972,
-0.07485098391771317,
-0.01780569553375244,
-0.2040504515171051,
-0.1108662486076355,
0.02034648135304451,
-0.024086060002446175,
0.060253776609897614,
0.03663964569568634,
0.058161649852991104,
0.0682457759976387,
-0.062000345438718796,
0.022308720275759697,
-0.09429476410150528,
-0.007400434464216232,
-0.12971387803554535,
-0.2131022810935974,
-0.073513463139534,
-0.0251644067466259,
0.15673165023326874,
-0.22283674776554108,
0.03073832206428051,
-0.017235958948731422,
0.12094975262880325,
0.026999209076166153,
-0.03536355495452881,
-0.052345335483551025,
0.0717862993478775,
-0.005649642553180456,
-0.0719388946890831,
0.04999411851167679,
0.018876362591981888,
-0.07384568452835083,
-0.09523084759712219,
-0.10673393309116364,
0.06381641328334808,
0.10398558527231216,
-0.025945361703634262,
-0.07303068786859512,
0.001836166251450777,
-0.05525669455528259,
-0.027649739757180214,
-0.06463546305894852,
-0.02207126095890999,
0.16969908773899078,
-0.0005715456791222095,
0.14356115460395813,
-0.0834343284368515,
-0.057597678154706955,
0.008779426105320454,
-0.0056413281708955765,
0.03297480195760727,
0.05996577814221382,
0.10462567955255508,
-0.06412754207849503,
0.09202875196933746,
0.11539918184280396,
-0.10949275642633438,
0.13345466554164886,
-0.0510312020778656,
-0.07367164641618729,
-0.016281625255942345,
-0.02774122729897499,
-0.017194394022226334,
0.17173506319522858,
-0.14972227811813354,
-0.006646972615271807,
0.012911605648696423,
0.012013557367026806,
0.04539842903614044,
-0.19925051927566528,
0.006969705689698458,
0.011818370781838894,
-0.030858034268021584,
-0.029095986858010292,
-0.03958989679813385,
0.004633236210793257,
0.09556369483470917,
0.026797957718372345,
-0.034531351178884506,
0.013430490158498287,
-0.0012521425960585475,
-0.06617479026317596,
0.20497533679008484,
-0.11308403313159943,
-0.06042194366455078,
-0.12783050537109375,
-0.015663601458072662,
-0.07389295101165771,
-0.0159149132668972,
0.013500211760401726,
-0.10371751338243484,
-0.04563552513718605,
-0.078755222260952,
0.04030437394976616,
0.0031310939230024815,
-0.0025760014541447163,
0.03783976286649704,
-0.002397438045591116,
0.12309616059064865,
-0.1548866629600525,
0.018054286018013954,
-0.05457179993391037,
-0.12588155269622803,
-0.04096965864300728,
0.07165742665529251,
0.09822451323270798,
0.1370978206396103,
-0.03311805799603462,
0.004704192280769348,
-0.026723917573690414,
0.2510181963443756,
-0.05612584203481674,
-0.010892143473029137,
0.11281119287014008,
-0.0039106872864067554,
0.030800575390458107,
0.11791612207889557,
0.07621391117572784,
-0.10629865527153015,
0.025683723390102386,
0.10323770344257355,
-0.02779451012611389,
-0.24046795070171356,
-0.020435314625501633,
-0.02495889738202095,
-0.0850997269153595,
0.06194235384464264,
0.04131773114204407,
-0.02128034271299839,
0.07306015491485596,
0.020880216732621193,
0.0944453701376915,
-0.044887788593769073,
0.08008521795272827,
0.10952842980623245,
0.032991450279951096,
0.1288934201002121,
-0.039412133395671844,
-0.031477008014917374,
0.053933169692754745,
0.02060777135193348,
0.2563970983028412,
-0.006702064070850611,
0.08229359239339828,
0.07529397308826447,
0.13708147406578064,
-0.02783367782831192,
0.047326959669589996,
-0.023130914196372032,
-0.03089766763150692,
-0.0028362509328871965,
-0.06246305629611015,
0.007889976724982262,
0.0017466360004618764,
-0.08093895763158798,
0.07434961199760437,
-0.07212928682565689,
-0.00026523962151259184,
0.030730990692973137,
0.27088338136672974,
0.02704533189535141,
-0.2838734984397888,
-0.09228469431400299,
0.012429178692400455,
-0.022867759689688683,
-0.059011735022068024,
0.015006457455456257,
0.12166035175323486,
-0.07500919699668884,
0.032772328704595566,
-0.05938720703125,
0.10608158260583878,
0.012071660719811916,
0.013474592007696629,
0.09098143875598907,
0.1775290071964264,
-0.010476050898432732,
0.06074105575680733,
-0.22530651092529297,
0.2271723449230194,
0.03957974538207054,
0.12575487792491913,
-0.030568398535251617,
0.02801329642534256,
0.022621585056185722,
0.0932193323969841,
0.03988900035619736,
-0.017212726175785065,
-0.01226562075316906,
-0.1920415461063385,
-0.025080343708395958,
0.06109541282057762,
0.1048564538359642,
-0.026767568662762642,
0.09680061787366867,
-0.030404577031731606,
0.01997925154864788,
0.06957114487886429,
-0.038007672876119614,
-0.1784280687570572,
-0.08210169523954391,
-0.0367114283144474,
0.00778612494468689,
-0.06378093361854553,
-0.08418510109186172,
-0.10180138051509857,
-0.046901635825634,
0.12216600775718689,
0.007340844254940748,
-0.03202234208583832,
-0.11664906144142151,
0.120026595890522,
0.09516461938619614,
-0.04054797440767288,
0.02630266174674034,
0.01088206097483635,
0.12704631686210632,
0.04111802950501442,
-0.08254344761371613,
0.08136151731014252,
-0.07600247114896774,
-0.14332561194896698,
-0.05326191335916519,
0.08616351336240768,
0.06544451415538788,
0.04254426434636116,
-0.0013655818765982985,
0.012316019274294376,
0.0160522498190403,
-0.08891723304986954,
-0.008044606074690819,
0.006184715311974287,
0.05793992057442665,
0.02548087388277054,
-0.10274190455675125,
0.024264542385935783,
-0.022527310997247696,
0.009523293003439903,
0.0890611857175827,
0.21170632541179657,
-0.07336940616369247,
0.006509089842438698,
0.11334212124347687,
-0.09791427105665207,
-0.18121740221977234,
0.1334703415632248,
0.1132139042019844,
0.012898596003651619,
0.04509252309799194,
-0.22071856260299683,
0.1851789504289627,
0.12617170810699463,
-0.03256872668862343,
0.07460881024599075,
-0.2654837667942047,
-0.1474827527999878,
0.10224820673465729,
0.15116846561431885,
0.05368407443165779,
-0.14694513380527496,
-0.02621578425168991,
-0.05052242428064346,
-0.14533227682113647,
0.14154241979122162,
-0.19056135416030884,
0.08906607329845428,
0.014819380827248096,
0.09551923722028732,
0.006784106604754925,
-0.024546166881918907,
0.1309857964515686,
0.02715943194925785,
0.10910353064537048,
-0.03516397997736931,
0.03359880670905113,
0.13203492760658264,
-0.039924994111061096,
0.01794857159256935,
-0.02929198555648327,
0.04023260250687599,
-0.11362937092781067,
-0.008446554653346539,
-0.06219552829861641,
0.04685792699456215,
-0.05529642477631569,
-0.06668093800544739,
-0.04733619838953018,
0.045771170407533646,
-0.0040003578178584576,
-0.03432123363018036,
0.10127580165863037,
0.0064048185013234615,
0.15282893180847168,
0.10222841054201126,
0.10581778734922409,
-0.06717903167009354,
-0.05917458236217499,
0.04630143195390701,
-0.021769218146800995,
0.08766922354698181,
-0.16409912705421448,
0.01701395772397518,
0.11053027212619781,
0.039991602301597595,
0.1208009347319603,
0.06129561737179756,
-0.06598833948373795,
0.02284175530076027,
0.056810248643159866,
-0.09240669012069702,
-0.14526401460170746,
0.006260847672820091,
-0.013805711641907692,
-0.14134013652801514,
0.07663756608963013,
0.11131893843412399,
-0.07164061814546585,
-0.00865525659173727,
0.0033417341765016317,
-0.012486434541642666,
-0.07176554203033447,
0.19657829403877258,
0.07280092686414719,
0.05766695365309715,
-0.09194426983594894,
0.10539798438549042,
0.053739674389362335,
-0.039857879281044006,
0.026020463556051254,
0.04224343225359917,
-0.07298851013183594,
-0.00490296445786953,
0.0636269748210907,
0.21466051042079926,
-0.06202057749032974,
-0.0530448779463768,
-0.11492492258548737,
-0.10204654932022095,
0.029275143519043922,
0.17394870519638062,
0.07375438511371613,
-0.04017915576696396,
-0.04791403189301491,
0.07268666476011276,
-0.13075780868530273,
0.08954650163650513,
0.010646265000104904,
0.09050412476062775,
-0.11434698104858398,
0.10019921511411667,
0.017825566232204437,
0.022118736058473587,
-0.027783771976828575,
0.031443264335393906,
-0.10861753672361374,
-0.013250855728983879,
-0.203622967004776,
-0.03563598915934563,
0.004955441225320101,
0.007441971451044083,
-0.004023309331387281,
-0.04924299940466881,
-0.07411099970340729,
0.04614689201116562,
-0.09768941253423691,
-0.03390924632549286,
0.0502750426530838,
0.03908810392022133,
-0.13807711005210876,
0.020212432369589806,
0.01450246013700962,
-0.07342683523893356,
0.057397179305553436,
0.028698327019810677,
0.040927864611148834,
0.07322116196155548,
-0.12187693268060684,
-0.004099969286471605,
0.016296107321977615,
0.031060604378581047,
0.09078742563724518,
-0.04262184724211693,
-0.00876342598348856,
-0.02874162420630455,
0.09689350426197052,
0.001769997994415462,
0.08794352412223816,
-0.11255102604627609,
-0.026786508038640022,
-0.052619583904743195,
-0.04238247498869896,
-0.047606054693460464,
0.03907421603798866,
0.09794692695140839,
0.07675346732139587,
0.1607373058795929,
-0.0746709555387497,
-0.01608264446258545,
-0.2046947032213211,
-0.023801056668162346,
-0.003368314588442445,
-0.04843796044588089,
-0.08697528392076492,
-0.05931956693530083,
0.06216065585613251,
-0.07049544155597687,
0.12023050338029861,
0.013809438794851303,
0.10502126812934875,
0.04993043094873428,
-0.019413992762565613,
-0.017407825216650963,
0.01433313637971878,
0.19931857287883759,
0.05462876707315445,
0.0030950140208005905,
0.058262526988983154,
0.03433223441243172,
0.07704102247953415,
0.08397378772497177,
0.18304769694805145,
0.13153022527694702,
-0.05198739469051361,
0.07940490543842316,
0.10450825095176697,
-0.07266988605260849,
-0.1281593143939972,
0.1381160169839859,
-0.03086724691092968,
0.11068011820316315,
-0.06563221663236618,
0.20413078367710114,
0.09172802418470383,
-0.1607627123594284,
0.0405319444835186,
-0.07329678535461426,
-0.09700877219438553,
-0.12059926241636276,
-0.011454121209681034,
-0.06691760569810867,
-0.1670144945383072,
0.016583798453211784,
-0.14034408330917358,
0.006806403398513794,
0.1507534682750702,
0.004717755597084761,
0.008966171182692051,
0.1774207055568695,
-0.035564061254262924,
0.028064725920557976,
0.047920748591423035,
0.001135868369601667,
-0.017773697152733803,
-0.09012981504201889,
-0.06925975531339645,
0.04668596759438515,
0.00937720574438572,
0.05201723054051399,
-0.06028207391500473,
-0.02692568488419056,
0.03271709755063057,
0.007802112493664026,
-0.06326821446418762,
0.022630054503679276,
0.027217920869588852,
0.049524590373039246,
0.04289171099662781,
0.01239149458706379,
0.005296844057738781,
-0.03831743821501732,
0.2716510593891144,
-0.08075346797704697,
-0.07905786484479904,
-0.13885734975337982,
0.20278200507164001,
0.007796280086040497,
-0.008208172395825386,
0.041871100664138794,
-0.11041989922523499,
-0.028591029345989227,
0.18185295164585114,
0.1470005214214325,
-0.09548846632242203,
-0.033376242965459824,
0.001700010965578258,
-0.029779953882098198,
-0.10044853389263153,
0.1543399840593338,
0.10845950245857239,
0.01951836235821247,
-0.06288507580757141,
-0.039232634007930756,
-0.018952952697873116,
-0.024431919679045677,
-0.04801316559314728,
0.05137334391474724,
0.04162292554974556,
0.0018375141080468893,
-0.03666456788778305,
0.05851592868566513,
-0.004868046380579472,
-0.22034989297389984,
0.03841280937194824,
-0.1300404816865921,
-0.17887189984321594,
-0.042549263685941696,
0.07995288074016571,
-0.024034623056650162,
0.05384025722742081,
-0.042238492518663406,
0.013402975164353848,
0.1251058578491211,
-0.04099695757031441,
-0.015027342364192009,
-0.15958727896213531,
0.13320569694042206,
-0.10422777384519577,
0.22585272789001465,
-0.028796929866075516,
0.061766576021909714,
0.11895111948251724,
0.01946345530450344,
-0.1029658392071724,
0.035318247973918915,
0.047606173902750015,
-0.10642577707767487,
-0.0024260547943413258,
0.12722280621528625,
-0.05226152762770653,
0.0987548902630806,
0.022451704367995262,
-0.1359514743089676,
0.004536112304776907,
-0.07490644603967667,
-0.056435421109199524,
-0.07986103743314743,
-0.01941641978919506,
-0.10971599817276001,
0.1294938325881958,
0.20704592764377594,
-0.025367362424731255,
0.0032631023786962032,
-0.08827781677246094,
0.03990548476576805,
0.061341963708400726,
0.08971542119979858,
-0.037925638258457184,
-0.24031659960746765,
0.027869602665305138,
0.07191156595945358,
-0.03394339978694916,
-0.2725347876548767,
-0.06316375732421875,
0.07039590924978256,
-0.03616834059357643,
-0.04962202534079552,
0.08719946444034576,
0.11355262249708176,
0.05517735332250595,
-0.05876823142170906,
-0.16644208133220673,
-0.07593192160129547,
0.14567328989505768,
-0.14290936291217804,
-0.06946080178022385
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.